Alert button
Picture for Yifei Cheng

Yifei Cheng

Alert button

Communication-Efficient Distributed Learning with Local Immediate Error Compensation

Add code
Bookmark button
Alert button
Feb 19, 2024
Yifei Cheng, Li Shen, Linli Xu, Xun Qian, Shiwei Wu, Yiming Zhou, Tie Zhang, Dacheng Tao, Enhong Chen

Viaarxiv icon

DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training

Add code
Bookmark button
Alert button
Feb 28, 2022
Joya Chen, Kai Xu, Yifei Cheng, Angela Yao

Figure 1 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 2 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 3 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Figure 4 for DropIT: Dropping Intermediate Tensors for Memory-Efficient DNN Training
Viaarxiv icon

STL-SGD: Speeding Up Local SGD with Stagewise Communication Period

Add code
Bookmark button
Alert button
Jun 11, 2020
Shuheng Shen, Yifei Cheng, Jingchang Liu, Linli Xu

Figure 1 for STL-SGD: Speeding Up Local SGD with Stagewise Communication Period
Figure 2 for STL-SGD: Speeding Up Local SGD with Stagewise Communication Period
Figure 3 for STL-SGD: Speeding Up Local SGD with Stagewise Communication Period
Figure 4 for STL-SGD: Speeding Up Local SGD with Stagewise Communication Period
Viaarxiv icon

Variance Reduced Local SGD with Lower Communication Complexity

Add code
Bookmark button
Alert button
Dec 30, 2019
Xianfeng Liang, Shuheng Shen, Jingchang Liu, Zhen Pan, Enhong Chen, Yifei Cheng

Figure 1 for Variance Reduced Local SGD with Lower Communication Complexity
Figure 2 for Variance Reduced Local SGD with Lower Communication Complexity
Figure 3 for Variance Reduced Local SGD with Lower Communication Complexity
Figure 4 for Variance Reduced Local SGD with Lower Communication Complexity
Viaarxiv icon

Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent

Add code
Bookmark button
Alert button
Jun 28, 2019
Shuheng Shen, Linli Xu, Jingchang Liu, Xianfeng Liang, Yifei Cheng

Figure 1 for Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent
Figure 2 for Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent
Figure 3 for Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent
Figure 4 for Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent
Viaarxiv icon