Alert button
Picture for Zhiquan Lai

Zhiquan Lai

Alert button

Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models

Add code
Bookmark button
Alert button
Jun 21, 2022
Zhiquan Lai, Shengwei Li, Xudong Tang, Keshi Ge, Weijie Liu, Yabo Duan, Linbo Qiao, Dongsheng Li

Figure 1 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 2 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 3 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Figure 4 for Merak: An Efficient Distributed DNN Training Framework with Automated 3D Parallelism for Giant Foundation Models
Viaarxiv icon

DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation

Add code
Bookmark button
Alert button
Mar 30, 2022
Yu Tang, Chenyu Wang, Yufan Zhang, Yuliang Liu, Xingcheng Zhang, Linbo Qiao, Zhiquan Lai, Dongsheng Li

Figure 1 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 2 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 3 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 4 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Viaarxiv icon

EmbRace: Accelerating Sparse Communication for Distributed Training of NLP Neural Networks

Add code
Bookmark button
Alert button
Oct 18, 2021
Shengwei Li, Zhiquan Lai, Dongsheng Li, Xiangyu Ye, Yabo Duan

Figure 1 for EmbRace: Accelerating Sparse Communication for Distributed Training of NLP Neural Networks
Figure 2 for EmbRace: Accelerating Sparse Communication for Distributed Training of NLP Neural Networks
Figure 3 for EmbRace: Accelerating Sparse Communication for Distributed Training of NLP Neural Networks
Figure 4 for EmbRace: Accelerating Sparse Communication for Distributed Training of NLP Neural Networks
Viaarxiv icon

S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning

Add code
Bookmark button
Alert button
Oct 05, 2021
Keshi Ge, Yongquan Fu, Zhiquan Lai, Xiaoge Deng, Dongsheng Li

Figure 1 for S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning
Figure 2 for S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning
Figure 3 for S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning
Figure 4 for S2 Reducer: High-Performance Sparse Communication to Accelerate Distributed Deep Learning
Viaarxiv icon

Hierarchical Adaptive Pooling by Capturing High-order Dependency for Graph Representation Learning

Add code
Bookmark button
Alert button
Apr 13, 2021
Ning Liu, Songlei Jian, Dongsheng Li, Yiming Zhang, Zhiquan Lai, Hongzuo Xu

Figure 1 for Hierarchical Adaptive Pooling by Capturing High-order Dependency for Graph Representation Learning
Figure 2 for Hierarchical Adaptive Pooling by Capturing High-order Dependency for Graph Representation Learning
Figure 3 for Hierarchical Adaptive Pooling by Capturing High-order Dependency for Graph Representation Learning
Figure 4 for Hierarchical Adaptive Pooling by Capturing High-order Dependency for Graph Representation Learning
Viaarxiv icon

ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach

Add code
Bookmark button
Alert button
Jun 17, 2020
Yu Tang, Zhigang Kan, Dequan Sun, Linbo Qiao, Jingjing Xiao, Zhiquan Lai, Dongsheng Li

Figure 1 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 2 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 3 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Figure 4 for ADMMiRNN: Training RNN with Stable Convergence via An Efficient ADMM Approach
Viaarxiv icon