Alert button
Picture for Shaohuai Shi

Shaohuai Shi

Alert button

FedImpro: Measuring and Improving Client Update in Federated Learning

Add code
Bookmark button
Alert button
Feb 10, 2024
Zhenheng Tang, Yonggang Zhang, Shaohuai Shi, Xinmei Tian, Tongliang Liu, Bo Han, Xiaowen Chu

Viaarxiv icon

Dissecting the Runtime Performance of the Training, Fine-tuning, and Inference of Large Language Models

Add code
Bookmark button
Alert button
Nov 07, 2023
Longteng Zhang, Xiang Liu, Zeyu Li, Xinglin Pan, Peijie Dong, Ruibo Fan, Rui Guo, Xin Wang, Qiong Luo, Shaohuai Shi, Xiaowen Chu

Viaarxiv icon

FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs

Add code
Bookmark button
Alert button
Sep 03, 2023
Zhenheng Tang, Yuxin Wang, Xin He, Longteng Zhang, Xinglin Pan, Qiang Wang, Rongfei Zeng, Kaiyong Zhao, Shaohuai Shi, Bingsheng He, Xiaowen Chu

Figure 1 for FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs
Figure 2 for FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs
Figure 3 for FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs
Figure 4 for FusionAI: Decentralized Training and Deploying LLMs with Massive Consumer-Level GPUs
Viaarxiv icon

LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning

Add code
Bookmark button
Alert button
Aug 07, 2023
Longteng Zhang, Lin Zhang, Shaohuai Shi, Xiaowen Chu, Bo Li

Figure 1 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 2 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 3 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 4 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Viaarxiv icon

Eva: A General Vectorized Approximation Framework for Second-order Optimization

Add code
Bookmark button
Alert button
Aug 04, 2023
Lin Zhang, Shaohuai Shi, Bo Li

Figure 1 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 2 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 3 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 4 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Viaarxiv icon

Evaluation and Optimization of Gradient Compression for Distributed Deep Learning

Add code
Bookmark button
Alert button
Jun 15, 2023
Lin Zhang, Longteng Zhang, Shaohuai Shi, Xiaowen Chu, Bo Li

Figure 1 for Evaluation and Optimization of Gradient Compression for Distributed Deep Learning
Figure 2 for Evaluation and Optimization of Gradient Compression for Distributed Deep Learning
Figure 3 for Evaluation and Optimization of Gradient Compression for Distributed Deep Learning
Figure 4 for Evaluation and Optimization of Gradient Compression for Distributed Deep Learning
Viaarxiv icon

FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training

Add code
Bookmark button
Alert button
Mar 03, 2023
Zhenheng Tang, Xiaowen Chu, Ryan Yide Ran, Sunwoo Lee, Shaohuai Shi, Yonggang Zhang, Yuxin Wang, Alex Qiaozhong Liang, Salman Avestimehr, Chaoyang He

Figure 1 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 2 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 3 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 4 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Viaarxiv icon

Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning

Add code
Bookmark button
Alert button
Feb 24, 2023
Lin Zhang, Shaohuai Shi, Xiaowen Chu, Wei Wang, Bo Li, Chengjian Liu

Figure 1 for Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning
Figure 2 for Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning
Figure 3 for Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning
Figure 4 for Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning
Viaarxiv icon

An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning

Add code
Bookmark button
Alert button
Nov 30, 2022
Shaohuai Shi, Qing Yang, Yang Xiang, Shuhan Qi, Xuan Wang

Figure 1 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 2 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 3 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 4 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Viaarxiv icon

EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching

Add code
Bookmark button
Alert button
Jul 20, 2022
Qiang Wang, Shaohuai Shi, Kaiyong Zhao, Xiaowen Chu

Figure 1 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 2 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 3 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 4 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Viaarxiv icon