Picture for Shaohuai Shi

Shaohuai Shi

LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning

Add code
Aug 07, 2023
Figure 1 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 2 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 3 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Figure 4 for LoRA-FA: Memory-efficient Low-rank Adaptation for Large Language Models Fine-tuning
Viaarxiv icon

Eva: A General Vectorized Approximation Framework for Second-order Optimization

Add code
Aug 04, 2023
Figure 1 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 2 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 3 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Figure 4 for Eva: A General Vectorized Approximation Framework for Second-order Optimization
Viaarxiv icon

Evaluation and Optimization of Gradient Compression for Distributed Deep Learning

Add code
Jun 15, 2023
Viaarxiv icon

FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training

Add code
Mar 03, 2023
Figure 1 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 2 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 3 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Figure 4 for FedML Parrot: A Scalable Federated Learning System via Heterogeneity-aware Scheduling on Sequential and Hierarchical Training
Viaarxiv icon

Decoupling the All-Reduce Primitive for Accelerating Distributed Deep Learning

Add code
Feb 24, 2023
Viaarxiv icon

An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning

Add code
Nov 30, 2022
Figure 1 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 2 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 3 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Figure 4 for An Efficient Split Fine-tuning Framework for Edge and Cloud Collaborative Learning
Viaarxiv icon

EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching

Add code
Jul 20, 2022
Figure 1 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 2 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 3 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Figure 4 for EASNet: Searching Elastic and Accurate Network Architecture for Stereo Matching
Viaarxiv icon

Scalable K-FAC Training for Deep Neural Networks with Distributed Preconditioning

Add code
Jun 30, 2022
Figure 1 for Scalable K-FAC Training for Deep Neural Networks with Distributed Preconditioning
Figure 2 for Scalable K-FAC Training for Deep Neural Networks with Distributed Preconditioning
Figure 3 for Scalable K-FAC Training for Deep Neural Networks with Distributed Preconditioning
Figure 4 for Scalable K-FAC Training for Deep Neural Networks with Distributed Preconditioning
Viaarxiv icon

Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning

Add code
Jun 06, 2022
Figure 1 for Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning
Figure 2 for Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning
Figure 3 for Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning
Figure 4 for Virtual Homogeneity Learning: Defending against Data Heterogeneity in Federated Learning
Viaarxiv icon

Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters

Add code
May 19, 2022
Figure 1 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 2 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 3 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Figure 4 for Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters
Viaarxiv icon