Picture for Yangrui Chen

Yangrui Chen

The University of Hong Kong, ByteDance

Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation

Add code
Aug 07, 2024
Viaarxiv icon

MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs

Add code
Feb 23, 2024
Figure 1 for MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs
Figure 2 for MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs
Figure 3 for MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs
Figure 4 for MegaScale: Scaling Large Language Model Training to More Than 10,000 GPUs
Viaarxiv icon

BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing

Add code
Dec 16, 2021
Figure 1 for BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Figure 2 for BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Figure 3 for BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Figure 4 for BGL: GPU-Efficient GNN Training by Optimizing Graph Data I/O and Preprocessing
Viaarxiv icon

DL2: A Deep Learning-driven Scheduler for Deep Learning Clusters

Add code
Sep 13, 2019
Figure 1 for DL2: A Deep Learning-driven Scheduler for Deep Learning Clusters
Figure 2 for DL2: A Deep Learning-driven Scheduler for Deep Learning Clusters
Figure 3 for DL2: A Deep Learning-driven Scheduler for Deep Learning Clusters
Figure 4 for DL2: A Deep Learning-driven Scheduler for Deep Learning Clusters
Viaarxiv icon