Picture for Hongsun Jang

Hongsun Jang

Pipette: Automatic Fine-grained Large Language Model Training Configurator for Real-World Clusters

Add code
May 28, 2024
Figure 1 for Pipette: Automatic Fine-grained Large Language Model Training Configurator for Real-World Clusters
Figure 2 for Pipette: Automatic Fine-grained Large Language Model Training Configurator for Real-World Clusters
Figure 3 for Pipette: Automatic Fine-grained Large Language Model Training Configurator for Real-World Clusters
Figure 4 for Pipette: Automatic Fine-grained Large Language Model Training Configurator for Real-World Clusters
Viaarxiv icon

Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System

Add code
Mar 11, 2024
Figure 1 for Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System
Figure 2 for Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System
Figure 3 for Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System
Figure 4 for Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System
Viaarxiv icon

PeerAiD: Improving Adversarial Distillation from a Specialized Peer Tutor

Add code
Mar 11, 2024
Viaarxiv icon

GraNNDis: Efficient Unified Distributed Training Framework for Deep GNNs on Large Clusters

Add code
Nov 12, 2023
Figure 1 for GraNNDis: Efficient Unified Distributed Training Framework for Deep GNNs on Large Clusters
Figure 2 for GraNNDis: Efficient Unified Distributed Training Framework for Deep GNNs on Large Clusters
Figure 3 for GraNNDis: Efficient Unified Distributed Training Framework for Deep GNNs on Large Clusters
Figure 4 for GraNNDis: Efficient Unified Distributed Training Framework for Deep GNNs on Large Clusters
Viaarxiv icon

Pipe-BD: Pipelined Parallel Blockwise Distillation

Add code
Jan 29, 2023
Viaarxiv icon

Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression

Add code
Jan 24, 2023
Figure 1 for Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
Figure 2 for Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
Figure 3 for Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
Figure 4 for Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
Viaarxiv icon