Alert button
Picture for Sangkug Lym

Sangkug Lym

Alert button

Reducing Activation Recomputation in Large Transformer Models

Add code
Bookmark button
Alert button
May 10, 2022
Vijay Korthikanti, Jared Casper, Sangkug Lym, Lawrence McAfee, Michael Andersch, Mohammad Shoeybi, Bryan Catanzaro

Figure 1 for Reducing Activation Recomputation in Large Transformer Models
Figure 2 for Reducing Activation Recomputation in Large Transformer Models
Figure 3 for Reducing Activation Recomputation in Large Transformer Models
Figure 4 for Reducing Activation Recomputation in Large Transformer Models
Viaarxiv icon

FlexSA: Flexible Systolic Array Architecture for Efficient Pruned DNN Model Training

Add code
Bookmark button
Alert button
Apr 27, 2020
Sangkug Lym, Mattan Erez

Figure 1 for FlexSA: Flexible Systolic Array Architecture for Efficient Pruned DNN Model Training
Figure 2 for FlexSA: Flexible Systolic Array Architecture for Efficient Pruned DNN Model Training
Figure 3 for FlexSA: Flexible Systolic Array Architecture for Efficient Pruned DNN Model Training
Figure 4 for FlexSA: Flexible Systolic Array Architecture for Efficient Pruned DNN Model Training
Viaarxiv icon

DeLTA: GPU Performance Model for Deep Learning Applications with In-depth Memory System Traffic Analysis

Add code
Bookmark button
Alert button
Apr 02, 2019
Sangkug Lym, Donghyuk Lee, Mike O'Connor, Niladrish Chatterjee, Mattan Erez

Figure 1 for DeLTA: GPU Performance Model for Deep Learning Applications with In-depth Memory System Traffic Analysis
Figure 2 for DeLTA: GPU Performance Model for Deep Learning Applications with In-depth Memory System Traffic Analysis
Figure 3 for DeLTA: GPU Performance Model for Deep Learning Applications with In-depth Memory System Traffic Analysis
Figure 4 for DeLTA: GPU Performance Model for Deep Learning Applications with In-depth Memory System Traffic Analysis
Viaarxiv icon

PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training

Add code
Bookmark button
Alert button
Jan 26, 2019
Sangkug Lym, Esha Choukse, Siavash Zangeneh, Wei Wen, Mattan Erez, Sujay Shanghavi

Figure 1 for PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training
Figure 2 for PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training
Figure 3 for PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training
Figure 4 for PruneTrain: Gradual Structured Pruning from Scratch for Faster Neural Network Training
Viaarxiv icon

Mini-batch Serialization: CNN Training with Inter-layer Data Reuse

Add code
Bookmark button
Alert button
Sep 30, 2018
Sangkug Lym, Armand Behroozi, Wei Wen, Ge Li, Yongkee Kwon, Mattan Erez

Figure 1 for Mini-batch Serialization: CNN Training with Inter-layer Data Reuse
Figure 2 for Mini-batch Serialization: CNN Training with Inter-layer Data Reuse
Figure 3 for Mini-batch Serialization: CNN Training with Inter-layer Data Reuse
Figure 4 for Mini-batch Serialization: CNN Training with Inter-layer Data Reuse
Viaarxiv icon