Picture for Nikoli Dryden

Nikoli Dryden

Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks

Add code
Jan 31, 2021
Figure 1 for Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
Figure 2 for Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
Figure 3 for Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
Figure 4 for Sparsity in Deep Learning: Pruning and growth for efficient inference and training in neural networks
Viaarxiv icon

Clairvoyant Prefetching for Distributed Machine Learning I/O

Add code
Jan 21, 2021
Figure 1 for Clairvoyant Prefetching for Distributed Machine Learning I/O
Figure 2 for Clairvoyant Prefetching for Distributed Machine Learning I/O
Figure 3 for Clairvoyant Prefetching for Distributed Machine Learning I/O
Figure 4 for Clairvoyant Prefetching for Distributed Machine Learning I/O
Viaarxiv icon

The Case for Strong Scaling in Deep Learning: Training Large 3D CNNs with Hybrid Parallelism

Add code
Jul 25, 2020
Viaarxiv icon

Data Movement Is All You Need: A Case Study on Optimizing Transformers

Add code
Jul 02, 2020
Figure 1 for Data Movement Is All You Need: A Case Study on Optimizing Transformers
Figure 2 for Data Movement Is All You Need: A Case Study on Optimizing Transformers
Figure 3 for Data Movement Is All You Need: A Case Study on Optimizing Transformers
Figure 4 for Data Movement Is All You Need: A Case Study on Optimizing Transformers
Viaarxiv icon

Shapeshifter Networks: Cross-layer Parameter Sharing for Scalable and Effective Deep Learning

Add code
Jun 18, 2020
Figure 1 for Shapeshifter Networks: Cross-layer Parameter Sharing for Scalable and Effective Deep Learning
Figure 2 for Shapeshifter Networks: Cross-layer Parameter Sharing for Scalable and Effective Deep Learning
Figure 3 for Shapeshifter Networks: Cross-layer Parameter Sharing for Scalable and Effective Deep Learning
Figure 4 for Shapeshifter Networks: Cross-layer Parameter Sharing for Scalable and Effective Deep Learning
Viaarxiv icon

Deep Learning for Post-Processing Ensemble Weather Forecasts

Add code
May 18, 2020
Figure 1 for Deep Learning for Post-Processing Ensemble Weather Forecasts
Figure 2 for Deep Learning for Post-Processing Ensemble Weather Forecasts
Figure 3 for Deep Learning for Post-Processing Ensemble Weather Forecasts
Figure 4 for Deep Learning for Post-Processing Ensemble Weather Forecasts
Viaarxiv icon

Breaking (Global) Barriers in Parallel Stochastic Optimization with Wait-Avoiding Group Averaging

Add code
Apr 30, 2020
Figure 1 for Breaking (Global) Barriers in Parallel Stochastic Optimization with Wait-Avoiding Group Averaging
Figure 2 for Breaking (Global) Barriers in Parallel Stochastic Optimization with Wait-Avoiding Group Averaging
Figure 3 for Breaking (Global) Barriers in Parallel Stochastic Optimization with Wait-Avoiding Group Averaging
Figure 4 for Breaking (Global) Barriers in Parallel Stochastic Optimization with Wait-Avoiding Group Averaging
Viaarxiv icon

Predicting Weather Uncertainty with Deep Convnets

Add code
Dec 04, 2019
Figure 1 for Predicting Weather Uncertainty with Deep Convnets
Figure 2 for Predicting Weather Uncertainty with Deep Convnets
Figure 3 for Predicting Weather Uncertainty with Deep Convnets
Figure 4 for Predicting Weather Uncertainty with Deep Convnets
Viaarxiv icon

Improving Strong-Scaling of CNN Training by Exploiting Finer-Grained Parallelism

Add code
Mar 15, 2019
Figure 1 for Improving Strong-Scaling of CNN Training by Exploiting Finer-Grained Parallelism
Figure 2 for Improving Strong-Scaling of CNN Training by Exploiting Finer-Grained Parallelism
Figure 3 for Improving Strong-Scaling of CNN Training by Exploiting Finer-Grained Parallelism
Figure 4 for Improving Strong-Scaling of CNN Training by Exploiting Finer-Grained Parallelism
Viaarxiv icon