Alert button
Picture for Vikram Saletore

Vikram Saletore

Alert button

Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case

Add code
Bookmark button
Alert button
Mar 18, 2021
Florian Rehm, Sofia Vallecorsa, Vikram Saletore, Hans Pabst, Adel Chaibi, Valeriu Codreanu, Kerstin Borras, Dirk Krücker

Figure 1 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 2 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 3 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 4 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Viaarxiv icon

Training Multiscale-CNN for Large Microscopy Image Classification in One Hour

Add code
Bookmark button
Alert button
Oct 03, 2019
Kushal Datta, Imtiaz Hossain, Sun Choi, Vikram Saletore, Kyle Ambert, William J. Godinez, Xian Zhang

Figure 1 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 2 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 3 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 4 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Viaarxiv icon

Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model

Add code
Bookmark button
Alert button
Jun 07, 2019
Aishwarya Bhandare, Vamsi Sripathi, Deepthi Karkada, Vivek Menon, Sun Choi, Kushal Datta, Vikram Saletore

Figure 1 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 2 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 3 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 4 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Viaarxiv icon

Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models

Add code
Bookmark button
Alert button
May 10, 2019
Derya Cavdar, Valeriu Codreanu, Can Karakus, John A. Lockman III, Damian Podareanu, Vikram Saletore, Alexander Sergeev, Don D. Smith II, Victor Suthichai, Quy Ta, Srinivas Varadharajan, Lucas A. Wilson, Rengan Xu, Pei Yang

Figure 1 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 2 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 3 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 4 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Viaarxiv icon

Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train

Add code
Bookmark button
Alert button
Nov 15, 2017
Valeriu Codreanu, Damian Podareanu, Vikram Saletore

Figure 1 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 2 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 3 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 4 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Viaarxiv icon