Alert button
Picture for Kailash Gopalakrishnan

Kailash Gopalakrishnan

Alert button

Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization

Add code
Bookmark button
Alert button
Jun 16, 2022
Andrea Fasoli, Chia-Yu Chen, Mauricio Serrano, Swagath Venkataramani, George Saon, Xiaodong Cui, Brian Kingsbury, Kailash Gopalakrishnan

Figure 1 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Figure 2 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Figure 3 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Viaarxiv icon

4-bit Quantization of LSTM-based Speech Recognition Models

Add code
Bookmark button
Alert button
Aug 27, 2021
Andrea Fasoli, Chia-Yu Chen, Mauricio Serrano, Xiao Sun, Naigang Wang, Swagath Venkataramani, George Saon, Xiaodong Cui, Brian Kingsbury, Wei Zhang, Zoltán Tüske, Kailash Gopalakrishnan

Figure 1 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 2 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 3 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 4 for 4-bit Quantization of LSTM-based Speech Recognition Models
Viaarxiv icon

ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training

Add code
Bookmark button
Alert button
Apr 21, 2021
Chia-Yu Chen, Jiamin Ni, Songtao Lu, Xiaodong Cui, Pin-Yu Chen, Xiao Sun, Naigang Wang, Swagath Venkataramani, Vijayalakshmi Srinivasan, Wei Zhang, Kailash Gopalakrishnan

Figure 1 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 2 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 3 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 4 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Viaarxiv icon

FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training

Add code
Bookmark button
Alert button
Dec 24, 2020
Yonggan Fu, Haoran You, Yang Zhao, Yue Wang, Chaojian Li, Kailash Gopalakrishnan, Zhangyang Wang, Yingyan Lin

Figure 1 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 2 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 3 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 4 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Viaarxiv icon

Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks

Add code
Bookmark button
Alert button
Jan 19, 2019
Charbel Sakr, Naigang Wang, Chia-Yu Chen, Jungwook Choi, Ankur Agrawal, Naresh Shanbhag, Kailash Gopalakrishnan

Figure 1 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 2 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 3 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 4 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Viaarxiv icon

Training Deep Neural Networks with 8-bit Floating Point Numbers

Add code
Bookmark button
Alert button
Dec 19, 2018
Naigang Wang, Jungwook Choi, Daniel Brand, Chia-Yu Chen, Kailash Gopalakrishnan

Figure 1 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 2 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 3 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 4 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Viaarxiv icon

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)

Add code
Bookmark button
Alert button
Jul 17, 2018
Jungwook Choi, Pierce I-Jen Chuang, Zhuo Wang, Swagath Venkataramani, Vijayalakshmi Srinivasan, Kailash Gopalakrishnan

Figure 1 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Figure 2 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Figure 3 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Figure 4 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks (QNN)
Viaarxiv icon

PACT: Parameterized Clipping Activation for Quantized Neural Networks

Add code
Bookmark button
Alert button
Jul 17, 2018
Jungwook Choi, Zhuo Wang, Swagath Venkataramani, Pierce I-Jen Chuang, Vijayalakshmi Srinivasan, Kailash Gopalakrishnan

Figure 1 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 2 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 3 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 4 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Viaarxiv icon

AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training

Add code
Bookmark button
Alert button
Dec 07, 2017
Chia-Yu Chen, Jungwook Choi, Daniel Brand, Ankur Agrawal, Wei Zhang, Kailash Gopalakrishnan

Figure 1 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 2 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 3 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 4 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Viaarxiv icon

Deep Learning with Limited Numerical Precision

Add code
Bookmark button
Alert button
Feb 09, 2015
Suyog Gupta, Ankur Agrawal, Kailash Gopalakrishnan, Pritish Narayanan

Figure 1 for Deep Learning with Limited Numerical Precision
Figure 2 for Deep Learning with Limited Numerical Precision
Figure 3 for Deep Learning with Limited Numerical Precision
Figure 4 for Deep Learning with Limited Numerical Precision
Viaarxiv icon