Picture for Kailash Gopalakrishnan

Kailash Gopalakrishnan

Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization

Add code
Jun 16, 2022
Figure 1 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Figure 2 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Figure 3 for Accelerating Inference and Language Model Fusion of Recurrent Neural Network Transducers via End-to-End 4-bit Quantization
Viaarxiv icon

4-bit Quantization of LSTM-based Speech Recognition Models

Add code
Aug 27, 2021
Figure 1 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 2 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 3 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 4 for 4-bit Quantization of LSTM-based Speech Recognition Models
Viaarxiv icon

ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training

Add code
Apr 21, 2021
Figure 1 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 2 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 3 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 4 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Viaarxiv icon

FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training

Add code
Dec 24, 2020
Figure 1 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 2 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 3 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Figure 4 for FracTrain: Fractionally Squeezing Bit Savings Both Temporally and Spatially for Efficient DNN Training
Viaarxiv icon

Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks

Add code
Jan 19, 2019
Figure 1 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 2 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 3 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 4 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Viaarxiv icon

Training Deep Neural Networks with 8-bit Floating Point Numbers

Add code
Dec 19, 2018
Figure 1 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 2 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 3 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 4 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Viaarxiv icon

Bridging the Accuracy Gap for 2-bit Quantized Neural Networks

Add code
Jul 17, 2018
Figure 1 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks
Figure 2 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks
Figure 3 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks
Figure 4 for Bridging the Accuracy Gap for 2-bit Quantized Neural Networks
Viaarxiv icon

PACT: Parameterized Clipping Activation for Quantized Neural Networks

Add code
Jul 17, 2018
Figure 1 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 2 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 3 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Figure 4 for PACT: Parameterized Clipping Activation for Quantized Neural Networks
Viaarxiv icon

AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training

Add code
Dec 07, 2017
Figure 1 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 2 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 3 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Figure 4 for AdaComp : Adaptive Residual Gradient Compression for Data-Parallel Distributed Training
Viaarxiv icon

Deep Learning with Limited Numerical Precision

Add code
Feb 09, 2015
Figure 1 for Deep Learning with Limited Numerical Precision
Figure 2 for Deep Learning with Limited Numerical Precision
Figure 3 for Deep Learning with Limited Numerical Precision
Figure 4 for Deep Learning with Limited Numerical Precision
Viaarxiv icon