Picture for Naigang Wang

Naigang Wang

MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization

Add code
Jun 02, 2024
Figure 1 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 2 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 3 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 4 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Viaarxiv icon

A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts

Add code
May 28, 2024
Viaarxiv icon

Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization

Add code
Apr 04, 2024
Figure 1 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 2 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 3 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 4 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Viaarxiv icon

COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization

Add code
Mar 11, 2024
Figure 1 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 2 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 3 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 4 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Viaarxiv icon

4-bit Quantization of LSTM-based Speech Recognition Models

Add code
Aug 27, 2021
Figure 1 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 2 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 3 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 4 for 4-bit Quantization of LSTM-based Speech Recognition Models
Viaarxiv icon

ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training

Add code
Apr 21, 2021
Figure 1 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 2 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 3 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Figure 4 for ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
Viaarxiv icon

All at Once Network Quantization via Collaborative Knowledge Transfer

Add code
Mar 02, 2021
Figure 1 for All at Once Network Quantization via Collaborative Knowledge Transfer
Figure 2 for All at Once Network Quantization via Collaborative Knowledge Transfer
Figure 3 for All at Once Network Quantization via Collaborative Knowledge Transfer
Figure 4 for All at Once Network Quantization via Collaborative Knowledge Transfer
Viaarxiv icon

A Comprehensive Survey on Hardware-Aware Neural Architecture Search

Add code
Jan 22, 2021
Figure 1 for A Comprehensive Survey on Hardware-Aware Neural Architecture Search
Figure 2 for A Comprehensive Survey on Hardware-Aware Neural Architecture Search
Figure 3 for A Comprehensive Survey on Hardware-Aware Neural Architecture Search
Figure 4 for A Comprehensive Survey on Hardware-Aware Neural Architecture Search
Viaarxiv icon

Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks

Add code
Jan 19, 2019
Figure 1 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 2 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 3 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Figure 4 for Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks
Viaarxiv icon

Training Deep Neural Networks with 8-bit Floating Point Numbers

Add code
Dec 19, 2018
Figure 1 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 2 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 3 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Figure 4 for Training Deep Neural Networks with 8-bit Floating Point Numbers
Viaarxiv icon