Picture for Naigang Wang

Naigang Wang

Is Finer Better? The Limits of Microscaling Formats in Large Language Models

Add code
Jan 26, 2026
Viaarxiv icon

Compressed Decentralized Momentum Stochastic Gradient Methods for Nonconvex Optimization

Add code
Aug 07, 2025
Viaarxiv icon

CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization

Add code
Jan 30, 2025
Figure 1 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 2 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 3 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 4 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Viaarxiv icon

Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing

Add code
Oct 09, 2024
Figure 1 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 2 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 3 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 4 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Viaarxiv icon

Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging

Add code
Oct 01, 2024
Figure 1 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 2 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 3 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 4 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Viaarxiv icon

MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization

Add code
Jun 02, 2024
Figure 1 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 2 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 3 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 4 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Viaarxiv icon

A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts

Add code
May 28, 2024
Figure 1 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 2 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 3 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 4 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Viaarxiv icon

Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization

Add code
Apr 04, 2024
Figure 1 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 2 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 3 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 4 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Viaarxiv icon

COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization

Add code
Mar 11, 2024
Figure 1 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 2 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 3 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Figure 4 for COMQ: A Backpropagation-Free Algorithm for Post-Training Quantization
Viaarxiv icon

4-bit Quantization of LSTM-based Speech Recognition Models

Add code
Aug 27, 2021
Figure 1 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 2 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 3 for 4-bit Quantization of LSTM-based Speech Recognition Models
Figure 4 for 4-bit Quantization of LSTM-based Speech Recognition Models
Viaarxiv icon