Picture for Naigang Wang

Naigang Wang

LoDAdaC: a unified local training-based decentralized framework with adaptive gradients and compressed communication

Add code
Apr 11, 2026
Viaarxiv icon

Efficient Quantization of Mixture-of-Experts with Theoretical Generalization Guarantees

Add code
Apr 07, 2026
Viaarxiv icon

Is Finer Better? The Limits of Microscaling Formats in Large Language Models

Add code
Jan 26, 2026
Viaarxiv icon

Compressed Decentralized Momentum Stochastic Gradient Methods for Nonconvex Optimization

Add code
Aug 07, 2025
Viaarxiv icon

CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization

Add code
Jan 30, 2025
Figure 1 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 2 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 3 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Figure 4 for CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization
Viaarxiv icon

Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing

Add code
Oct 09, 2024
Figure 1 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 2 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 3 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Figure 4 for Unlocking Real-Time Fluorescence Lifetime Imaging: Multi-Pixel Parallelism for FPGA-Accelerated Processing
Viaarxiv icon

Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging

Add code
Oct 01, 2024
Figure 1 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 2 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 3 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Figure 4 for Compressing Recurrent Neural Networks for FPGA-accelerated Implementation in Fluorescence Lifetime Imaging
Viaarxiv icon

MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization

Add code
Jun 02, 2024
Figure 1 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 2 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 3 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Figure 4 for MagR: Weight Magnitude Reduction for Enhancing Post-Training Quantization
Viaarxiv icon

A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts

Add code
May 28, 2024
Figure 1 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 2 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 3 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Figure 4 for A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Viaarxiv icon

Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization

Add code
Apr 04, 2024
Figure 1 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 2 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 3 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 4 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Viaarxiv icon