Picture for Vahid Partovi Nia

Vahid Partovi Nia

Huawei Noah's Ark Lab

OAC: Output-adaptive Calibration for Accurate Post-training Quantization

Add code
May 23, 2024
Figure 1 for OAC: Output-adaptive Calibration for Accurate Post-training Quantization
Figure 2 for OAC: Output-adaptive Calibration for Accurate Post-training Quantization
Figure 3 for OAC: Output-adaptive Calibration for Accurate Post-training Quantization
Figure 4 for OAC: Output-adaptive Calibration for Accurate Post-training Quantization
Viaarxiv icon

AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs

Add code
May 22, 2024
Figure 1 for AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs
Figure 2 for AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs
Figure 3 for AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs
Figure 4 for AdpQ: A Zero-shot Calibration Free Adaptive Post Training Quantization Method for LLMs
Viaarxiv icon

Understanding Neural Network Binarization with Forward and Backward Proximal Quantizers

Add code
Feb 27, 2024
Viaarxiv icon

Mitigating Outlier Activations in Low-Precision Fine-Tuning of Language Models

Add code
Dec 15, 2023
Viaarxiv icon

Mathematical Challenges in Deep Learning

Add code
Mar 24, 2023
Figure 1 for Mathematical Challenges in Deep Learning
Figure 2 for Mathematical Challenges in Deep Learning
Figure 3 for Mathematical Challenges in Deep Learning
Figure 4 for Mathematical Challenges in Deep Learning
Viaarxiv icon

Scaling Deep Networks with the Mesh Adaptive Direct Search algorithm

Add code
Jan 17, 2023
Figure 1 for Scaling Deep Networks with the Mesh Adaptive Direct Search algorithm
Figure 2 for Scaling Deep Networks with the Mesh Adaptive Direct Search algorithm
Figure 3 for Scaling Deep Networks with the Mesh Adaptive Direct Search algorithm
Viaarxiv icon

On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats

Add code
Jan 09, 2023
Figure 1 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 2 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 3 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 4 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Viaarxiv icon

Training Integer-Only Deep Recurrent Neural Networks

Add code
Dec 22, 2022
Figure 1 for Training Integer-Only Deep Recurrent Neural Networks
Figure 2 for Training Integer-Only Deep Recurrent Neural Networks
Figure 3 for Training Integer-Only Deep Recurrent Neural Networks
Figure 4 for Training Integer-Only Deep Recurrent Neural Networks
Viaarxiv icon

EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models

Add code
Dec 22, 2022
Figure 1 for EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models
Figure 2 for EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models
Figure 3 for EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models
Figure 4 for EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models
Viaarxiv icon

KronA: Parameter Efficient Tuning with Kronecker Adapter

Add code
Dec 20, 2022
Figure 1 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 2 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 3 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 4 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Viaarxiv icon