Picture for Vahid Partovi Nia

Vahid Partovi Nia

Huawei Noah's Ark Lab

On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats

Add code
Jan 09, 2023
Figure 1 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 2 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 3 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Figure 4 for On the Convergence of Stochastic Gradient Descent in Low-precision Number Formats
Viaarxiv icon

Training Integer-Only Deep Recurrent Neural Networks

Add code
Dec 22, 2022
Viaarxiv icon

EuclidNets: An Alternative Operation for Efficient Inference of Deep Learning Models

Add code
Dec 22, 2022
Viaarxiv icon

KronA: Parameter Efficient Tuning with Kronecker Adapter

Add code
Dec 20, 2022
Figure 1 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 2 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 3 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Figure 4 for KronA: Parameter Efficient Tuning with Kronecker Adapter
Viaarxiv icon

SeKron: A Decomposition Method Supporting Many Factorization Structures

Add code
Oct 12, 2022
Figure 1 for SeKron: A Decomposition Method Supporting Many Factorization Structures
Figure 2 for SeKron: A Decomposition Method Supporting Many Factorization Structures
Figure 3 for SeKron: A Decomposition Method Supporting Many Factorization Structures
Figure 4 for SeKron: A Decomposition Method Supporting Many Factorization Structures
Viaarxiv icon

Integer Fine-tuning of Transformer-based Models

Add code
Sep 20, 2022
Figure 1 for Integer Fine-tuning of Transformer-based Models
Figure 2 for Integer Fine-tuning of Transformer-based Models
Figure 3 for Integer Fine-tuning of Transformer-based Models
Figure 4 for Integer Fine-tuning of Transformer-based Models
Viaarxiv icon

DenseShift: Towards Accurate and Transferable Low-Bit Shift Network

Add code
Aug 20, 2022
Figure 1 for DenseShift: Towards Accurate and Transferable Low-Bit Shift Network
Figure 2 for DenseShift: Towards Accurate and Transferable Low-Bit Shift Network
Figure 3 for DenseShift: Towards Accurate and Transferable Low-Bit Shift Network
Figure 4 for DenseShift: Towards Accurate and Transferable Low-Bit Shift Network
Viaarxiv icon

Is Integer Arithmetic Enough for Deep Learning Training?

Add code
Jul 18, 2022
Figure 1 for Is Integer Arithmetic Enough for Deep Learning Training?
Figure 2 for Is Integer Arithmetic Enough for Deep Learning Training?
Figure 3 for Is Integer Arithmetic Enough for Deep Learning Training?
Figure 4 for Is Integer Arithmetic Enough for Deep Learning Training?
Viaarxiv icon

Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks

Add code
Feb 18, 2022
Figure 1 for Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks
Figure 2 for Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks
Figure 3 for Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks
Figure 4 for Rethinking Pareto Frontier for Performance Evaluation of Deep Neural Networks
Viaarxiv icon

Demystifying and Generalizing BinaryConnect

Add code
Oct 25, 2021
Figure 1 for Demystifying and Generalizing BinaryConnect
Figure 2 for Demystifying and Generalizing BinaryConnect
Figure 3 for Demystifying and Generalizing BinaryConnect
Figure 4 for Demystifying and Generalizing BinaryConnect
Viaarxiv icon