Alert button
Picture for Amir Gholami

Amir Gholami

Alert button

Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization

Add code
Bookmark button
Alert button
Oct 07, 2019
Paras Jain, Ajay Jain, Aniruddha Nrusimha, Amir Gholami, Pieter Abbeel, Kurt Keutzer, Ion Stoica, Joseph E. Gonzalez

Figure 1 for Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
Figure 2 for Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
Figure 3 for Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
Figure 4 for Checkmate: Breaking the Memory Wall with Optimal Tensor Rematerialization
Viaarxiv icon

Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT

Add code
Bookmark button
Alert button
Sep 25, 2019
Sheng Shen, Zhen Dong, Jiayu Ye, Linjian Ma, Zhewei Yao, Amir Gholami, Michael W. Mahoney, Kurt Keutzer

Figure 1 for Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Figure 2 for Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Figure 3 for Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Figure 4 for Q-BERT: Hessian Based Ultra Low Precision Quantization of BERT
Viaarxiv icon

ANODEV2: A Coupled Neural ODE Evolution Framework

Add code
Bookmark button
Alert button
Jun 10, 2019
Tianjun Zhang, Zhewei Yao, Amir Gholami, Kurt Keutzer, Joseph Gonzalez, George Biros, Michael Mahoney

Figure 1 for ANODEV2: A Coupled Neural ODE Evolution Framework
Figure 2 for ANODEV2: A Coupled Neural ODE Evolution Framework
Figure 3 for ANODEV2: A Coupled Neural ODE Evolution Framework
Figure 4 for ANODEV2: A Coupled Neural ODE Evolution Framework
Viaarxiv icon

HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision

Add code
Bookmark button
Alert button
Apr 29, 2019
Zhen Dong, Zhewei Yao, Amir Gholami, Michael Mahoney, Kurt Keutzer

Figure 1 for HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
Figure 2 for HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
Figure 3 for HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
Figure 4 for HAWQ: Hessian AWare Quantization of Neural Networks with Mixed-Precision
Viaarxiv icon

Inefficiency of K-FAC for Large Batch Size Training

Add code
Bookmark button
Alert button
Mar 14, 2019
Linjian Ma, Gabe Montague, Jiayu Ye, Zhewei Yao, Amir Gholami, Kurt Keutzer, Michael W. Mahoney

Figure 1 for Inefficiency of K-FAC for Large Batch Size Training
Figure 2 for Inefficiency of K-FAC for Large Batch Size Training
Figure 3 for Inefficiency of K-FAC for Large Batch Size Training
Figure 4 for Inefficiency of K-FAC for Large Batch Size Training
Viaarxiv icon

ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs

Add code
Bookmark button
Alert button
Feb 27, 2019
Amir Gholami, Kurt Keutzer, George Biros

Figure 1 for ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs
Figure 2 for ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs
Figure 3 for ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs
Figure 4 for ANODE: Unconditionally Accurate Memory-Efficient Gradients for Neural ODEs
Viaarxiv icon

Trust Region Based Adversarial Attack on Neural Networks

Add code
Bookmark button
Alert button
Dec 16, 2018
Zhewei Yao, Amir Gholami, Peng Xu, Kurt Keutzer, Michael Mahoney

Figure 1 for Trust Region Based Adversarial Attack on Neural Networks
Figure 2 for Trust Region Based Adversarial Attack on Neural Networks
Figure 3 for Trust Region Based Adversarial Attack on Neural Networks
Figure 4 for Trust Region Based Adversarial Attack on Neural Networks
Viaarxiv icon

Parameter Re-Initialization through Cyclical Batch Size Schedules

Add code
Bookmark button
Alert button
Dec 04, 2018
Norman Mu, Zhewei Yao, Amir Gholami, Kurt Keutzer, Michael Mahoney

Figure 1 for Parameter Re-Initialization through Cyclical Batch Size Schedules
Figure 2 for Parameter Re-Initialization through Cyclical Batch Size Schedules
Figure 3 for Parameter Re-Initialization through Cyclical Batch Size Schedules
Figure 4 for Parameter Re-Initialization through Cyclical Batch Size Schedules
Viaarxiv icon

On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent

Add code
Bookmark button
Alert button
Nov 30, 2018
Noah Golmant, Nikita Vemuri, Zhewei Yao, Vladimir Feinberg, Amir Gholami, Kai Rothauge, Michael W. Mahoney, Joseph Gonzalez

Figure 1 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 2 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 3 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Figure 4 for On the Computational Inefficiency of Large Batch Sizes for Stochastic Gradient Descent
Viaarxiv icon