Picture for Yaohui Cai

Yaohui Cai

Trainable Fixed-Point Quantization for Deep Learning Acceleration on FPGAs

Add code
Jan 31, 2024
Viaarxiv icon

Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference

Add code
Dec 23, 2023
Figure 1 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 2 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 3 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Figure 4 for Understanding the Potential of FPGA-Based Spatial Acceleration for Large Language Model Inference
Viaarxiv icon

QuIP: 2-Bit Quantization of Large Language Models With Guarantees

Add code
Jul 25, 2023
Figure 1 for QuIP: 2-Bit Quantization of Large Language Models With Guarantees
Figure 2 for QuIP: 2-Bit Quantization of Large Language Models With Guarantees
Figure 3 for QuIP: 2-Bit Quantization of Large Language Models With Guarantees
Figure 4 for QuIP: 2-Bit Quantization of Large Language Models With Guarantees
Viaarxiv icon

Structured Pruning is All You Need for Pruning CNNs at Initialization

Add code
Mar 04, 2022
Figure 1 for Structured Pruning is All You Need for Pruning CNNs at Initialization
Figure 2 for Structured Pruning is All You Need for Pruning CNNs at Initialization
Figure 3 for Structured Pruning is All You Need for Pruning CNNs at Initialization
Figure 4 for Structured Pruning is All You Need for Pruning CNNs at Initialization
Viaarxiv icon

SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation

Add code
Feb 07, 2021
Figure 1 for SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation
Figure 2 for SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation
Figure 3 for SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation
Figure 4 for SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation
Viaarxiv icon

CoDeNet: Algorithm-hardware Co-design for Deformable Convolution

Add code
Jun 12, 2020
Figure 1 for CoDeNet: Algorithm-hardware Co-design for Deformable Convolution
Figure 2 for CoDeNet: Algorithm-hardware Co-design for Deformable Convolution
Figure 3 for CoDeNet: Algorithm-hardware Co-design for Deformable Convolution
Figure 4 for CoDeNet: Algorithm-hardware Co-design for Deformable Convolution
Viaarxiv icon

Algorithm-hardware Co-design for Deformable Convolution

Add code
Feb 19, 2020
Figure 1 for Algorithm-hardware Co-design for Deformable Convolution
Figure 2 for Algorithm-hardware Co-design for Deformable Convolution
Viaarxiv icon

ZeroQ: A Novel Zero Shot Quantization Framework

Add code
Jan 01, 2020
Figure 1 for ZeroQ: A Novel Zero Shot Quantization Framework
Figure 2 for ZeroQ: A Novel Zero Shot Quantization Framework
Figure 3 for ZeroQ: A Novel Zero Shot Quantization Framework
Figure 4 for ZeroQ: A Novel Zero Shot Quantization Framework
Viaarxiv icon

HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks

Add code
Nov 10, 2019
Figure 1 for HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Figure 2 for HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Figure 3 for HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Figure 4 for HAWQ-V2: Hessian Aware trace-Weighted Quantization of Neural Networks
Viaarxiv icon