Picture for Ritchie Zhao

Ritchie Zhao

Microscaling Data Formats for Deep Learning

Add code
Oct 19, 2023
Figure 1 for Microscaling Data Formats for Deep Learning
Figure 2 for Microscaling Data Formats for Deep Learning
Figure 3 for Microscaling Data Formats for Deep Learning
Figure 4 for Microscaling Data Formats for Deep Learning
Viaarxiv icon

Shared Microexponents: A Little Shifting Goes a Long Way

Add code
Feb 16, 2023
Figure 1 for Shared Microexponents: A Little Shifting Goes a Long Way
Figure 2 for Shared Microexponents: A Little Shifting Goes a Long Way
Figure 3 for Shared Microexponents: A Little Shifting Goes a Long Way
Figure 4 for Shared Microexponents: A Little Shifting Goes a Long Way
Viaarxiv icon

Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations

Add code
Feb 17, 2020
Figure 1 for Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Figure 2 for Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Figure 3 for Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Figure 4 for Precision Gating: Improving Neural Network Efficiency with Dynamic Dual-Precision Activations
Viaarxiv icon

Overwrite Quantization: Opportunistic Outlier Handling for Neural Network Accelerators

Add code
Oct 13, 2019
Figure 1 for Overwrite Quantization: Opportunistic Outlier Handling for Neural Network Accelerators
Figure 2 for Overwrite Quantization: Opportunistic Outlier Handling for Neural Network Accelerators
Figure 3 for Overwrite Quantization: Opportunistic Outlier Handling for Neural Network Accelerators
Figure 4 for Overwrite Quantization: Opportunistic Outlier Handling for Neural Network Accelerators
Viaarxiv icon

Improving Neural Network Quantization without Retraining using Outlier Channel Splitting

Add code
Jan 30, 2019
Figure 1 for Improving Neural Network Quantization without Retraining using Outlier Channel Splitting
Figure 2 for Improving Neural Network Quantization without Retraining using Outlier Channel Splitting
Figure 3 for Improving Neural Network Quantization without Retraining using Outlier Channel Splitting
Figure 4 for Improving Neural Network Quantization without Retraining using Outlier Channel Splitting
Viaarxiv icon

Building Efficient Deep Neural Networks with Unitary Group Convolutions

Add code
Nov 19, 2018
Figure 1 for Building Efficient Deep Neural Networks with Unitary Group Convolutions
Figure 2 for Building Efficient Deep Neural Networks with Unitary Group Convolutions
Figure 3 for Building Efficient Deep Neural Networks with Unitary Group Convolutions
Figure 4 for Building Efficient Deep Neural Networks with Unitary Group Convolutions
Viaarxiv icon

Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration

Add code
Jul 15, 2017
Figure 1 for Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
Figure 2 for Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
Figure 3 for Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
Figure 4 for Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
Viaarxiv icon