Alert button
Picture for Patrick Judd

Patrick Judd

Alert button

FP8 Formats for Deep Learning

Add code
Bookmark button
Alert button
Sep 12, 2022
Paulius Micikevicius, Dusan Stosic, Neil Burgess, Marius Cornea, Pradeep Dubey, Richard Grisenthwaite, Sangwon Ha, Alexander Heinecke, Patrick Judd, John Kamalu, Naveen Mellempudi, Stuart Oberman, Mohammad Shoeybi, Michael Siu, Hao Wu

Figure 1 for FP8 Formats for Deep Learning
Figure 2 for FP8 Formats for Deep Learning
Figure 3 for FP8 Formats for Deep Learning
Figure 4 for FP8 Formats for Deep Learning
Viaarxiv icon

Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation

Add code
Bookmark button
Alert button
Apr 20, 2020
Hao Wu, Patrick Judd, Xiaojie Zhang, Mikhail Isaev, Paulius Micikevicius

Figure 1 for Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation
Figure 2 for Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation
Figure 3 for Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation
Figure 4 for Integer Quantization for Deep Learning Inference: Principles and Empirical Evaluation
Viaarxiv icon

Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks

Add code
Bookmark button
Alert button
May 16, 2018
Sayeh Sharify, Alberto Delmas Lascorz, Kevin Siu, Patrick Judd, Andreas Moshovos

Figure 1 for Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks
Figure 2 for Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks
Figure 3 for Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks
Figure 4 for Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks
Viaarxiv icon

DPRed: Making Typical Activation Values Matter In Deep Learning Computing

Add code
Bookmark button
Alert button
May 15, 2018
Alberto Delmas, Sayeh Sharify, Patrick Judd, Kevin Siu, Milos Nikolic, Andreas Moshovos

Figure 1 for DPRed: Making Typical Activation Values Matter In Deep Learning Computing
Figure 2 for DPRed: Making Typical Activation Values Matter In Deep Learning Computing
Figure 3 for DPRed: Making Typical Activation Values Matter In Deep Learning Computing
Figure 4 for DPRed: Making Typical Activation Values Matter In Deep Learning Computing
Viaarxiv icon

Bit-Tactical: Exploiting Ineffectual Computations in Convolutional Neural Networks: Which, Why, and How

Add code
Bookmark button
Alert button
Mar 09, 2018
Alberto Delmas, Patrick Judd, Dylan Malone Stuart, Zissis Poulos, Mostafa Mahmoud, Sayeh Sharify, Milos Nikolic, Andreas Moshovos

Figure 1 for Bit-Tactical: Exploiting Ineffectual Computations in Convolutional Neural Networks: Which, Why, and How
Figure 2 for Bit-Tactical: Exploiting Ineffectual Computations in Convolutional Neural Networks: Which, Why, and How
Figure 3 for Bit-Tactical: Exploiting Ineffectual Computations in Convolutional Neural Networks: Which, Why, and How
Figure 4 for Bit-Tactical: Exploiting Ineffectual Computations in Convolutional Neural Networks: Which, Why, and How
Viaarxiv icon

Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability

Add code
Bookmark button
Alert button
Jul 27, 2017
Alberto Delmas, Sayeh Sharify, Patrick Judd, Andreas Moshovos

Figure 1 for Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability
Figure 2 for Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability
Figure 3 for Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability
Figure 4 for Tartan: Accelerating Fully-Connected and Convolutional Layers in Deep Learning Networks by Exploiting Numerical Precision Variability
Viaarxiv icon

Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks

Add code
Bookmark button
Alert button
Jun 01, 2017
Alberto Delmas, Patrick Judd, Sayeh Sharify, Andreas Moshovos

Figure 1 for Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks
Figure 2 for Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks
Figure 3 for Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks
Figure 4 for Dynamic Stripes: Exploiting the Dynamic Precision Requirements of Activation Values in Neural Networks
Viaarxiv icon

Cnvlutin2: Ineffectual-Activation-and-Weight-Free Deep Neural Network Computing

Add code
Bookmark button
Alert button
Apr 29, 2017
Patrick Judd, Alberto Delmas, Sayeh Sharify, Andreas Moshovos

Figure 1 for Cnvlutin2: Ineffectual-Activation-and-Weight-Free Deep Neural Network Computing
Figure 2 for Cnvlutin2: Ineffectual-Activation-and-Weight-Free Deep Neural Network Computing
Figure 3 for Cnvlutin2: Ineffectual-Activation-and-Weight-Free Deep Neural Network Computing
Figure 4 for Cnvlutin2: Ineffectual-Activation-and-Weight-Free Deep Neural Network Computing
Viaarxiv icon

Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets

Add code
Bookmark button
Alert button
Jan 08, 2016
Patrick Judd, Jorge Albericio, Tayler Hetherington, Tor Aamodt, Natalie Enright Jerger, Raquel Urtasun, Andreas Moshovos

Figure 1 for Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets
Figure 2 for Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets
Figure 3 for Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets
Figure 4 for Reduced-Precision Strategies for Bounded Memory in Deep Neural Nets
Viaarxiv icon