Alert button
Picture for Yaman Umuroglu

Yaman Umuroglu

Alert button

A2Q+: Improving Accumulator-Aware Weight Quantization

Add code
Bookmark button
Alert button
Jan 19, 2024
Ian Colbert, Alessandro Pappalardo, Jakoba Petri-Koenig, Yaman Umuroglu

Viaarxiv icon

Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark

Add code
Bookmark button
Alert button
Jun 23, 2022
Hendrik Borras, Giuseppe Di Guglielmo, Javier Duarte, Nicolò Ghielmetti, Ben Hawks, Scott Hauck, Shih-Chieh Hsu, Ryan Kastner, Jason Liang, Andres Meza, Jules Muhizi, Tai Nguyen, Rushil Roy, Nhan Tran, Yaman Umuroglu, Olivia Weng, Aidan Yokuda, Michaela Blott

Figure 1 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 2 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 3 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Figure 4 for Open-source FPGA-ML codesign for the MLPerf Tiny Benchmark
Viaarxiv icon

QONNX: Representing Arbitrary-Precision Quantized Neural Networks

Add code
Bookmark button
Alert button
Jun 17, 2022
Alessandro Pappalardo, Yaman Umuroglu, Michaela Blott, Jovan Mitrevski, Ben Hawks, Nhan Tran, Vladimir Loncar, Sioni Summers, Hendrik Borras, Jules Muhizi, Matthew Trahms, Shih-Chieh Hsu, Scott Hauck, Javier Duarte

Figure 1 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 2 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 3 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Figure 4 for QONNX: Representing Arbitrary-Precision Quantized Neural Networks
Viaarxiv icon

EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators

Add code
Bookmark button
Alert button
Feb 04, 2022
Lois Orosa, Skanda Koppula, Yaman Umuroglu, Konstantinos Kanellopoulos, Juan Gomez-Luna, Michaela Blott, Kees Vissers, Onur Mutlu

Figure 1 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 2 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 3 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Figure 4 for EcoFlow: Efficient Convolutional Dataflows for Low-Power Neural Network Accelerators
Viaarxiv icon

Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference

Add code
Bookmark button
Alert button
Feb 22, 2021
Benjamin Hawks, Javier Duarte, Nicholas J. Fraser, Alessandro Pappalardo, Nhan Tran, Yaman Umuroglu

Figure 1 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 2 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 3 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Figure 4 for Ps and Qs: Quantization-aware pruning for efficient low latency neural network inference
Viaarxiv icon

LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications

Add code
Bookmark button
Alert button
Apr 06, 2020
Yaman Umuroglu, Yash Akhauri, Nicholas J. Fraser, Michaela Blott

Figure 1 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 2 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 3 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Figure 4 for LogicNets: Co-Designed Neural Networks and Circuits for Extreme-Throughput Applications
Viaarxiv icon

Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic

Add code
Bookmark button
Alert button
Jun 26, 2018
Michaela Blott, Thomas B. Preusser, Nicholas Fraser, Giulio Gambardella, Kenneth OBrien, Yaman Umuroglu, Miriam Leeser

Figure 1 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 2 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 3 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Figure 4 for Scaling Neural Network Performance through Customized Hardware Architectures on Reconfigurable Logic
Viaarxiv icon

Streamlined Deployment for Quantized Neural Networks

Add code
Bookmark button
Alert button
May 30, 2018
Yaman Umuroglu, Magnus Jahre

Figure 1 for Streamlined Deployment for Quantized Neural Networks
Figure 2 for Streamlined Deployment for Quantized Neural Networks
Figure 3 for Streamlined Deployment for Quantized Neural Networks
Figure 4 for Streamlined Deployment for Quantized Neural Networks
Viaarxiv icon

Scaling Binarized Neural Networks on Reconfigurable Logic

Add code
Bookmark button
Alert button
Jan 27, 2017
Nicholas J. Fraser, Yaman Umuroglu, Giulio Gambardella, Michaela Blott, Philip Leong, Magnus Jahre, Kees Vissers

Figure 1 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 2 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 3 for Scaling Binarized Neural Networks on Reconfigurable Logic
Figure 4 for Scaling Binarized Neural Networks on Reconfigurable Logic
Viaarxiv icon

FINN: A Framework for Fast, Scalable Binarized Neural Network Inference

Add code
Bookmark button
Alert button
Dec 01, 2016
Yaman Umuroglu, Nicholas J. Fraser, Giulio Gambardella, Michaela Blott, Philip Leong, Magnus Jahre, Kees Vissers

Figure 1 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 2 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 3 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Figure 4 for FINN: A Framework for Fast, Scalable Binarized Neural Network Inference
Viaarxiv icon