Picture for Peter Y. K. Cheung

Peter Y. K. Cheung

Logic Shrinkage: Learned FPGA Netlist Sparsity for Efficient Neural Network Inference

Add code
Jan 02, 2022
Figure 1 for Logic Shrinkage: Learned FPGA Netlist Sparsity for Efficient Neural Network Inference
Figure 2 for Logic Shrinkage: Learned FPGA Netlist Sparsity for Efficient Neural Network Inference
Figure 3 for Logic Shrinkage: Learned FPGA Netlist Sparsity for Efficient Neural Network Inference
Figure 4 for Logic Shrinkage: Learned FPGA Netlist Sparsity for Efficient Neural Network Inference
Viaarxiv icon

Enabling Binary Neural Network Training on the Edge

Add code
Feb 10, 2021
Figure 1 for Enabling Binary Neural Network Training on the Edge
Figure 2 for Enabling Binary Neural Network Training on the Edge
Figure 3 for Enabling Binary Neural Network Training on the Edge
Figure 4 for Enabling Binary Neural Network Training on the Edge
Viaarxiv icon

LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference

Add code
Oct 24, 2019
Figure 1 for LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference
Figure 2 for LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference
Figure 3 for LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference
Figure 4 for LUTNet: Learning FPGA Configurations for Highly Efficient Neural Network Inference
Viaarxiv icon

Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs

Add code
Oct 21, 2019
Figure 1 for Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs
Figure 2 for Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs
Figure 3 for Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs
Figure 4 for Automatic Generation of Multi-precision Multi-arithmetic CNN Accelerators for FPGAs
Viaarxiv icon

LUTNet: Rethinking Inference in FPGA Soft Logic

Add code
Apr 01, 2019
Figure 1 for LUTNet: Rethinking Inference in FPGA Soft Logic
Figure 2 for LUTNet: Rethinking Inference in FPGA Soft Logic
Figure 3 for LUTNet: Rethinking Inference in FPGA Soft Logic
Figure 4 for LUTNet: Rethinking Inference in FPGA Soft Logic
Viaarxiv icon

Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going

Add code
Jan 21, 2019
Figure 1 for Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going
Figure 2 for Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going
Figure 3 for Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going
Figure 4 for Deep Neural Network Approximation for Custom Hardware: Where We've Been, Where We're Going
Viaarxiv icon

Accuracy to Throughput Trade-offs for Reduced Precision Neural Networks on Reconfigurable Logic

Add code
Jul 17, 2018
Figure 1 for Accuracy to Throughput Trade-offs for Reduced Precision Neural Networks on Reconfigurable Logic
Figure 2 for Accuracy to Throughput Trade-offs for Reduced Precision Neural Networks on Reconfigurable Logic
Figure 3 for Accuracy to Throughput Trade-offs for Reduced Precision Neural Networks on Reconfigurable Logic
Figure 4 for Accuracy to Throughput Trade-offs for Reduced Precision Neural Networks on Reconfigurable Logic
Viaarxiv icon