Picture for Massoud Pedram

Massoud Pedram

A Fast Training-Free Compression Framework for Vision Transformers

Add code
Mar 04, 2023
Viaarxiv icon

Efficient Compilation and Mapping of Fixed Function Combinational Logic onto Digital Signal Processors Targeting Neural Network Inference and Utilizing High-level Synthesis

Add code
Jul 30, 2022
Figure 1 for Efficient Compilation and Mapping of Fixed Function Combinational Logic onto Digital Signal Processors Targeting Neural Network Inference and Utilizing High-level Synthesis
Figure 2 for Efficient Compilation and Mapping of Fixed Function Combinational Logic onto Digital Signal Processors Targeting Neural Network Inference and Utilizing High-level Synthesis
Figure 3 for Efficient Compilation and Mapping of Fixed Function Combinational Logic onto Digital Signal Processors Targeting Neural Network Inference and Utilizing High-level Synthesis
Figure 4 for Efficient Compilation and Mapping of Fixed Function Combinational Logic onto Digital Signal Processors Targeting Neural Network Inference and Utilizing High-level Synthesis
Viaarxiv icon

Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators

Add code
Jun 30, 2022
Figure 1 for Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators
Figure 2 for Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators
Figure 3 for Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators
Figure 4 for Sparse Periodic Systolic Dataflow for Lowering Latency and Power Dissipation of Convolutional Neural Network Accelerators
Viaarxiv icon

A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness

Add code
Mar 28, 2022
Figure 1 for A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness
Figure 2 for A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness
Figure 3 for A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness
Figure 4 for A Fast and Efficient Conditional Learning for Tunable Trade-Off between Accuracy and Robustness
Viaarxiv icon

BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch

Add code
Dec 24, 2021
Figure 1 for BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Figure 2 for BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Figure 3 for BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Figure 4 for BMPQ: Bit-Gradient Sensitivity Driven Mixed-Precision Quantization of DNNs from Scratch
Viaarxiv icon

Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression

Add code
Jul 16, 2021
Figure 1 for Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression
Figure 2 for Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression
Figure 3 for Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression
Figure 4 for Towards Low-Latency Energy-Efficient Deep SNNs via Attention-Guided Compression
Viaarxiv icon

NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function Combinational Logic

Add code
Apr 07, 2021
Figure 1 for NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function Combinational Logic
Figure 2 for NullaNet Tiny: Ultra-low-latency DNN Inference Through Fixed-function Combinational Logic
Viaarxiv icon

A2P-MANN: Adaptive Attention Inference Hops Pruned Memory-Augmented Neural Networks

Add code
Jan 24, 2021
Figure 1 for A2P-MANN: Adaptive Attention Inference Hops Pruned Memory-Augmented Neural Networks
Figure 2 for A2P-MANN: Adaptive Attention Inference Hops Pruned Memory-Augmented Neural Networks
Figure 3 for A2P-MANN: Adaptive Attention Inference Hops Pruned Memory-Augmented Neural Networks
Figure 4 for A2P-MANN: Adaptive Attention Inference Hops Pruned Memory-Augmented Neural Networks
Viaarxiv icon

BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification

Add code
Jan 07, 2021
Figure 1 for BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Figure 2 for BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Figure 3 for BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Figure 4 for BRDS: An FPGA-based LSTM Accelerator with Row-Balanced Dual-Ratio Sparsification
Viaarxiv icon

A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs

Add code
Nov 03, 2020
Figure 1 for A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs
Figure 2 for A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs
Figure 3 for A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs
Figure 4 for A Tunable Robust Pruning Framework Through Dynamic Network Rewiring of DNNs
Viaarxiv icon