Picture for Zhengang Li

Zhengang Li

Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time Mobile Acceleration

Add code
Nov 22, 2021
Figure 1 for Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time Mobile Acceleration
Figure 2 for Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time Mobile Acceleration
Figure 3 for Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time Mobile Acceleration
Figure 4 for Automatic Mapping of the Best-Suited DNN Pruning Schemes for Real-Time Mobile Acceleration
Viaarxiv icon

MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge

Add code
Oct 26, 2021
Figure 1 for MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge
Figure 2 for MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge
Figure 3 for MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge
Figure 4 for MEST: Accurate and Fast Memory-Economic Sparse Training Framework on the Edge
Viaarxiv icon

GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity

Add code
Aug 25, 2021
Figure 1 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 2 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 3 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 4 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Viaarxiv icon

Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI

Add code
Jun 18, 2021
Figure 1 for Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI
Figure 2 for Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI
Figure 3 for Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI
Figure 4 for Improving DNN Fault Tolerance using Weight Pruning and Differential Crossbar Mapping for ReRAM-based Edge AI
Viaarxiv icon

FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator

Add code
Jun 16, 2021
Figure 1 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 2 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 3 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Figure 4 for FORMS: Fine-grained Polarized ReRAM-based In-situ Computation for Mixed-signal DNN Accelerator
Viaarxiv icon

Towards Fast and Accurate Multi-Person Pose Estimation on Mobile Devices

Add code
Jun 06, 2021
Figure 1 for Towards Fast and Accurate Multi-Person Pose Estimation on Mobile Devices
Figure 2 for Towards Fast and Accurate Multi-Person Pose Estimation on Mobile Devices
Figure 3 for Towards Fast and Accurate Multi-Person Pose Estimation on Mobile Devices
Figure 4 for Towards Fast and Accurate Multi-Person Pose Estimation on Mobile Devices
Viaarxiv icon

6.7ms on Mobile with over 78% ImageNet Accuracy: Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration

Add code
Dec 01, 2020
Figure 1 for 6.7ms on Mobile with over 78% ImageNet Accuracy: Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration
Figure 2 for 6.7ms on Mobile with over 78% ImageNet Accuracy: Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration
Figure 3 for 6.7ms on Mobile with over 78% ImageNet Accuracy: Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration
Figure 4 for 6.7ms on Mobile with over 78% ImageNet Accuracy: Unified Network Pruning and Architecture Search for Beyond Real-Time Mobile Acceleration
Viaarxiv icon

Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning

Add code
Oct 08, 2020
Figure 1 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 2 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 3 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 4 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Viaarxiv icon

Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices

Add code
Jul 20, 2020
Figure 1 for Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices
Figure 2 for Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices
Figure 3 for Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices
Figure 4 for Achieving Real-Time Execution of 3D Convolutional Neural Networks on Mobile Devices
Viaarxiv icon

A Privacy-Preserving DNN Pruning and Mobile Acceleration Framework

Add code
Mar 13, 2020
Figure 1 for A Privacy-Preserving DNN Pruning and Mobile Acceleration Framework
Figure 2 for A Privacy-Preserving DNN Pruning and Mobile Acceleration Framework
Figure 3 for A Privacy-Preserving DNN Pruning and Mobile Acceleration Framework
Figure 4 for A Privacy-Preserving DNN Pruning and Mobile Acceleration Framework
Viaarxiv icon