Alert button
Picture for Tianyun Zhang

Tianyun Zhang

Alert button

Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning

Add code
Bookmark button
Alert button
Oct 08, 2020
Bingbing Li, Zhenglun Kong, Tianyun Zhang, Ji Li, Zhengang Li, Hang Liu, Caiwen Ding

Figure 1 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 2 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 3 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Figure 4 for Efficient Transformer-based Large Scale Language Representations using Hardware-friendly Block Structured Pruning
Viaarxiv icon

Computation on Sparse Neural Networks: an Inspiration for Future Hardware

Add code
Bookmark button
Alert button
Apr 24, 2020
Fei Sun, Minghai Qin, Tianyun Zhang, Liu Liu, Yen-Kuang Chen, Yuan Xie

Figure 1 for Computation on Sparse Neural Networks: an Inspiration for Future Hardware
Figure 2 for Computation on Sparse Neural Networks: an Inspiration for Future Hardware
Figure 3 for Computation on Sparse Neural Networks: an Inspiration for Future Hardware
Viaarxiv icon

A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods

Add code
Bookmark button
Alert button
Apr 12, 2020
Tianyun Zhang, Xiaolong Ma, Zheng Zhan, Shanglin Zhou, Minghai Qin, Fei Sun, Yen-Kuang Chen, Caiwen Ding, Makan Fardad, Yanzhi Wang

Figure 1 for A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods
Figure 2 for A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods
Figure 3 for A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods
Figure 4 for A Unified DNN Weight Compression Framework Using Reweighted Optimization Methods
Viaarxiv icon

BLK-REW: A Unified Block-based DNN Pruning Framework using Reweighted Regularization Method

Add code
Bookmark button
Alert button
Feb 22, 2020
Xiaolong Ma, Zhengang Li, Yifan Gong, Tianyun Zhang, Wei Niu, Zheng Zhan, Pu Zhao, Jian Tang, Xue Lin, Bin Ren, Yanzhi Wang

Figure 1 for BLK-REW: A Unified Block-based DNN Pruning Framework using Reweighted Regularization Method
Figure 2 for BLK-REW: A Unified Block-based DNN Pruning Framework using Reweighted Regularization Method
Figure 3 for BLK-REW: A Unified Block-based DNN Pruning Framework using Reweighted Regularization Method
Figure 4 for BLK-REW: A Unified Block-based DNN Pruning Framework using Reweighted Regularization Method
Viaarxiv icon

An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices

Add code
Bookmark button
Alert button
Feb 22, 2020
Xiaolong Ma, Wei Niu, Tianyun Zhang, Sijia Liu, Sheng Lin, Hongjia Li, Xiang Chen, Jian Tang, Kaisheng Ma, Bin Ren, Yanzhi Wang

Figure 1 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 2 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 3 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Figure 4 for An Image Enhancing Pattern-based Sparsity for Real-time Inference on Mobile Devices
Viaarxiv icon

An Ultra-Efficient Memristor-Based DNN Framework with Structured Weight Pruning and Quantization Using ADMM

Add code
Bookmark button
Alert button
Aug 29, 2019
Geng Yuan, Xiaolong Ma, Caiwen Ding, Sheng Lin, Tianyun Zhang, Zeinab S. Jalali, Yilong Zhao, Li Jiang, Sucheta Soundarajan, Yanzhi Wang

Figure 1 for An Ultra-Efficient Memristor-Based DNN Framework with Structured Weight Pruning and Quantization Using ADMM
Figure 2 for An Ultra-Efficient Memristor-Based DNN Framework with Structured Weight Pruning and Quantization Using ADMM
Figure 3 for An Ultra-Efficient Memristor-Based DNN Framework with Structured Weight Pruning and Quantization Using ADMM
Figure 4 for An Ultra-Efficient Memristor-Based DNN Framework with Structured Weight Pruning and Quantization Using ADMM
Viaarxiv icon