Alert button
Picture for Peiyan Dong

Peiyan Dong

Alert button

EdgeQAT: Entropy and Distribution Guided Quantization-Aware Training for the Acceleration of Lightweight LLMs on the Edge

Add code
Bookmark button
Alert button
Feb 16, 2024
Xuan Shen, Zhenglun Kong, Changdi Yang, Zhaoyang Han, Lei Lu, Peiyan Dong, Cheng Lyu, Chih-hsiang Li, Xuehang Guo, Zhihao Shu, Wei Niu, Miriam Leeser, Pu Zhao, Yanzhi Wang

Viaarxiv icon

Agile-Quant: Activation-Guided Quantization for Faster Inference of LLMs on the Edge

Add code
Bookmark button
Alert button
Dec 09, 2023
Xuan Shen, Peiyan Dong, Lei Lu, Zhenglun Kong, Zhengang Li, Ming Lin, Chao Wu, Yanzhi Wang

Viaarxiv icon

SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices

Add code
Bookmark button
Alert button
Sep 21, 2023
Zhengang Li, Geng Yuan, Tomoharu Yamauchi, Zabihi Masoud, Yanyue Xie, Peiyan Dong, Xulong Tang, Nobuyuki Yoshikawa, Devesh Tiwari, Yanzhi Wang, Olivia Chen

Figure 1 for SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices
Figure 2 for SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices
Figure 3 for SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices
Figure 4 for SupeRBNN: Randomized Binary Neural Network Using Adiabatic Superconductor Josephson Devices
Viaarxiv icon

Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training

Add code
Bookmark button
Alert button
Nov 19, 2022
Zhenglun Kong, Haoyu Ma, Geng Yuan, Mengshu Sun, Yanyue Xie, Peiyan Dong, Xin Meng, Xuan Shen, Hao Tang, Minghai Qin, Tianlong Chen, Xiaolong Ma, Xiaohui Xie, Zhangyang Wang, Yanzhi Wang

Figure 1 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 2 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 3 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Figure 4 for Peeling the Onion: Hierarchical Reduction of Data Redundancy for Efficient Vision Transformer Training
Viaarxiv icon

HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers

Add code
Bookmark button
Alert button
Nov 15, 2022
Peiyan Dong, Mengshu Sun, Alec Lu, Yanyue Xie, Kenneth Liu, Zhenglun Kong, Xin Meng, Zhengang Li, Xue Lin, Zhenman Fang, Yanzhi Wang

Figure 1 for HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers
Figure 2 for HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers
Figure 3 for HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers
Figure 4 for HeatViT: Hardware-Efficient Adaptive Token Pruning for Vision Transformers
Viaarxiv icon

The Lottery Ticket Hypothesis for Vision Transformers

Add code
Bookmark button
Alert button
Nov 02, 2022
Xuan Shen, Zhenglun Kong, Minghai Qin, Peiyan Dong, Geng Yuan, Xin Meng, Hao Tang, Xiaolong Ma, Yanzhi Wang

Figure 1 for The Lottery Ticket Hypothesis for Vision Transformers
Figure 2 for The Lottery Ticket Hypothesis for Vision Transformers
Figure 3 for The Lottery Ticket Hypothesis for Vision Transformers
Figure 4 for The Lottery Ticket Hypothesis for Vision Transformers
Viaarxiv icon

Quantum Neural Network Compression

Add code
Bookmark button
Alert button
Jul 05, 2022
Zhirui Hu, Peiyan Dong, Zhepeng Wang, Youzuo Lin, Yanzhi Wang, Weiwen Jiang

Figure 1 for Quantum Neural Network Compression
Figure 2 for Quantum Neural Network Compression
Figure 3 for Quantum Neural Network Compression
Figure 4 for Quantum Neural Network Compression
Viaarxiv icon

SPViT: Enabling Faster Vision Transformers via Soft Token Pruning

Add code
Bookmark button
Alert button
Dec 27, 2021
Zhenglun Kong, Peiyan Dong, Xiaolong Ma, Xin Meng, Wei Niu, Mengshu Sun, Bin Ren, Minghai Qin, Hao Tang, Yanzhi Wang

Figure 1 for SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
Figure 2 for SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
Figure 3 for SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
Figure 4 for SPViT: Enabling Faster Vision Transformers via Soft Token Pruning
Viaarxiv icon

GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity

Add code
Bookmark button
Alert button
Aug 25, 2021
Wei Niu, Zhengang Li, Xiaolong Ma, Peiyan Dong, Gang Zhou, Xuehai Qian, Xue Lin, Yanzhi Wang, Bin Ren

Figure 1 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 2 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 3 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Figure 4 for GRIM: A General, Real-Time Deep Learning Inference Framework for Mobile Devices based on Fine-Grained Structured Weight Sparsity
Viaarxiv icon

RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition

Add code
Bookmark button
Alert button
Feb 19, 2020
Peiyan Dong, Siyue Wang, Wei Niu, Chengming Zhang, Sheng Lin, Zhengang Li, Yifan Gong, Bin Ren, Xue Lin, Yanzhi Wang, Dingwen Tao

Figure 1 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 2 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 3 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Figure 4 for RTMobile: Beyond Real-Time Mobile Acceleration of RNNs for Speech Recognition
Viaarxiv icon