Alert button
Picture for Guangli Li

Guangli Li

Alert button

ProSparse: Introducing and Enhancing Intrinsic Activation Sparsity within Large Language Models

Add code
Bookmark button
Alert button
Feb 27, 2024
Chenyang Song, Xu Han, Zhengyan Zhang, Shengding Hu, Xiyu Shi, Kuai Li, Chen Chen, Zhiyuan Liu, Guangli Li, Tao Yang, Maosong Sun

Viaarxiv icon

Pinpointing the Memory Behaviors of DNN Training

Add code
Bookmark button
Alert button
Apr 01, 2021
Jiansong Li, Xiao Dong, Guangli Li, Peng Zhao, Xueying Wang, Xiaobing Chen, Xianzhi Yu, Yongxin Yang, Zihan Jiang, Wei Cao, Lei Liu, Xiaobing Feng

Figure 1 for Pinpointing the Memory Behaviors of DNN Training
Figure 2 for Pinpointing the Memory Behaviors of DNN Training
Figure 3 for Pinpointing the Memory Behaviors of DNN Training
Figure 4 for Pinpointing the Memory Behaviors of DNN Training
Viaarxiv icon

Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices

Add code
Bookmark button
Alert button
Oct 30, 2020
Guangli Li, Xiu Ma, Xueying Wang, Lei Liu, Jingling Xue, Xiaobing Feng

Figure 1 for Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices
Figure 2 for Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices
Figure 3 for Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices
Figure 4 for Fusion-Catalyzed Pruning for Optimizing Deep Learning on Intelligent Edge Devices
Viaarxiv icon

LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units

Add code
Bookmark button
Alert button
Mar 20, 2020
Guangli Li, Lei Liu, Xueying Wang, Xiu Ma, Xiaobing Feng

Figure 1 for LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
Figure 2 for LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
Figure 3 for LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
Figure 4 for LANCE: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
Viaarxiv icon

LANCE: efficient low-precision quantized Winograd convolution for neural networks based on graphics processing units

Add code
Bookmark button
Alert button
Mar 19, 2020
Guangli Li, Lei Liu, Xueying Wang, Xiu Ma, Xiaobing Feng

Figure 1 for LANCE: efficient low-precision quantized Winograd convolution for neural networks based on graphics processing units
Figure 2 for LANCE: efficient low-precision quantized Winograd convolution for neural networks based on graphics processing units
Figure 3 for LANCE: efficient low-precision quantized Winograd convolution for neural networks based on graphics processing units
Figure 4 for LANCE: efficient low-precision quantized Winograd convolution for neural networks based on graphics processing units
Viaarxiv icon

Background subtraction on depth videos with convolutional neural networks

Add code
Bookmark button
Alert button
Jan 17, 2019
Xueying Wang, Lei Liu, Guangli Li, Xiao Dong, Peng Zhao, Xiaobing Feng

Figure 1 for Background subtraction on depth videos with convolutional neural networks
Figure 2 for Background subtraction on depth videos with convolutional neural networks
Figure 3 for Background subtraction on depth videos with convolutional neural networks
Figure 4 for Background subtraction on depth videos with convolutional neural networks
Viaarxiv icon

Auto-tuning Neural Network Quantization Framework for Collaborative Inference Between the Cloud and Edge

Add code
Bookmark button
Alert button
Dec 16, 2018
Guangli Li, Lei Liu, Xueying Wang, Xiao Dong, Peng Zhao, Xiaobing Feng

Figure 1 for Auto-tuning Neural Network Quantization Framework for Collaborative Inference Between the Cloud and Edge
Figure 2 for Auto-tuning Neural Network Quantization Framework for Collaborative Inference Between the Cloud and Edge
Figure 3 for Auto-tuning Neural Network Quantization Framework for Collaborative Inference Between the Cloud and Edge
Figure 4 for Auto-tuning Neural Network Quantization Framework for Collaborative Inference Between the Cloud and Edge
Viaarxiv icon