Alert button
Picture for Yuchao Li

Yuchao Li

Alert button

Most Likely Sequence Generation for $n$-Grams, Transformers, HMMs, and Markov Chains, by Using Rollout Algorithms

Add code
Bookmark button
Alert button
Mar 19, 2024
Yuchao Li, Dimitri Bertsekas

Viaarxiv icon

Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity

Add code
Bookmark button
Alert button
Sep 19, 2023
Haojun Xia, Zhen Zheng, Yuchao Li, Donglin Zhuang, Zhongzhu Zhou, Xiafei Qiu, Yong Li, Wei Lin, Shuaiwen Leon Song

Figure 1 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 2 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 3 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 4 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Viaarxiv icon

Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning

Add code
Bookmark button
Alert button
Nov 15, 2022
William Emanuelsson, Alejandro Penacho Riveiros, Yuchao Li, Karl H. Johansson, Jonas Mårtensson

Figure 1 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Figure 2 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Figure 3 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Viaarxiv icon

Parameter-Efficient Sparsity for Large Language Models Fine-Tuning

Add code
Bookmark button
Alert button
May 23, 2022
Yuchao Li, Fuli Luo, Chuanqi Tan, Mengdi Wang, Songfang Huang, Shen Li, Junjie Bai

Figure 1 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 2 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 3 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 4 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Viaarxiv icon

An Information Theory-inspired Strategy for Automatic Network Pruning

Add code
Bookmark button
Alert button
Aug 19, 2021
Xiawu Zheng, Yuexiao Ma, Teng Xi, Gang Zhang, Errui Ding, Yuchao Li, Jie Chen, Yonghong Tian, Rongrong Ji

Figure 1 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 2 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 3 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 4 for An Information Theory-inspired Strategy for Automatic Network Pruning
Viaarxiv icon

1$\times$N Block Pattern for Network Sparsity

Add code
Bookmark button
Alert button
Jun 15, 2021
Mingbao Lin, Yuchao Li, Yuxin Zhang, Bohong Chen, Fei Chao, Mengdi Wang, Shen Li, Jun Yang, Rongrong Ji

Figure 1 for 1$\times$N Block Pattern for Network Sparsity
Figure 2 for 1$\times$N Block Pattern for Network Sparsity
Figure 3 for 1$\times$N Block Pattern for Network Sparsity
Figure 4 for 1$\times$N Block Pattern for Network Sparsity
Viaarxiv icon

You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient

Add code
Bookmark button
Alert button
Jun 04, 2021
Shaokun Zhang, Xiawu Zheng, Chenyi Yang, Yuchao Li, Yan Wang, Fei Chao, Mengdi Wang, Shen Li, Jun Yang, Rongrong Ji

Figure 1 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 2 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 3 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 4 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Viaarxiv icon

Towards Compact CNNs via Collaborative Compression

Add code
Bookmark button
Alert button
May 24, 2021
Yuchao Li, Shaohui Lin, Jianzhuang Liu, Qixiang Ye, Mengdi Wang, Fei Chao, Fan Yang, Jincheng Ma, Qi Tian, Rongrong Ji

Figure 1 for Towards Compact CNNs via Collaborative Compression
Figure 2 for Towards Compact CNNs via Collaborative Compression
Figure 3 for Towards Compact CNNs via Collaborative Compression
Figure 4 for Towards Compact CNNs via Collaborative Compression
Viaarxiv icon

PAMS: Quantized Super-Resolution via Parameterized Max Scale

Add code
Bookmark button
Alert button
Nov 09, 2020
Huixia Li, Chenqian Yan, Shaohui Lin, Xiawu Zheng, Yuchao Li, Baochang Zhang, Fan Yang, Rongrong Ji

Figure 1 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 2 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 3 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 4 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Viaarxiv icon