Picture for Yuchao Li

Yuchao Li

An Approximate Dynamic Programming Framework for Occlusion-Robust Multi-Object Tracking

Add code
May 24, 2024
Viaarxiv icon

Most Likely Sequence Generation for $n$-Grams, Transformers, HMMs, and Markov Chains, by Using Rollout Algorithms

Add code
Mar 19, 2024
Viaarxiv icon

Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity

Add code
Sep 19, 2023
Figure 1 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 2 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 3 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Figure 4 for Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity
Viaarxiv icon

Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning

Add code
Nov 15, 2022
Figure 1 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Figure 2 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Figure 3 for Multiagent Rollout with Reshuffling for Warehouse Robots Path Planning
Viaarxiv icon

Parameter-Efficient Sparsity for Large Language Models Fine-Tuning

Add code
May 23, 2022
Figure 1 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 2 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 3 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Figure 4 for Parameter-Efficient Sparsity for Large Language Models Fine-Tuning
Viaarxiv icon

An Information Theory-inspired Strategy for Automatic Network Pruning

Add code
Aug 19, 2021
Figure 1 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 2 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 3 for An Information Theory-inspired Strategy for Automatic Network Pruning
Figure 4 for An Information Theory-inspired Strategy for Automatic Network Pruning
Viaarxiv icon

1$\times$N Block Pattern for Network Sparsity

Add code
Jun 15, 2021
Figure 1 for 1$\times$N Block Pattern for Network Sparsity
Figure 2 for 1$\times$N Block Pattern for Network Sparsity
Figure 3 for 1$\times$N Block Pattern for Network Sparsity
Figure 4 for 1$\times$N Block Pattern for Network Sparsity
Viaarxiv icon

You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient

Add code
Jun 04, 2021
Figure 1 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 2 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 3 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Figure 4 for You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Nature Gradient
Viaarxiv icon

Towards Compact CNNs via Collaborative Compression

Add code
May 24, 2021
Figure 1 for Towards Compact CNNs via Collaborative Compression
Figure 2 for Towards Compact CNNs via Collaborative Compression
Figure 3 for Towards Compact CNNs via Collaborative Compression
Figure 4 for Towards Compact CNNs via Collaborative Compression
Viaarxiv icon

PAMS: Quantized Super-Resolution via Parameterized Max Scale

Add code
Nov 09, 2020
Figure 1 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 2 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 3 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Figure 4 for PAMS: Quantized Super-Resolution via Parameterized Max Scale
Viaarxiv icon