Picture for Shao Tang

Shao Tang

OTPrune: Distribution-Aligned Visual Token Pruning via Optimal Transport

Add code
Feb 25, 2026
Viaarxiv icon

Bayesian Preference Learning for Test-Time Steerable Reward Models

Add code
Feb 09, 2026
Viaarxiv icon

Distilling the Essence: Efficient Reasoning Distillation via Sequence Truncation

Add code
Dec 24, 2025
Viaarxiv icon

Reasoning Models Can be Accurately Pruned Via Chain-of-Thought Reconstruction

Add code
Sep 15, 2025
Figure 1 for Reasoning Models Can be Accurately Pruned Via Chain-of-Thought Reconstruction
Figure 2 for Reasoning Models Can be Accurately Pruned Via Chain-of-Thought Reconstruction
Figure 3 for Reasoning Models Can be Accurately Pruned Via Chain-of-Thought Reconstruction
Figure 4 for Reasoning Models Can be Accurately Pruned Via Chain-of-Thought Reconstruction
Viaarxiv icon

LiDDA: Data Driven Attribution at LinkedIn

Add code
May 14, 2025
Viaarxiv icon

On the Robustness of Reward Models for Language Model Alignment

Add code
May 12, 2025
Viaarxiv icon

Efficient AI in Practice: Training and Deployment of Efficient LLMs for Industry Applications

Add code
Feb 20, 2025
Viaarxiv icon

From Features to Transformers: Redefining Ranking for Scalable Impact

Add code
Feb 05, 2025
Viaarxiv icon

AlphaPO -- Reward shape matters for LLM alignment

Add code
Jan 07, 2025
Viaarxiv icon

Liger Kernel: Efficient Triton Kernels for LLM Training

Add code
Oct 14, 2024
Figure 1 for Liger Kernel: Efficient Triton Kernels for LLM Training
Figure 2 for Liger Kernel: Efficient Triton Kernels for LLM Training
Figure 3 for Liger Kernel: Efficient Triton Kernels for LLM Training
Figure 4 for Liger Kernel: Efficient Triton Kernels for LLM Training
Viaarxiv icon