Picture for Beidi Chen

Beidi Chen

MonarchRT: Efficient Attention for Real-Time Video Generation

Add code
Feb 12, 2026
Viaarxiv icon

MedVerse: Efficient and Reliable Medical Reasoning via DAG-Structured Parallel Execution

Add code
Feb 07, 2026
Viaarxiv icon

Jackpot: Optimal Budgeted Rejection Sampling for Extreme Actor-Policy Mismatch Reinforcement Learning

Add code
Feb 05, 2026
Viaarxiv icon

Reliable and Responsible Foundation Models: A Comprehensive Survey

Add code
Feb 04, 2026
Viaarxiv icon

STEM: Scaling Transformers with Embedding Modules

Add code
Jan 15, 2026
Viaarxiv icon

RLBoost: Harvesting Preemptible Resources for Cost-Efficient Reinforcement Learning on LLMs

Add code
Oct 22, 2025
Figure 1 for RLBoost: Harvesting Preemptible Resources for Cost-Efficient Reinforcement Learning on LLMs
Figure 2 for RLBoost: Harvesting Preemptible Resources for Cost-Efficient Reinforcement Learning on LLMs
Figure 3 for RLBoost: Harvesting Preemptible Resources for Cost-Efficient Reinforcement Learning on LLMs
Figure 4 for RLBoost: Harvesting Preemptible Resources for Cost-Efficient Reinforcement Learning on LLMs
Viaarxiv icon

Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation

Add code
Jun 11, 2025
Viaarxiv icon

Kinetics: Rethinking Test-Time Scaling Laws

Add code
Jun 06, 2025
Viaarxiv icon

Scalable LLM Math Reasoning Acceleration with Low-rank Distillation

Add code
May 08, 2025
Viaarxiv icon

HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading

Add code
Feb 18, 2025
Figure 1 for HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading
Figure 2 for HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading
Figure 3 for HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading
Figure 4 for HeadInfer: Memory-Efficient LLM Inference by Head-wise Offloading
Viaarxiv icon