Picture for Haibin Lin

Haibin Lin

FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable Reasoning

Add code
Oct 26, 2025
Viaarxiv icon

Laminar: A Scalable Asynchronous RL Post-Training Framework

Add code
Oct 14, 2025
Viaarxiv icon

Boosting Embodied AI Agents through Perception-Generation Disaggregation and Asynchronous Pipeline Execution

Add code
Sep 11, 2025
Viaarxiv icon

SwiftSpec: Ultra-Low Latency LLM Decoding by Scaling Asynchronous Speculative Decoding

Add code
Jun 12, 2025
Viaarxiv icon

MegaScale-MoE: Large-Scale Communication-Efficient Training of Mixture-of-Experts Models in Production

Add code
May 19, 2025
Viaarxiv icon

Seed1.5-VL Technical Report

Add code
May 11, 2025
Viaarxiv icon

Understanding Stragglers in Large Model Training Using What-if Analysis

Add code
May 09, 2025
Figure 1 for Understanding Stragglers in Large Model Training Using What-if Analysis
Figure 2 for Understanding Stragglers in Large Model Training Using What-if Analysis
Figure 3 for Understanding Stragglers in Large Model Training Using What-if Analysis
Figure 4 for Understanding Stragglers in Large Model Training Using What-if Analysis
Viaarxiv icon

OVERLORD: Ultimate Scaling of DataLoader for Multi-Source Large Foundation Model Training

Add code
Apr 14, 2025
Viaarxiv icon

VAPO: Efficient and Reliable Reinforcement Learning for Advanced Reasoning Tasks

Add code
Apr 08, 2025
Figure 1 for VAPO: Efficient and Reliable Reinforcement Learning for Advanced Reasoning Tasks
Figure 2 for VAPO: Efficient and Reliable Reinforcement Learning for Advanced Reasoning Tasks
Figure 3 for VAPO: Efficient and Reliable Reinforcement Learning for Advanced Reasoning Tasks
Viaarxiv icon

MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism

Add code
Apr 03, 2025
Figure 1 for MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism
Figure 2 for MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism
Figure 3 for MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism
Figure 4 for MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism
Viaarxiv icon