Picture for Guohao Dai

Guohao Dai

Infinigence-AI, Shanghai Jiao Tong University

SJD++: Improved Speculative Jacobi Decoding for Training-free Acceleration of Discrete Auto-regressive Text-to-Image Generation

Add code
Dec 08, 2025
Viaarxiv icon

BitSnap: Checkpoint Sparsification and Quantization in LLM Training

Add code
Nov 18, 2025
Viaarxiv icon

Think-at-Hard: Selective Latent Iterations to Improve Reasoning Language Models

Add code
Nov 11, 2025
Viaarxiv icon

TASP: Topology-aware Sequence Parallelism

Add code
Sep 30, 2025
Figure 1 for TASP: Topology-aware Sequence Parallelism
Figure 2 for TASP: Topology-aware Sequence Parallelism
Figure 3 for TASP: Topology-aware Sequence Parallelism
Figure 4 for TASP: Topology-aware Sequence Parallelism
Viaarxiv icon

RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation

Add code
Sep 19, 2025
Figure 1 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 2 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 3 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Figure 4 for RLinf: Flexible and Efficient Large-scale Reinforcement Learning via Macro-to-Micro Flow Transformation
Viaarxiv icon

SpecDiff: Accelerating Diffusion Model Inference with Self-Speculation

Add code
Sep 17, 2025
Viaarxiv icon

VocabTailor: Dynamic Vocabulary Selection for Downstream Tasks in Small Language Models

Add code
Aug 21, 2025
Viaarxiv icon

Megrez2 Technical Report

Add code
Jul 23, 2025
Viaarxiv icon

R2R: Efficiently Navigating Divergent Reasoning Paths with Small-Large Model Token Routing

Add code
May 27, 2025
Viaarxiv icon

PM-KVQ: Progressive Mixed-precision KV Cache Quantization for Long-CoT LLMs

Add code
May 24, 2025
Figure 1 for PM-KVQ: Progressive Mixed-precision KV Cache Quantization for Long-CoT LLMs
Figure 2 for PM-KVQ: Progressive Mixed-precision KV Cache Quantization for Long-CoT LLMs
Figure 3 for PM-KVQ: Progressive Mixed-precision KV Cache Quantization for Long-CoT LLMs
Figure 4 for PM-KVQ: Progressive Mixed-precision KV Cache Quantization for Long-CoT LLMs
Viaarxiv icon