Picture for Xiangang Li

Xiangang Li

Not All Correct Answers Are Equal: Why Your Distillation Source Matters

Add code
May 20, 2025
Viaarxiv icon

AM-Thinking-v1: Advancing the Frontier of Reasoning at 32B Scale

Add code
May 13, 2025
Viaarxiv icon

Exploring the Potential of Offline RL for Reasoning in LLMs: A Preliminary Study

Add code
May 04, 2025
Viaarxiv icon

DeepDistill: Enhancing LLM Reasoning Capabilities via Large-Scale Difficulty-Graded Data Training

Add code
Apr 24, 2025
Viaarxiv icon

SARI: Structured Audio Reasoning via Curriculum-Guided Reinforcement Learning

Add code
Apr 22, 2025
Viaarxiv icon

Leveraging Reasoning Model Answers to Enhance Non-Reasoning Model Capability

Add code
Apr 13, 2025
Viaarxiv icon

How Difficulty-Aware Staged Reinforcement Learning Enhances LLMs' Reasoning Capabilities: A Preliminary Experimental Study

Add code
Apr 01, 2025
Viaarxiv icon

Memorizing is Not Enough: Deep Knowledge Injection Through Reasoning

Add code
Apr 01, 2025
Viaarxiv icon

1.4 Million Open-Source Distilled Reasoning Dataset to Empower Large Language Model Training

Add code
Mar 25, 2025
Viaarxiv icon

Think Twice: Enhancing LLM Reasoning by Scaling Multi-round Test-time Thinking

Add code
Mar 25, 2025
Viaarxiv icon