Picture for Qingxiu Dong

Qingxiu Dong

The Era of Agentic Organization: Learning to Organize with Language Models

Add code
Oct 30, 2025
Figure 1 for The Era of Agentic Organization: Learning to Organize with Language Models
Figure 2 for The Era of Agentic Organization: Learning to Organize with Language Models
Figure 3 for The Era of Agentic Organization: Learning to Organize with Language Models
Figure 4 for The Era of Agentic Organization: Learning to Organize with Language Models
Viaarxiv icon

Reinforcement Pre-Training

Add code
Jun 09, 2025
Viaarxiv icon

Think Only When You Need with Large Hybrid-Reasoning Models

Add code
May 21, 2025
Viaarxiv icon

Reward Reasoning Model

Add code
May 20, 2025
Figure 1 for Reward Reasoning Model
Figure 2 for Reward Reasoning Model
Figure 3 for Reward Reasoning Model
Figure 4 for Reward Reasoning Model
Viaarxiv icon

RICo: Refined In-Context Contribution for Automatic Instruction-Tuning Data Selection

Add code
May 18, 2025
Viaarxiv icon

SelfBudgeter: Adaptive Token Allocation for Efficient LLM Reasoning

Add code
May 16, 2025
Viaarxiv icon

ICon: In-Context Contribution for Automatic Data Selection

Add code
May 08, 2025
Viaarxiv icon

Scaling Laws of Synthetic Data for Language Models

Add code
Mar 26, 2025
Figure 1 for Scaling Laws of Synthetic Data for Language Models
Figure 2 for Scaling Laws of Synthetic Data for Language Models
Figure 3 for Scaling Laws of Synthetic Data for Language Models
Figure 4 for Scaling Laws of Synthetic Data for Language Models
Viaarxiv icon

MPO: Boosting LLM Agents with Meta Plan Optimization

Add code
Mar 04, 2025
Figure 1 for MPO: Boosting LLM Agents with Meta Plan Optimization
Figure 2 for MPO: Boosting LLM Agents with Meta Plan Optimization
Figure 3 for MPO: Boosting LLM Agents with Meta Plan Optimization
Figure 4 for MPO: Boosting LLM Agents with Meta Plan Optimization
Viaarxiv icon

How Far are LLMs from Being Our Digital Twins? A Benchmark for Persona-Based Behavior Chain Simulation

Add code
Feb 20, 2025
Viaarxiv icon