Picture for Dong Yu

Dong Yu

The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models

Add code
Mar 04, 2025
Figure 1 for The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models
Figure 2 for The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models
Figure 3 for The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models
Figure 4 for The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models
Viaarxiv icon

Improving LLM General Preference Alignment via Optimistic Online Mirror Descent

Add code
Feb 24, 2025
Figure 1 for Improving LLM General Preference Alignment via Optimistic Online Mirror Descent
Figure 2 for Improving LLM General Preference Alignment via Optimistic Online Mirror Descent
Figure 3 for Improving LLM General Preference Alignment via Optimistic Online Mirror Descent
Viaarxiv icon

Balancing Speech Understanding and Generation Using Continual Pre-training for Codec-based Speech LLM

Add code
Feb 24, 2025
Viaarxiv icon

LLM-Enhanced Dialogue Management for Full-Duplex Spoken Dialogue Systems

Add code
Feb 19, 2025
Figure 1 for LLM-Enhanced Dialogue Management for Full-Duplex Spoken Dialogue Systems
Figure 2 for LLM-Enhanced Dialogue Management for Full-Duplex Spoken Dialogue Systems
Figure 3 for LLM-Enhanced Dialogue Management for Full-Duplex Spoken Dialogue Systems
Figure 4 for LLM-Enhanced Dialogue Management for Full-Duplex Spoken Dialogue Systems
Viaarxiv icon

Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs

Add code
Jan 30, 2025
Figure 1 for Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs
Figure 2 for Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs
Figure 3 for Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs
Figure 4 for Thoughts Are All Over the Place: On the Underthinking of o1-Like LLMs
Viaarxiv icon

OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas

Add code
Jan 26, 2025
Figure 1 for OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas
Figure 2 for OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas
Figure 3 for OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas
Figure 4 for OpenCharacter: Training Customizable Role-Playing LLMs with Large-Scale Synthetic Personas
Viaarxiv icon

Lifelong Learning of Large Language Model based Agents: A Roadmap

Add code
Jan 13, 2025
Viaarxiv icon

Disambiguation of Chinese Polyphones in an End-to-End Framework with Semantic Features Extracted by Pre-trained BERT

Add code
Jan 02, 2025
Figure 1 for Disambiguation of Chinese Polyphones in an End-to-End Framework with Semantic Features Extracted by Pre-trained BERT
Figure 2 for Disambiguation of Chinese Polyphones in an End-to-End Framework with Semantic Features Extracted by Pre-trained BERT
Figure 3 for Disambiguation of Chinese Polyphones in an End-to-End Framework with Semantic Features Extracted by Pre-trained BERT
Figure 4 for Disambiguation of Chinese Polyphones in an End-to-End Framework with Semantic Features Extracted by Pre-trained BERT
Viaarxiv icon

Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs

Add code
Dec 30, 2024
Viaarxiv icon

A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression

Add code
Dec 23, 2024
Figure 1 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 2 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 3 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 4 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Viaarxiv icon