Picture for Yeyun Gong

Yeyun Gong

Recycling Pretrained Checkpoints: Orthogonal Growth of Mixture-of-Experts for Efficient Large Language Model Pre-Training

Add code
Oct 09, 2025
Viaarxiv icon

Training Matryoshka Mixture-of-Experts for Elastic Inference-Time Expert Utilization

Add code
Sep 30, 2025
Viaarxiv icon

PeRL: Permutation-Enhanced Reinforcement Learning for Interleaved Vision-Language Reasoning

Add code
Jun 17, 2025
Viaarxiv icon

SwS: Self-aware Weakness-driven Problem Synthesis in Reinforcement Learning for LLM Reasoning

Add code
Jun 10, 2025
Viaarxiv icon

HiCaM: A Hierarchical-Causal Modification Framework for Long-Form Text Modification

Add code
May 30, 2025
Viaarxiv icon

How does Alignment Enhance LLMs' Multilingual Capabilities? A Language Neurons Perspective

Add code
May 27, 2025
Viaarxiv icon

Overcoming Vocabulary Mismatch: Vocabulary-agnostic Teacher Guided Language Modeling

Add code
Mar 24, 2025
Viaarxiv icon

Process-based Self-Rewarding Language Models

Add code
Mar 05, 2025
Figure 1 for Process-based Self-Rewarding Language Models
Figure 2 for Process-based Self-Rewarding Language Models
Figure 3 for Process-based Self-Rewarding Language Models
Figure 4 for Process-based Self-Rewarding Language Models
Viaarxiv icon

DeepThink: Aligning Language Models with Domain-Specific User Intents

Add code
Feb 08, 2025
Viaarxiv icon

Optimizing Large Language Model Training Using FP4 Quantization

Add code
Jan 28, 2025
Figure 1 for Optimizing Large Language Model Training Using FP4 Quantization
Figure 2 for Optimizing Large Language Model Training Using FP4 Quantization
Figure 3 for Optimizing Large Language Model Training Using FP4 Quantization
Figure 4 for Optimizing Large Language Model Training Using FP4 Quantization
Viaarxiv icon