Picture for Yingbin Liang

Yingbin Liang

Bridging Online and Offline RL: Contextual Bandit Learning for Multi-Turn Code Generation

Add code
Feb 03, 2026
Viaarxiv icon

ConvexBench: Can LLMs Recognize Convex Functions?

Add code
Feb 01, 2026
Viaarxiv icon

Mixture-of-Transformers Learn Faster: A Theoretical Study on Classification Problems

Add code
Oct 30, 2025
Viaarxiv icon

Monitoring State Transitions in Markovian Systems with Sampling Cost

Add code
Oct 25, 2025
Viaarxiv icon

Large Language Models Achieve Gold Medal Performance at International Astronomy & Astrophysics Olympiad

Add code
Oct 06, 2025
Viaarxiv icon

Multi-head Transformers Provably Learn Symbolic Multi-step Reasoning via Gradient Descent

Add code
Aug 11, 2025
Viaarxiv icon

Provable In-Context Learning of Nonlinear Regression with Transformers

Add code
Jul 28, 2025
Figure 1 for Provable In-Context Learning of Nonlinear Regression with Transformers
Viaarxiv icon

How Transformers Learn Regular Language Recognition: A Theoretical Study on Training Dynamics and Implicit Bias

Add code
May 02, 2025
Viaarxiv icon

Contrastive and Variational Approaches in Self-Supervised Learning for Complex Data Mining

Add code
Apr 05, 2025
Viaarxiv icon

Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining

Add code
Feb 10, 2025
Figure 1 for Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining
Figure 2 for Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining
Figure 3 for Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining
Figure 4 for Dynamic Loss-Based Sample Reweighting for Improved Large Language Model Pretraining
Viaarxiv icon