Picture for Ziyang Cai

Ziyang Cai

Extrapolation by Association: Length Generalization Transfer in Transformers

Add code
Jun 10, 2025
Viaarxiv icon

R&B: Domain Regrouping and Data Mixture Balancing for Efficient Foundation Model Training

Add code
May 01, 2025
Viaarxiv icon

Self-Improving Transformers Overcome Easy-to-Hard and Length Generalization Challenges

Add code
Feb 03, 2025
Viaarxiv icon

Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition

Add code
Oct 08, 2024
Figure 1 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 2 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 3 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 4 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Viaarxiv icon

Delving into Out-of-Distribution Detection with Vision-Language Representations

Add code
Nov 24, 2022
Viaarxiv icon