Picture for Yushi Bai

Yushi Bai

IndexCache: Accelerating Sparse Attention via Cross-Layer Index Reuse

Add code
Mar 12, 2026
Viaarxiv icon

GLM-5: from Vibe Coding to Agentic Engineering

Add code
Feb 17, 2026
Viaarxiv icon

GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models

Add code
Aug 08, 2025
Viaarxiv icon

LongWriter-Zero: Mastering Ultra-Long Text Generation via Reinforcement Learning

Add code
Jun 23, 2025
Viaarxiv icon

SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models

Add code
Jun 04, 2025
Figure 1 for SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models
Figure 2 for SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models
Figure 3 for SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models
Figure 4 for SuperWriter: Reflection-Driven Long-Form Generation with Large Language Models
Viaarxiv icon

How does Transformer Learn Implicit Reasoning?

Add code
May 29, 2025
Figure 1 for How does Transformer Learn Implicit Reasoning?
Figure 2 for How does Transformer Learn Implicit Reasoning?
Figure 3 for How does Transformer Learn Implicit Reasoning?
Figure 4 for How does Transformer Learn Implicit Reasoning?
Viaarxiv icon

Hard Negative Contrastive Learning for Fine-Grained Geometric Understanding in Large Multimodal Models

Add code
May 26, 2025
Viaarxiv icon

An LMM for Efficient Video Understanding via Reinforced Compression of Video Cubes

Add code
Apr 21, 2025
Figure 1 for An LMM for Efficient Video Understanding via Reinforced Compression of Video Cubes
Figure 2 for An LMM for Efficient Video Understanding via Reinforced Compression of Video Cubes
Figure 3 for An LMM for Efficient Video Understanding via Reinforced Compression of Video Cubes
Figure 4 for An LMM for Efficient Video Understanding via Reinforced Compression of Video Cubes
Viaarxiv icon

Shifting Long-Context LLMs Research from Input to Output

Add code
Mar 07, 2025
Figure 1 for Shifting Long-Context LLMs Research from Input to Output
Figure 2 for Shifting Long-Context LLMs Research from Input to Output
Figure 3 for Shifting Long-Context LLMs Research from Input to Output
Figure 4 for Shifting Long-Context LLMs Research from Input to Output
Viaarxiv icon

LongWriter-V: Enabling Ultra-Long and High-Fidelity Generation in Vision-Language Models

Add code
Feb 20, 2025
Viaarxiv icon