Picture for Xinting Huang

Xinting Huang

Discovering Interpretable Algorithms by Decompiling Transformers to RASP

Add code
Feb 09, 2026
Viaarxiv icon

Free(): Learning to Forget in Malloc-Only Reasoning Models

Add code
Feb 08, 2026
Viaarxiv icon

Dual-Phase LLM Reasoning: Self-Evolved Mathematical Frameworks

Add code
Jan 09, 2026
Viaarxiv icon

The End of Manual Decoding: Towards Truly End-to-End Language Models

Add code
Oct 30, 2025
Figure 1 for The End of Manual Decoding: Towards Truly End-to-End Language Models
Figure 2 for The End of Manual Decoding: Towards Truly End-to-End Language Models
Figure 3 for The End of Manual Decoding: Towards Truly End-to-End Language Models
Figure 4 for The End of Manual Decoding: Towards Truly End-to-End Language Models
Viaarxiv icon

UNCLE: Uncertainty Expressions in Long-Form Generation

Add code
May 22, 2025
Viaarxiv icon

Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought

Add code
May 21, 2025
Viaarxiv icon

Low-hallucination Synthetic Captions for Large-Scale Vision-Language Model Pre-training

Add code
Apr 17, 2025
Viaarxiv icon

Contextualize-then-Aggregate: Circuits for In-Context Learning in Gemma-2 2B

Add code
Mar 31, 2025
Figure 1 for Contextualize-then-Aggregate: Circuits for In-Context Learning in Gemma-2 2B
Figure 2 for Contextualize-then-Aggregate: Circuits for In-Context Learning in Gemma-2 2B
Figure 3 for Contextualize-then-Aggregate: Circuits for In-Context Learning in Gemma-2 2B
Figure 4 for Contextualize-then-Aggregate: Circuits for In-Context Learning in Gemma-2 2B
Viaarxiv icon

Lower Bounds for Chain-of-Thought Reasoning in Hard-Attention Transformers

Add code
Feb 04, 2025
Viaarxiv icon

A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression

Add code
Dec 23, 2024
Figure 1 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 2 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 3 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 4 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Viaarxiv icon