Picture for Alisa Liu

Alisa Liu

Nemotron 3 Super: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Apr 14, 2026
Viaarxiv icon

Are Language Models Sensitive to Morally Irrelevant Distractors?

Add code
Feb 10, 2026
Viaarxiv icon

Are you going to finish that? A Practical Study of the Tokenization Boundary Problem

Add code
Jan 30, 2026
Viaarxiv icon

NVIDIA Nemotron 3: Efficient and Open Intelligence

Add code
Dec 24, 2025
Viaarxiv icon

Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning

Add code
Dec 23, 2025
Viaarxiv icon

Olmo 3

Add code
Dec 15, 2025
Viaarxiv icon

Broken Tokens? Your Language Model can Secretly Handle Non-Canonical Tokenizations

Add code
Jun 23, 2025
Viaarxiv icon

Sampling from Your Language Model One Byte at a Time

Add code
Jun 17, 2025
Viaarxiv icon

LLAMAPIE: Proactive In-Ear Conversation Assistants

Add code
May 07, 2025
Figure 1 for LLAMAPIE: Proactive In-Ear Conversation Assistants
Figure 2 for LLAMAPIE: Proactive In-Ear Conversation Assistants
Figure 3 for LLAMAPIE: Proactive In-Ear Conversation Assistants
Figure 4 for LLAMAPIE: Proactive In-Ear Conversation Assistants
Viaarxiv icon

SuperBPE: Space Travel for Language Models

Add code
Mar 17, 2025
Figure 1 for SuperBPE: Space Travel for Language Models
Figure 2 for SuperBPE: Space Travel for Language Models
Figure 3 for SuperBPE: Space Travel for Language Models
Figure 4 for SuperBPE: Space Travel for Language Models
Viaarxiv icon