Picture for Dan Alistarh

Dan Alistarh

Behemoth: Benchmarking Unlearning in LLMs Using Fully Synthetic Data

Add code
Jan 30, 2026
Viaarxiv icon

Quartet II: Accurate LLM Pre-Training in NVFP4 by Improved Unbiased Gradient Estimation

Add code
Jan 30, 2026
Viaarxiv icon

ECO: Quantized Training without Full-Precision Master Weights

Add code
Jan 29, 2026
Viaarxiv icon

LLMQ: Efficient Lower-Precision Pretraining for Consumer GPUs

Add code
Dec 17, 2025
Viaarxiv icon

Speculative Decoding Speed-of-Light: Optimal Lower Bounds via Branching Random Walks

Add code
Dec 12, 2025
Viaarxiv icon

Expand Neurons, Not Parameters

Add code
Oct 06, 2025
Figure 1 for Expand Neurons, Not Parameters
Figure 2 for Expand Neurons, Not Parameters
Figure 3 for Expand Neurons, Not Parameters
Figure 4 for Expand Neurons, Not Parameters
Viaarxiv icon

The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM

Add code
Oct 02, 2025
Viaarxiv icon

Optimizers Qualitatively Alter Solutions And We Should Leverage This

Add code
Jul 16, 2025
Viaarxiv icon

Efficient Data Selection at Scale via Influence Distillation

Add code
May 25, 2025
Viaarxiv icon

SVD-Free Low-Rank Adaptive Gradient Optimization for Large Language Models

Add code
May 23, 2025
Viaarxiv icon