Picture for Dan Alistarh

Dan Alistarh

GSQ: Highly-Accurate Low-Precision Scalar Quantization for LLMs via Gumbel-Softmax Sampling

Add code
Apr 20, 2026
Viaarxiv icon

Towards Robust Scaling Laws for Optimizers

Add code
Feb 07, 2026
Viaarxiv icon

LoRDO: Distributed Low-Rank Optimization with Infrequent Communication

Add code
Feb 04, 2026
Viaarxiv icon

MatGPTQ: Accurate and Efficient Post-Training Matryoshka Quantization

Add code
Feb 03, 2026
Viaarxiv icon

DASH: Faster Shampoo via Batched Block Preconditioning and Efficient Inverse-Root Solvers

Add code
Feb 02, 2026
Viaarxiv icon

Quartet II: Accurate LLM Pre-Training in NVFP4 by Improved Unbiased Gradient Estimation

Add code
Jan 30, 2026
Viaarxiv icon

Behemoth: Benchmarking Unlearning in LLMs Using Fully Synthetic Data

Add code
Jan 30, 2026
Viaarxiv icon

ECO: Quantized Training without Full-Precision Master Weights

Add code
Jan 29, 2026
Viaarxiv icon

LLMQ: Efficient Lower-Precision Pretraining for Consumer GPUs

Add code
Dec 17, 2025
Viaarxiv icon

Speculative Decoding Speed-of-Light: Optimal Lower Bounds via Branching Random Walks

Add code
Dec 12, 2025
Viaarxiv icon