Picture for Dan Alistarh

Dan Alistarh

Expand Neurons, Not Parameters

Add code
Oct 06, 2025
Viaarxiv icon

The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM

Add code
Oct 02, 2025
Viaarxiv icon

Optimizers Qualitatively Alter Solutions And We Should Leverage This

Add code
Jul 16, 2025
Viaarxiv icon

Efficient Data Selection at Scale via Influence Distillation

Add code
May 25, 2025
Viaarxiv icon

SVD-Free Low-Rank Adaptive Gradient Optimization for Large Language Models

Add code
May 23, 2025
Viaarxiv icon

Layer-wise Quantization for Quantized Optimistic Dual Averaging

Add code
May 20, 2025
Viaarxiv icon

Quartet: Native FP4 Training Can Be Optimal for Large Language Models

Add code
May 20, 2025
Viaarxiv icon

Towards Combinatorial Interpretability of Neural Computation

Add code
Apr 10, 2025
Viaarxiv icon

Hogwild! Inference: Parallel LLM Generation via Concurrent Attention

Add code
Apr 09, 2025
Figure 1 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 2 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 3 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 4 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Viaarxiv icon

Compression Scaling Laws:Unifying Sparsity and Quantization

Add code
Feb 23, 2025
Figure 1 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 2 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 3 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 4 for Compression Scaling Laws:Unifying Sparsity and Quantization
Viaarxiv icon