Picture for Dan Alistarh

Dan Alistarh

Optimizers Qualitatively Alter Solutions And We Should Leverage This

Add code
Jul 16, 2025
Viaarxiv icon

Efficient Data Selection at Scale via Influence Distillation

Add code
May 25, 2025
Viaarxiv icon

SVD-Free Low-Rank Adaptive Gradient Optimization for Large Language Models

Add code
May 23, 2025
Viaarxiv icon

Quartet: Native FP4 Training Can Be Optimal for Large Language Models

Add code
May 20, 2025
Viaarxiv icon

Layer-wise Quantization for Quantized Optimistic Dual Averaging

Add code
May 20, 2025
Viaarxiv icon

Towards Combinatorial Interpretability of Neural Computation

Add code
Apr 10, 2025
Viaarxiv icon

Hogwild! Inference: Parallel LLM Generation via Concurrent Attention

Add code
Apr 09, 2025
Figure 1 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 2 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 3 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Figure 4 for Hogwild! Inference: Parallel LLM Generation via Concurrent Attention
Viaarxiv icon

Compression Scaling Laws:Unifying Sparsity and Quantization

Add code
Feb 23, 2025
Figure 1 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 2 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 3 for Compression Scaling Laws:Unifying Sparsity and Quantization
Figure 4 for Compression Scaling Laws:Unifying Sparsity and Quantization
Viaarxiv icon

DarwinLM: Evolutionary Structured Pruning of Large Language Models

Add code
Feb 11, 2025
Figure 1 for DarwinLM: Evolutionary Structured Pruning of Large Language Models
Figure 2 for DarwinLM: Evolutionary Structured Pruning of Large Language Models
Figure 3 for DarwinLM: Evolutionary Structured Pruning of Large Language Models
Figure 4 for DarwinLM: Evolutionary Structured Pruning of Large Language Models
Viaarxiv icon

Position: It's Time to Act on the Risk of Efficient Personalized Text Generation

Add code
Feb 10, 2025
Viaarxiv icon