Picture for Dan Alistarh

Dan Alistarh

Mathador-LM: A Dynamic Benchmark for Mathematical Reasoning on Large Language Models

Add code
Jun 18, 2024
Viaarxiv icon

Sparse Expansion and Neuronal Disentanglement

Add code
May 24, 2024
Figure 1 for Sparse Expansion and Neuronal Disentanglement
Figure 2 for Sparse Expansion and Neuronal Disentanglement
Figure 3 for Sparse Expansion and Neuronal Disentanglement
Figure 4 for Sparse Expansion and Neuronal Disentanglement
Viaarxiv icon

MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence

Add code
May 24, 2024
Figure 1 for MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
Figure 2 for MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
Figure 3 for MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
Figure 4 for MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
Viaarxiv icon

PV-Tuning: Beyond Straight-Through Estimation for Extreme LLM Compression

Add code
May 23, 2024
Viaarxiv icon

Enabling High-Sparsity Foundational Llama Models with Efficient Pretraining and Deployment

Add code
May 06, 2024
Viaarxiv icon

Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization

Add code
Apr 04, 2024
Figure 1 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 2 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 3 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Figure 4 for Mitigating the Impact of Outlier Channels for Language Model Quantization with Activation Regularization
Viaarxiv icon

QuaRot: Outlier-Free 4-Bit Inference in Rotated LLMs

Add code
Mar 30, 2024
Viaarxiv icon

RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation

Add code
Jan 12, 2024
Figure 1 for RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
Figure 2 for RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
Figure 3 for RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
Figure 4 for RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
Viaarxiv icon

Extreme Compression of Large Language Models via Additive Quantization

Add code
Jan 11, 2024
Figure 1 for Extreme Compression of Large Language Models via Additive Quantization
Figure 2 for Extreme Compression of Large Language Models via Additive Quantization
Figure 3 for Extreme Compression of Large Language Models via Additive Quantization
Figure 4 for Extreme Compression of Large Language Models via Additive Quantization
Viaarxiv icon

How to Prune Your Language Model: Recovering Accuracy on the "Sparsity May Cry'' Benchmark

Add code
Dec 21, 2023
Figure 1 for How to Prune Your Language Model: Recovering Accuracy on the "Sparsity May Cry'' Benchmark
Figure 2 for How to Prune Your Language Model: Recovering Accuracy on the "Sparsity May Cry'' Benchmark
Figure 3 for How to Prune Your Language Model: Recovering Accuracy on the "Sparsity May Cry'' Benchmark
Figure 4 for How to Prune Your Language Model: Recovering Accuracy on the "Sparsity May Cry'' Benchmark
Viaarxiv icon