Alert button
Picture for Dan Alistarh

Dan Alistarh

Alert button

Sparse Finetuning for Inference Acceleration of Large Language Models

Add code
Bookmark button
Alert button
Oct 10, 2023
Eldar Kurtic, Denis Kuznedelev, Elias Frantar, Michael Goin, Dan Alistarh

Viaarxiv icon

SPADE: Sparsity-Guided Debugging for Deep Neural Networks

Add code
Bookmark button
Alert button
Oct 06, 2023
Arshia Soltani Moakhar, Eugenia Iofinova, Dan Alistarh

Figure 1 for SPADE: Sparsity-Guided Debugging for Deep Neural Networks
Figure 2 for SPADE: Sparsity-Guided Debugging for Deep Neural Networks
Figure 3 for SPADE: Sparsity-Guided Debugging for Deep Neural Networks
Figure 4 for SPADE: Sparsity-Guided Debugging for Deep Neural Networks
Viaarxiv icon

Scaling Laws for Sparsely-Connected Foundation Models

Add code
Bookmark button
Alert button
Sep 15, 2023
Elias Frantar, Carlos Riquelme, Neil Houlsby, Dan Alistarh, Utku Evci

Figure 1 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 2 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 3 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 4 for Scaling Laws for Sparsely-Connected Foundation Models
Viaarxiv icon

Accurate Neural Network Pruning Requires Rethinking Sparse Optimization

Add code
Bookmark button
Alert button
Aug 03, 2023
Denis Kuznedelev, Eldar Kurtic, Eugenia Iofinova, Elias Frantar, Alexandra Peste, Dan Alistarh

Figure 1 for Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Figure 2 for Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Figure 3 for Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Figure 4 for Accurate Neural Network Pruning Requires Rethinking Sparse Optimization
Viaarxiv icon

QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models

Add code
Bookmark button
Alert button
Jul 07, 2023
Tommaso Pegolotti, Elias Frantar, Dan Alistarh, Markus Püschel

Figure 1 for QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models
Figure 2 for QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models
Figure 3 for QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models
Figure 4 for QIGen: Generating Efficient Kernels for Quantized Inference on Large Language Models
Viaarxiv icon

Error Feedback Can Accurately Compress Preconditioners

Add code
Bookmark button
Alert button
Jun 16, 2023
Ionut-Vlad Modoranu, Aleksei Kalinov, Eldar Kurtic, Dan Alistarh

Figure 1 for Error Feedback Can Accurately Compress Preconditioners
Figure 2 for Error Feedback Can Accurately Compress Preconditioners
Figure 3 for Error Feedback Can Accurately Compress Preconditioners
Figure 4 for Error Feedback Can Accurately Compress Preconditioners
Viaarxiv icon

Decentralized Learning Dynamics in the Gossip Model

Add code
Bookmark button
Alert button
Jun 14, 2023
John Lazarsfeld, Dan Alistarh

Figure 1 for Decentralized Learning Dynamics in the Gossip Model
Figure 2 for Decentralized Learning Dynamics in the Gossip Model
Figure 3 for Decentralized Learning Dynamics in the Gossip Model
Figure 4 for Decentralized Learning Dynamics in the Gossip Model
Viaarxiv icon

SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression

Add code
Bookmark button
Alert button
Jun 05, 2023
Tim Dettmers, Ruslan Svirschevski, Vage Egiazarian, Denis Kuznedelev, Elias Frantar, Saleh Ashkboos, Alexander Borzunov, Torsten Hoefler, Dan Alistarh

Figure 1 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 2 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 3 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Figure 4 for SpQR: A Sparse-Quantized Representation for Near-Lossless LLM Weight Compression
Viaarxiv icon