Alert button
Picture for Utku Evci

Utku Evci

Alert button

Progressive Gradient Flow for Robust N:M Sparsity Training in Transformers

Add code
Bookmark button
Alert button
Feb 07, 2024
Abhimanyu Rajeshkumar Bambhaniya, Amir Yazdanbakhsh, Suvinay Subramanian, Sheng-Chun Kao, Shivani Agrawal, Utku Evci, Tushar Krishna

Viaarxiv icon

Scaling Laws for Sparsely-Connected Foundation Models

Add code
Bookmark button
Alert button
Sep 15, 2023
Elias Frantar, Carlos Riquelme, Neil Houlsby, Dan Alistarh, Utku Evci

Figure 1 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 2 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 3 for Scaling Laws for Sparsely-Connected Foundation Models
Figure 4 for Scaling Laws for Sparsely-Connected Foundation Models
Viaarxiv icon

Dynamic Sparse Training with Structured Sparsity

Add code
Bookmark button
Alert button
May 03, 2023
Mike Lasby, Anna Golubeva, Utku Evci, Mihai Nica, Yani Ioannou

Figure 1 for Dynamic Sparse Training with Structured Sparsity
Figure 2 for Dynamic Sparse Training with Structured Sparsity
Figure 3 for Dynamic Sparse Training with Structured Sparsity
Figure 4 for Dynamic Sparse Training with Structured Sparsity
Viaarxiv icon

JaxPruner: A concise library for sparsity research

Add code
Bookmark button
Alert button
May 02, 2023
Joo Hyung Lee, Wonpyo Park, Nicole Mitchell, Jonathan Pilault, Johan Obando-Ceron, Han-Byul Kim, Namhoon Lee, Elias Frantar, Yun Long, Amir Yazdanbakhsh, Shivani Agrawal, Suvinay Subramanian, Xin Wang, Sheng-Chun Kao, Xingyao Zhang, Trevor Gale, Aart Bik, Woohyun Han, Milen Ferev, Zhonglin Han, Hong-Seok Kim, Yann Dauphin, Gintare Karolina Dziugaite, Pablo Samuel Castro, Utku Evci

Figure 1 for JaxPruner: A concise library for sparsity research
Figure 2 for JaxPruner: A concise library for sparsity research
Viaarxiv icon

The Dormant Neuron Phenomenon in Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 24, 2023
Ghada Sokar, Rishabh Agarwal, Pablo Samuel Castro, Utku Evci

Figure 1 for The Dormant Neuron Phenomenon in Deep Reinforcement Learning
Figure 2 for The Dormant Neuron Phenomenon in Deep Reinforcement Learning
Figure 3 for The Dormant Neuron Phenomenon in Deep Reinforcement Learning
Figure 4 for The Dormant Neuron Phenomenon in Deep Reinforcement Learning
Viaarxiv icon

Scaling Vision Transformers to 22 Billion Parameters

Add code
Bookmark button
Alert button
Feb 10, 2023
Mostafa Dehghani, Josip Djolonga, Basil Mustafa, Piotr Padlewski, Jonathan Heek, Justin Gilmer, Andreas Steiner, Mathilde Caron, Robert Geirhos, Ibrahim Alabdulmohsin, Rodolphe Jenatton, Lucas Beyer, Michael Tschannen, Anurag Arnab, Xiao Wang, Carlos Riquelme, Matthias Minderer, Joan Puigcerver, Utku Evci, Manoj Kumar, Sjoerd van Steenkiste, Gamaleldin F. Elsayed, Aravindh Mahendran, Fisher Yu, Avital Oliver, Fantine Huot, Jasmijn Bastings, Mark Patrick Collier, Alexey Gritsenko, Vighnesh Birodkar, Cristina Vasconcelos, Yi Tay, Thomas Mensink, Alexander Kolesnikov, Filip Pavetić, Dustin Tran, Thomas Kipf, Mario Lučić, Xiaohua Zhai, Daniel Keysers, Jeremiah Harmsen, Neil Houlsby

Figure 1 for Scaling Vision Transformers to 22 Billion Parameters
Figure 2 for Scaling Vision Transformers to 22 Billion Parameters
Figure 3 for Scaling Vision Transformers to 22 Billion Parameters
Figure 4 for Scaling Vision Transformers to 22 Billion Parameters
Viaarxiv icon

Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask

Add code
Bookmark button
Alert button
Sep 15, 2022
Sheng-Chun Kao, Amir Yazdanbakhsh, Suvinay Subramanian, Shivani Agrawal, Utku Evci, Tushar Krishna

Figure 1 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 2 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 3 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Figure 4 for Training Recipe for N:M Structured Sparsity with Decaying Pruning Mask
Viaarxiv icon

The State of Sparse Training in Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 17, 2022
Laura Graesser, Utku Evci, Erich Elsen, Pablo Samuel Castro

Figure 1 for The State of Sparse Training in Deep Reinforcement Learning
Figure 2 for The State of Sparse Training in Deep Reinforcement Learning
Figure 3 for The State of Sparse Training in Deep Reinforcement Learning
Figure 4 for The State of Sparse Training in Deep Reinforcement Learning
Viaarxiv icon

GradMax: Growing Neural Networks using Gradient Information

Add code
Bookmark button
Alert button
Jan 13, 2022
Utku Evci, Max Vladymyrov, Thomas Unterthiner, Bart van Merriënboer, Fabian Pedregosa

Figure 1 for GradMax: Growing Neural Networks using Gradient Information
Figure 2 for GradMax: Growing Neural Networks using Gradient Information
Figure 3 for GradMax: Growing Neural Networks using Gradient Information
Figure 4 for GradMax: Growing Neural Networks using Gradient Information
Viaarxiv icon