Alert button
Picture for Behnam Neyshabur

Behnam Neyshabur

Alert button

Google Research

The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks

Add code
Bookmark button
Alert button
Oct 12, 2021
Rahim Entezari, Hanie Sedghi, Olga Saukh, Behnam Neyshabur

Figure 1 for The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks
Figure 2 for The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks
Figure 3 for The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks
Figure 4 for The Role of Permutation Invariance in Linear Mode Connectivity of Neural Networks
Viaarxiv icon

A Loss Curvature Perspective on Training Instability in Deep Learning

Add code
Bookmark button
Alert button
Oct 08, 2021
Justin Gilmer, Behrooz Ghorbani, Ankush Garg, Sneha Kudugunta, Behnam Neyshabur, David Cardoze, George Dahl, Zachary Nado, Orhan Firat

Figure 1 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 2 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 3 for A Loss Curvature Perspective on Training Instability in Deep Learning
Figure 4 for A Loss Curvature Perspective on Training Instability in Deep Learning
Viaarxiv icon

Exploring the Limits of Large Scale Pre-training

Add code
Bookmark button
Alert button
Oct 05, 2021
Samira Abnar, Mostafa Dehghani, Behnam Neyshabur, Hanie Sedghi

Figure 1 for Exploring the Limits of Large Scale Pre-training
Figure 2 for Exploring the Limits of Large Scale Pre-training
Figure 3 for Exploring the Limits of Large Scale Pre-training
Figure 4 for Exploring the Limits of Large Scale Pre-training
Viaarxiv icon

The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning

Add code
Bookmark button
Alert button
Jun 30, 2021
Anders Andreassen, Yasaman Bahri, Behnam Neyshabur, Rebecca Roelofs

Figure 1 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 2 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 3 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Figure 4 for The Evolution of Out-of-Distribution Robustness Throughout Fine-Tuning
Viaarxiv icon

Deep Learning Through the Lens of Example Difficulty

Add code
Bookmark button
Alert button
Jun 18, 2021
Robert J. N. Baldock, Hartmut Maennel, Behnam Neyshabur

Figure 1 for Deep Learning Through the Lens of Example Difficulty
Figure 2 for Deep Learning Through the Lens of Example Difficulty
Figure 3 for Deep Learning Through the Lens of Example Difficulty
Figure 4 for Deep Learning Through the Lens of Example Difficulty
Viaarxiv icon

NeurIPS 2020 Competition: Predicting Generalization in Deep Learning

Add code
Bookmark button
Alert button
Dec 14, 2020
Yiding Jiang, Pierre Foret, Scott Yak, Daniel M. Roy, Hossein Mobahi, Gintare Karolina Dziugaite, Samy Bengio, Suriya Gunasekar, Isabelle Guyon, Behnam Neyshabur

Figure 1 for NeurIPS 2020 Competition: Predicting Generalization in Deep Learning
Viaarxiv icon

When Do Curricula Work?

Add code
Bookmark button
Alert button
Dec 05, 2020
Xiaoxia Wu, Ethan Dyer, Behnam Neyshabur

Figure 1 for When Do Curricula Work?
Figure 2 for When Do Curricula Work?
Figure 3 for When Do Curricula Work?
Figure 4 for When Do Curricula Work?
Viaarxiv icon

Understanding the Failure Modes of Out-of-Distribution Generalization

Add code
Bookmark button
Alert button
Oct 29, 2020
Vaishnavh Nagarajan, Anders Andreassen, Behnam Neyshabur

Figure 1 for Understanding the Failure Modes of Out-of-Distribution Generalization
Figure 2 for Understanding the Failure Modes of Out-of-Distribution Generalization
Figure 3 for Understanding the Failure Modes of Out-of-Distribution Generalization
Figure 4 for Understanding the Failure Modes of Out-of-Distribution Generalization
Viaarxiv icon

Are wider nets better given the same number of parameters?

Add code
Bookmark button
Alert button
Oct 27, 2020
Anna Golubeva, Behnam Neyshabur, Guy Gur-Ari

Figure 1 for Are wider nets better given the same number of parameters?
Figure 2 for Are wider nets better given the same number of parameters?
Figure 3 for Are wider nets better given the same number of parameters?
Figure 4 for Are wider nets better given the same number of parameters?
Viaarxiv icon

The Deep Bootstrap: Good Online Learners are Good Offline Generalizers

Add code
Bookmark button
Alert button
Oct 16, 2020
Preetum Nakkiran, Behnam Neyshabur, Hanie Sedghi

Figure 1 for The Deep Bootstrap: Good Online Learners are Good Offline Generalizers
Figure 2 for The Deep Bootstrap: Good Online Learners are Good Offline Generalizers
Figure 3 for The Deep Bootstrap: Good Online Learners are Good Offline Generalizers
Figure 4 for The Deep Bootstrap: Good Online Learners are Good Offline Generalizers
Viaarxiv icon