Alert button
Picture for Mor Shpigel Nacson

Mor Shpigel Nacson

Alert button

How Uniform Random Weights Induce Non-uniform Bias: Typical Interpolating Neural Networks Generalize with Narrow Teachers

Add code
Bookmark button
Alert button
Feb 09, 2024
Gon Buzaglo, Itamar Harel, Mor Shpigel Nacson, Alon Brutzkus, Nathan Srebro, Daniel Soudry

Viaarxiv icon

The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks

Add code
Bookmark button
Alert button
Jun 30, 2023
Mor Shpigel Nacson, Rotem Mulayoff, Greg Ongie, Tomer Michaeli, Daniel Soudry

Figure 1 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 2 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 3 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 4 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Viaarxiv icon

Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond

Add code
Bookmark button
Alert button
May 22, 2023
Itai Kreisler, Mor Shpigel Nacson, Daniel Soudry, Yair Carmon

Figure 1 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 2 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 3 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 4 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Viaarxiv icon

On the Implicit Bias of Initialization Shape: Beyond Infinitesimal Mirror Descent

Add code
Bookmark button
Alert button
Feb 19, 2021
Shahar Azulay, Edward Moroshko, Mor Shpigel Nacson, Blake Woodworth, Nathan Srebro, Amir Globerson, Daniel Soudry

Figure 1 for On the Implicit Bias of Initialization Shape: Beyond Infinitesimal Mirror Descent
Figure 2 for On the Implicit Bias of Initialization Shape: Beyond Infinitesimal Mirror Descent
Viaarxiv icon

At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?

Add code
Bookmark button
Alert button
Sep 26, 2019
Niv Giladi, Mor Shpigel Nacson, Elad Hoffer, Daniel Soudry

Figure 1 for At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?
Figure 2 for At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?
Figure 3 for At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?
Figure 4 for At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?
Viaarxiv icon

Lexicographic and Depth-Sensitive Margins in Homogeneous and Non-Homogeneous Deep Models

Add code
Bookmark button
Alert button
May 17, 2019
Mor Shpigel Nacson, Suriya Gunasekar, Jason D. Lee, Nathan Srebro, Daniel Soudry

Viaarxiv icon

Convergence of Gradient Descent on Separable Data

Add code
Bookmark button
Alert button
Jun 12, 2018
Mor Shpigel Nacson, Jason Lee, Suriya Gunasekar, Pedro H. P. Savarese, Nathan Srebro, Daniel Soudry

Figure 1 for Convergence of Gradient Descent on Separable Data
Figure 2 for Convergence of Gradient Descent on Separable Data
Figure 3 for Convergence of Gradient Descent on Separable Data
Figure 4 for Convergence of Gradient Descent on Separable Data
Viaarxiv icon

Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate

Add code
Bookmark button
Alert button
Jun 05, 2018
Mor Shpigel Nacson, Nathan Srebro, Daniel Soudry

Figure 1 for Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate
Figure 2 for Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate
Figure 3 for Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate
Viaarxiv icon

The Implicit Bias of Gradient Descent on Separable Data

Add code
Bookmark button
Alert button
Mar 21, 2018
Daniel Soudry, Elad Hoffer, Mor Shpigel Nacson, Suriya Gunasekar, Nathan Srebro

Figure 1 for The Implicit Bias of Gradient Descent on Separable Data
Figure 2 for The Implicit Bias of Gradient Descent on Separable Data
Figure 3 for The Implicit Bias of Gradient Descent on Separable Data
Figure 4 for The Implicit Bias of Gradient Descent on Separable Data
Viaarxiv icon