Alert button
Picture for Daniel Soudry

Daniel Soudry

Alert button

How Uniform Random Weights Induce Non-uniform Bias: Typical Interpolating Neural Networks Generalize with Narrow Teachers

Add code
Bookmark button
Alert button
Feb 09, 2024
Gon Buzaglo, Itamar Harel, Mor Shpigel Nacson, Alon Brutzkus, Nathan Srebro, Daniel Soudry

Viaarxiv icon

Towards Cheaper Inference in Deep Networks with Lower Bit-Width Accumulators

Add code
Bookmark button
Alert button
Jan 25, 2024
Yaniv Blumenfeld, Itay Hubara, Daniel Soudry

Viaarxiv icon

The Joint Effect of Task Similarity and Overparameterization on Catastrophic Forgetting -- An Analytical Model

Add code
Bookmark button
Alert button
Jan 24, 2024
Daniel Goldfarb, Itay Evron, Nir Weinberger, Daniel Soudry, Paul Hand

Viaarxiv icon

How do Minimum-Norm Shallow Denoisers Look in Function Space?

Add code
Bookmark button
Alert button
Nov 12, 2023
Chen Zeno, Greg Ongie, Yaniv Blumenfeld, Nir Weinberger, Daniel Soudry

Figure 1 for How do Minimum-Norm Shallow Denoisers Look in Function Space?
Figure 2 for How do Minimum-Norm Shallow Denoisers Look in Function Space?
Figure 3 for How do Minimum-Norm Shallow Denoisers Look in Function Space?
Figure 4 for How do Minimum-Norm Shallow Denoisers Look in Function Space?
Viaarxiv icon

The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks

Add code
Bookmark button
Alert button
Jun 30, 2023
Mor Shpigel Nacson, Rotem Mulayoff, Greg Ongie, Tomer Michaeli, Daniel Soudry

Figure 1 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 2 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 3 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Figure 4 for The Implicit Bias of Minima Stability in Multivariate Shallow ReLU Networks
Viaarxiv icon

DropCompute: simple and more robust distributed synchronous training via compute variance reduction

Add code
Bookmark button
Alert button
Jun 18, 2023
Niv Giladi, Shahar Gottlieb, Moran Shkolnik, Asaf Karnieli, Ron Banner, Elad Hoffer, Kfir Yehuda Levy, Daniel Soudry

Figure 1 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 2 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 3 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Figure 4 for DropCompute: simple and more robust distributed synchronous training via compute variance reduction
Viaarxiv icon

Continual Learning in Linear Classification on Separable Data

Add code
Bookmark button
Alert button
Jun 06, 2023
Itay Evron, Edward Moroshko, Gon Buzaglo, Maroun Khriesh, Badea Marjieh, Nathan Srebro, Daniel Soudry

Figure 1 for Continual Learning in Linear Classification on Separable Data
Figure 2 for Continual Learning in Linear Classification on Separable Data
Figure 3 for Continual Learning in Linear Classification on Separable Data
Figure 4 for Continual Learning in Linear Classification on Separable Data
Viaarxiv icon

Explore to Generalize in Zero-Shot RL

Add code
Bookmark button
Alert button
Jun 05, 2023
Ev Zisselman, Itai Lavie, Daniel Soudry, Aviv Tamar

Figure 1 for Explore to Generalize in Zero-Shot RL
Figure 2 for Explore to Generalize in Zero-Shot RL
Figure 3 for Explore to Generalize in Zero-Shot RL
Figure 4 for Explore to Generalize in Zero-Shot RL
Viaarxiv icon

Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond

Add code
Bookmark button
Alert button
May 22, 2023
Itai Kreisler, Mor Shpigel Nacson, Daniel Soudry, Yair Carmon

Figure 1 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 2 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 3 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Figure 4 for Gradient Descent Monotonically Decreases the Sharpness of Gradient Flow Solutions in Scalar Networks and Beyond
Viaarxiv icon

Alias-Free Convnets: Fractional Shift Invariance via Polynomial Activations

Add code
Bookmark button
Alert button
Mar 15, 2023
Hagay Michaeli, Tomer Michaeli, Daniel Soudry

Figure 1 for Alias-Free Convnets: Fractional Shift Invariance via Polynomial Activations
Figure 2 for Alias-Free Convnets: Fractional Shift Invariance via Polynomial Activations
Figure 3 for Alias-Free Convnets: Fractional Shift Invariance via Polynomial Activations
Figure 4 for Alias-Free Convnets: Fractional Shift Invariance via Polynomial Activations
Viaarxiv icon