Alert button
Picture for Hadi Daneshmand

Hadi Daneshmand

Alert button

Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion

Add code
Bookmark button
Alert button
Oct 03, 2023
Alexandru Meterez, Amir Joudaki, Francesco Orabona, Alexander Immer, Gunnar Rätsch, Hadi Daneshmand

Figure 1 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 2 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 3 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 4 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Viaarxiv icon

Transformers learn to implement preconditioned gradient descent for in-context learning

Add code
Bookmark button
Alert button
Jun 01, 2023
Kwangjun Ahn, Xiang Cheng, Hadi Daneshmand, Suvrit Sra

Figure 1 for Transformers learn to implement preconditioned gradient descent for in-context learning
Figure 2 for Transformers learn to implement preconditioned gradient descent for in-context learning
Figure 3 for Transformers learn to implement preconditioned gradient descent for in-context learning
Figure 4 for Transformers learn to implement preconditioned gradient descent for in-context learning
Viaarxiv icon

On the impact of activation and normalization in obtaining isometric embeddings at initialization

Add code
Bookmark button
Alert button
May 28, 2023
Amir Joudaki, Hadi Daneshmand, Francis Bach

Figure 1 for On the impact of activation and normalization in obtaining isometric embeddings at initialization
Figure 2 for On the impact of activation and normalization in obtaining isometric embeddings at initialization
Figure 3 for On the impact of activation and normalization in obtaining isometric embeddings at initialization
Figure 4 for On the impact of activation and normalization in obtaining isometric embeddings at initialization
Viaarxiv icon

Efficient displacement convex optimization with particle gradient descent

Add code
Bookmark button
Alert button
Feb 09, 2023
Hadi Daneshmand, Jason D. Lee, Chi Jin

Figure 1 for Efficient displacement convex optimization with particle gradient descent
Figure 2 for Efficient displacement convex optimization with particle gradient descent
Figure 3 for Efficient displacement convex optimization with particle gradient descent
Figure 4 for Efficient displacement convex optimization with particle gradient descent
Viaarxiv icon

Entropy Maximization with Depth: A Variational Principle for Random Neural Networks

Add code
Bookmark button
Alert button
May 25, 2022
Amir Joudaki, Hadi Daneshmand, Francis Bach

Figure 1 for Entropy Maximization with Depth: A Variational Principle for Random Neural Networks
Viaarxiv icon

Polynomial-time sparse measure recovery

Add code
Bookmark button
Alert button
Apr 16, 2022
Hadi Daneshmand, Francis Bach

Figure 1 for Polynomial-time sparse measure recovery
Figure 2 for Polynomial-time sparse measure recovery
Figure 3 for Polynomial-time sparse measure recovery
Viaarxiv icon

Batch Normalization Orthogonalizes Representations in Deep Random Networks

Add code
Bookmark button
Alert button
Jun 07, 2021
Hadi Daneshmand, Amir Joudaki, Francis Bach

Figure 1 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 2 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 3 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 4 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Viaarxiv icon

Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization

Add code
Bookmark button
Alert button
Feb 23, 2021
Peiyuan Zhang, Antonio Orvieto, Hadi Daneshmand, Thomas Hofmann, Roy Smith

Figure 1 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 2 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 3 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 4 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Viaarxiv icon

Theoretical Understanding of Batch-normalization: A Markov Chain Perspective

Add code
Bookmark button
Alert button
Mar 09, 2020
Hadi Daneshmand, Jonas Kohler, Francis Bach, Thomas Hofmann, Aurelien Lucchi

Figure 1 for Theoretical Understanding of Batch-normalization: A Markov Chain Perspective
Figure 2 for Theoretical Understanding of Batch-normalization: A Markov Chain Perspective
Figure 3 for Theoretical Understanding of Batch-normalization: A Markov Chain Perspective
Figure 4 for Theoretical Understanding of Batch-normalization: A Markov Chain Perspective
Viaarxiv icon