Alert button
Picture for Boris Hanin

Boris Hanin

Alert button

Maximal Initial Learning Rates in Deep ReLU Networks

Add code
Bookmark button
Alert button
Dec 14, 2022
Gaurav Iyer, Boris Hanin, David Rolnick

Figure 1 for Maximal Initial Learning Rates in Deep ReLU Networks
Figure 2 for Maximal Initial Learning Rates in Deep ReLU Networks
Figure 3 for Maximal Initial Learning Rates in Deep ReLU Networks
Figure 4 for Maximal Initial Learning Rates in Deep ReLU Networks
Viaarxiv icon

Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis

Add code
Bookmark button
Alert button
May 11, 2022
Wuyang Chen, Wei Huang, Xinyu Gong, Boris Hanin, Zhangyang Wang

Figure 1 for Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis
Figure 2 for Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis
Figure 3 for Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis
Figure 4 for Deep Architecture Connectivity Matters for Its Convergence: A Fine-Grained Analysis
Viaarxiv icon

Correlation Functions in Random Fully Connected Neural Networks at Finite Width

Add code
Bookmark button
Alert button
Apr 03, 2022
Boris Hanin

Viaarxiv icon

Ridgeless Interpolation with Shallow ReLU Networks in $1D$ is Nearest Neighbor Curvature Extrapolation and Provably Generalizes on Lipschitz Functions

Add code
Bookmark button
Alert button
Sep 27, 2021
Boris Hanin

Figure 1 for Ridgeless Interpolation with Shallow ReLU Networks in $1D$ is Nearest Neighbor Curvature Extrapolation and Provably Generalizes on Lipschitz Functions
Figure 2 for Ridgeless Interpolation with Shallow ReLU Networks in $1D$ is Nearest Neighbor Curvature Extrapolation and Provably Generalizes on Lipschitz Functions
Figure 3 for Ridgeless Interpolation with Shallow ReLU Networks in $1D$ is Nearest Neighbor Curvature Extrapolation and Provably Generalizes on Lipschitz Functions
Figure 4 for Ridgeless Interpolation with Shallow ReLU Networks in $1D$ is Nearest Neighbor Curvature Extrapolation and Provably Generalizes on Lipschitz Functions
Viaarxiv icon

Random Neural Networks in the Infinite Width Limit as Gaussian Processes

Add code
Bookmark button
Alert button
Jul 04, 2021
Boris Hanin

Viaarxiv icon

The Principles of Deep Learning Theory

Add code
Bookmark button
Alert button
Jun 18, 2021
Daniel A. Roberts, Sho Yaida, Boris Hanin

Viaarxiv icon

Deep ReLU Networks Preserve Expected Length

Add code
Bookmark button
Alert button
Feb 21, 2021
Boris Hanin, Ryan Jeong, David Rolnick

Figure 1 for Deep ReLU Networks Preserve Expected Length
Figure 2 for Deep ReLU Networks Preserve Expected Length
Figure 3 for Deep ReLU Networks Preserve Expected Length
Figure 4 for Deep ReLU Networks Preserve Expected Length
Viaarxiv icon

Data augmentation as stochastic optimization

Add code
Bookmark button
Alert button
Oct 21, 2020
Boris Hanin, Yi Sun

Figure 1 for Data augmentation as stochastic optimization
Figure 2 for Data augmentation as stochastic optimization
Viaarxiv icon

Finite Depth and Width Corrections to the Neural Tangent Kernel

Add code
Bookmark button
Alert button
Sep 13, 2019
Boris Hanin, Mihai Nica

Figure 1 for Finite Depth and Width Corrections to the Neural Tangent Kernel
Figure 2 for Finite Depth and Width Corrections to the Neural Tangent Kernel
Viaarxiv icon

Deep ReLU Networks Have Surprisingly Few Activation Patterns

Add code
Bookmark button
Alert button
Jun 03, 2019
Boris Hanin, David Rolnick

Figure 1 for Deep ReLU Networks Have Surprisingly Few Activation Patterns
Figure 2 for Deep ReLU Networks Have Surprisingly Few Activation Patterns
Figure 3 for Deep ReLU Networks Have Surprisingly Few Activation Patterns
Figure 4 for Deep ReLU Networks Have Surprisingly Few Activation Patterns
Viaarxiv icon