Alert button
Picture for Sho Yaida

Sho Yaida

Alert button

Effective Theory of Transformers at Initialization

Add code
Bookmark button
Alert button
Apr 04, 2023
Emily Dinan, Sho Yaida, Susan Zhang

Figure 1 for Effective Theory of Transformers at Initialization
Figure 2 for Effective Theory of Transformers at Initialization
Figure 3 for Effective Theory of Transformers at Initialization
Figure 4 for Effective Theory of Transformers at Initialization
Viaarxiv icon

Meta-Principled Family of Hyperparameter Scaling Strategies

Add code
Bookmark button
Alert button
Oct 10, 2022
Sho Yaida

Viaarxiv icon

The Principles of Deep Learning Theory

Add code
Bookmark button
Alert button
Jun 18, 2021
Daniel A. Roberts, Sho Yaida, Boris Hanin

Viaarxiv icon

Non-Gaussian processes and neural networks at finite widths

Add code
Bookmark button
Alert button
Sep 30, 2019
Sho Yaida

Figure 1 for Non-Gaussian processes and neural networks at finite widths
Viaarxiv icon

Robust Learning with Jacobian Regularization

Add code
Bookmark button
Alert button
Aug 07, 2019
Judy Hoffman, Daniel A. Roberts, Sho Yaida

Figure 1 for Robust Learning with Jacobian Regularization
Figure 2 for Robust Learning with Jacobian Regularization
Figure 3 for Robust Learning with Jacobian Regularization
Figure 4 for Robust Learning with Jacobian Regularization
Viaarxiv icon

Fluctuation-dissipation relations for stochastic gradient descent

Add code
Bookmark button
Alert button
Sep 28, 2018
Sho Yaida

Figure 1 for Fluctuation-dissipation relations for stochastic gradient descent
Figure 2 for Fluctuation-dissipation relations for stochastic gradient descent
Figure 3 for Fluctuation-dissipation relations for stochastic gradient descent
Figure 4 for Fluctuation-dissipation relations for stochastic gradient descent
Viaarxiv icon