Alert button
Picture for Jeffrey Pennington

Jeffrey Pennington

Alert button

Disentangling trainability and generalization in deep learning

Add code
Bookmark button
Alert button
Dec 30, 2019
Lechao Xiao, Jeffrey Pennington, Samuel S. Schoenholz

Figure 1 for Disentangling trainability and generalization in deep learning
Figure 2 for Disentangling trainability and generalization in deep learning
Figure 3 for Disentangling trainability and generalization in deep learning
Figure 4 for Disentangling trainability and generalization in deep learning
Viaarxiv icon

A Random Matrix Perspective on Mixtures of Nonlinearities for Deep Learning

Add code
Bookmark button
Alert button
Dec 02, 2019
Ben Adlam, Jake Levinson, Jeffrey Pennington

Figure 1 for A Random Matrix Perspective on Mixtures of Nonlinearities for Deep Learning
Figure 2 for A Random Matrix Perspective on Mixtures of Nonlinearities for Deep Learning
Figure 3 for A Random Matrix Perspective on Mixtures of Nonlinearities for Deep Learning
Figure 4 for A Random Matrix Perspective on Mixtures of Nonlinearities for Deep Learning
Viaarxiv icon

A Mean Field Theory of Batch Normalization

Add code
Bookmark button
Alert button
Mar 05, 2019
Greg Yang, Jeffrey Pennington, Vinay Rao, Jascha Sohl-Dickstein, Samuel S. Schoenholz

Figure 1 for A Mean Field Theory of Batch Normalization
Figure 2 for A Mean Field Theory of Batch Normalization
Figure 3 for A Mean Field Theory of Batch Normalization
Figure 4 for A Mean Field Theory of Batch Normalization
Viaarxiv icon

Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent

Add code
Bookmark button
Alert button
Feb 18, 2019
Jaehoon Lee, Lechao Xiao, Samuel S. Schoenholz, Yasaman Bahri, Jascha Sohl-Dickstein, Jeffrey Pennington

Figure 1 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 2 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 3 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Figure 4 for Wide Neural Networks of Any Depth Evolve as Linear Models Under Gradient Descent
Viaarxiv icon

Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs

Add code
Bookmark button
Alert button
Jan 25, 2019
Dar Gilboa, Bo Chang, Minmin Chen, Greg Yang, Samuel S. Schoenholz, Ed H. Chi, Jeffrey Pennington

Figure 1 for Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs
Figure 2 for Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs
Figure 3 for Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs
Figure 4 for Dynamical Isometry and a Mean Field Theory of LSTMs and GRUs
Viaarxiv icon

Bayesian Convolutional Neural Networks with Many Channels are Gaussian Processes

Add code
Bookmark button
Alert button
Oct 11, 2018
Roman Novak, Lechao Xiao, Jaehoon Lee, Yasaman Bahri, Daniel A. Abolafia, Jeffrey Pennington, Jascha Sohl-Dickstein

Figure 1 for Bayesian Convolutional Neural Networks with Many Channels are Gaussian Processes
Figure 2 for Bayesian Convolutional Neural Networks with Many Channels are Gaussian Processes
Figure 3 for Bayesian Convolutional Neural Networks with Many Channels are Gaussian Processes
Figure 4 for Bayesian Convolutional Neural Networks with Many Channels are Gaussian Processes
Viaarxiv icon

Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks

Add code
Bookmark button
Alert button
Aug 15, 2018
Minmin Chen, Jeffrey Pennington, Samuel S. Schoenholz

Figure 1 for Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks
Figure 2 for Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks
Figure 3 for Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks
Figure 4 for Dynamical Isometry and a Mean Field Theory of RNNs: Gating Enables Signal Propagation in Recurrent Neural Networks
Viaarxiv icon

Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks

Add code
Bookmark button
Alert button
Jul 10, 2018
Lechao Xiao, Yasaman Bahri, Jascha Sohl-Dickstein, Samuel S. Schoenholz, Jeffrey Pennington

Figure 1 for Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks
Figure 2 for Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks
Figure 3 for Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks
Figure 4 for Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10,000-Layer Vanilla Convolutional Neural Networks
Viaarxiv icon

Sensitivity and Generalization in Neural Networks: an Empirical Study

Add code
Bookmark button
Alert button
Jun 18, 2018
Roman Novak, Yasaman Bahri, Daniel A. Abolafia, Jeffrey Pennington, Jascha Sohl-Dickstein

Figure 1 for Sensitivity and Generalization in Neural Networks: an Empirical Study
Figure 2 for Sensitivity and Generalization in Neural Networks: an Empirical Study
Figure 3 for Sensitivity and Generalization in Neural Networks: an Empirical Study
Figure 4 for Sensitivity and Generalization in Neural Networks: an Empirical Study
Viaarxiv icon