Alert button
Picture for Jascha Sohl-Dickstein

Jascha Sohl-Dickstein

Alert button

Towards NNGP-guided Neural Architecture Search

Add code
Bookmark button
Alert button
Nov 11, 2020
Daniel S. Park, Jaehoon Lee, Daiyi Peng, Yuan Cao, Jascha Sohl-Dickstein

Figure 1 for Towards NNGP-guided Neural Architecture Search
Figure 2 for Towards NNGP-guided Neural Architecture Search
Figure 3 for Towards NNGP-guided Neural Architecture Search
Figure 4 for Towards NNGP-guided Neural Architecture Search
Viaarxiv icon

Reverse engineering learned optimizers reveals known and novel mechanisms

Add code
Bookmark button
Alert button
Nov 04, 2020
Niru Maheswaranathan, David Sussillo, Luke Metz, Ruoxi Sun, Jascha Sohl-Dickstein

Figure 1 for Reverse engineering learned optimizers reveals known and novel mechanisms
Figure 2 for Reverse engineering learned optimizers reveals known and novel mechanisms
Figure 3 for Reverse engineering learned optimizers reveals known and novel mechanisms
Figure 4 for Reverse engineering learned optimizers reveals known and novel mechanisms
Viaarxiv icon

Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence

Add code
Bookmark button
Alert button
Oct 21, 2020
Vinay Rao, Jascha Sohl-Dickstein

Figure 1 for Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence
Figure 2 for Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence
Figure 3 for Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence
Figure 4 for Is Batch Norm unique? An empirical investigation and prescription to emulate the best properties of common normalizers without batch dependence
Viaarxiv icon

Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves

Add code
Bookmark button
Alert button
Sep 23, 2020
Luke Metz, Niru Maheswaranathan, C. Daniel Freeman, Ben Poole, Jascha Sohl-Dickstein

Figure 1 for Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves
Figure 2 for Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves
Figure 3 for Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves
Figure 4 for Tasks, stability, architecture, and compute: Training more effective learned optimizers, and using them to train themselves
Viaarxiv icon

Finite Versus Infinite Neural Networks: an Empirical Study

Add code
Bookmark button
Alert button
Sep 08, 2020
Jaehoon Lee, Samuel S. Schoenholz, Jeffrey Pennington, Ben Adlam, Lechao Xiao, Roman Novak, Jascha Sohl-Dickstein

Figure 1 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 2 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 3 for Finite Versus Infinite Neural Networks: an Empirical Study
Figure 4 for Finite Versus Infinite Neural Networks: an Empirical Study
Viaarxiv icon

Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible

Add code
Bookmark button
Alert button
Aug 25, 2020
Neha S. Wadia, Daniel Duckworth, Samuel S. Schoenholz, Ethan Dyer, Jascha Sohl-Dickstein

Figure 1 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 2 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 3 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Figure 4 for Whitening and second order optimization both destroy information about the dataset, and can make generalization impossible
Viaarxiv icon

A new method for parameter estimation in probabilistic models: Minimum probability flow

Add code
Bookmark button
Alert button
Jul 17, 2020
Jascha Sohl-Dickstein, Peter Battaglino, Michael R. DeWeese

Figure 1 for A new method for parameter estimation in probabilistic models: Minimum probability flow
Figure 2 for A new method for parameter estimation in probabilistic models: Minimum probability flow
Figure 3 for A new method for parameter estimation in probabilistic models: Minimum probability flow
Viaarxiv icon

Exact posterior distributions of wide Bayesian neural networks

Add code
Bookmark button
Alert button
Jun 18, 2020
Jiri Hron, Yasaman Bahri, Roman Novak, Jeffrey Pennington, Jascha Sohl-Dickstein

Figure 1 for Exact posterior distributions of wide Bayesian neural networks
Figure 2 for Exact posterior distributions of wide Bayesian neural networks
Viaarxiv icon