Picture for Jascha Sohl-Dickstein

Jascha Sohl-Dickstein

Shammie

Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling

Add code
Mar 24, 2020
Figure 1 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 2 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 3 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 4 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Viaarxiv icon

Using a thousand optimization tasks to learn hyperparameter search strategies

Add code
Mar 11, 2020
Figure 1 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 2 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 3 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 4 for Using a thousand optimization tasks to learn hyperparameter search strategies
Viaarxiv icon

The large learning rate phase of deep learning: the catapult mechanism

Add code
Mar 04, 2020
Figure 1 for The large learning rate phase of deep learning: the catapult mechanism
Figure 2 for The large learning rate phase of deep learning: the catapult mechanism
Figure 3 for The large learning rate phase of deep learning: the catapult mechanism
Figure 4 for The large learning rate phase of deep learning: the catapult mechanism
Viaarxiv icon

On the infinite width limit of neural networks with a standard parameterization

Add code
Jan 25, 2020
Figure 1 for On the infinite width limit of neural networks with a standard parameterization
Figure 2 for On the infinite width limit of neural networks with a standard parameterization
Figure 3 for On the infinite width limit of neural networks with a standard parameterization
Figure 4 for On the infinite width limit of neural networks with a standard parameterization
Viaarxiv icon

Neural Tangents: Fast and Easy Infinite Neural Networks in Python

Add code
Dec 05, 2019
Figure 1 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 2 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 3 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 4 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Viaarxiv icon

Neural reparameterization improves structural optimization

Add code
Sep 14, 2019
Figure 1 for Neural reparameterization improves structural optimization
Figure 2 for Neural reparameterization improves structural optimization
Figure 3 for Neural reparameterization improves structural optimization
Figure 4 for Neural reparameterization improves structural optimization
Viaarxiv icon

Using learned optimizers to make models robust to input noise

Add code
Jun 08, 2019
Figure 1 for Using learned optimizers to make models robust to input noise
Figure 2 for Using learned optimizers to make models robust to input noise
Figure 3 for Using learned optimizers to make models robust to input noise
Figure 4 for Using learned optimizers to make models robust to input noise
Viaarxiv icon

The Effect of Network Width on Stochastic Gradient Descent and Generalization: an Empirical Study

Add code
May 09, 2019
Figure 1 for The Effect of Network Width on Stochastic Gradient Descent and Generalization: an Empirical Study
Figure 2 for The Effect of Network Width on Stochastic Gradient Descent and Generalization: an Empirical Study
Figure 3 for The Effect of Network Width on Stochastic Gradient Descent and Generalization: an Empirical Study
Figure 4 for The Effect of Network Width on Stochastic Gradient Descent and Generalization: an Empirical Study
Viaarxiv icon

A RAD approach to deep mixture models

Add code
Mar 18, 2019
Figure 1 for A RAD approach to deep mixture models
Figure 2 for A RAD approach to deep mixture models
Figure 3 for A RAD approach to deep mixture models
Figure 4 for A RAD approach to deep mixture models
Viaarxiv icon

A Mean Field Theory of Batch Normalization

Add code
Mar 05, 2019
Figure 1 for A Mean Field Theory of Batch Normalization
Figure 2 for A Mean Field Theory of Batch Normalization
Figure 3 for A Mean Field Theory of Batch Normalization
Figure 4 for A Mean Field Theory of Batch Normalization
Viaarxiv icon