Alert button
Picture for Jascha Sohl-Dickstein

Jascha Sohl-Dickstein

Alert button

Infinite attention: NNGP and NTK for deep attention networks

Add code
Bookmark button
Alert button
Jun 18, 2020
Jiri Hron, Yasaman Bahri, Jascha Sohl-Dickstein, Roman Novak

Figure 1 for Infinite attention: NNGP and NTK for deep attention networks
Figure 2 for Infinite attention: NNGP and NTK for deep attention networks
Figure 3 for Infinite attention: NNGP and NTK for deep attention networks
Figure 4 for Infinite attention: NNGP and NTK for deep attention networks
Viaarxiv icon

Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling

Add code
Bookmark button
Alert button
Mar 24, 2020
Tong Che, Ruixiang Zhang, Jascha Sohl-Dickstein, Hugo Larochelle, Liam Paull, Yuan Cao, Yoshua Bengio

Figure 1 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 2 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 3 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Figure 4 for Your GAN is Secretly an Energy-based Model and You Should use Discriminator Driven Latent Sampling
Viaarxiv icon

Using a thousand optimization tasks to learn hyperparameter search strategies

Add code
Bookmark button
Alert button
Mar 11, 2020
Luke Metz, Niru Maheswaranathan, Ruoxi Sun, C. Daniel Freeman, Ben Poole, Jascha Sohl-Dickstein

Figure 1 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 2 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 3 for Using a thousand optimization tasks to learn hyperparameter search strategies
Figure 4 for Using a thousand optimization tasks to learn hyperparameter search strategies
Viaarxiv icon

The large learning rate phase of deep learning: the catapult mechanism

Add code
Bookmark button
Alert button
Mar 04, 2020
Aitor Lewkowycz, Yasaman Bahri, Ethan Dyer, Jascha Sohl-Dickstein, Guy Gur-Ari

Figure 1 for The large learning rate phase of deep learning: the catapult mechanism
Figure 2 for The large learning rate phase of deep learning: the catapult mechanism
Figure 3 for The large learning rate phase of deep learning: the catapult mechanism
Figure 4 for The large learning rate phase of deep learning: the catapult mechanism
Viaarxiv icon

On the infinite width limit of neural networks with a standard parameterization

Add code
Bookmark button
Alert button
Jan 25, 2020
Jascha Sohl-Dickstein, Roman Novak, Samuel S. Schoenholz, Jaehoon Lee

Figure 1 for On the infinite width limit of neural networks with a standard parameterization
Figure 2 for On the infinite width limit of neural networks with a standard parameterization
Figure 3 for On the infinite width limit of neural networks with a standard parameterization
Figure 4 for On the infinite width limit of neural networks with a standard parameterization
Viaarxiv icon

Neural Tangents: Fast and Easy Infinite Neural Networks in Python

Add code
Bookmark button
Alert button
Dec 05, 2019
Roman Novak, Lechao Xiao, Jiri Hron, Jaehoon Lee, Alexander A. Alemi, Jascha Sohl-Dickstein, Samuel S. Schoenholz

Figure 1 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 2 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 3 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Figure 4 for Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Viaarxiv icon

Neural reparameterization improves structural optimization

Add code
Bookmark button
Alert button
Sep 14, 2019
Stephan Hoyer, Jascha Sohl-Dickstein, Sam Greydanus

Figure 1 for Neural reparameterization improves structural optimization
Figure 2 for Neural reparameterization improves structural optimization
Figure 3 for Neural reparameterization improves structural optimization
Figure 4 for Neural reparameterization improves structural optimization
Viaarxiv icon