Alert button
Picture for Aidan N. Gomez

Aidan N. Gomez

Alert button

Exploring Low Rank Training of Deep Neural Networks

Add code
Bookmark button
Alert button
Sep 27, 2022
Siddhartha Rao Kamalakara, Acyr Locatelli, Bharat Venkitesh, Jimmy Ba, Yarin Gal, Aidan N. Gomez

Figure 1 for Exploring Low Rank Training of Deep Neural Networks
Figure 2 for Exploring Low Rank Training of Deep Neural Networks
Figure 3 for Exploring Low Rank Training of Deep Neural Networks
Figure 4 for Exploring Low Rank Training of Deep Neural Networks
Viaarxiv icon

Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt

Add code
Bookmark button
Alert button
Jun 16, 2022
Sören Mindermann, Jan Brauner, Muhammed Razzak, Mrinank Sharma, Andreas Kirsch, Winnie Xu, Benedikt Höltgen, Aidan N. Gomez, Adrien Morisot, Sebastian Farquhar, Yarin Gal

Figure 1 for Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt
Figure 2 for Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt
Figure 3 for Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt
Figure 4 for Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt
Viaarxiv icon

Prioritized training on points that are learnable, worth learning, and not yet learned

Add code
Bookmark button
Alert button
Jul 06, 2021
Sören Mindermann, Muhammed Razzak, Winnie Xu, Andreas Kirsch, Mrinank Sharma, Adrien Morisot, Aidan N. Gomez, Sebastian Farquhar, Jan Brauner, Yarin Gal

Figure 1 for Prioritized training on points that are learnable, worth learning, and not yet learned
Figure 2 for Prioritized training on points that are learnable, worth learning, and not yet learned
Figure 3 for Prioritized training on points that are learnable, worth learning, and not yet learned
Figure 4 for Prioritized training on points that are learnable, worth learning, and not yet learned
Viaarxiv icon

Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning

Add code
Bookmark button
Alert button
Jun 04, 2021
Jannik Kossen, Neil Band, Clare Lyle, Aidan N. Gomez, Tom Rainforth, Yarin Gal

Figure 1 for Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning
Figure 2 for Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning
Figure 3 for Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning
Figure 4 for Self-Attention Between Datapoints: Going Beyond Individual Input-Output Pairs in Deep Learning
Viaarxiv icon

Robustness to Pruning Predicts Generalization in Deep Neural Networks

Add code
Bookmark button
Alert button
Mar 10, 2021
Lorenz Kuhn, Clare Lyle, Aidan N. Gomez, Jonas Rothfuss, Yarin Gal

Figure 1 for Robustness to Pruning Predicts Generalization in Deep Neural Networks
Figure 2 for Robustness to Pruning Predicts Generalization in Deep Neural Networks
Figure 3 for Robustness to Pruning Predicts Generalization in Deep Neural Networks
Figure 4 for Robustness to Pruning Predicts Generalization in Deep Neural Networks
Viaarxiv icon

Interlocking Backpropagation: Improving depthwise model-parallelism

Add code
Bookmark button
Alert button
Oct 08, 2020
Aidan N. Gomez, Oscar Key, Stephen Gou, Nick Frosst, Jeff Dean, Yarin Gal

Figure 1 for Interlocking Backpropagation: Improving depthwise model-parallelism
Figure 2 for Interlocking Backpropagation: Improving depthwise model-parallelism
Figure 3 for Interlocking Backpropagation: Improving depthwise model-parallelism
Figure 4 for Interlocking Backpropagation: Improving depthwise model-parallelism
Viaarxiv icon

SliceOut: Training Transformers and CNNs faster while using less memory

Add code
Bookmark button
Alert button
Jul 21, 2020
Pascal Notin, Aidan N. Gomez, Joanna Yoo, Yarin Gal

Figure 1 for SliceOut: Training Transformers and CNNs faster while using less memory
Figure 2 for SliceOut: Training Transformers and CNNs faster while using less memory
Figure 3 for SliceOut: Training Transformers and CNNs faster while using less memory
Figure 4 for SliceOut: Training Transformers and CNNs faster while using less memory
Viaarxiv icon

Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers

Add code
Bookmark button
Alert button
Jun 08, 2020
Tim Z. Xiao, Aidan N. Gomez, Yarin Gal

Figure 1 for Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
Figure 2 for Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
Figure 3 for Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
Figure 4 for Wat zei je? Detecting Out-of-Distribution Translations with Variational Transformers
Viaarxiv icon

A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks

Add code
Bookmark button
Alert button
Dec 22, 2019
Angelos Filos, Sebastian Farquhar, Aidan N. Gomez, Tim G. J. Rudner, Zachary Kenton, Lewis Smith, Milad Alizadeh, Arnoud de Kroon, Yarin Gal

Figure 1 for A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks
Figure 2 for A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks
Figure 3 for A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks
Figure 4 for A Systematic Comparison of Bayesian Deep Learning Robustness in Diabetic Retinopathy Tasks
Viaarxiv icon

Learning Sparse Networks Using Targeted Dropout

Add code
Bookmark button
Alert button
Jun 05, 2019
Aidan N. Gomez, Ivan Zhang, Kevin Swersky, Yarin Gal, Geoffrey E. Hinton

Figure 1 for Learning Sparse Networks Using Targeted Dropout
Figure 2 for Learning Sparse Networks Using Targeted Dropout
Figure 3 for Learning Sparse Networks Using Targeted Dropout
Figure 4 for Learning Sparse Networks Using Targeted Dropout
Viaarxiv icon