Alert button
Picture for Will Dabney

Will Dabney

Alert button

Counterfactual Credit Assignment in Model-Free Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 18, 2020
Thomas Mesnard, Théophane Weber, Fabio Viola, Shantanu Thakoor, Alaa Saade, Anna Harutyunyan, Will Dabney, Tom Stepleton, Nicolas Heess, Arthur Guez, Marcus Hutter, Lars Buesing, Rémi Munos

Figure 1 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 2 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 3 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 4 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Viaarxiv icon

Revisiting Fundamentals of Experience Replay

Add code
Bookmark button
Alert button
Jul 13, 2020
William Fedus, Prajit Ramachandran, Rishabh Agarwal, Yoshua Bengio, Hugo Larochelle, Mark Rowland, Will Dabney

Figure 1 for Revisiting Fundamentals of Experience Replay
Figure 2 for Revisiting Fundamentals of Experience Replay
Figure 3 for Revisiting Fundamentals of Experience Replay
Figure 4 for Revisiting Fundamentals of Experience Replay
Viaarxiv icon

Deep Reinforcement Learning and its Neuroscientific Implications

Add code
Bookmark button
Alert button
Jul 07, 2020
Matthew Botvinick, Jane X. Wang, Will Dabney, Kevin J. Miller, Zeb Kurth-Nelson

Figure 1 for Deep Reinforcement Learning and its Neuroscientific Implications
Figure 2 for Deep Reinforcement Learning and its Neuroscientific Implications
Figure 3 for Deep Reinforcement Learning and its Neuroscientific Implications
Figure 4 for Deep Reinforcement Learning and its Neuroscientific Implications
Viaarxiv icon

The Value-Improvement Path: Towards Better Representations for Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 03, 2020
Will Dabney, André Barreto, Mark Rowland, Robert Dadashi, John Quan, Marc G. Bellemare, David Silver

Figure 1 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 2 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 3 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 4 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Viaarxiv icon

Temporally-Extended ε-Greedy Exploration

Add code
Bookmark button
Alert button
Jun 02, 2020
Will Dabney, Georg Ostrovski, André Barreto

Figure 1 for Temporally-Extended ε-Greedy Exploration
Figure 2 for Temporally-Extended ε-Greedy Exploration
Figure 3 for Temporally-Extended ε-Greedy Exploration
Figure 4 for Temporally-Extended ε-Greedy Exploration
Viaarxiv icon

Adapting Behaviour for Learning Progress

Add code
Bookmark button
Alert button
Dec 14, 2019
Tom Schaul, Diana Borsa, David Ding, David Szepesvari, Georg Ostrovski, Will Dabney, Simon Osindero

Figure 1 for Adapting Behaviour for Learning Progress
Figure 2 for Adapting Behaviour for Learning Progress
Figure 3 for Adapting Behaviour for Learning Progress
Figure 4 for Adapting Behaviour for Learning Progress
Viaarxiv icon

Hindsight Credit Assignment

Add code
Bookmark button
Alert button
Dec 05, 2019
Anna Harutyunyan, Will Dabney, Thomas Mesnard, Mohammad Azar, Bilal Piot, Nicolas Heess, Hado van Hasselt, Greg Wayne, Satinder Singh, Doina Precup, Remi Munos

Figure 1 for Hindsight Credit Assignment
Figure 2 for Hindsight Credit Assignment
Figure 3 for Hindsight Credit Assignment
Figure 4 for Hindsight Credit Assignment
Viaarxiv icon

Conditional Importance Sampling for Off-Policy Learning

Add code
Bookmark button
Alert button
Oct 16, 2019
Mark Rowland, Anna Harutyunyan, Hado van Hasselt, Diana Borsa, Tom Schaul, Rémi Munos, Will Dabney

Figure 1 for Conditional Importance Sampling for Off-Policy Learning
Figure 2 for Conditional Importance Sampling for Off-Policy Learning
Figure 3 for Conditional Importance Sampling for Off-Policy Learning
Figure 4 for Conditional Importance Sampling for Off-Policy Learning
Viaarxiv icon

Adaptive Trade-Offs in Off-Policy Learning

Add code
Bookmark button
Alert button
Oct 16, 2019
Mark Rowland, Will Dabney, Rémi Munos

Figure 1 for Adaptive Trade-Offs in Off-Policy Learning
Figure 2 for Adaptive Trade-Offs in Off-Policy Learning
Figure 3 for Adaptive Trade-Offs in Off-Policy Learning
Figure 4 for Adaptive Trade-Offs in Off-Policy Learning
Viaarxiv icon

Fast Task Inference with Variational Intrinsic Successor Features

Add code
Bookmark button
Alert button
Jun 12, 2019
Steven Hansen, Will Dabney, Andre Barreto, Tom Van de Wiele, David Warde-Farley, Volodymyr Mnih

Figure 1 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 2 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 3 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 4 for Fast Task Inference with Variational Intrinsic Successor Features
Viaarxiv icon