Alert button
Picture for Will Dabney

Will Dabney

Alert button

Settling the Reward Hypothesis

Add code
Bookmark button
Alert button
Dec 20, 2022
Michael Bowling, John D. Martin, David Abel, Will Dabney

Figure 1 for Settling the Reward Hypothesis
Figure 2 for Settling the Reward Hypothesis
Viaarxiv icon

Understanding Self-Predictive Learning for Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 06, 2022
Yunhao Tang, Zhaohan Daniel Guo, Pierre Harvey Richemond, Bernardo Ávila Pires, Yash Chandak, Rémi Munos, Mark Rowland, Mohammad Gheshlaghi Azar, Charline Le Lan, Clare Lyle, András György, Shantanu Thakoor, Will Dabney, Bilal Piot, Daniele Calandriello, Michal Valko

Figure 1 for Understanding Self-Predictive Learning for Reinforcement Learning
Figure 2 for Understanding Self-Predictive Learning for Reinforcement Learning
Figure 3 for Understanding Self-Predictive Learning for Reinforcement Learning
Figure 4 for Understanding Self-Predictive Learning for Reinforcement Learning
Viaarxiv icon

The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 15, 2022
Yunhao Tang, Mark Rowland, Rémi Munos, Bernardo Ávila Pires, Will Dabney, Marc G. Bellemare

Figure 1 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 2 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 3 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 4 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Viaarxiv icon

Generalised Policy Improvement with Geometric Policy Composition

Add code
Bookmark button
Alert button
Jun 17, 2022
Shantanu Thakoor, Mark Rowland, Diana Borsa, Will Dabney, Rémi Munos, André Barreto

Figure 1 for Generalised Policy Improvement with Geometric Policy Composition
Figure 2 for Generalised Policy Improvement with Geometric Policy Composition
Figure 3 for Generalised Policy Improvement with Geometric Policy Composition
Figure 4 for Generalised Policy Improvement with Geometric Policy Composition
Viaarxiv icon

Learning Dynamics and Generalization in Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 05, 2022
Clare Lyle, Mark Rowland, Will Dabney, Marta Kwiatkowska, Yarin Gal

Figure 1 for Learning Dynamics and Generalization in Reinforcement Learning
Figure 2 for Learning Dynamics and Generalization in Reinforcement Learning
Figure 3 for Learning Dynamics and Generalization in Reinforcement Learning
Figure 4 for Learning Dynamics and Generalization in Reinforcement Learning
Viaarxiv icon

Understanding and Preventing Capacity Loss in Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 20, 2022
Clare Lyle, Mark Rowland, Will Dabney

Figure 1 for Understanding and Preventing Capacity Loss in Reinforcement Learning
Figure 2 for Understanding and Preventing Capacity Loss in Reinforcement Learning
Figure 3 for Understanding and Preventing Capacity Loss in Reinforcement Learning
Figure 4 for Understanding and Preventing Capacity Loss in Reinforcement Learning
Viaarxiv icon

On the Expressivity of Markov Reward

Add code
Bookmark button
Alert button
Nov 01, 2021
David Abel, Will Dabney, Anna Harutyunyan, Mark K. Ho, Michael L. Littman, Doina Precup, Satinder Singh

Figure 1 for On the Expressivity of Markov Reward
Figure 2 for On the Expressivity of Markov Reward
Figure 3 for On the Expressivity of Markov Reward
Figure 4 for On the Expressivity of Markov Reward
Viaarxiv icon

The Difficulty of Passive Learning in Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 26, 2021
Georg Ostrovski, Pablo Samuel Castro, Will Dabney

Figure 1 for The Difficulty of Passive Learning in Deep Reinforcement Learning
Figure 2 for The Difficulty of Passive Learning in Deep Reinforcement Learning
Figure 3 for The Difficulty of Passive Learning in Deep Reinforcement Learning
Figure 4 for The Difficulty of Passive Learning in Deep Reinforcement Learning
Viaarxiv icon

Revisiting Peng's Q($λ$) for Modern Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 27, 2021
Tadashi Kozuno, Yunhao Tang, Mark Rowland, Rémi Munos, Steven Kapturowski, Will Dabney, Michal Valko, David Abel

Figure 1 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 2 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 3 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 4 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Viaarxiv icon

On The Effect of Auxiliary Tasks on Representation Dynamics

Add code
Bookmark button
Alert button
Feb 25, 2021
Clare Lyle, Mark Rowland, Georg Ostrovski, Will Dabney

Figure 1 for On The Effect of Auxiliary Tasks on Representation Dynamics
Figure 2 for On The Effect of Auxiliary Tasks on Representation Dynamics
Figure 3 for On The Effect of Auxiliary Tasks on Representation Dynamics
Figure 4 for On The Effect of Auxiliary Tasks on Representation Dynamics
Viaarxiv icon