Alert button
Picture for David Silver

David Silver

Alert button

Discovery of Options via Meta-Learned Subgoals

Add code
Bookmark button
Alert button
Feb 12, 2021
Vivek Veeriah, Tom Zahavy, Matteo Hessel, Zhongwen Xu, Junhyuk Oh, Iurii Kemaev, Hado van Hasselt, David Silver, Satinder Singh

Figure 1 for Discovery of Options via Meta-Learned Subgoals
Figure 2 for Discovery of Options via Meta-Learned Subgoals
Figure 3 for Discovery of Options via Meta-Learned Subgoals
Figure 4 for Discovery of Options via Meta-Learned Subgoals
Viaarxiv icon

The Value Equivalence Principle for Model-Based Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 06, 2020
Christopher Grimm, André Barreto, Satinder Singh, David Silver

Figure 1 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 2 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 3 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 4 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Viaarxiv icon

Discovering Reinforcement Learning Algorithms

Add code
Bookmark button
Alert button
Jul 17, 2020
Junhyuk Oh, Matteo Hessel, Wojciech M. Czarnecki, Zhongwen Xu, Hado van Hasselt, Satinder Singh, David Silver

Figure 1 for Discovering Reinforcement Learning Algorithms
Figure 2 for Discovering Reinforcement Learning Algorithms
Figure 3 for Discovering Reinforcement Learning Algorithms
Figure 4 for Discovering Reinforcement Learning Algorithms
Viaarxiv icon

Meta-Gradient Reinforcement Learning with an Objective Discovered Online

Add code
Bookmark button
Alert button
Jul 16, 2020
Zhongwen Xu, Hado van Hasselt, Matteo Hessel, Junhyuk Oh, Satinder Singh, David Silver

Figure 1 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 2 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 3 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 4 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Viaarxiv icon

Expected Eligibility Traces

Add code
Bookmark button
Alert button
Jul 03, 2020
Hado van Hasselt, Sephora Madjiheurem, Matteo Hessel, David Silver, André Barreto, Diana Borsa

Figure 1 for Expected Eligibility Traces
Figure 2 for Expected Eligibility Traces
Figure 3 for Expected Eligibility Traces
Figure 4 for Expected Eligibility Traces
Viaarxiv icon

The Value-Improvement Path: Towards Better Representations for Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 03, 2020
Will Dabney, André Barreto, Mark Rowland, Robert Dadashi, John Quan, Marc G. Bellemare, David Silver

Figure 1 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 2 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 3 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 4 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Viaarxiv icon

Self-Tuning Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 02, 2020
Tom Zahavy, Zhongwen Xu, Vivek Veeriah, Matteo Hessel, Junhyuk Oh, Hado van Hasselt, David Silver, Satinder Singh

Figure 1 for Self-Tuning Deep Reinforcement Learning
Figure 2 for Self-Tuning Deep Reinforcement Learning
Figure 3 for Self-Tuning Deep Reinforcement Learning
Figure 4 for Self-Tuning Deep Reinforcement Learning
Viaarxiv icon

Value-driven Hindsight Modelling

Add code
Bookmark button
Alert button
Feb 19, 2020
Arthur Guez, Fabio Viola, Théophane Weber, Lars Buesing, Steven Kapturowski, Doina Precup, David Silver, Nicolas Heess

Figure 1 for Value-driven Hindsight Modelling
Figure 2 for Value-driven Hindsight Modelling
Figure 3 for Value-driven Hindsight Modelling
Figure 4 for Value-driven Hindsight Modelling
Viaarxiv icon

What Can Learned Intrinsic Rewards Capture?

Add code
Bookmark button
Alert button
Dec 11, 2019
Zeyu Zheng, Junhyuk Oh, Matteo Hessel, Zhongwen Xu, Manuel Kroiss, Hado van Hasselt, David Silver, Satinder Singh

Figure 1 for What Can Learned Intrinsic Rewards Capture?
Figure 2 for What Can Learned Intrinsic Rewards Capture?
Figure 3 for What Can Learned Intrinsic Rewards Capture?
Figure 4 for What Can Learned Intrinsic Rewards Capture?
Viaarxiv icon