Picture for David Silver

David Silver

University College London

Muesli: Combining Improvements in Policy Optimization

Add code
Apr 13, 2021
Figure 1 for Muesli: Combining Improvements in Policy Optimization
Figure 2 for Muesli: Combining Improvements in Policy Optimization
Figure 3 for Muesli: Combining Improvements in Policy Optimization
Figure 4 for Muesli: Combining Improvements in Policy Optimization
Viaarxiv icon

Discovery of Options via Meta-Learned Subgoals

Add code
Feb 12, 2021
Figure 1 for Discovery of Options via Meta-Learned Subgoals
Figure 2 for Discovery of Options via Meta-Learned Subgoals
Figure 3 for Discovery of Options via Meta-Learned Subgoals
Figure 4 for Discovery of Options via Meta-Learned Subgoals
Viaarxiv icon

The Value Equivalence Principle for Model-Based Reinforcement Learning

Add code
Nov 06, 2020
Figure 1 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 2 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 3 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 4 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Viaarxiv icon

Discovering Reinforcement Learning Algorithms

Add code
Jul 17, 2020
Figure 1 for Discovering Reinforcement Learning Algorithms
Figure 2 for Discovering Reinforcement Learning Algorithms
Figure 3 for Discovering Reinforcement Learning Algorithms
Figure 4 for Discovering Reinforcement Learning Algorithms
Viaarxiv icon

Meta-Gradient Reinforcement Learning with an Objective Discovered Online

Add code
Jul 16, 2020
Figure 1 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 2 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 3 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 4 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Viaarxiv icon

Expected Eligibility Traces

Add code
Jul 03, 2020
Figure 1 for Expected Eligibility Traces
Figure 2 for Expected Eligibility Traces
Figure 3 for Expected Eligibility Traces
Figure 4 for Expected Eligibility Traces
Viaarxiv icon

The Value-Improvement Path: Towards Better Representations for Reinforcement Learning

Add code
Jun 03, 2020
Figure 1 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 2 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 3 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 4 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Viaarxiv icon

Self-Tuning Deep Reinforcement Learning

Add code
Mar 02, 2020
Figure 1 for Self-Tuning Deep Reinforcement Learning
Figure 2 for Self-Tuning Deep Reinforcement Learning
Figure 3 for Self-Tuning Deep Reinforcement Learning
Figure 4 for Self-Tuning Deep Reinforcement Learning
Viaarxiv icon

Value-driven Hindsight Modelling

Add code
Feb 19, 2020
Figure 1 for Value-driven Hindsight Modelling
Figure 2 for Value-driven Hindsight Modelling
Figure 3 for Value-driven Hindsight Modelling
Figure 4 for Value-driven Hindsight Modelling
Viaarxiv icon

What Can Learned Intrinsic Rewards Capture?

Add code
Dec 11, 2019
Figure 1 for What Can Learned Intrinsic Rewards Capture?
Figure 2 for What Can Learned Intrinsic Rewards Capture?
Figure 3 for What Can Learned Intrinsic Rewards Capture?
Figure 4 for What Can Learned Intrinsic Rewards Capture?
Viaarxiv icon