Picture for Volodymyr Mnih

Volodymyr Mnih

Unsupervised Control Through Non-Parametric Discriminative Rewards

Add code
Nov 28, 2018
Figure 1 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 2 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 3 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 4 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Viaarxiv icon

The Uncertainty Bellman Equation and Exploration

Add code
Oct 22, 2018
Figure 1 for The Uncertainty Bellman Equation and Exploration
Figure 2 for The Uncertainty Bellman Equation and Exploration
Figure 3 for The Uncertainty Bellman Equation and Exploration
Figure 4 for The Uncertainty Bellman Equation and Exploration
Viaarxiv icon

Learning by Playing - Solving Sparse Reward Tasks from Scratch

Add code
Feb 28, 2018
Figure 1 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 2 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 3 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 4 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Viaarxiv icon

Sample Efficient Actor-Critic with Experience Replay

Add code
Jul 10, 2017
Figure 1 for Sample Efficient Actor-Critic with Experience Replay
Figure 2 for Sample Efficient Actor-Critic with Experience Replay
Figure 3 for Sample Efficient Actor-Critic with Experience Replay
Figure 4 for Sample Efficient Actor-Critic with Experience Replay
Viaarxiv icon

Combining policy gradient and Q-learning

Add code
Apr 07, 2017
Figure 1 for Combining policy gradient and Q-learning
Figure 2 for Combining policy gradient and Q-learning
Figure 3 for Combining policy gradient and Q-learning
Figure 4 for Combining policy gradient and Q-learning
Viaarxiv icon

Using Fast Weights to Attend to the Recent Past

Add code
Dec 05, 2016
Figure 1 for Using Fast Weights to Attend to the Recent Past
Figure 2 for Using Fast Weights to Attend to the Recent Past
Figure 3 for Using Fast Weights to Attend to the Recent Past
Figure 4 for Using Fast Weights to Attend to the Recent Past
Viaarxiv icon

Reinforcement Learning with Unsupervised Auxiliary Tasks

Add code
Nov 16, 2016
Figure 1 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 2 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 3 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 4 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Viaarxiv icon

Learning values across many orders of magnitude

Add code
Aug 16, 2016
Figure 1 for Learning values across many orders of magnitude
Figure 2 for Learning values across many orders of magnitude
Figure 3 for Learning values across many orders of magnitude
Viaarxiv icon

Asynchronous Methods for Deep Reinforcement Learning

Add code
Jun 16, 2016
Figure 1 for Asynchronous Methods for Deep Reinforcement Learning
Figure 2 for Asynchronous Methods for Deep Reinforcement Learning
Figure 3 for Asynchronous Methods for Deep Reinforcement Learning
Figure 4 for Asynchronous Methods for Deep Reinforcement Learning
Viaarxiv icon

Strategic Attentive Writer for Learning Macro-Actions

Add code
Jun 15, 2016
Figure 1 for Strategic Attentive Writer for Learning Macro-Actions
Figure 2 for Strategic Attentive Writer for Learning Macro-Actions
Figure 3 for Strategic Attentive Writer for Learning Macro-Actions
Figure 4 for Strategic Attentive Writer for Learning Macro-Actions
Viaarxiv icon