Alert button
Picture for Volodymyr Mnih

Volodymyr Mnih

Alert button

Learning by Playing - Solving Sparse Reward Tasks from Scratch

Add code
Bookmark button
Alert button
Feb 28, 2018
Martin Riedmiller, Roland Hafner, Thomas Lampe, Michael Neunert, Jonas Degrave, Tom Van de Wiele, Volodymyr Mnih, Nicolas Heess, Jost Tobias Springenberg

Figure 1 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 2 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 3 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Figure 4 for Learning by Playing - Solving Sparse Reward Tasks from Scratch
Viaarxiv icon

Sample Efficient Actor-Critic with Experience Replay

Add code
Bookmark button
Alert button
Jul 10, 2017
Ziyu Wang, Victor Bapst, Nicolas Heess, Volodymyr Mnih, Remi Munos, Koray Kavukcuoglu, Nando de Freitas

Figure 1 for Sample Efficient Actor-Critic with Experience Replay
Figure 2 for Sample Efficient Actor-Critic with Experience Replay
Figure 3 for Sample Efficient Actor-Critic with Experience Replay
Figure 4 for Sample Efficient Actor-Critic with Experience Replay
Viaarxiv icon

Combining policy gradient and Q-learning

Add code
Bookmark button
Alert button
Apr 07, 2017
Brendan O'Donoghue, Remi Munos, Koray Kavukcuoglu, Volodymyr Mnih

Figure 1 for Combining policy gradient and Q-learning
Figure 2 for Combining policy gradient and Q-learning
Figure 3 for Combining policy gradient and Q-learning
Figure 4 for Combining policy gradient and Q-learning
Viaarxiv icon

Using Fast Weights to Attend to the Recent Past

Add code
Bookmark button
Alert button
Dec 05, 2016
Jimmy Ba, Geoffrey Hinton, Volodymyr Mnih, Joel Z. Leibo, Catalin Ionescu

Figure 1 for Using Fast Weights to Attend to the Recent Past
Figure 2 for Using Fast Weights to Attend to the Recent Past
Figure 3 for Using Fast Weights to Attend to the Recent Past
Figure 4 for Using Fast Weights to Attend to the Recent Past
Viaarxiv icon

Reinforcement Learning with Unsupervised Auxiliary Tasks

Add code
Bookmark button
Alert button
Nov 16, 2016
Max Jaderberg, Volodymyr Mnih, Wojciech Marian Czarnecki, Tom Schaul, Joel Z Leibo, David Silver, Koray Kavukcuoglu

Figure 1 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 2 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 3 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Figure 4 for Reinforcement Learning with Unsupervised Auxiliary Tasks
Viaarxiv icon

Learning values across many orders of magnitude

Add code
Bookmark button
Alert button
Aug 16, 2016
Hado van Hasselt, Arthur Guez, Matteo Hessel, Volodymyr Mnih, David Silver

Figure 1 for Learning values across many orders of magnitude
Figure 2 for Learning values across many orders of magnitude
Figure 3 for Learning values across many orders of magnitude
Viaarxiv icon

Asynchronous Methods for Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 16, 2016
Volodymyr Mnih, Adrià Puigdomènech Badia, Mehdi Mirza, Alex Graves, Timothy P. Lillicrap, Tim Harley, David Silver, Koray Kavukcuoglu

Figure 1 for Asynchronous Methods for Deep Reinforcement Learning
Figure 2 for Asynchronous Methods for Deep Reinforcement Learning
Figure 3 for Asynchronous Methods for Deep Reinforcement Learning
Figure 4 for Asynchronous Methods for Deep Reinforcement Learning
Viaarxiv icon

Strategic Attentive Writer for Learning Macro-Actions

Add code
Bookmark button
Alert button
Jun 15, 2016
Alexander, Vezhnevets, Volodymyr Mnih, John Agapiou, Simon Osindero, Alex Graves, Oriol Vinyals, Koray Kavukcuoglu

Figure 1 for Strategic Attentive Writer for Learning Macro-Actions
Figure 2 for Strategic Attentive Writer for Learning Macro-Actions
Figure 3 for Strategic Attentive Writer for Learning Macro-Actions
Figure 4 for Strategic Attentive Writer for Learning Macro-Actions
Viaarxiv icon

Policy Distillation

Add code
Bookmark button
Alert button
Jan 07, 2016
Andrei A. Rusu, Sergio Gomez Colmenarejo, Caglar Gulcehre, Guillaume Desjardins, James Kirkpatrick, Razvan Pascanu, Volodymyr Mnih, Koray Kavukcuoglu, Raia Hadsell

Figure 1 for Policy Distillation
Figure 2 for Policy Distillation
Figure 3 for Policy Distillation
Figure 4 for Policy Distillation
Viaarxiv icon

Massively Parallel Methods for Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 16, 2015
Arun Nair, Praveen Srinivasan, Sam Blackwell, Cagdas Alcicek, Rory Fearon, Alessandro De Maria, Vedavyas Panneershelvam, Mustafa Suleyman, Charles Beattie, Stig Petersen, Shane Legg, Volodymyr Mnih, Koray Kavukcuoglu, David Silver

Figure 1 for Massively Parallel Methods for Deep Reinforcement Learning
Figure 2 for Massively Parallel Methods for Deep Reinforcement Learning
Figure 3 for Massively Parallel Methods for Deep Reinforcement Learning
Figure 4 for Massively Parallel Methods for Deep Reinforcement Learning
Viaarxiv icon