Alert button
Picture for Volodymyr Mnih

Volodymyr Mnih

Alert button

Vision-Language Models as a Source of Rewards

Dec 14, 2023
Kate Baumli, Satinder Baveja, Feryal Behbahani, Harris Chan, Gheorghe Comanici, Sebastian Flennerhag, Maxime Gazeau, Kristian Holsheimer, Dan Horgan, Michael Laskin, Clare Lyle, Hussain Masoom, Kay McKinney, Volodymyr Mnih, Alexander Neitz, Fabio Pardo, Jack Parker-Holder, John Quan, Tim Rocktäschel, Himanshu Sahni, Tom Schaul, Yannick Schroecker, Stephen Spencer, Richie Steigerwald, Luyu Wang, Lei Zhang

Viaarxiv icon

In-context Reinforcement Learning with Algorithm Distillation

Oct 25, 2022
Michael Laskin, Luyu Wang, Junhyuk Oh, Emilio Parisotto, Stephen Spencer, Richie Steigerwald, DJ Strouse, Steven Hansen, Angelos Filos, Ethan Brooks, Maxime Gazeau, Himanshu Sahni, Satinder Singh, Volodymyr Mnih

Figure 1 for In-context Reinforcement Learning with Algorithm Distillation
Figure 2 for In-context Reinforcement Learning with Algorithm Distillation
Figure 3 for In-context Reinforcement Learning with Algorithm Distillation
Figure 4 for In-context Reinforcement Learning with Algorithm Distillation
Viaarxiv icon

Wasserstein Distance Maximizing Intrinsic Control

Oct 28, 2021
Ishan Durugkar, Steven Hansen, Stephen Spencer, Volodymyr Mnih

Figure 1 for Wasserstein Distance Maximizing Intrinsic Control
Figure 2 for Wasserstein Distance Maximizing Intrinsic Control
Figure 3 for Wasserstein Distance Maximizing Intrinsic Control
Viaarxiv icon

Discovering Diverse Nearly Optimal Policies withSuccessor Features

Jun 01, 2021
Tom Zahavy, Brendan O'Donoghue, Andre Barreto, Volodymyr Mnih, Sebastian Flennerhag, Satinder Singh

Figure 1 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 2 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 3 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 4 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Viaarxiv icon

Relative Variational Intrinsic Control

Dec 14, 2020
Kate Baumli, David Warde-Farley, Steven Hansen, Volodymyr Mnih

Figure 1 for Relative Variational Intrinsic Control
Figure 2 for Relative Variational Intrinsic Control
Figure 3 for Relative Variational Intrinsic Control
Figure 4 for Relative Variational Intrinsic Control
Viaarxiv icon

Q-Learning in enormous action spaces via amortized approximate maximization

Jan 22, 2020
Tom Van de Wiele, David Warde-Farley, Andriy Mnih, Volodymyr Mnih

Figure 1 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 2 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 3 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 4 for Q-Learning in enormous action spaces via amortized approximate maximization
Viaarxiv icon

Unsupervised Learning of Object Keypoints for Perception and Control

Jun 19, 2019
Tejas Kulkarni, Ankush Gupta, Catalin Ionescu, Sebastian Borgeaud, Malcolm Reynolds, Andrew Zisserman, Volodymyr Mnih

Figure 1 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 2 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 3 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 4 for Unsupervised Learning of Object Keypoints for Perception and Control
Viaarxiv icon

Fast Task Inference with Variational Intrinsic Successor Features

Jun 12, 2019
Steven Hansen, Will Dabney, Andre Barreto, Tom Van de Wiele, David Warde-Farley, Volodymyr Mnih

Figure 1 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 2 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 3 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 4 for Fast Task Inference with Variational Intrinsic Successor Features
Viaarxiv icon

Unsupervised Control Through Non-Parametric Discriminative Rewards

Nov 28, 2018
David Warde-Farley, Tom Van de Wiele, Tejas Kulkarni, Catalin Ionescu, Steven Hansen, Volodymyr Mnih

Figure 1 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 2 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 3 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 4 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Viaarxiv icon

The Uncertainty Bellman Equation and Exploration

Oct 22, 2018
Brendan O'Donoghue, Ian Osband, Remi Munos, Volodymyr Mnih

Figure 1 for The Uncertainty Bellman Equation and Exploration
Figure 2 for The Uncertainty Bellman Equation and Exploration
Figure 3 for The Uncertainty Bellman Equation and Exploration
Figure 4 for The Uncertainty Bellman Equation and Exploration
Viaarxiv icon