Picture for Volodymyr Mnih

Volodymyr Mnih

Vision-Language Models as a Source of Rewards

Add code
Dec 14, 2023
Figure 1 for Vision-Language Models as a Source of Rewards
Figure 2 for Vision-Language Models as a Source of Rewards
Figure 3 for Vision-Language Models as a Source of Rewards
Figure 4 for Vision-Language Models as a Source of Rewards
Viaarxiv icon

In-context Reinforcement Learning with Algorithm Distillation

Add code
Oct 25, 2022
Figure 1 for In-context Reinforcement Learning with Algorithm Distillation
Figure 2 for In-context Reinforcement Learning with Algorithm Distillation
Figure 3 for In-context Reinforcement Learning with Algorithm Distillation
Figure 4 for In-context Reinforcement Learning with Algorithm Distillation
Viaarxiv icon

Wasserstein Distance Maximizing Intrinsic Control

Add code
Oct 28, 2021
Figure 1 for Wasserstein Distance Maximizing Intrinsic Control
Figure 2 for Wasserstein Distance Maximizing Intrinsic Control
Figure 3 for Wasserstein Distance Maximizing Intrinsic Control
Viaarxiv icon

Discovering Diverse Nearly Optimal Policies withSuccessor Features

Add code
Jun 01, 2021
Figure 1 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 2 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 3 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 4 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Viaarxiv icon

Relative Variational Intrinsic Control

Add code
Dec 14, 2020
Figure 1 for Relative Variational Intrinsic Control
Figure 2 for Relative Variational Intrinsic Control
Figure 3 for Relative Variational Intrinsic Control
Figure 4 for Relative Variational Intrinsic Control
Viaarxiv icon

Q-Learning in enormous action spaces via amortized approximate maximization

Add code
Jan 22, 2020
Figure 1 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 2 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 3 for Q-Learning in enormous action spaces via amortized approximate maximization
Figure 4 for Q-Learning in enormous action spaces via amortized approximate maximization
Viaarxiv icon

Unsupervised Learning of Object Keypoints for Perception and Control

Add code
Jun 19, 2019
Figure 1 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 2 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 3 for Unsupervised Learning of Object Keypoints for Perception and Control
Figure 4 for Unsupervised Learning of Object Keypoints for Perception and Control
Viaarxiv icon

Fast Task Inference with Variational Intrinsic Successor Features

Add code
Jun 12, 2019
Figure 1 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 2 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 3 for Fast Task Inference with Variational Intrinsic Successor Features
Figure 4 for Fast Task Inference with Variational Intrinsic Successor Features
Viaarxiv icon

Unsupervised Control Through Non-Parametric Discriminative Rewards

Add code
Nov 28, 2018
Figure 1 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 2 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 3 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Figure 4 for Unsupervised Control Through Non-Parametric Discriminative Rewards
Viaarxiv icon

The Uncertainty Bellman Equation and Exploration

Add code
Oct 22, 2018
Figure 1 for The Uncertainty Bellman Equation and Exploration
Figure 2 for The Uncertainty Bellman Equation and Exploration
Figure 3 for The Uncertainty Bellman Equation and Exploration
Figure 4 for The Uncertainty Bellman Equation and Exploration
Viaarxiv icon