Alert button
Picture for Nicolas Heess

Nicolas Heess

Alert button

Critic Regularized Regression

Add code
Bookmark button
Alert button
Jun 26, 2020
Ziyu Wang, Alexander Novikov, Konrad Żołna, Jost Tobias Springenberg, Scott Reed, Bobak Shahriari, Noah Siegel, Josh Merel, Caglar Gulcehre, Nicolas Heess, Nando de Freitas

Figure 1 for Critic Regularized Regression
Figure 2 for Critic Regularized Regression
Figure 3 for Critic Regularized Regression
Figure 4 for Critic Regularized Regression
Viaarxiv icon

RL Unplugged: Benchmarks for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 24, 2020
Caglar Gulcehre, Ziyu Wang, Alexander Novikov, Tom Le Paine, Sergio Gómez Colmenarejo, Konrad Zolna, Rishabh Agarwal, Josh Merel, Daniel Mankowitz, Cosmin Paduraru, Gabriel Dulac-Arnold, Jerry Li, Mohammad Norouzi, Matt Hoffman, Ofir Nachum, George Tucker, Nicolas Heess, Nando de Freitas

Figure 1 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 2 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 3 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 4 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Viaarxiv icon

dm_control: Software and Tasks for Continuous Control

Add code
Bookmark button
Alert button
Jun 22, 2020
Yuval Tassa, Saran Tunyasuvunakool, Alistair Muldal, Yotam Doron, Siqi Liu, Steven Bohez, Josh Merel, Tom Erez, Timothy Lillicrap, Nicolas Heess

Figure 1 for dm_control: Software and Tasks for Continuous Control
Figure 2 for dm_control: Software and Tasks for Continuous Control
Figure 3 for dm_control: Software and Tasks for Continuous Control
Figure 4 for dm_control: Software and Tasks for Continuous Control
Viaarxiv icon

Simple Sensor Intentions for Exploration

Add code
Bookmark button
Alert button
May 15, 2020
Tim Hertweck, Martin Riedmiller, Michael Bloesch, Jost Tobias Springenberg, Noah Siegel, Markus Wulfmeier, Roland Hafner, Nicolas Heess

Figure 1 for Simple Sensor Intentions for Exploration
Figure 2 for Simple Sensor Intentions for Exploration
Figure 3 for Simple Sensor Intentions for Exploration
Figure 4 for Simple Sensor Intentions for Exploration
Viaarxiv icon

A Distributional View on Multi-Objective Policy Optimization

Add code
Bookmark button
Alert button
May 15, 2020
Abbas Abdolmaleki, Sandy H. Huang, Leonard Hasenclever, Michael Neunert, H. Francis Song, Martina Zambelli, Murilo F. Martins, Nicolas Heess, Raia Hadsell, Martin Riedmiller

Figure 1 for A Distributional View on Multi-Objective Policy Optimization
Figure 2 for A Distributional View on Multi-Objective Policy Optimization
Figure 3 for A Distributional View on Multi-Objective Policy Optimization
Figure 4 for A Distributional View on Multi-Objective Policy Optimization
Viaarxiv icon

Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning

Add code
Bookmark button
Alert button
Apr 23, 2020
Giambattista Parascandolo, Lars Buesing, Josh Merel, Leonard Hasenclever, John Aslanides, Jessica B. Hamrick, Nicolas Heess, Alexander Neitz, Theophane Weber

Figure 1 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 2 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 3 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 4 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Viaarxiv icon

Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 23, 2020
Noah Y. Siegel, Jost Tobias Springenberg, Felix Berkenkamp, Abbas Abdolmaleki, Michael Neunert, Thomas Lampe, Roland Hafner, Nicolas Heess, Martin Riedmiller

Figure 1 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 2 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 3 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 4 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Viaarxiv icon

Value-driven Hindsight Modelling

Add code
Bookmark button
Alert button
Feb 19, 2020
Arthur Guez, Fabio Viola, Théophane Weber, Lars Buesing, Steven Kapturowski, Doina Precup, David Silver, Nicolas Heess

Figure 1 for Value-driven Hindsight Modelling
Figure 2 for Value-driven Hindsight Modelling
Figure 3 for Value-driven Hindsight Modelling
Figure 4 for Value-driven Hindsight Modelling
Viaarxiv icon

Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics

Add code
Bookmark button
Alert button
Jan 02, 2020
Michael Neunert, Abbas Abdolmaleki, Markus Wulfmeier, Thomas Lampe, Jost Tobias Springenberg, Roland Hafner, Francesco Romano, Jonas Buchli, Nicolas Heess, Martin Riedmiller

Figure 1 for Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics
Figure 2 for Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics
Figure 3 for Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics
Figure 4 for Continuous-Discrete Reinforcement Learning for Hybrid Control in Robotics
Viaarxiv icon

Hindsight Credit Assignment

Add code
Bookmark button
Alert button
Dec 05, 2019
Anna Harutyunyan, Will Dabney, Thomas Mesnard, Mohammad Azar, Bilal Piot, Nicolas Heess, Hado van Hasselt, Greg Wayne, Satinder Singh, Doina Precup, Remi Munos

Figure 1 for Hindsight Credit Assignment
Figure 2 for Hindsight Credit Assignment
Figure 3 for Hindsight Credit Assignment
Figure 4 for Hindsight Credit Assignment
Viaarxiv icon