Alert button
Picture for Nicolas Heess

Nicolas Heess

Alert button

Temporal Difference Uncertainties as a Signal for Exploration

Add code
Bookmark button
Alert button
Oct 05, 2020
Sebastian Flennerhag, Jane X. Wang, Pablo Sprechmann, Francesco Visin, Alexandre Galashov, Steven Kapturowski, Diana L. Borsa, Nicolas Heess, Andre Barreto, Razvan Pascanu

Figure 1 for Temporal Difference Uncertainties as a Signal for Exploration
Figure 2 for Temporal Difference Uncertainties as a Signal for Exploration
Figure 3 for Temporal Difference Uncertainties as a Signal for Exploration
Figure 4 for Temporal Difference Uncertainties as a Signal for Exploration
Viaarxiv icon

Action and Perception as Divergence Minimization

Add code
Bookmark button
Alert button
Oct 05, 2020
Danijar Hafner, Pedro A. Ortega, Jimmy Ba, Thomas Parr, Karl Friston, Nicolas Heess

Figure 1 for Action and Perception as Divergence Minimization
Figure 2 for Action and Perception as Divergence Minimization
Figure 3 for Action and Perception as Divergence Minimization
Figure 4 for Action and Perception as Divergence Minimization
Viaarxiv icon

Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban

Add code
Bookmark button
Alert button
Oct 03, 2020
Peter Karkus, Mehdi Mirza, Arthur Guez, Andrew Jaegle, Timothy Lillicrap, Lars Buesing, Nicolas Heess, Theophane Weber

Figure 1 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 2 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 3 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 4 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Viaarxiv icon

Learning to swim in potential flow

Add code
Bookmark button
Alert button
Sep 30, 2020
Yusheng Jiao, Feng Ling, Sina Heydari, Nicolas Heess, Josh Merel, Eva Kanso

Figure 1 for Learning to swim in potential flow
Figure 2 for Learning to swim in potential flow
Figure 3 for Learning to swim in potential flow
Figure 4 for Learning to swim in potential flow
Viaarxiv icon

Physically Embedded Planning Problems: New Challenges for Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 11, 2020
Mehdi Mirza, Andrew Jaegle, Jonathan J. Hunt, Arthur Guez, Saran Tunyasuvunakool, Alistair Muldal, Théophane Weber, Peter Karkus, Sébastien Racanière, Lars Buesing, Timothy Lillicrap, Nicolas Heess

Figure 1 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 2 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 3 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 4 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Viaarxiv icon

Importance Weighted Policy Learning and Adaption

Add code
Bookmark button
Alert button
Sep 10, 2020
Alexandre Galashov, Jakub Sygnowski, Guillaume Desjardins, Jan Humplik, Leonard Hasenclever, Rae Jeong, Yee Whye Teh, Nicolas Heess

Figure 1 for Importance Weighted Policy Learning and Adaption
Figure 2 for Importance Weighted Policy Learning and Adaption
Figure 3 for Importance Weighted Policy Learning and Adaption
Figure 4 for Importance Weighted Policy Learning and Adaption
Viaarxiv icon

Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion

Add code
Bookmark button
Alert button
Aug 06, 2020
Roland Hafner, Tim Hertweck, Philipp Klöppner, Michael Bloesch, Michael Neunert, Markus Wulfmeier, Saran Tunyasuvunakool, Nicolas Heess, Martin Riedmiller

Figure 1 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 2 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 3 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 4 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Viaarxiv icon

Data-efficient Hindsight Off-policy Option Learning

Add code
Bookmark button
Alert button
Jul 30, 2020
Markus Wulfmeier, Dushyant Rao, Roland Hafner, Thomas Lampe, Abbas Abdolmaleki, Tim Hertweck, Michael Neunert, Dhruva Tirumala, Noah Siegel, Nicolas Heess, Martin Riedmiller

Figure 1 for Data-efficient Hindsight Off-policy Option Learning
Figure 2 for Data-efficient Hindsight Off-policy Option Learning
Figure 3 for Data-efficient Hindsight Off-policy Option Learning
Figure 4 for Data-efficient Hindsight Off-policy Option Learning
Viaarxiv icon

RL Unplugged: Benchmarks for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 02, 2020
Caglar Gulcehre, Ziyu Wang, Alexander Novikov, Tom Le Paine, Sergio Gomez Colmenarejo, Konrad Zolna, Rishabh Agarwal, Josh Merel, Daniel Mankowitz, Cosmin Paduraru, Gabriel Dulac-Arnold, Jerry Li, Mohammad Norouzi, Matt Hoffman, Ofir Nachum, George Tucker, Nicolas Heess, Nando de Freitas

Figure 1 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 2 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 3 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 4 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Viaarxiv icon