Alert button
Picture for Martin Riedmiller

Martin Riedmiller

Alert button

Representation Matters: Improving Perception and Exploration for Robotics

Add code
Bookmark button
Alert button
Nov 03, 2020
Markus Wulfmeier, Arunkumar Byravan, Tim Hertweck, Irina Higgins, Ankush Gupta, Tejas Kulkarni, Malcolm Reynolds, Denis Teplyashin, Roland Hafner, Thomas Lampe, Martin Riedmiller

Figure 1 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 2 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 3 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 4 for Representation Matters: Improving Perception and Exploration for Robotics
Viaarxiv icon

"What, not how": Solving an under-actuated insertion task from scratch

Add code
Bookmark button
Alert button
Oct 30, 2020
Giulia Vezzani, Michael Neunert, Markus Wulfmeier, Rae Jeong, Thomas Lampe, Noah Siegel, Roland Hafner, Abbas Abdolmaleki, Martin Riedmiller, Francesco Nori

Figure 1 for "What, not how": Solving an under-actuated insertion task from scratch
Figure 2 for "What, not how": Solving an under-actuated insertion task from scratch
Figure 3 for "What, not how": Solving an under-actuated insertion task from scratch
Figure 4 for "What, not how": Solving an under-actuated insertion task from scratch
Viaarxiv icon

Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification

Add code
Bookmark button
Alert button
Oct 20, 2020
Daniel J. Mankowitz, Dan A. Calian, Rae Jeong, Cosmin Paduraru, Nicolas Heess, Sumanth Dathathri, Martin Riedmiller, Timothy Mann

Figure 1 for Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification
Figure 2 for Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification
Figure 3 for Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification
Figure 4 for Robust Constrained Reinforcement Learning for Continuous Control with Model Misspecification
Viaarxiv icon

Local Search for Policy Iteration in Continuous Control

Add code
Bookmark button
Alert button
Oct 12, 2020
Jost Tobias Springenberg, Nicolas Heess, Daniel Mankowitz, Josh Merel, Arunkumar Byravan, Abbas Abdolmaleki, Jackie Kay, Jonas Degrave, Julian Schrittwieser, Yuval Tassa, Jonas Buchli, Dan Belov, Martin Riedmiller

Figure 1 for Local Search for Policy Iteration in Continuous Control
Figure 2 for Local Search for Policy Iteration in Continuous Control
Figure 3 for Local Search for Policy Iteration in Continuous Control
Figure 4 for Local Search for Policy Iteration in Continuous Control
Viaarxiv icon

Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion

Add code
Bookmark button
Alert button
Aug 06, 2020
Roland Hafner, Tim Hertweck, Philipp Klöppner, Michael Bloesch, Michael Neunert, Markus Wulfmeier, Saran Tunyasuvunakool, Nicolas Heess, Martin Riedmiller

Figure 1 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 2 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 3 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Figure 4 for Towards General and Autonomous Learning of Core Skills: A Case Study in Locomotion
Viaarxiv icon

Data-efficient Hindsight Off-policy Option Learning

Add code
Bookmark button
Alert button
Jul 30, 2020
Markus Wulfmeier, Dushyant Rao, Roland Hafner, Thomas Lampe, Abbas Abdolmaleki, Tim Hertweck, Michael Neunert, Dhruva Tirumala, Noah Siegel, Nicolas Heess, Martin Riedmiller

Figure 1 for Data-efficient Hindsight Off-policy Option Learning
Figure 2 for Data-efficient Hindsight Off-policy Option Learning
Figure 3 for Data-efficient Hindsight Off-policy Option Learning
Figure 4 for Data-efficient Hindsight Off-policy Option Learning
Viaarxiv icon

Simple Sensor Intentions for Exploration

Add code
Bookmark button
Alert button
May 15, 2020
Tim Hertweck, Martin Riedmiller, Michael Bloesch, Jost Tobias Springenberg, Noah Siegel, Markus Wulfmeier, Roland Hafner, Nicolas Heess

Figure 1 for Simple Sensor Intentions for Exploration
Figure 2 for Simple Sensor Intentions for Exploration
Figure 3 for Simple Sensor Intentions for Exploration
Figure 4 for Simple Sensor Intentions for Exploration
Viaarxiv icon

A Distributional View on Multi-Objective Policy Optimization

Add code
Bookmark button
Alert button
May 15, 2020
Abbas Abdolmaleki, Sandy H. Huang, Leonard Hasenclever, Michael Neunert, H. Francis Song, Martina Zambelli, Murilo F. Martins, Nicolas Heess, Raia Hadsell, Martin Riedmiller

Figure 1 for A Distributional View on Multi-Objective Policy Optimization
Figure 2 for A Distributional View on Multi-Objective Policy Optimization
Figure 3 for A Distributional View on Multi-Objective Policy Optimization
Figure 4 for A Distributional View on Multi-Objective Policy Optimization
Viaarxiv icon

Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 23, 2020
Noah Y. Siegel, Jost Tobias Springenberg, Felix Berkenkamp, Abbas Abdolmaleki, Michael Neunert, Thomas Lampe, Roland Hafner, Nicolas Heess, Martin Riedmiller

Figure 1 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 2 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 3 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Figure 4 for Keep Doing What Worked: Behavioral Modelling Priors for Offline Reinforcement Learning
Viaarxiv icon