Alert button
Picture for Martin Riedmiller

Martin Riedmiller

Alert button

Towards practical reinforcement learning for tokamak magnetic control

Add code
Bookmark button
Alert button
Jul 21, 2023
Brendan D. Tracey, Andrea Michi, Yuri Chervonyi, Ian Davies, Cosmin Paduraru, Nevena Lazic, Federico Felici, Timo Ewalds, Craig Donner, Cristian Galperti, Jonas Buchli, Michael Neunert, Andrea Huber, Jonathan Evens, Paula Kurylowicz, Daniel J. Mankowitz, Martin Riedmiller, The TCV Team

Figure 1 for Towards practical reinforcement learning for tokamak magnetic control
Figure 2 for Towards practical reinforcement learning for tokamak magnetic control
Figure 3 for Towards practical reinforcement learning for tokamak magnetic control
Figure 4 for Towards practical reinforcement learning for tokamak magnetic control
Viaarxiv icon

Towards A Unified Agent with Foundation Models

Add code
Bookmark button
Alert button
Jul 18, 2023
Norman Di Palo, Arunkumar Byravan, Leonard Hasenclever, Markus Wulfmeier, Nicolas Heess, Martin Riedmiller

Figure 1 for Towards A Unified Agent with Foundation Models
Figure 2 for Towards A Unified Agent with Foundation Models
Figure 3 for Towards A Unified Agent with Foundation Models
Figure 4 for Towards A Unified Agent with Foundation Models
Viaarxiv icon

RoboCat: A Self-Improving Foundation Agent for Robotic Manipulation

Add code
Bookmark button
Alert button
Jun 20, 2023
Konstantinos Bousmalis, Giulia Vezzani, Dushyant Rao, Coline Devin, Alex X. Lee, Maria Bauza, Todor Davchev, Yuxiang Zhou, Agrim Gupta, Akhil Raju, Antoine Laurens, Claudio Fantacci, Valentin Dalibard, Martina Zambelli, Murilo Martins, Rugile Pevceviciute, Michiel Blokzijl, Misha Denil, Nathan Batchelor, Thomas Lampe, Emilio Parisotto, Konrad Żołna, Scott Reed, Sergio Gómez Colmenarejo, Jon Scholz, Abbas Abdolmaleki, Oliver Groth, Jean-Baptiste Regli, Oleg Sushkov, Tom Rothörl, José Enrique Chen, Yusuf Aytar, Dave Barker, Joy Ortiz, Martin Riedmiller, Jost Tobias Springenberg, Raia Hadsell, Francesco Nori, Nicolas Heess

Viaarxiv icon

A Generalist Dynamics Model for Control

Add code
Bookmark button
Alert button
May 18, 2023
Ingmar Schubert, Jingwei Zhang, Jake Bruce, Sarah Bechtle, Emilio Parisotto, Martin Riedmiller, Jost Tobias Springenberg, Arunkumar Byravan, Leonard Hasenclever, Nicolas Heess

Figure 1 for A Generalist Dynamics Model for Control
Figure 2 for A Generalist Dynamics Model for Control
Figure 3 for A Generalist Dynamics Model for Control
Figure 4 for A Generalist Dynamics Model for Control
Viaarxiv icon

Leveraging Jumpy Models for Planning and Fast Learning in Robotic Domains

Add code
Bookmark button
Alert button
Feb 24, 2023
Jingwei Zhang, Jost Tobias Springenberg, Arunkumar Byravan, Leonard Hasenclever, Abbas Abdolmaleki, Dushyant Rao, Nicolas Heess, Martin Riedmiller

Figure 1 for Leveraging Jumpy Models for Planning and Fast Learning in Robotic Domains
Figure 2 for Leveraging Jumpy Models for Planning and Fast Learning in Robotic Domains
Figure 3 for Leveraging Jumpy Models for Planning and Fast Learning in Robotic Domains
Figure 4 for Leveraging Jumpy Models for Planning and Fast Learning in Robotic Domains
Viaarxiv icon

SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration

Add code
Bookmark button
Alert button
Dec 03, 2022
Giulia Vezzani, Dhruva Tirumala, Markus Wulfmeier, Dushyant Rao, Abbas Abdolmaleki, Ben Moran, Tuomas Haarnoja, Jan Humplik, Roland Hafner, Michael Neunert, Claudio Fantacci, Tim Hertweck, Thomas Lampe, Fereshteh Sadeghi, Nicolas Heess, Martin Riedmiller

Figure 1 for SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration
Figure 2 for SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration
Figure 3 for SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration
Figure 4 for SkillS: Adaptive Skill Sequencing for Efficient Temporally-Extended Exploration
Viaarxiv icon

Solving Continuous Control via Q-learning

Add code
Bookmark button
Alert button
Oct 22, 2022
Tim Seyde, Peter Werner, Wilko Schwarting, Igor Gilitschenski, Martin Riedmiller, Daniela Rus, Markus Wulfmeier

Figure 1 for Solving Continuous Control via Q-learning
Figure 2 for Solving Continuous Control via Q-learning
Figure 3 for Solving Continuous Control via Q-learning
Figure 4 for Solving Continuous Control via Q-learning
Viaarxiv icon

MO2: Model-Based Offline Options

Add code
Bookmark button
Alert button
Sep 05, 2022
Sasha Salter, Markus Wulfmeier, Dhruva Tirumala, Nicolas Heess, Martin Riedmiller, Raia Hadsell, Dushyant Rao

Figure 1 for MO2: Model-Based Offline Options
Figure 2 for MO2: Model-Based Offline Options
Figure 3 for MO2: Model-Based Offline Options
Figure 4 for MO2: Model-Based Offline Options
Viaarxiv icon

Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach

Add code
Bookmark button
Alert button
Apr 22, 2022
Bobak Shahriari, Abbas Abdolmaleki, Arunkumar Byravan, Abe Friesen, Siqi Liu, Jost Tobias Springenberg, Nicolas Heess, Matt Hoffman, Martin Riedmiller

Figure 1 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 2 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 3 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 4 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Viaarxiv icon