Alert button
Picture for Arunkumar Byravan

Arunkumar Byravan

Alert button

The Challenges of Exploration for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Jan 27, 2022
Nathan Lambert, Markus Wulfmeier, William Whitney, Arunkumar Byravan, Michael Bloesch, Vibhavari Dasagi, Tim Hertweck, Martin Riedmiller

Figure 1 for The Challenges of Exploration for Offline Reinforcement Learning
Figure 2 for The Challenges of Exploration for Offline Reinforcement Learning
Figure 3 for The Challenges of Exploration for Offline Reinforcement Learning
Figure 4 for The Challenges of Exploration for Offline Reinforcement Learning
Viaarxiv icon

Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes

Add code
Bookmark button
Alert button
Nov 03, 2021
Alex X. Lee, Coline Devin, Yuxiang Zhou, Thomas Lampe, Konstantinos Bousmalis, Jost Tobias Springenberg, Arunkumar Byravan, Abbas Abdolmaleki, Nimrod Gileadi, David Khosid, Claudio Fantacci, Jose Enrique Chen, Akhil Raju, Rae Jeong, Michael Neunert, Antoine Laurens, Stefano Saliceti, Federico Casarini, Martin Riedmiller, Raia Hadsell, Francesco Nori

Figure 1 for Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
Figure 2 for Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
Figure 3 for Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
Figure 4 for Beyond Pick-and-Place: Tackling Robotic Stacking of Diverse Shapes
Viaarxiv icon

Evaluating model-based planning and planner amortization for continuous control

Add code
Bookmark button
Alert button
Oct 07, 2021
Arunkumar Byravan, Leonard Hasenclever, Piotr Trochim, Mehdi Mirza, Alessandro Davide Ialongo, Yuval Tassa, Jost Tobias Springenberg, Abbas Abdolmaleki, Nicolas Heess, Josh Merel, Martin Riedmiller

Figure 1 for Evaluating model-based planning and planner amortization for continuous control
Figure 2 for Evaluating model-based planning and planner amortization for continuous control
Figure 3 for Evaluating model-based planning and planner amortization for continuous control
Figure 4 for Evaluating model-based planning and planner amortization for continuous control
Viaarxiv icon

Learning Dynamics Models for Model Predictive Agents

Add code
Bookmark button
Alert button
Sep 29, 2021
Michael Lutter, Leonard Hasenclever, Arunkumar Byravan, Gabriel Dulac-Arnold, Piotr Trochim, Nicolas Heess, Josh Merel, Yuval Tassa

Figure 1 for Learning Dynamics Models for Model Predictive Agents
Figure 2 for Learning Dynamics Models for Model Predictive Agents
Figure 3 for Learning Dynamics Models for Model Predictive Agents
Figure 4 for Learning Dynamics Models for Model Predictive Agents
Viaarxiv icon

On Multi-objective Policy Optimization as a Tool for Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 15, 2021
Abbas Abdolmaleki, Sandy H. Huang, Giulia Vezzani, Bobak Shahriari, Jost Tobias Springenberg, Shruti Mishra, Dhruva TB, Arunkumar Byravan, Konstantinos Bousmalis, Andras Gyorgy, Csaba Szepesvari, Raia Hadsell, Nicolas Heess, Martin Riedmiller

Figure 1 for On Multi-objective Policy Optimization as a Tool for Reinforcement Learning
Figure 2 for On Multi-objective Policy Optimization as a Tool for Reinforcement Learning
Figure 3 for On Multi-objective Policy Optimization as a Tool for Reinforcement Learning
Figure 4 for On Multi-objective Policy Optimization as a Tool for Reinforcement Learning
Viaarxiv icon

Representation Matters: Improving Perception and Exploration for Robotics

Add code
Bookmark button
Alert button
Nov 03, 2020
Markus Wulfmeier, Arunkumar Byravan, Tim Hertweck, Irina Higgins, Ankush Gupta, Tejas Kulkarni, Malcolm Reynolds, Denis Teplyashin, Roland Hafner, Thomas Lampe, Martin Riedmiller

Figure 1 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 2 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 3 for Representation Matters: Improving Perception and Exploration for Robotics
Figure 4 for Representation Matters: Improving Perception and Exploration for Robotics
Viaarxiv icon

Local Search for Policy Iteration in Continuous Control

Add code
Bookmark button
Alert button
Oct 12, 2020
Jost Tobias Springenberg, Nicolas Heess, Daniel Mankowitz, Josh Merel, Arunkumar Byravan, Abbas Abdolmaleki, Jackie Kay, Jonas Degrave, Julian Schrittwieser, Yuval Tassa, Jonas Buchli, Dan Belov, Martin Riedmiller

Figure 1 for Local Search for Policy Iteration in Continuous Control
Figure 2 for Local Search for Policy Iteration in Continuous Control
Figure 3 for Local Search for Policy Iteration in Continuous Control
Figure 4 for Local Search for Policy Iteration in Continuous Control
Viaarxiv icon

Motion-Nets: 6D Tracking of Unknown Objects in Unseen Environments using RGB

Add code
Bookmark button
Alert button
Oct 30, 2019
Felix Leeb, Arunkumar Byravan, Dieter Fox

Figure 1 for Motion-Nets: 6D Tracking of Unknown Objects in Unseen Environments using RGB
Figure 2 for Motion-Nets: 6D Tracking of Unknown Objects in Unseen Environments using RGB
Figure 3 for Motion-Nets: 6D Tracking of Unknown Objects in Unseen Environments using RGB
Figure 4 for Motion-Nets: 6D Tracking of Unknown Objects in Unseen Environments using RGB
Viaarxiv icon

Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models

Add code
Bookmark button
Alert button
Oct 09, 2019
Arunkumar Byravan, Jost Tobias Springenberg, Abbas Abdolmaleki, Roland Hafner, Michael Neunert, Thomas Lampe, Noah Siegel, Nicolas Heess, Martin Riedmiller

Figure 1 for Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models
Figure 2 for Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models
Figure 3 for Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models
Figure 4 for Imagined Value Gradients: Model-Based Policy Optimization with Transferable Latent Dynamics Models
Viaarxiv icon