Alert button
Picture for Nicolas Heess

Nicolas Heess

Alert button

Coordinating Policies Among Multiple Agents via an Intelligent Communication Channel

May 25, 2022
Dianbo Liu, Vedant Shah, Oussama Boussif, Cristian Meo, Anirudh Goyal, Tianmin Shu, Michael Mozer, Nicolas Heess, Yoshua Bengio

Figure 1 for Coordinating Policies Among Multiple Agents via an Intelligent Communication Channel
Figure 2 for Coordinating Policies Among Multiple Agents via an Intelligent Communication Channel
Figure 3 for Coordinating Policies Among Multiple Agents via an Intelligent Communication Channel
Figure 4 for Coordinating Policies Among Multiple Agents via an Intelligent Communication Channel
Viaarxiv icon

Data augmentation for efficient learning from parametric experts

May 23, 2022
Alexandre Galashov, Josh Merel, Nicolas Heess

Figure 1 for Data augmentation for efficient learning from parametric experts
Figure 2 for Data augmentation for efficient learning from parametric experts
Figure 3 for Data augmentation for efficient learning from parametric experts
Figure 4 for Data augmentation for efficient learning from parametric experts
Viaarxiv icon

A Generalist Agent

May 19, 2022
Scott Reed, Konrad Zolna, Emilio Parisotto, Sergio Gomez Colmenarejo, Alexander Novikov, Gabriel Barth-Maron, Mai Gimenez, Yury Sulsky, Jackie Kay, Jost Tobias Springenberg, Tom Eccles, Jake Bruce, Ali Razavi, Ashley Edwards, Nicolas Heess, Yutian Chen, Raia Hadsell, Oriol Vinyals, Mahyar Bordbar, Nando de Freitas

Figure 1 for A Generalist Agent
Figure 2 for A Generalist Agent
Figure 3 for A Generalist Agent
Figure 4 for A Generalist Agent
Viaarxiv icon

Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach

Apr 22, 2022
Bobak Shahriari, Abbas Abdolmaleki, Arunkumar Byravan, Abe Friesen, Siqi Liu, Jost Tobias Springenberg, Nicolas Heess, Matt Hoffman, Martin Riedmiller

Figure 1 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 2 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 3 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Figure 4 for Revisiting Gaussian mixture critics in off-policy reinforcement learning: a sample-based approach
Viaarxiv icon

Revisiting Gaussian mixture critic in off-policy reinforcement learning: a sample-based approach

Apr 21, 2022
Bobak Shahriari, Abbas Abdolmaleki, Arunkumar Byravan, Abe Friesen, Siqi Liu, Jost Tobias Springenberg, Nicolas Heess, Matt Hoffman, Martin Riedmiller

Figure 1 for Revisiting Gaussian mixture critic in off-policy reinforcement learning: a sample-based approach
Figure 2 for Revisiting Gaussian mixture critic in off-policy reinforcement learning: a sample-based approach
Figure 3 for Revisiting Gaussian mixture critic in off-policy reinforcement learning: a sample-based approach
Figure 4 for Revisiting Gaussian mixture critic in off-policy reinforcement learning: a sample-based approach
Viaarxiv icon

COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation

Apr 19, 2022
Jongmin Lee, Cosmin Paduraru, Daniel J. Mankowitz, Nicolas Heess, Doina Precup, Kee-Eung Kim, Arthur Guez

Figure 1 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 2 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 3 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 4 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Viaarxiv icon

Offline Distillation for Robot Lifelong Learning with Imbalanced Experience

Apr 12, 2022
Wenxuan Zhou, Steven Bohez, Jan Humplik, Abbas Abdolmaleki, Dushyant Rao, Markus Wulfmeier, Tuomas Haarnoja, Nicolas Heess

Figure 1 for Offline Distillation for Robot Lifelong Learning with Imbalanced Experience
Figure 2 for Offline Distillation for Robot Lifelong Learning with Imbalanced Experience
Figure 3 for Offline Distillation for Robot Lifelong Learning with Imbalanced Experience
Figure 4 for Offline Distillation for Robot Lifelong Learning with Imbalanced Experience
Viaarxiv icon

Imitate and Repurpose: Learning Reusable Robot Movement Skills From Human and Animal Behaviors

Mar 31, 2022
Steven Bohez, Saran Tunyasuvunakool, Philemon Brakel, Fereshteh Sadeghi, Leonard Hasenclever, Yuval Tassa, Emilio Parisotto, Jan Humplik, Tuomas Haarnoja, Roland Hafner, Markus Wulfmeier, Michael Neunert, Ben Moran, Noah Siegel, Andrea Huber, Francesco Romano, Nathan Batchelor, Federico Casarini, Josh Merel, Raia Hadsell, Nicolas Heess

Figure 1 for Imitate and Repurpose: Learning Reusable Robot Movement Skills From Human and Animal Behaviors
Figure 2 for Imitate and Repurpose: Learning Reusable Robot Movement Skills From Human and Animal Behaviors
Figure 3 for Imitate and Repurpose: Learning Reusable Robot Movement Skills From Human and Animal Behaviors
Figure 4 for Imitate and Repurpose: Learning Reusable Robot Movement Skills From Human and Animal Behaviors
Viaarxiv icon