Alert button
Picture for Marc G. Bellemare

Marc G. Bellemare

Alert button

A Distributional Analogue to the Successor Representation

Add code
Bookmark button
Alert button
Feb 13, 2024
Harley Wiltzer, Jesse Farebrother, Arthur Gretton, Yunhao Tang, André Barreto, Will Dabney, Marc G. Bellemare, Mark Rowland

Viaarxiv icon

Learning and Controlling Silicon Dopant Transitions in Graphene using Scanning Transmission Electron Microscopy

Add code
Bookmark button
Alert button
Nov 21, 2023
Max Schwarzer, Jesse Farebrother, Joshua Greaves, Ekin Dogus Cubuk, Rishabh Agarwal, Aaron Courville, Marc G. Bellemare, Sergei Kalinin, Igor Mordatch, Pablo Samuel Castro, Kevin M. Roccapriore

Viaarxiv icon

Small batch deep reinforcement learning

Add code
Bookmark button
Alert button
Oct 05, 2023
Johan Obando-Ceron, Marc G. Bellemare, Pablo Samuel Castro

Viaarxiv icon

Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control

Add code
Bookmark button
Alert button
Sep 26, 2023
Nate Rahn, Pierluca D'Oro, Harley Wiltzer, Pierre-Luc Bacon, Marc G. Bellemare

Figure 1 for Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control
Figure 2 for Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control
Figure 3 for Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control
Figure 4 for Policy Optimization in a Noisy Neighborhood: On Return Landscapes in Continuous Control
Viaarxiv icon

Bootstrapped Representations in Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 16, 2023
Charline Le Lan, Stephen Tu, Mark Rowland, Anna Harutyunyan, Rishabh Agarwal, Marc G. Bellemare, Will Dabney

Figure 1 for Bootstrapped Representations in Reinforcement Learning
Figure 2 for Bootstrapped Representations in Reinforcement Learning
Figure 3 for Bootstrapped Representations in Reinforcement Learning
Figure 4 for Bootstrapped Representations in Reinforcement Learning
Viaarxiv icon

The Statistical Benefits of Quantile Temporal-Difference Learning for Value Estimation

Add code
Bookmark button
Alert button
May 28, 2023
Mark Rowland, Yunhao Tang, Clare Lyle, Rémi Munos, Marc G. Bellemare, Will Dabney

Figure 1 for The Statistical Benefits of Quantile Temporal-Difference Learning for Value Estimation
Figure 2 for The Statistical Benefits of Quantile Temporal-Difference Learning for Value Estimation
Figure 3 for The Statistical Benefits of Quantile Temporal-Difference Learning for Value Estimation
Figure 4 for The Statistical Benefits of Quantile Temporal-Difference Learning for Value Estimation
Viaarxiv icon

Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks

Add code
Bookmark button
Alert button
Apr 25, 2023
Jesse Farebrother, Joshua Greaves, Rishabh Agarwal, Charline Le Lan, Ross Goroshin, Pablo Samuel Castro, Marc G. Bellemare

Figure 1 for Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks
Figure 2 for Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks
Figure 3 for Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks
Figure 4 for Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks
Viaarxiv icon

An Analysis of Quantile Temporal-Difference Learning

Add code
Bookmark button
Alert button
Jan 11, 2023
Mark Rowland, Rémi Munos, Mohammad Gheshlaghi Azar, Yunhao Tang, Georg Ostrovski, Anna Harutyunyan, Karl Tuyls, Marc G. Bellemare, Will Dabney

Figure 1 for An Analysis of Quantile Temporal-Difference Learning
Figure 2 for An Analysis of Quantile Temporal-Difference Learning
Figure 3 for An Analysis of Quantile Temporal-Difference Learning
Figure 4 for An Analysis of Quantile Temporal-Difference Learning
Viaarxiv icon

A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces

Add code
Bookmark button
Alert button
Dec 08, 2022
Charline Le Lan, Joshua Greaves, Jesse Farebrother, Mark Rowland, Fabian Pedregosa, Rishabh Agarwal, Marc G. Bellemare

Figure 1 for A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces
Figure 2 for A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces
Figure 3 for A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces
Figure 4 for A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces
Viaarxiv icon

The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 15, 2022
Yunhao Tang, Mark Rowland, Rémi Munos, Bernardo Ávila Pires, Will Dabney, Marc G. Bellemare

Figure 1 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 2 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 3 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 4 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Viaarxiv icon