Picture for Marc G. Bellemare

Marc G. Bellemare

Proto-Value Networks: Scaling Representation Learning with Auxiliary Tasks

Add code
Apr 25, 2023
Viaarxiv icon

An Analysis of Quantile Temporal-Difference Learning

Add code
Jan 11, 2023
Figure 1 for An Analysis of Quantile Temporal-Difference Learning
Figure 2 for An Analysis of Quantile Temporal-Difference Learning
Figure 3 for An Analysis of Quantile Temporal-Difference Learning
Figure 4 for An Analysis of Quantile Temporal-Difference Learning
Viaarxiv icon

A Novel Stochastic Gradient Descent Algorithm for Learning Principal Subspaces

Add code
Dec 08, 2022
Viaarxiv icon

The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning

Add code
Jul 15, 2022
Figure 1 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 2 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 3 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Figure 4 for The Nature of Temporal Difference Errors in Multi-step Distributional Reinforcement Learning
Viaarxiv icon

Beyond Tabula Rasa: Reincarnating Reinforcement Learning

Add code
Jun 03, 2022
Figure 1 for Beyond Tabula Rasa: Reincarnating Reinforcement Learning
Figure 2 for Beyond Tabula Rasa: Reincarnating Reinforcement Learning
Figure 3 for Beyond Tabula Rasa: Reincarnating Reinforcement Learning
Figure 4 for Beyond Tabula Rasa: Reincarnating Reinforcement Learning
Viaarxiv icon

Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time Reinforcement Learning

Add code
May 24, 2022
Figure 1 for Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time Reinforcement Learning
Figure 2 for Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time Reinforcement Learning
Figure 3 for Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time Reinforcement Learning
Figure 4 for Distributional Hamilton-Jacobi-Bellman Equations for Continuous-Time Reinforcement Learning
Viaarxiv icon

On the Generalization of Representations in Reinforcement Learning

Add code
Mar 01, 2022
Figure 1 for On the Generalization of Representations in Reinforcement Learning
Figure 2 for On the Generalization of Representations in Reinforcement Learning
Figure 3 for On the Generalization of Representations in Reinforcement Learning
Figure 4 for On the Generalization of Representations in Reinforcement Learning
Viaarxiv icon

On Bonus-Based Exploration Methods in the Arcade Learning Environment

Add code
Sep 22, 2021
Figure 1 for On Bonus-Based Exploration Methods in the Arcade Learning Environment
Figure 2 for On Bonus-Based Exploration Methods in the Arcade Learning Environment
Figure 3 for On Bonus-Based Exploration Methods in the Arcade Learning Environment
Figure 4 for On Bonus-Based Exploration Methods in the Arcade Learning Environment
Viaarxiv icon

Deep Reinforcement Learning at the Edge of the Statistical Precipice

Add code
Aug 30, 2021
Figure 1 for Deep Reinforcement Learning at the Edge of the Statistical Precipice
Figure 2 for Deep Reinforcement Learning at the Edge of the Statistical Precipice
Figure 3 for Deep Reinforcement Learning at the Edge of the Statistical Precipice
Figure 4 for Deep Reinforcement Learning at the Edge of the Statistical Precipice
Viaarxiv icon

Metrics and continuity in reinforcement learning

Add code
Feb 02, 2021
Figure 1 for Metrics and continuity in reinforcement learning
Figure 2 for Metrics and continuity in reinforcement learning
Figure 3 for Metrics and continuity in reinforcement learning
Figure 4 for Metrics and continuity in reinforcement learning
Viaarxiv icon