Picture for Doina Precup

Doina Precup

McGill University, Mila- Quebec Artificial Intelligence Institute

A Survey of Exploration Methods in Reinforcement Learning

Add code
Sep 02, 2021
Figure 1 for A Survey of Exploration Methods in Reinforcement Learning
Figure 2 for A Survey of Exploration Methods in Reinforcement Learning
Figure 3 for A Survey of Exploration Methods in Reinforcement Learning
Figure 4 for A Survey of Exploration Methods in Reinforcement Learning
Viaarxiv icon

Policy Gradients Incorporating the Future

Add code
Aug 11, 2021
Figure 1 for Policy Gradients Incorporating the Future
Figure 2 for Policy Gradients Incorporating the Future
Figure 3 for Policy Gradients Incorporating the Future
Figure 4 for Policy Gradients Incorporating the Future
Viaarxiv icon

Temporally Abstract Partial Models

Add code
Aug 06, 2021
Figure 1 for Temporally Abstract Partial Models
Figure 2 for Temporally Abstract Partial Models
Figure 3 for Temporally Abstract Partial Models
Figure 4 for Temporally Abstract Partial Models
Viaarxiv icon

The Option Keyboard: Combining Skills in Reinforcement Learning

Add code
Jun 24, 2021
Figure 1 for The Option Keyboard: Combining Skills in Reinforcement Learning
Figure 2 for The Option Keyboard: Combining Skills in Reinforcement Learning
Figure 3 for The Option Keyboard: Combining Skills in Reinforcement Learning
Figure 4 for The Option Keyboard: Combining Skills in Reinforcement Learning
Viaarxiv icon

Randomized Exploration for Reinforcement Learning with General Value Function Approximation

Add code
Jun 15, 2021
Figure 1 for Randomized Exploration for Reinforcement Learning with General Value Function Approximation
Figure 2 for Randomized Exploration for Reinforcement Learning with General Value Function Approximation
Figure 3 for Randomized Exploration for Reinforcement Learning with General Value Function Approximation
Figure 4 for Randomized Exploration for Reinforcement Learning with General Value Function Approximation
Viaarxiv icon

A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation

Add code
Jun 12, 2021
Figure 1 for A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation
Figure 2 for A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation
Figure 3 for A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation
Figure 4 for A Deep Reinforcement Learning Approach to Marginalized Importance Sampling with the Successor Representation
Viaarxiv icon

Preferential Temporal Difference Learning

Add code
Jun 11, 2021
Figure 1 for Preferential Temporal Difference Learning
Figure 2 for Preferential Temporal Difference Learning
Figure 3 for Preferential Temporal Difference Learning
Figure 4 for Preferential Temporal Difference Learning
Viaarxiv icon

Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation

Add code
Jun 08, 2021
Figure 1 for Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation
Figure 2 for Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation
Figure 3 for Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation
Figure 4 for Flow Network based Generative Models for Non-Iterative Diverse Candidate Generation
Viaarxiv icon

Correcting Momentum in Temporal Difference Learning

Add code
Jun 07, 2021
Figure 1 for Correcting Momentum in Temporal Difference Learning
Figure 2 for Correcting Momentum in Temporal Difference Learning
Figure 3 for Correcting Momentum in Temporal Difference Learning
Figure 4 for Correcting Momentum in Temporal Difference Learning
Viaarxiv icon

A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning

Add code
Jun 03, 2021
Figure 1 for A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning
Figure 2 for A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning
Figure 3 for A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning
Figure 4 for A Consciousness-Inspired Planning Agent for Model-Based Reinforcement Learning
Viaarxiv icon