Picture for Harsh Satija

Harsh Satija

McGill University, Mila- Quebec Artificial Intelligence Institute

A Survey of Exploration Methods in Reinforcement Learning

Add code
Sep 02, 2021
Figure 1 for A Survey of Exploration Methods in Reinforcement Learning
Figure 2 for A Survey of Exploration Methods in Reinforcement Learning
Figure 3 for A Survey of Exploration Methods in Reinforcement Learning
Figure 4 for A Survey of Exploration Methods in Reinforcement Learning
Viaarxiv icon

Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs

Add code
May 31, 2021
Figure 1 for Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs
Figure 2 for Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs
Figure 3 for Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs
Figure 4 for Multi-Objective SPIBB: Seldonian Offline Policy Improvement with Safety Constraints in Finite MDPs
Viaarxiv icon

Locally Persistent Exploration in Continuous Control Tasks with Sparse Rewards

Add code
Dec 26, 2020
Figure 1 for Locally Persistent Exploration in Continuous Control Tasks with Sparse Rewards
Figure 2 for Locally Persistent Exploration in Continuous Control Tasks with Sparse Rewards
Figure 3 for Locally Persistent Exploration in Continuous Control Tasks with Sparse Rewards
Figure 4 for Locally Persistent Exploration in Continuous Control Tasks with Sparse Rewards
Viaarxiv icon

Constrained Markov Decision Processes via Backward Value Functions

Add code
Aug 26, 2020
Figure 1 for Constrained Markov Decision Processes via Backward Value Functions
Figure 2 for Constrained Markov Decision Processes via Backward Value Functions
Figure 3 for Constrained Markov Decision Processes via Backward Value Functions
Viaarxiv icon

Randomized Value Functions via Multiplicative Normalizing Flows

Add code
Oct 22, 2018
Figure 1 for Randomized Value Functions via Multiplicative Normalizing Flows
Figure 2 for Randomized Value Functions via Multiplicative Normalizing Flows
Figure 3 for Randomized Value Functions via Multiplicative Normalizing Flows
Figure 4 for Randomized Value Functions via Multiplicative Normalizing Flows
Viaarxiv icon

Decoupling Dynamics and Reward for Transfer Learning

Add code
May 09, 2018
Figure 1 for Decoupling Dynamics and Reward for Transfer Learning
Figure 2 for Decoupling Dynamics and Reward for Transfer Learning
Figure 3 for Decoupling Dynamics and Reward for Transfer Learning
Figure 4 for Decoupling Dynamics and Reward for Transfer Learning
Viaarxiv icon