Picture for Philip S. Thomas

Philip S. Thomas

A Compression-Inspired Framework for Macro Discovery

Add code
Feb 22, 2019
Figure 1 for A Compression-Inspired Framework for Macro Discovery
Figure 2 for A Compression-Inspired Framework for Macro Discovery
Figure 3 for A Compression-Inspired Framework for Macro Discovery
Figure 4 for A Compression-Inspired Framework for Macro Discovery
Viaarxiv icon

Asynchronous Coagent Networks: Stochastic Networks for Reinforcement Learning without Backpropagation or a Clock

Add code
Feb 21, 2019
Figure 1 for Asynchronous Coagent Networks: Stochastic Networks for Reinforcement Learning without Backpropagation or a Clock
Figure 2 for Asynchronous Coagent Networks: Stochastic Networks for Reinforcement Learning without Backpropagation or a Clock
Figure 3 for Asynchronous Coagent Networks: Stochastic Networks for Reinforcement Learning without Backpropagation or a Clock
Figure 4 for Asynchronous Coagent Networks: Stochastic Networks for Reinforcement Learning without Backpropagation or a Clock
Viaarxiv icon

A Meta-MDP Approach to Exploration for Lifelong Reinforcement Learning

Add code
Feb 03, 2019
Figure 1 for A Meta-MDP Approach to Exploration for Lifelong Reinforcement Learning
Figure 2 for A Meta-MDP Approach to Exploration for Lifelong Reinforcement Learning
Figure 3 for A Meta-MDP Approach to Exploration for Lifelong Reinforcement Learning
Viaarxiv icon

Learning Action Representations for Reinforcement Learning

Add code
Feb 01, 2019
Figure 1 for Learning Action Representations for Reinforcement Learning
Figure 2 for Learning Action Representations for Reinforcement Learning
Figure 3 for Learning Action Representations for Reinforcement Learning
Figure 4 for Learning Action Representations for Reinforcement Learning
Viaarxiv icon

Privacy Preserving Off-Policy Evaluation

Add code
Feb 01, 2019
Figure 1 for Privacy Preserving Off-Policy Evaluation
Figure 2 for Privacy Preserving Off-Policy Evaluation
Figure 3 for Privacy Preserving Off-Policy Evaluation
Figure 4 for Privacy Preserving Off-Policy Evaluation
Viaarxiv icon

Natural Option Critic

Add code
Dec 04, 2018
Figure 1 for Natural Option Critic
Figure 2 for Natural Option Critic
Figure 3 for Natural Option Critic
Figure 4 for Natural Option Critic
Viaarxiv icon

Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation

Add code
Dec 05, 2017
Figure 1 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 2 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 3 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 4 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Viaarxiv icon

On Ensuring that Intelligent Machines Are Well-Behaved

Add code
Aug 17, 2017
Figure 1 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 2 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 3 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 4 for On Ensuring that Intelligent Machines Are Well-Behaved
Viaarxiv icon

Policy Gradient Methods for Reinforcement Learning with Function Approximation and Action-Dependent Baselines

Add code
Jun 20, 2017
Viaarxiv icon

Data-Efficient Policy Evaluation Through Behavior Policy Search

Add code
Jun 12, 2017
Figure 1 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Figure 2 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Figure 3 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Viaarxiv icon