Alert button
Picture for Philip S. Thomas

Philip S. Thomas

Alert button

Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation

Add code
Bookmark button
Alert button
Dec 05, 2017
Zhaohan Daniel Guo, Philip S. Thomas, Emma Brunskill

Figure 1 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 2 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 3 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Figure 4 for Using Options and Covariance Testing for Long Horizon Off-Policy Policy Evaluation
Viaarxiv icon

On Ensuring that Intelligent Machines Are Well-Behaved

Add code
Bookmark button
Alert button
Aug 17, 2017
Philip S. Thomas, Bruno Castro da Silva, Andrew G. Barto, Emma Brunskill

Figure 1 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 2 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 3 for On Ensuring that Intelligent Machines Are Well-Behaved
Figure 4 for On Ensuring that Intelligent Machines Are Well-Behaved
Viaarxiv icon

Policy Gradient Methods for Reinforcement Learning with Function Approximation and Action-Dependent Baselines

Add code
Bookmark button
Alert button
Jun 20, 2017
Philip S. Thomas, Emma Brunskill

Viaarxiv icon

Data-Efficient Policy Evaluation Through Behavior Policy Search

Add code
Bookmark button
Alert button
Jun 12, 2017
Josiah P. Hanna, Philip S. Thomas, Peter Stone, Scott Niekum

Figure 1 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Figure 2 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Figure 3 for Data-Efficient Policy Evaluation Through Behavior Policy Search
Viaarxiv icon

Decoupling Learning Rules from Representations

Add code
Bookmark button
Alert button
Jun 09, 2017
Philip S. Thomas, Christoph Dann, Emma Brunskill

Figure 1 for Decoupling Learning Rules from Representations
Figure 2 for Decoupling Learning Rules from Representations
Viaarxiv icon

Importance Sampling with Unequal Support

Add code
Bookmark button
Alert button
Nov 10, 2016
Philip S. Thomas, Emma Brunskill

Figure 1 for Importance Sampling with Unequal Support
Figure 2 for Importance Sampling with Unequal Support
Figure 3 for Importance Sampling with Unequal Support
Figure 4 for Importance Sampling with Unequal Support
Viaarxiv icon

A Notation for Markov Decision Processes

Add code
Bookmark button
Alert button
Sep 08, 2016
Philip S. Thomas, Billy Okal

Viaarxiv icon

Data-Efficient Off-Policy Policy Evaluation for Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 04, 2016
Philip S. Thomas, Emma Brunskill

Figure 1 for Data-Efficient Off-Policy Policy Evaluation for Reinforcement Learning
Figure 2 for Data-Efficient Off-Policy Policy Evaluation for Reinforcement Learning
Viaarxiv icon

Increasing the Action Gap: New Operators for Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 15, 2015
Marc G. Bellemare, Georg Ostrovski, Arthur Guez, Philip S. Thomas, Rémi Munos

Figure 1 for Increasing the Action Gap: New Operators for Reinforcement Learning
Figure 2 for Increasing the Action Gap: New Operators for Reinforcement Learning
Figure 3 for Increasing the Action Gap: New Operators for Reinforcement Learning
Figure 4 for Increasing the Action Gap: New Operators for Reinforcement Learning
Viaarxiv icon