Picture for Emma Brunskill

Emma Brunskill

Stanford University

Online Model Selection for Reinforcement Learning with Function Approximation

Add code
Nov 19, 2020
Figure 1 for Online Model Selection for Reinforcement Learning with Function Approximation
Viaarxiv icon

Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration

Add code
Aug 18, 2020
Figure 1 for Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Figure 2 for Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Viaarxiv icon

Provably Good Batch Reinforcement Learning Without Great Exploration

Add code
Jul 22, 2020
Figure 1 for Provably Good Batch Reinforcement Learning Without Great Exploration
Figure 2 for Provably Good Batch Reinforcement Learning Without Great Exploration
Figure 3 for Provably Good Batch Reinforcement Learning Without Great Exploration
Figure 4 for Provably Good Batch Reinforcement Learning Without Great Exploration
Viaarxiv icon

Learning Abstract Models for Strategic Exploration and Fast Reward Transfer

Add code
Jul 12, 2020
Figure 1 for Learning Abstract Models for Strategic Exploration and Fast Reward Transfer
Figure 2 for Learning Abstract Models for Strategic Exploration and Fast Reward Transfer
Figure 3 for Learning Abstract Models for Strategic Exploration and Fast Reward Transfer
Figure 4 for Learning Abstract Models for Strategic Exploration and Fast Reward Transfer
Viaarxiv icon

Power-Constrained Bandits

Add code
Apr 13, 2020
Figure 1 for Power-Constrained Bandits
Figure 2 for Power-Constrained Bandits
Figure 3 for Power-Constrained Bandits
Figure 4 for Power-Constrained Bandits
Viaarxiv icon

Value Driven Representation for Human-in-the-Loop Reinforcement Learning

Add code
Apr 02, 2020
Figure 1 for Value Driven Representation for Human-in-the-Loop Reinforcement Learning
Figure 2 for Value Driven Representation for Human-in-the-Loop Reinforcement Learning
Figure 3 for Value Driven Representation for Human-in-the-Loop Reinforcement Learning
Viaarxiv icon

Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding

Add code
Mar 12, 2020
Figure 1 for Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding
Figure 2 for Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding
Figure 3 for Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding
Figure 4 for Off-policy Policy Evaluation For Sequential Decisions Under Unobserved Confounding
Viaarxiv icon

Learning Near Optimal Policies with Low Inherent Bellman Error

Add code
Mar 05, 2020
Figure 1 for Learning Near Optimal Policies with Low Inherent Bellman Error
Viaarxiv icon

Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions

Add code
Feb 14, 2020
Figure 1 for Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions
Figure 2 for Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions
Figure 3 for Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions
Figure 4 for Interpretable Off-Policy Evaluation in Reinforcement Learning by Highlighting Influential Transitions
Viaarxiv icon

Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning

Add code
Jan 31, 2020
Figure 1 for Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning
Figure 2 for Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning
Figure 3 for Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning
Figure 4 for Towards the Systematic Reporting of the Energy and Carbon Footprints of Machine Learning
Viaarxiv icon