Picture for Martha White

Martha White

Maximizing Information Gain in Partially Observable Environments via Prediction Reward

Add code
May 11, 2020
Figure 1 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 2 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 3 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Figure 4 for Maximizing Information Gain in Partially Observable Environments via Prediction Reward
Viaarxiv icon

Maxmin Q-learning: Controlling the Estimation Bias of Q-learning

Add code
Feb 16, 2020
Figure 1 for Maxmin Q-learning: Controlling the Estimation Bias of Q-learning
Figure 2 for Maxmin Q-learning: Controlling the Estimation Bias of Q-learning
Figure 3 for Maxmin Q-learning: Controlling the Estimation Bias of Q-learning
Figure 4 for Maxmin Q-learning: Controlling the Estimation Bias of Q-learning
Viaarxiv icon

An implicit function learning approach for parametric modal regression

Add code
Feb 14, 2020
Figure 1 for An implicit function learning approach for parametric modal regression
Figure 2 for An implicit function learning approach for parametric modal regression
Figure 3 for An implicit function learning approach for parametric modal regression
Figure 4 for An implicit function learning approach for parametric modal regression
Viaarxiv icon

Is Fast Adaptation All You Need?

Add code
Oct 03, 2019
Figure 1 for Is Fast Adaptation All You Need?
Figure 2 for Is Fast Adaptation All You Need?
Figure 3 for Is Fast Adaptation All You Need?
Viaarxiv icon

Meta-descent for Online, Continual Prediction

Add code
Jul 17, 2019
Figure 1 for Meta-descent for Online, Continual Prediction
Figure 2 for Meta-descent for Online, Continual Prediction
Figure 3 for Meta-descent for Online, Continual Prediction
Figure 4 for Meta-descent for Online, Continual Prediction
Viaarxiv icon

Hill Climbing on Value Estimates for Search-control in Dyna

Add code
Jul 04, 2019
Figure 1 for Hill Climbing on Value Estimates for Search-control in Dyna
Figure 2 for Hill Climbing on Value Estimates for Search-control in Dyna
Figure 3 for Hill Climbing on Value Estimates for Search-control in Dyna
Figure 4 for Hill Climbing on Value Estimates for Search-control in Dyna
Viaarxiv icon

Adapting Behaviour via Intrinsic Reward: A Survey and Empirical Study

Add code
Jun 19, 2019
Figure 1 for Adapting Behaviour via Intrinsic Reward: A Survey and Empirical Study
Figure 2 for Adapting Behaviour via Intrinsic Reward: A Survey and Empirical Study
Figure 3 for Adapting Behaviour via Intrinsic Reward: A Survey and Empirical Study
Figure 4 for Adapting Behaviour via Intrinsic Reward: A Survey and Empirical Study
Viaarxiv icon

Importance Resampling for Off-policy Prediction

Add code
Jun 11, 2019
Figure 1 for Importance Resampling for Off-policy Prediction
Figure 2 for Importance Resampling for Off-policy Prediction
Figure 3 for Importance Resampling for Off-policy Prediction
Figure 4 for Importance Resampling for Off-policy Prediction
Viaarxiv icon

Meta-Learning Representations for Continual Learning

Add code
May 29, 2019
Figure 1 for Meta-Learning Representations for Continual Learning
Figure 2 for Meta-Learning Representations for Continual Learning
Figure 3 for Meta-Learning Representations for Continual Learning
Figure 4 for Meta-Learning Representations for Continual Learning
Viaarxiv icon

Planning with Expectation Models

Add code
Apr 03, 2019
Figure 1 for Planning with Expectation Models
Figure 2 for Planning with Expectation Models
Figure 3 for Planning with Expectation Models
Viaarxiv icon