Picture for Martha White

Martha White

Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling

Add code
Dec 03, 2018
Figure 1 for Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Figure 2 for Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Figure 3 for Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Figure 4 for Accelerating Large Scale Knowledge Distillation via Dynamic Importance Sampling
Viaarxiv icon

An Off-policy Policy Gradient Theorem Using Emphatic Weightings

Add code
Nov 22, 2018
Figure 1 for An Off-policy Policy Gradient Theorem Using Emphatic Weightings
Figure 2 for An Off-policy Policy Gradient Theorem Using Emphatic Weightings
Figure 3 for An Off-policy Policy Gradient Theorem Using Emphatic Weightings
Figure 4 for An Off-policy Policy Gradient Theorem Using Emphatic Weightings
Viaarxiv icon

The Barbados 2018 List of Open Issues in Continual Learning

Add code
Nov 16, 2018
Viaarxiv icon

Context-Dependent Upper-Confidence Bounds for Directed Exploration

Add code
Nov 15, 2018
Figure 1 for Context-Dependent Upper-Confidence Bounds for Directed Exploration
Figure 2 for Context-Dependent Upper-Confidence Bounds for Directed Exploration
Viaarxiv icon

The Utility of Sparse Representations for Control in Reinforcement Learning

Add code
Nov 15, 2018
Figure 1 for The Utility of Sparse Representations for Control in Reinforcement Learning
Figure 2 for The Utility of Sparse Representations for Control in Reinforcement Learning
Figure 3 for The Utility of Sparse Representations for Control in Reinforcement Learning
Figure 4 for The Utility of Sparse Representations for Control in Reinforcement Learning
Viaarxiv icon

Online Off-policy Prediction

Add code
Nov 06, 2018
Figure 1 for Online Off-policy Prediction
Figure 2 for Online Off-policy Prediction
Figure 3 for Online Off-policy Prediction
Figure 4 for Online Off-policy Prediction
Viaarxiv icon

Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces

Add code
Oct 22, 2018
Figure 1 for Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces
Figure 2 for Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces
Figure 3 for Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces
Figure 4 for Actor-Expert: A Framework for using Action-Value Methods in Continuous Action Spaces
Viaarxiv icon

High-confidence error estimates for learned value functions

Add code
Aug 28, 2018
Figure 1 for High-confidence error estimates for learned value functions
Figure 2 for High-confidence error estimates for learned value functions
Viaarxiv icon

General Value Function Networks

Add code
Jul 18, 2018
Figure 1 for General Value Function Networks
Figure 2 for General Value Function Networks
Figure 3 for General Value Function Networks
Figure 4 for General Value Function Networks
Viaarxiv icon

Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control

Add code
Jun 13, 2018
Figure 1 for Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control
Figure 2 for Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control
Figure 3 for Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control
Figure 4 for Reinforcement Learning with Function-Valued Action Spaces for Partial Differential Equation Control
Viaarxiv icon