Picture for Nicolas Heess

Nicolas Heess

Informatics

Credit Assignment Techniques in Stochastic Computation Graphs

Add code
Jan 07, 2019
Figure 1 for Credit Assignment Techniques in Stochastic Computation Graphs
Figure 2 for Credit Assignment Techniques in Stochastic Computation Graphs
Figure 3 for Credit Assignment Techniques in Stochastic Computation Graphs
Figure 4 for Credit Assignment Techniques in Stochastic Computation Graphs
Viaarxiv icon

Self-supervised Learning of Image Embedding for Continuous Control

Add code
Jan 03, 2019
Figure 1 for Self-supervised Learning of Image Embedding for Continuous Control
Figure 2 for Self-supervised Learning of Image Embedding for Continuous Control
Figure 3 for Self-supervised Learning of Image Embedding for Continuous Control
Figure 4 for Self-supervised Learning of Image Embedding for Continuous Control
Viaarxiv icon

Relative Entropy Regularized Policy Iteration

Add code
Dec 05, 2018
Figure 1 for Relative Entropy Regularized Policy Iteration
Figure 2 for Relative Entropy Regularized Policy Iteration
Figure 3 for Relative Entropy Regularized Policy Iteration
Figure 4 for Relative Entropy Regularized Policy Iteration
Viaarxiv icon

Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction

Add code
Dec 05, 2018
Figure 1 for Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction
Figure 2 for Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction
Figure 3 for Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction
Figure 4 for Entropic Policy Composition with Generalized Policy Improvement and Divergence Correction
Viaarxiv icon

Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures

Add code
Dec 04, 2018
Figure 1 for Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures
Figure 2 for Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures
Figure 3 for Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures
Figure 4 for Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures
Viaarxiv icon

Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search

Add code
Nov 15, 2018
Figure 1 for Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search
Figure 2 for Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search
Figure 3 for Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search
Figure 4 for Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search
Viaarxiv icon

Relational inductive biases, deep learning, and graph networks

Add code
Oct 17, 2018
Figure 1 for Relational inductive biases, deep learning, and graph networks
Figure 2 for Relational inductive biases, deep learning, and graph networks
Figure 3 for Relational inductive biases, deep learning, and graph networks
Figure 4 for Relational inductive biases, deep learning, and graph networks
Viaarxiv icon

Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards

Add code
Oct 08, 2018
Figure 1 for Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards
Figure 2 for Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards
Figure 3 for Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards
Figure 4 for Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards
Viaarxiv icon

Unsupervised Learning of 3D Structure from Images

Add code
Jun 19, 2018
Figure 1 for Unsupervised Learning of 3D Structure from Images
Figure 2 for Unsupervised Learning of 3D Structure from Images
Figure 3 for Unsupervised Learning of 3D Structure from Images
Figure 4 for Unsupervised Learning of 3D Structure from Images
Viaarxiv icon

Maximum a Posteriori Policy Optimisation

Add code
Jun 14, 2018
Figure 1 for Maximum a Posteriori Policy Optimisation
Figure 2 for Maximum a Posteriori Policy Optimisation
Figure 3 for Maximum a Posteriori Policy Optimisation
Figure 4 for Maximum a Posteriori Policy Optimisation
Viaarxiv icon