Picture for Satinder Singh

Satinder Singh

Discovery of Options via Meta-Learned Subgoals

Add code
Feb 12, 2021
Figure 1 for Discovery of Options via Meta-Learned Subgoals
Figure 2 for Discovery of Options via Meta-Learned Subgoals
Figure 3 for Discovery of Options via Meta-Learned Subgoals
Figure 4 for Discovery of Options via Meta-Learned Subgoals
Viaarxiv icon

Pairwise Weights for Temporal Credit Assignment

Add code
Feb 09, 2021
Figure 1 for Pairwise Weights for Temporal Credit Assignment
Figure 2 for Pairwise Weights for Temporal Credit Assignment
Figure 3 for Pairwise Weights for Temporal Credit Assignment
Figure 4 for Pairwise Weights for Temporal Credit Assignment
Viaarxiv icon

Learning State Representations from Random Deep Action-conditional Predictions

Add code
Feb 09, 2021
Figure 1 for Learning State Representations from Random Deep Action-conditional Predictions
Figure 2 for Learning State Representations from Random Deep Action-conditional Predictions
Figure 3 for Learning State Representations from Random Deep Action-conditional Predictions
Figure 4 for Learning State Representations from Random Deep Action-conditional Predictions
Viaarxiv icon

Efficient Querying for Cooperative Probabilistic Commitments

Add code
Dec 14, 2020
Figure 1 for Efficient Querying for Cooperative Probabilistic Commitments
Figure 2 for Efficient Querying for Cooperative Probabilistic Commitments
Figure 3 for Efficient Querying for Cooperative Probabilistic Commitments
Figure 4 for Efficient Querying for Cooperative Probabilistic Commitments
Viaarxiv icon

The Value Equivalence Principle for Model-Based Reinforcement Learning

Add code
Nov 06, 2020
Figure 1 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 2 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 3 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Figure 4 for The Value Equivalence Principle for Model-Based Reinforcement Learning
Viaarxiv icon

Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in First-person Simulated 3D Environments

Add code
Oct 28, 2020
Figure 1 for Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in First-person Simulated 3D Environments
Figure 2 for Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in First-person Simulated 3D Environments
Figure 3 for Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in First-person Simulated 3D Environments
Figure 4 for Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in First-person Simulated 3D Environments
Viaarxiv icon

Discovering Reinforcement Learning Algorithms

Add code
Jul 17, 2020
Figure 1 for Discovering Reinforcement Learning Algorithms
Figure 2 for Discovering Reinforcement Learning Algorithms
Figure 3 for Discovering Reinforcement Learning Algorithms
Figure 4 for Discovering Reinforcement Learning Algorithms
Viaarxiv icon

Meta-Gradient Reinforcement Learning with an Objective Discovered Online

Add code
Jul 16, 2020
Figure 1 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 2 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 3 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Figure 4 for Meta-Gradient Reinforcement Learning with an Objective Discovered Online
Viaarxiv icon

Learning to Play No-Press Diplomacy with Best Response Policy Iteration

Add code
Jun 17, 2020
Figure 1 for Learning to Play No-Press Diplomacy with Best Response Policy Iteration
Figure 2 for Learning to Play No-Press Diplomacy with Best Response Policy Iteration
Figure 3 for Learning to Play No-Press Diplomacy with Best Response Policy Iteration
Figure 4 for Learning to Play No-Press Diplomacy with Best Response Policy Iteration
Viaarxiv icon

Self-Tuning Deep Reinforcement Learning

Add code
Mar 02, 2020
Figure 1 for Self-Tuning Deep Reinforcement Learning
Figure 2 for Self-Tuning Deep Reinforcement Learning
Figure 3 for Self-Tuning Deep Reinforcement Learning
Figure 4 for Self-Tuning Deep Reinforcement Learning
Viaarxiv icon