Picture for Shimon Whiteson

Shimon Whiteson

University of Oxford

On the Practical Consistency of Meta-Reinforcement Learning Algorithms

Add code
Dec 01, 2021
Figure 1 for On the Practical Consistency of Meta-Reinforcement Learning Algorithms
Figure 2 for On the Practical Consistency of Meta-Reinforcement Learning Algorithms
Figure 3 for On the Practical Consistency of Meta-Reinforcement Learning Algorithms
Figure 4 for On the Practical Consistency of Meta-Reinforcement Learning Algorithms
Viaarxiv icon

Reinforcement Learning in Factored Action Spaces using Tensor Decompositions

Add code
Oct 27, 2021
Figure 1 for Reinforcement Learning in Factored Action Spaces using Tensor Decompositions
Figure 2 for Reinforcement Learning in Factored Action Spaces using Tensor Decompositions
Figure 3 for Reinforcement Learning in Factored Action Spaces using Tensor Decompositions
Figure 4 for Reinforcement Learning in Factored Action Spaces using Tensor Decompositions
Viaarxiv icon

Model based Multi-agent Reinforcement Learning with Tensor Decompositions

Add code
Oct 27, 2021
Figure 1 for Model based Multi-agent Reinforcement Learning with Tensor Decompositions
Figure 2 for Model based Multi-agent Reinforcement Learning with Tensor Decompositions
Figure 3 for Model based Multi-agent Reinforcement Learning with Tensor Decompositions
Figure 4 for Model based Multi-agent Reinforcement Learning with Tensor Decompositions
Viaarxiv icon

Truncated Emphatic Temporal Difference Methods for Prediction and Control

Add code
Aug 11, 2021
Figure 1 for Truncated Emphatic Temporal Difference Methods for Prediction and Control
Figure 2 for Truncated Emphatic Temporal Difference Methods for Prediction and Control
Figure 3 for Truncated Emphatic Temporal Difference Methods for Prediction and Control
Figure 4 for Truncated Emphatic Temporal Difference Methods for Prediction and Control
Viaarxiv icon

Implicit Communication as Minimum Entropy Coupling

Add code
Jul 17, 2021
Figure 1 for Implicit Communication as Minimum Entropy Coupling
Figure 2 for Implicit Communication as Minimum Entropy Coupling
Figure 3 for Implicit Communication as Minimum Entropy Coupling
Figure 4 for Implicit Communication as Minimum Entropy Coupling
Viaarxiv icon

Bayesian Bellman Operators

Add code
Jun 15, 2021
Figure 1 for Bayesian Bellman Operators
Figure 2 for Bayesian Bellman Operators
Figure 3 for Bayesian Bellman Operators
Figure 4 for Bayesian Bellman Operators
Viaarxiv icon

SoftDICE for Imitation Learning: Rethinking Off-policy Distribution Matching

Add code
Jun 06, 2021
Figure 1 for SoftDICE for Imitation Learning: Rethinking Off-policy Distribution Matching
Figure 2 for SoftDICE for Imitation Learning: Rethinking Off-policy Distribution Matching
Figure 3 for SoftDICE for Imitation Learning: Rethinking Off-policy Distribution Matching
Figure 4 for SoftDICE for Imitation Learning: Rethinking Off-policy Distribution Matching
Viaarxiv icon

Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning

Add code
May 31, 2021
Figure 1 for Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning
Figure 2 for Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning
Figure 3 for Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning
Figure 4 for Tesseract: Tensorised Actors for Multi-Agent Reinforcement Learning
Viaarxiv icon

Semi-On-Policy Training for Sample Efficient Multi-Agent Policy Gradients

Add code
May 06, 2021
Figure 1 for Semi-On-Policy Training for Sample Efficient Multi-Agent Policy Gradients
Figure 2 for Semi-On-Policy Training for Sample Efficient Multi-Agent Policy Gradients
Figure 3 for Semi-On-Policy Training for Sample Efficient Multi-Agent Policy Gradients
Figure 4 for Semi-On-Policy Training for Sample Efficient Multi-Agent Policy Gradients
Viaarxiv icon

Softmax with Regularization: Better Value Estimation in Multi-Agent Reinforcement Learning

Add code
Mar 22, 2021
Figure 1 for Softmax with Regularization: Better Value Estimation in Multi-Agent Reinforcement Learning
Figure 2 for Softmax with Regularization: Better Value Estimation in Multi-Agent Reinforcement Learning
Figure 3 for Softmax with Regularization: Better Value Estimation in Multi-Agent Reinforcement Learning
Figure 4 for Softmax with Regularization: Better Value Estimation in Multi-Agent Reinforcement Learning
Viaarxiv icon