Picture for John Quan

John Quan

Vision-Language Models as a Source of Rewards

Add code
Dec 14, 2023
Figure 1 for Vision-Language Models as a Source of Rewards
Figure 2 for Vision-Language Models as a Source of Rewards
Figure 3 for Vision-Language Models as a Source of Rewards
Figure 4 for Vision-Language Models as a Source of Rewards
Viaarxiv icon

The Phenomenon of Policy Churn

Add code
Jun 09, 2022
Figure 1 for The Phenomenon of Policy Churn
Figure 2 for The Phenomenon of Policy Churn
Figure 3 for The Phenomenon of Policy Churn
Figure 4 for The Phenomenon of Policy Churn
Viaarxiv icon

Podracer architectures for scalable Reinforcement Learning

Add code
Apr 13, 2021
Figure 1 for Podracer architectures for scalable Reinforcement Learning
Figure 2 for Podracer architectures for scalable Reinforcement Learning
Figure 3 for Podracer architectures for scalable Reinforcement Learning
Figure 4 for Podracer architectures for scalable Reinforcement Learning
Viaarxiv icon

The Value-Improvement Path: Towards Better Representations for Reinforcement Learning

Add code
Jun 03, 2020
Figure 1 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 2 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 3 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Figure 4 for The Value-Improvement Path: Towards Better Representations for Reinforcement Learning
Viaarxiv icon

General non-linear Bellman equations

Add code
Jul 08, 2019
Figure 1 for General non-linear Bellman equations
Figure 2 for General non-linear Bellman equations
Viaarxiv icon

Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement

Add code
Jan 30, 2019
Figure 1 for Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement
Figure 2 for Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement
Figure 3 for Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement
Figure 4 for Transfer in Deep Reinforcement Learning Using Successor Features and Generalised Policy Improvement
Viaarxiv icon

Universal Successor Features Approximators

Add code
Dec 18, 2018
Figure 1 for Universal Successor Features Approximators
Figure 2 for Universal Successor Features Approximators
Figure 3 for Universal Successor Features Approximators
Figure 4 for Universal Successor Features Approximators
Viaarxiv icon

Unicorn: Continual Learning with a Universal, Off-policy Agent

Add code
Jul 03, 2018
Figure 1 for Unicorn: Continual Learning with a Universal, Off-policy Agent
Figure 2 for Unicorn: Continual Learning with a Universal, Off-policy Agent
Figure 3 for Unicorn: Continual Learning with a Universal, Off-policy Agent
Figure 4 for Unicorn: Continual Learning with a Universal, Off-policy Agent
Viaarxiv icon

Observe and Look Further: Achieving Consistent Performance on Atari

Add code
May 29, 2018
Figure 1 for Observe and Look Further: Achieving Consistent Performance on Atari
Figure 2 for Observe and Look Further: Achieving Consistent Performance on Atari
Figure 3 for Observe and Look Further: Achieving Consistent Performance on Atari
Figure 4 for Observe and Look Further: Achieving Consistent Performance on Atari
Viaarxiv icon

Distributed Prioritized Experience Replay

Add code
Mar 02, 2018
Figure 1 for Distributed Prioritized Experience Replay
Figure 2 for Distributed Prioritized Experience Replay
Figure 3 for Distributed Prioritized Experience Replay
Figure 4 for Distributed Prioritized Experience Replay
Viaarxiv icon