Picture for George Tucker

George Tucker

Coupled Gradient Estimators for Discrete Latent Variables

Add code
Jun 15, 2021
Figure 1 for Coupled Gradient Estimators for Discrete Latent Variables
Figure 2 for Coupled Gradient Estimators for Discrete Latent Variables
Figure 3 for Coupled Gradient Estimators for Discrete Latent Variables
Figure 4 for Coupled Gradient Estimators for Discrete Latent Variables
Viaarxiv icon

Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization

Add code
Apr 28, 2021
Figure 1 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 2 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 3 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 4 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Viaarxiv icon

Benchmarks for Deep Off-Policy Evaluation

Add code
Mar 30, 2021
Figure 1 for Benchmarks for Deep Off-Policy Evaluation
Figure 2 for Benchmarks for Deep Off-Policy Evaluation
Figure 3 for Benchmarks for Deep Off-Policy Evaluation
Figure 4 for Benchmarks for Deep Off-Policy Evaluation
Viaarxiv icon

Offline Policy Selection under Uncertainty

Add code
Dec 12, 2020
Figure 1 for Offline Policy Selection under Uncertainty
Figure 2 for Offline Policy Selection under Uncertainty
Figure 3 for Offline Policy Selection under Uncertainty
Figure 4 for Offline Policy Selection under Uncertainty
Viaarxiv icon

RL Unplugged: Benchmarks for Offline Reinforcement Learning

Add code
Jul 02, 2020
Figure 1 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 2 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 3 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Figure 4 for RL Unplugged: Benchmarks for Offline Reinforcement Learning
Viaarxiv icon

Conservative Q-Learning for Offline Reinforcement Learning

Add code
Jun 29, 2020
Figure 1 for Conservative Q-Learning for Offline Reinforcement Learning
Figure 2 for Conservative Q-Learning for Offline Reinforcement Learning
Figure 3 for Conservative Q-Learning for Offline Reinforcement Learning
Figure 4 for Conservative Q-Learning for Offline Reinforcement Learning
Viaarxiv icon

DisARM: An Antithetic Gradient Estimator for Binary Latent Variables

Add code
Jun 18, 2020
Figure 1 for DisARM: An Antithetic Gradient Estimator for Binary Latent Variables
Figure 2 for DisARM: An Antithetic Gradient Estimator for Binary Latent Variables
Figure 3 for DisARM: An Antithetic Gradient Estimator for Binary Latent Variables
Figure 4 for DisARM: An Antithetic Gradient Estimator for Binary Latent Variables
Viaarxiv icon

Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems

Add code
May 04, 2020
Figure 1 for Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Figure 2 for Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Figure 3 for Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Figure 4 for Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems
Viaarxiv icon

D4RL: Datasets for Deep Data-Driven Reinforcement Learning

Add code
Apr 20, 2020
Figure 1 for D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Figure 2 for D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Figure 3 for D4RL: Datasets for Deep Data-Driven Reinforcement Learning
Viaarxiv icon

Meta-Learning without Memorization

Add code
Dec 24, 2019
Figure 1 for Meta-Learning without Memorization
Figure 2 for Meta-Learning without Memorization
Figure 3 for Meta-Learning without Memorization
Figure 4 for Meta-Learning without Memorization
Viaarxiv icon