Picture for George Tucker

George Tucker

Behavior Regularized Offline Reinforcement Learning

Add code
Nov 26, 2019
Figure 1 for Behavior Regularized Offline Reinforcement Learning
Figure 2 for Behavior Regularized Offline Reinforcement Learning
Figure 3 for Behavior Regularized Offline Reinforcement Learning
Figure 4 for Behavior Regularized Offline Reinforcement Learning
Viaarxiv icon

Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse

Add code
Nov 06, 2019
Figure 1 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 2 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 3 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Figure 4 for Don't Blame the ELBO! A Linear VAE Perspective on Posterior Collapse
Viaarxiv icon

Energy-Inspired Models: Learning with Sampler-Induced Distributions

Add code
Oct 31, 2019
Figure 1 for Energy-Inspired Models: Learning with Sampler-Induced Distributions
Figure 2 for Energy-Inspired Models: Learning with Sampler-Induced Distributions
Figure 3 for Energy-Inspired Models: Learning with Sampler-Induced Distributions
Figure 4 for Energy-Inspired Models: Learning with Sampler-Induced Distributions
Viaarxiv icon

Reinforcement Learning Driven Heuristic Optimization

Add code
Jun 16, 2019
Figure 1 for Reinforcement Learning Driven Heuristic Optimization
Figure 2 for Reinforcement Learning Driven Heuristic Optimization
Figure 3 for Reinforcement Learning Driven Heuristic Optimization
Figure 4 for Reinforcement Learning Driven Heuristic Optimization
Viaarxiv icon

Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction

Add code
Jun 03, 2019
Figure 1 for Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
Figure 2 for Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
Figure 3 for Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
Figure 4 for Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
Viaarxiv icon

On Variational Bounds of Mutual Information

Add code
May 16, 2019
Figure 1 for On Variational Bounds of Mutual Information
Figure 2 for On Variational Bounds of Mutual Information
Figure 3 for On Variational Bounds of Mutual Information
Figure 4 for On Variational Bounds of Mutual Information
Viaarxiv icon

Learning to Walk via Deep Reinforcement Learning

Add code
Mar 25, 2019
Figure 1 for Learning to Walk via Deep Reinforcement Learning
Figure 2 for Learning to Walk via Deep Reinforcement Learning
Figure 3 for Learning to Walk via Deep Reinforcement Learning
Figure 4 for Learning to Walk via Deep Reinforcement Learning
Viaarxiv icon

Model-Based Reinforcement Learning for Atari

Add code
Mar 05, 2019
Figure 1 for Model-Based Reinforcement Learning for Atari
Figure 2 for Model-Based Reinforcement Learning for Atari
Figure 3 for Model-Based Reinforcement Learning for Atari
Figure 4 for Model-Based Reinforcement Learning for Atari
Viaarxiv icon

Soft Actor-Critic Algorithms and Applications

Add code
Jan 29, 2019
Figure 1 for Soft Actor-Critic Algorithms and Applications
Figure 2 for Soft Actor-Critic Algorithms and Applications
Figure 3 for Soft Actor-Critic Algorithms and Applications
Figure 4 for Soft Actor-Critic Algorithms and Applications
Viaarxiv icon

The Laplacian in RL: Learning Representations with Efficient Approximations

Add code
Oct 10, 2018
Figure 1 for The Laplacian in RL: Learning Representations with Efficient Approximations
Figure 2 for The Laplacian in RL: Learning Representations with Efficient Approximations
Figure 3 for The Laplacian in RL: Learning Representations with Efficient Approximations
Figure 4 for The Laplacian in RL: Learning Representations with Efficient Approximations
Viaarxiv icon