Picture for Ilya Kostrikov

Ilya Kostrikov

Tony

In Defense of the Unitary Scalarization for Deep Multi-Task Learning

Add code
Jan 20, 2022
Figure 1 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 2 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 3 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Figure 4 for In Defense of the Unitary Scalarization for Deep Multi-Task Learning
Viaarxiv icon

RvS: What is Essential for Offline RL via Supervised Learning?

Add code
Dec 20, 2021
Figure 1 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 2 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 3 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 4 for RvS: What is Essential for Offline RL via Supervised Learning?
Viaarxiv icon

Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions

Add code
Nov 29, 2021
Figure 1 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 2 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 3 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 4 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Viaarxiv icon

Offline Reinforcement Learning with Implicit Q-Learning

Add code
Oct 12, 2021
Figure 1 for Offline Reinforcement Learning with Implicit Q-Learning
Figure 2 for Offline Reinforcement Learning with Implicit Q-Learning
Figure 3 for Offline Reinforcement Learning with Implicit Q-Learning
Figure 4 for Offline Reinforcement Learning with Implicit Q-Learning
Viaarxiv icon

Offline Reinforcement Learning with Fisher Divergence Critic Regularization

Add code
Mar 14, 2021
Figure 1 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 2 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 3 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 4 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Viaarxiv icon

Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation

Add code
Jul 27, 2020
Figure 1 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Figure 2 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Figure 3 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Viaarxiv icon

Automatic Data Augmentation for Generalization in Deep Reinforcement Learning

Add code
Jun 23, 2020
Figure 1 for Automatic Data Augmentation for Generalization in Deep Reinforcement Learning
Figure 2 for Automatic Data Augmentation for Generalization in Deep Reinforcement Learning
Figure 3 for Automatic Data Augmentation for Generalization in Deep Reinforcement Learning
Figure 4 for Automatic Data Augmentation for Generalization in Deep Reinforcement Learning
Viaarxiv icon

Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels

Add code
Apr 28, 2020
Figure 1 for Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels
Figure 2 for Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels
Figure 3 for Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels
Figure 4 for Image Augmentation Is All You Need: Regularizing Deep Reinforcement Learning from Pixels
Viaarxiv icon

Imitation Learning via Off-Policy Distribution Matching

Add code
Dec 10, 2019
Figure 1 for Imitation Learning via Off-Policy Distribution Matching
Figure 2 for Imitation Learning via Off-Policy Distribution Matching
Figure 3 for Imitation Learning via Off-Policy Distribution Matching
Figure 4 for Imitation Learning via Off-Policy Distribution Matching
Viaarxiv icon

AlgaeDICE: Policy Gradient from Arbitrary Experience

Add code
Dec 04, 2019
Figure 1 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 2 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 3 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Figure 4 for AlgaeDICE: Policy Gradient from Arbitrary Experience
Viaarxiv icon