Picture for Ofir Nachum

Ofir Nachum

Tony

Multi-Game Decision Transformers

Add code
May 30, 2022
Figure 1 for Multi-Game Decision Transformers
Figure 2 for Multi-Game Decision Transformers
Figure 3 for Multi-Game Decision Transformers
Figure 4 for Multi-Game Decision Transformers
Viaarxiv icon

Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters

Add code
May 27, 2022
Figure 1 for Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
Figure 2 for Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
Figure 3 for Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
Figure 4 for Why So Pessimistic? Estimating Uncertainties for Offline RL through Ensembles, and Why Their Independence Matters
Viaarxiv icon

Chain of Thought Imitation with Procedure Cloning

Add code
May 22, 2022
Figure 1 for Chain of Thought Imitation with Procedure Cloning
Figure 2 for Chain of Thought Imitation with Procedure Cloning
Figure 3 for Chain of Thought Imitation with Procedure Cloning
Figure 4 for Chain of Thought Imitation with Procedure Cloning
Viaarxiv icon

Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error

Add code
Jan 28, 2022
Figure 1 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 2 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 3 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 4 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Viaarxiv icon

Model Selection in Batch Policy Optimization

Add code
Dec 23, 2021
Figure 1 for Model Selection in Batch Policy Optimization
Figure 2 for Model Selection in Batch Policy Optimization
Viaarxiv icon

Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions

Add code
Nov 29, 2021
Figure 1 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 2 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 3 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 4 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Viaarxiv icon

TRAIL: Near-Optimal Imitation Learning with Suboptimal Data

Add code
Oct 27, 2021
Figure 1 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 2 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 3 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 4 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Viaarxiv icon

Policy Gradients Incorporating the Future

Add code
Aug 11, 2021
Figure 1 for Policy Gradients Incorporating the Future
Figure 2 for Policy Gradients Incorporating the Future
Figure 3 for Policy Gradients Incorporating the Future
Figure 4 for Policy Gradients Incorporating the Future
Viaarxiv icon

Provable Representation Learning for Imitation with Contrastive Fourier Features

Add code
May 26, 2021
Figure 1 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 2 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 3 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 4 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Viaarxiv icon

Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization

Add code
Apr 28, 2021
Figure 1 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 2 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 3 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 4 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Viaarxiv icon