Alert button
Picture for Ofir Nachum

Ofir Nachum

Alert button

Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error

Add code
Bookmark button
Alert button
Jan 28, 2022
Scott Fujimoto, David Meger, Doina Precup, Ofir Nachum, Shixiang Shane Gu

Figure 1 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 2 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 3 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 4 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Viaarxiv icon

Model Selection in Batch Policy Optimization

Add code
Bookmark button
Alert button
Dec 23, 2021
Jonathan N. Lee, George Tucker, Ofir Nachum, Bo Dai

Figure 1 for Model Selection in Batch Policy Optimization
Figure 2 for Model Selection in Batch Policy Optimization
Viaarxiv icon

Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions

Add code
Bookmark button
Alert button
Nov 29, 2021
Bogdan Mazoure, Ilya Kostrikov, Ofir Nachum, Jonathan Tompson

Figure 1 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 2 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 3 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Figure 4 for Improving Zero-shot Generalization in Offline Reinforcement Learning using Generalized Similarity Functions
Viaarxiv icon

TRAIL: Near-Optimal Imitation Learning with Suboptimal Data

Add code
Bookmark button
Alert button
Oct 27, 2021
Mengjiao Yang, Sergey Levine, Ofir Nachum

Figure 1 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 2 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 3 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 4 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Viaarxiv icon

Policy Gradients Incorporating the Future

Add code
Bookmark button
Alert button
Aug 11, 2021
David Venuto, Elaine Lau, Doina Precup, Ofir Nachum

Figure 1 for Policy Gradients Incorporating the Future
Figure 2 for Policy Gradients Incorporating the Future
Figure 3 for Policy Gradients Incorporating the Future
Figure 4 for Policy Gradients Incorporating the Future
Viaarxiv icon

Provable Representation Learning for Imitation with Contrastive Fourier Features

Add code
Bookmark button
Alert button
May 26, 2021
Ofir Nachum, Mengjiao Yang

Figure 1 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 2 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 3 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Figure 4 for Provable Representation Learning for Imitation with Contrastive Fourier Features
Viaarxiv icon

Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization

Add code
Bookmark button
Alert button
Apr 28, 2021
Michael R. Zhang, Tom Le Paine, Ofir Nachum, Cosmin Paduraru, George Tucker, Ziyu Wang, Mohammad Norouzi

Figure 1 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 2 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 3 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Figure 4 for Autoregressive Dynamics Models for Offline Policy Evaluation and Optimization
Viaarxiv icon

Benchmarks for Deep Off-Policy Evaluation

Add code
Bookmark button
Alert button
Mar 30, 2021
Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R. Zhang, Yutian Chen, Aviral Kumar, Cosmin Paduraru, Sergey Levine, Tom Le Paine

Figure 1 for Benchmarks for Deep Off-Policy Evaluation
Figure 2 for Benchmarks for Deep Off-Policy Evaluation
Figure 3 for Benchmarks for Deep Off-Policy Evaluation
Figure 4 for Benchmarks for Deep Off-Policy Evaluation
Viaarxiv icon

Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 23, 2021
Hiroki Furuta, Tatsuya Matsushima, Tadashi Kozuno, Yutaka Matsuo, Sergey Levine, Ofir Nachum, Shixiang Shane Gu

Figure 1 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 2 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 3 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 4 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Viaarxiv icon