Picture for Ofir Nachum

Ofir Nachum

Tony

Benchmarks for Deep Off-Policy Evaluation

Add code
Mar 30, 2021
Figure 1 for Benchmarks for Deep Off-Policy Evaluation
Figure 2 for Benchmarks for Deep Off-Policy Evaluation
Figure 3 for Benchmarks for Deep Off-Policy Evaluation
Figure 4 for Benchmarks for Deep Off-Policy Evaluation
Viaarxiv icon

Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning

Add code
Mar 23, 2021
Figure 1 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 2 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 3 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Figure 4 for Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning
Viaarxiv icon

Near Optimal Policy Optimization via REPS

Add code
Mar 17, 2021
Viaarxiv icon

Offline Reinforcement Learning with Fisher Divergence Critic Regularization

Add code
Mar 14, 2021
Figure 1 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 2 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 3 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Figure 4 for Offline Reinforcement Learning with Fisher Divergence Critic Regularization
Viaarxiv icon

Representation Matters: Offline Pretraining for Sequential Decision Making

Add code
Feb 11, 2021
Figure 1 for Representation Matters: Offline Pretraining for Sequential Decision Making
Figure 2 for Representation Matters: Offline Pretraining for Sequential Decision Making
Figure 3 for Representation Matters: Offline Pretraining for Sequential Decision Making
Figure 4 for Representation Matters: Offline Pretraining for Sequential Decision Making
Viaarxiv icon

Offline Policy Selection under Uncertainty

Add code
Dec 12, 2020
Figure 1 for Offline Policy Selection under Uncertainty
Figure 2 for Offline Policy Selection under Uncertainty
Figure 3 for Offline Policy Selection under Uncertainty
Figure 4 for Offline Policy Selection under Uncertainty
Viaarxiv icon

OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning

Add code
Oct 27, 2020
Figure 1 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 2 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 3 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 4 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Viaarxiv icon

CoinDICE: Off-Policy Confidence Interval Estimation

Add code
Oct 22, 2020
Figure 1 for CoinDICE: Off-Policy Confidence Interval Estimation
Figure 2 for CoinDICE: Off-Policy Confidence Interval Estimation
Figure 3 for CoinDICE: Off-Policy Confidence Interval Estimation
Viaarxiv icon

Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation

Add code
Jul 27, 2020
Figure 1 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Figure 2 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Figure 3 for Statistical Bootstrapping for Uncertainty Estimation in Off-Policy Evaluation
Viaarxiv icon

Off-Policy Evaluation via the Regularized Lagrangian

Add code
Jul 07, 2020
Figure 1 for Off-Policy Evaluation via the Regularized Lagrangian
Figure 2 for Off-Policy Evaluation via the Regularized Lagrangian
Figure 3 for Off-Policy Evaluation via the Regularized Lagrangian
Figure 4 for Off-Policy Evaluation via the Regularized Lagrangian
Viaarxiv icon