Picture for Aviral Kumar

Aviral Kumar

How to Leverage Unlabeled Data in Offline Reinforcement Learning

Add code
Feb 03, 2022
Figure 1 for How to Leverage Unlabeled Data in Offline Reinforcement Learning
Figure 2 for How to Leverage Unlabeled Data in Offline Reinforcement Learning
Figure 3 for How to Leverage Unlabeled Data in Offline Reinforcement Learning
Figure 4 for How to Leverage Unlabeled Data in Offline Reinforcement Learning
Viaarxiv icon

DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization

Add code
Dec 09, 2021
Figure 1 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 2 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 3 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 4 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Viaarxiv icon

Data-Driven Offline Optimization For Architecting Hardware Accelerators

Add code
Oct 20, 2021
Figure 1 for Data-Driven Offline Optimization For Architecting Hardware Accelerators
Figure 2 for Data-Driven Offline Optimization For Architecting Hardware Accelerators
Figure 3 for Data-Driven Offline Optimization For Architecting Hardware Accelerators
Figure 4 for Data-Driven Offline Optimization For Architecting Hardware Accelerators
Viaarxiv icon

A Workflow for Offline Model-Free Robotic Reinforcement Learning

Add code
Sep 23, 2021
Figure 1 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 2 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 3 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 4 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Viaarxiv icon

Conservative Data Sharing for Multi-Task Offline Reinforcement Learning

Add code
Sep 16, 2021
Figure 1 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 2 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 3 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 4 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Viaarxiv icon

Conservative Objective Models for Effective Offline Model-Based Optimization

Add code
Jul 14, 2021
Figure 1 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 2 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 3 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 4 for Conservative Objective Models for Effective Offline Model-Based Optimization
Viaarxiv icon

Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability

Add code
Jul 13, 2021
Figure 1 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 2 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 3 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 4 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Viaarxiv icon

Benchmarks for Deep Off-Policy Evaluation

Add code
Mar 30, 2021
Figure 1 for Benchmarks for Deep Off-Policy Evaluation
Figure 2 for Benchmarks for Deep Off-Policy Evaluation
Figure 3 for Benchmarks for Deep Off-Policy Evaluation
Figure 4 for Benchmarks for Deep Off-Policy Evaluation
Viaarxiv icon

COMBO: Conservative Offline Model-Based Policy Optimization

Add code
Feb 16, 2021
Figure 1 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 2 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 3 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 4 for COMBO: Conservative Offline Model-Based Policy Optimization
Viaarxiv icon

COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning

Add code
Oct 27, 2020
Figure 1 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 2 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 3 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 4 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Viaarxiv icon