Alert button
Picture for Aviral Kumar

Aviral Kumar

Alert button

A Workflow for Offline Model-Free Robotic Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 23, 2021
Aviral Kumar, Anikait Singh, Stephen Tian, Chelsea Finn, Sergey Levine

Figure 1 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 2 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 3 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Figure 4 for A Workflow for Offline Model-Free Robotic Reinforcement Learning
Viaarxiv icon

Conservative Data Sharing for Multi-Task Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 16, 2021
Tianhe Yu, Aviral Kumar, Yevgen Chebotar, Karol Hausman, Sergey Levine, Chelsea Finn

Figure 1 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 2 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 3 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Figure 4 for Conservative Data Sharing for Multi-Task Offline Reinforcement Learning
Viaarxiv icon

Conservative Objective Models for Effective Offline Model-Based Optimization

Add code
Bookmark button
Alert button
Jul 14, 2021
Brandon Trabucco, Aviral Kumar, Xinyang Geng, Sergey Levine

Figure 1 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 2 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 3 for Conservative Objective Models for Effective Offline Model-Based Optimization
Figure 4 for Conservative Objective Models for Effective Offline Model-Based Optimization
Viaarxiv icon

Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability

Add code
Bookmark button
Alert button
Jul 13, 2021
Dibya Ghosh, Jad Rahme, Aviral Kumar, Amy Zhang, Ryan P. Adams, Sergey Levine

Figure 1 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 2 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 3 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Figure 4 for Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability
Viaarxiv icon

Benchmarks for Deep Off-Policy Evaluation

Add code
Bookmark button
Alert button
Mar 30, 2021
Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R. Zhang, Yutian Chen, Aviral Kumar, Cosmin Paduraru, Sergey Levine, Tom Le Paine

Figure 1 for Benchmarks for Deep Off-Policy Evaluation
Figure 2 for Benchmarks for Deep Off-Policy Evaluation
Figure 3 for Benchmarks for Deep Off-Policy Evaluation
Figure 4 for Benchmarks for Deep Off-Policy Evaluation
Viaarxiv icon

COMBO: Conservative Offline Model-Based Policy Optimization

Add code
Bookmark button
Alert button
Feb 16, 2021
Tianhe Yu, Aviral Kumar, Rafael Rafailov, Aravind Rajeswaran, Sergey Levine, Chelsea Finn

Figure 1 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 2 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 3 for COMBO: Conservative Offline Model-Based Policy Optimization
Figure 4 for COMBO: Conservative Offline Model-Based Policy Optimization
Viaarxiv icon

COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2020
Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine

Figure 1 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 2 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 3 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 4 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Viaarxiv icon

Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2020
Aviral Kumar, Rishabh Agarwal, Dibya Ghosh, Sergey Levine

Figure 1 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 2 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 3 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 4 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Viaarxiv icon

Conservative Safety Critics for Exploration

Add code
Bookmark button
Alert button
Oct 27, 2020
Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg

Figure 1 for Conservative Safety Critics for Exploration
Figure 2 for Conservative Safety Critics for Exploration
Figure 3 for Conservative Safety Critics for Exploration
Figure 4 for Conservative Safety Critics for Exploration
Viaarxiv icon

One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL

Add code
Bookmark button
Alert button
Oct 27, 2020
Saurabh Kumar, Aviral Kumar, Sergey Levine, Chelsea Finn

Figure 1 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Figure 2 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Figure 3 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Viaarxiv icon