Picture for Sergey Levine

Sergey Levine

Stanford University

RvS: What is Essential for Offline RL via Supervised Learning?

Add code
Dec 20, 2021
Figure 1 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 2 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 3 for RvS: What is Essential for Offline RL via Supervised Learning?
Figure 4 for RvS: What is Essential for Offline RL via Supervised Learning?
Viaarxiv icon

Autonomous Reinforcement Learning: Formalism and Benchmarking

Add code
Dec 17, 2021
Figure 1 for Autonomous Reinforcement Learning: Formalism and Benchmarking
Figure 2 for Autonomous Reinforcement Learning: Formalism and Benchmarking
Figure 3 for Autonomous Reinforcement Learning: Formalism and Benchmarking
Figure 4 for Autonomous Reinforcement Learning: Formalism and Benchmarking
Viaarxiv icon

Extending the WILDS Benchmark for Unsupervised Adaptation

Add code
Dec 09, 2021
Figure 1 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 2 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 3 for Extending the WILDS Benchmark for Unsupervised Adaptation
Figure 4 for Extending the WILDS Benchmark for Unsupervised Adaptation
Viaarxiv icon

DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization

Add code
Dec 09, 2021
Figure 1 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 2 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 3 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Figure 4 for DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization
Viaarxiv icon

CoMPS: Continual Meta Policy Search

Add code
Dec 08, 2021
Figure 1 for CoMPS: Continual Meta Policy Search
Figure 2 for CoMPS: Continual Meta Policy Search
Figure 3 for CoMPS: Continual Meta Policy Search
Figure 4 for CoMPS: Continual Meta Policy Search
Viaarxiv icon

Information is Power: Intrinsic Control via Information Capture

Add code
Dec 07, 2021
Figure 1 for Information is Power: Intrinsic Control via Information Capture
Figure 2 for Information is Power: Intrinsic Control via Information Capture
Figure 3 for Information is Power: Intrinsic Control via Information Capture
Figure 4 for Information is Power: Intrinsic Control via Information Capture
Viaarxiv icon

Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments

Add code
Nov 22, 2021
Figure 1 for Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments
Figure 2 for Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments
Figure 3 for Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments
Figure 4 for Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments
Viaarxiv icon

AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale

Add code
Nov 11, 2021
Figure 1 for AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale
Figure 2 for AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale
Figure 3 for AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale
Figure 4 for AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale
Viaarxiv icon

Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning

Add code
Nov 04, 2021
Figure 1 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 2 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 3 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Figure 4 for Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning
Viaarxiv icon

TRAIL: Near-Optimal Imitation Learning with Suboptimal Data

Add code
Oct 27, 2021
Figure 1 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 2 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 3 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Figure 4 for TRAIL: Near-Optimal Imitation Learning with Suboptimal Data
Viaarxiv icon