Alert button
Picture for Sergey Levine

Sergey Levine

Alert button

Reinforcement Learning with Videos: Combining Offline Observations with Interaction

Add code
Bookmark button
Alert button
Nov 12, 2020
Karl Schmeckpeper, Oleh Rybkin, Kostas Daniilidis, Sergey Levine, Chelsea Finn

Figure 1 for Reinforcement Learning with Videos: Combining Offline Observations with Interaction
Figure 2 for Reinforcement Learning with Videos: Combining Offline Observations with Interaction
Figure 3 for Reinforcement Learning with Videos: Combining Offline Observations with Interaction
Figure 4 for Reinforcement Learning with Videos: Combining Offline Observations with Interaction
Viaarxiv icon

Continual Learning of Control Primitives: Skill Discovery via Reset-Games

Add code
Bookmark button
Alert button
Nov 10, 2020
Kelvin Xu, Siddharth Verma, Chelsea Finn, Sergey Levine

Figure 1 for Continual Learning of Control Primitives: Skill Discovery via Reset-Games
Figure 2 for Continual Learning of Control Primitives: Skill Discovery via Reset-Games
Figure 3 for Continual Learning of Control Primitives: Skill Discovery via Reset-Games
Figure 4 for Continual Learning of Control Primitives: Skill Discovery via Reset-Games
Viaarxiv icon

Amortized Conditional Normalized Maximum Likelihood

Add code
Bookmark button
Alert button
Nov 05, 2020
Aurick Zhou, Sergey Levine

Figure 1 for Amortized Conditional Normalized Maximum Likelihood
Figure 2 for Amortized Conditional Normalized Maximum Likelihood
Figure 3 for Amortized Conditional Normalized Maximum Likelihood
Figure 4 for Amortized Conditional Normalized Maximum Likelihood
Viaarxiv icon

Rearrangement: A Challenge for Embodied AI

Add code
Bookmark button
Alert button
Nov 03, 2020
Dhruv Batra, Angel X. Chang, Sonia Chernova, Andrew J. Davison, Jia Deng, Vladlen Koltun, Sergey Levine, Jitendra Malik, Igor Mordatch, Roozbeh Mottaghi, Manolis Savva, Hao Su

Figure 1 for Rearrangement: A Challenge for Embodied AI
Figure 2 for Rearrangement: A Challenge for Embodied AI
Figure 3 for Rearrangement: A Challenge for Embodied AI
Figure 4 for Rearrangement: A Challenge for Embodied AI
Viaarxiv icon

COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2020
Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine

Figure 1 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 2 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 3 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 4 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Viaarxiv icon

Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2020
Aviral Kumar, Rishabh Agarwal, Dibya Ghosh, Sergey Levine

Figure 1 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 2 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 3 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Figure 4 for Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning
Viaarxiv icon

Conservative Safety Critics for Exploration

Add code
Bookmark button
Alert button
Oct 27, 2020
Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg

Figure 1 for Conservative Safety Critics for Exploration
Figure 2 for Conservative Safety Critics for Exploration
Figure 3 for Conservative Safety Critics for Exploration
Figure 4 for Conservative Safety Critics for Exploration
Viaarxiv icon

$γ$-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction

Add code
Bookmark button
Alert button
Oct 27, 2020
Michael Janner, Igor Mordatch, Sergey Levine

Figure 1 for $γ$-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction
Figure 2 for $γ$-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction
Figure 3 for $γ$-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction
Figure 4 for $γ$-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction
Viaarxiv icon

One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL

Add code
Bookmark button
Alert button
Oct 27, 2020
Saurabh Kumar, Aviral Kumar, Sergey Levine, Chelsea Finn

Figure 1 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Figure 2 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Figure 3 for One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL
Viaarxiv icon

OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 27, 2020
Anurag Ajay, Aviral Kumar, Pulkit Agrawal, Sergey Levine, Ofir Nachum

Figure 1 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 2 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 3 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 4 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Viaarxiv icon