Alert button
Picture for Sergey Levine

Sergey Levine

Alert button

High-Dimensional Continuous Control Using Generalized Advantage Estimation

Add code
Bookmark button
Alert button
Oct 20, 2018
John Schulman, Philipp Moritz, Sergey Levine, Michael Jordan, Pieter Abbeel

Figure 1 for High-Dimensional Continuous Control Using Generalized Advantage Estimation
Figure 2 for High-Dimensional Continuous Control Using Generalized Advantage Estimation
Figure 3 for High-Dimensional Continuous Control Using Generalized Advantage Estimation
Figure 4 for High-Dimensional Continuous Control Using Generalized Advantage Estimation
Viaarxiv icon

Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation

Add code
Bookmark button
Alert button
Oct 16, 2018
Gregory Kahn, Adam Villaflor, Pieter Abbeel, Sergey Levine

Figure 1 for Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation
Figure 2 for Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation
Figure 3 for Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation
Figure 4 for Composable Action-Conditioned Predictors: Flexible Off-Policy Learning for Robot Navigation
Viaarxiv icon

Deep Imitative Models for Flexible Inference, Planning, and Control

Add code
Bookmark button
Alert button
Oct 15, 2018
Nicholas Rhinehart, Rowan McAllister, Sergey Levine

Figure 1 for Deep Imitative Models for Flexible Inference, Planning, and Control
Figure 2 for Deep Imitative Models for Flexible Inference, Planning, and Control
Figure 3 for Deep Imitative Models for Flexible Inference, Planning, and Control
Figure 4 for Deep Imitative Models for Flexible Inference, Planning, and Control
Viaarxiv icon

Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning

Add code
Bookmark button
Alert button
Oct 15, 2018
Ilya Kostrikov, Kumar Krishna Agrawal, Debidatta Dwibedi, Sergey Levine, Jonathan Tompson

Figure 1 for Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Figure 2 for Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Figure 3 for Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Figure 4 for Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning
Viaarxiv icon

SFV: Reinforcement Learning of Physical Skills from Videos

Add code
Bookmark button
Alert button
Oct 15, 2018
Xue Bin Peng, Angjoo Kanazawa, Jitendra Malik, Pieter Abbeel, Sergey Levine

Figure 1 for SFV: Reinforcement Learning of Physical Skills from Videos
Figure 2 for SFV: Reinforcement Learning of Physical Skills from Videos
Figure 3 for SFV: Reinforcement Learning of Physical Skills from Videos
Figure 4 for SFV: Reinforcement Learning of Physical Skills from Videos
Viaarxiv icon

Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost

Add code
Bookmark button
Alert button
Oct 14, 2018
Henry Zhu, Abhishek Gupta, Aravind Rajeswaran, Sergey Levine, Vikash Kumar

Figure 1 for Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost
Figure 2 for Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost
Figure 3 for Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost
Figure 4 for Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost
Viaarxiv icon

Path Integral Guided Policy Search

Add code
Bookmark button
Alert button
Oct 11, 2018
Yevgen Chebotar, Mrinal Kalakrishnan, Ali Yahya, Adrian Li, Stefan Schaal, Sergey Levine

Figure 1 for Path Integral Guided Policy Search
Figure 2 for Path Integral Guided Policy Search
Figure 3 for Path Integral Guided Policy Search
Figure 4 for Path Integral Guided Policy Search
Viaarxiv icon

Learning a Prior over Intent via Meta-Inverse Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 10, 2018
Kelvin Xu, Ellis Ratner, Anca Dragan, Sergey Levine, Chelsea Finn

Figure 1 for Learning a Prior over Intent via Meta-Inverse Reinforcement Learning
Figure 2 for Learning a Prior over Intent via Meta-Inverse Reinforcement Learning
Figure 3 for Learning a Prior over Intent via Meta-Inverse Reinforcement Learning
Viaarxiv icon

Diversity is All You Need: Learning Skills without a Reward Function

Add code
Bookmark button
Alert button
Oct 09, 2018
Benjamin Eysenbach, Abhishek Gupta, Julian Ibarz, Sergey Levine

Figure 1 for Diversity is All You Need: Learning Skills without a Reward Function
Figure 2 for Diversity is All You Need: Learning Skills without a Reward Function
Figure 3 for Diversity is All You Need: Learning Skills without a Reward Function
Figure 4 for Diversity is All You Need: Learning Skills without a Reward Function
Viaarxiv icon

Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning

Add code
Bookmark button
Alert button
Oct 06, 2018
Frederik Ebert, Sudeep Dasari, Alex X. Lee, Sergey Levine, Chelsea Finn

Figure 1 for Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning
Figure 2 for Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning
Figure 3 for Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning
Figure 4 for Robustness via Retrying: Closed-Loop Robotic Manipulation with Self-Supervised Learning
Viaarxiv icon