Alert button
Picture for Sergey Levine

Sergey Levine

Alert button

Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators

May 05, 2023
Alexander Herzog, Kanishka Rao, Karol Hausman, Yao Lu, Paul Wohlhart, Mengyuan Yan, Jessica Lin, Montserrat Gonzalez Arenas, Ted Xiao, Daniel Kappler, Daniel Ho, Jarek Rettinghouse, Yevgen Chebotar, Kuang-Huei Lee, Keerthana Gopalakrishnan, Ryan Julian, Adrian Li, Chuyuan Kelly Fu, Bob Wei, Sangeetha Ramesh, Khem Holden, Kim Kleiven, David Rendleman, Sean Kirmani, Jeff Bingham, Jon Weisz, Ying Xu, Wenlong Lu, Matthew Bennice, Cody Fong, David Do, Jessica Lam, Yunfei Bai, Benjie Holson, Michael Quinlan, Noah Brown, Mrinal Kalakrishnan, Julian Ibarz, Peter Pastor, Sergey Levine

Figure 1 for Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators
Figure 2 for Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators
Figure 3 for Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators
Figure 4 for Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators
Viaarxiv icon

Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware

Apr 23, 2023
Tony Z. Zhao, Vikash Kumar, Sergey Levine, Chelsea Finn

Figure 1 for Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware
Figure 2 for Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware
Figure 3 for Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware
Figure 4 for Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware
Viaarxiv icon

IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies

Apr 20, 2023
Philippe Hansen-Estruch, Ilya Kostrikov, Michael Janner, Jakub Grudzien Kuba, Sergey Levine

Figure 1 for IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies
Figure 2 for IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies
Figure 3 for IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies
Figure 4 for IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies
Viaarxiv icon

Efficient Deep Reinforcement Learning Requires Regulating Overfitting

Apr 20, 2023
Qiyang Li, Aviral Kumar, Ilya Kostrikov, Sergey Levine

Figure 1 for Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Figure 2 for Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Figure 3 for Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Figure 4 for Efficient Deep Reinforcement Learning Requires Regulating Overfitting
Viaarxiv icon

Learning and Adapting Agile Locomotion Skills by Transferring Experience

Apr 19, 2023
Laura Smith, J. Chase Kew, Tianyu Li, Linda Luu, Xue Bin Peng, Sehoon Ha, Jie Tan, Sergey Levine

Figure 1 for Learning and Adapting Agile Locomotion Skills by Transferring Experience
Figure 2 for Learning and Adapting Agile Locomotion Skills by Transferring Experience
Figure 3 for Learning and Adapting Agile Locomotion Skills by Transferring Experience
Figure 4 for Learning and Adapting Agile Locomotion Skills by Transferring Experience
Viaarxiv icon

FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing

Apr 19, 2023
Kyle Stachowicz, Dhruv Shah, Arjun Bhorkar, Ilya Kostrikov, Sergey Levine

Figure 1 for FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing
Figure 2 for FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing
Figure 3 for FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing
Figure 4 for FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing
Viaarxiv icon

Reinforcement Learning from Passive Data via Latent Intentions

Apr 10, 2023
Dibya Ghosh, Chethan Bhateja, Sergey Levine

Figure 1 for Reinforcement Learning from Passive Data via Latent Intentions
Figure 2 for Reinforcement Learning from Passive Data via Latent Intentions
Figure 3 for Reinforcement Learning from Passive Data via Latent Intentions
Figure 4 for Reinforcement Learning from Passive Data via Latent Intentions
Viaarxiv icon

Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement

Mar 20, 2023
Michael Chang, Alyssa L. Dayan, Franziska Meier, Thomas L. Griffiths, Sergey Levine, Amy Zhang

Figure 1 for Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement
Figure 2 for Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement
Figure 3 for Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement
Figure 4 for Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement
Viaarxiv icon

Learning to Influence Human Behavior with Offline Reinforcement Learning

Mar 10, 2023
Joey Hong, Anca Dragan, Sergey Levine

Figure 1 for Learning to Influence Human Behavior with Offline Reinforcement Learning
Figure 2 for Learning to Influence Human Behavior with Offline Reinforcement Learning
Figure 3 for Learning to Influence Human Behavior with Offline Reinforcement Learning
Figure 4 for Learning to Influence Human Behavior with Offline Reinforcement Learning
Viaarxiv icon

Ignorance is Bliss: Robust Control via Information Gating

Mar 10, 2023
Manan Tomar, Riashat Islam, Sergey Levine, Philip Bachman

Figure 1 for Ignorance is Bliss: Robust Control via Information Gating
Figure 2 for Ignorance is Bliss: Robust Control via Information Gating
Figure 3 for Ignorance is Bliss: Robust Control via Information Gating
Figure 4 for Ignorance is Bliss: Robust Control via Information Gating
Viaarxiv icon