Alert button
Picture for Avi Singh

Avi Singh

Alert button

Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models

Dec 22, 2023
Avi Singh, John D. Co-Reyes, Rishabh Agarwal, Ankesh Anand, Piyush Patil, Xavier Garcia, Peter J. Liu, James Harrison, Jaehoon Lee, Kelvin Xu, Aaron Parisi, Abhishek Kumar, Alex Alemi, Alex Rizkowsky, Azade Nova, Ben Adlam, Bernd Bohnet, Gamaleldin Elsayed, Hanie Sedghi, Igor Mordatch, Isabelle Simpson, Izzeddin Gur, Jasper Snoek, Jeffrey Pennington, Jiri Hron, Kathleen Kenealy, Kevin Swersky, Kshiteej Mahajan, Laura Culp, Lechao Xiao, Maxwell L. Bileschi, Noah Constant, Roman Novak, Rosanne Liu, Tris Warkentin, Yundi Qian, Yamini Bansal, Ethan Dyer, Behnam Neyshabur, Jascha Sohl-Dickstein, Noah Fiedel

Figure 1 for Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models
Figure 2 for Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models
Figure 3 for Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models
Figure 4 for Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models
Viaarxiv icon

Improving Large Language Model Fine-tuning for Solving Math Problems

Oct 16, 2023
Yixin Liu, Avi Singh, C. Daniel Freeman, John D. Co-Reyes, Peter J. Liu

Viaarxiv icon

Robotic Table Tennis: A Case Study into a High Speed Learning System

Sep 06, 2023
David B. D'Ambrosio, Jonathan Abelian, Saminda Abeyruwan, Michael Ahn, Alex Bewley, Justin Boyd, Krzysztof Choromanski, Omar Cortes, Erwin Coumans, Tianli Ding, Wenbo Gao, Laura Graesser, Atil Iscen, Navdeep Jaitly, Deepali Jain, Juhana Kangaspunta, Satoshi Kataoka, Gus Kouretas, Yuheng Kuang, Nevena Lazic, Corey Lynch, Reza Mahjourian, Sherry Q. Moore, Thinh Nguyen, Ken Oslund, Barney J Reed, Krista Reymann, Pannag R. Sanketi, Anish Shankar, Pierre Sermanet, Vikas Sindhwani, Avi Singh, Vincent Vanhoucke, Grace Vesom, Peng Xu

Figure 1 for Robotic Table Tennis: A Case Study into a High Speed Learning System
Figure 2 for Robotic Table Tennis: A Case Study into a High Speed Learning System
Figure 3 for Robotic Table Tennis: A Case Study into a High Speed Learning System
Figure 4 for Robotic Table Tennis: A Case Study into a High Speed Learning System
Viaarxiv icon

Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning

Oct 05, 2022
David Brandfonbrener, Stephen Tu, Avi Singh, Stefan Welker, Chad Boodoo, Nikolai Matni, Jake Varley

Figure 1 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 2 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 3 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 4 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Viaarxiv icon

Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning

Jul 17, 2022
Homer Walke, Jonathan Yang, Albert Yu, Aviral Kumar, Jedrzej Orbik, Avi Singh, Sergey Levine

Figure 1 for Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning
Figure 2 for Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning
Figure 3 for Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning
Figure 4 for Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning
Viaarxiv icon

i-Sim2Real: Reinforcement Learning of Robotic Policies in Tight Human-Robot Interaction Loops

Jul 14, 2022
Saminda Abeyruwan, Laura Graesser, David B. D'Ambrosio, Avi Singh, Anish Shankar, Alex Bewley, Pannag R. Sanketi

Figure 1 for i-Sim2Real: Reinforcement Learning of Robotic Policies in Tight Human-Robot Interaction Loops
Figure 2 for i-Sim2Real: Reinforcement Learning of Robotic Policies in Tight Human-Robot Interaction Loops
Figure 3 for i-Sim2Real: Reinforcement Learning of Robotic Policies in Tight Human-Robot Interaction Loops
Figure 4 for i-Sim2Real: Reinforcement Learning of Robotic Policies in Tight Human-Robot Interaction Loops
Viaarxiv icon

Parrot: Data-Driven Behavioral Priors for Reinforcement Learning

Nov 19, 2020
Avi Singh, Huihan Liu, Gaoyue Zhou, Albert Yu, Nicholas Rhinehart, Sergey Levine

Figure 1 for Parrot: Data-Driven Behavioral Priors for Reinforcement Learning
Figure 2 for Parrot: Data-Driven Behavioral Priors for Reinforcement Learning
Figure 3 for Parrot: Data-Driven Behavioral Priors for Reinforcement Learning
Figure 4 for Parrot: Data-Driven Behavioral Priors for Reinforcement Learning
Viaarxiv icon

COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning

Oct 27, 2020
Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine

Figure 1 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 2 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 3 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Figure 4 for COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning
Viaarxiv icon

The Ingredients of Real-World Robotic Reinforcement Learning

Apr 27, 2020
Henry Zhu, Justin Yu, Abhishek Gupta, Dhruv Shah, Kristian Hartikainen, Avi Singh, Vikash Kumar, Sergey Levine

Figure 1 for The Ingredients of Real-World Robotic Reinforcement Learning
Figure 2 for The Ingredients of Real-World Robotic Reinforcement Learning
Figure 3 for The Ingredients of Real-World Robotic Reinforcement Learning
Figure 4 for The Ingredients of Real-World Robotic Reinforcement Learning
Viaarxiv icon

Scalable Multi-Task Imitation Learning with Autonomous Improvement

Feb 25, 2020
Avi Singh, Eric Jang, Alexander Irpan, Daniel Kappler, Murtaza Dalal, Sergey Levine, Mohi Khansari, Chelsea Finn

Figure 1 for Scalable Multi-Task Imitation Learning with Autonomous Improvement
Figure 2 for Scalable Multi-Task Imitation Learning with Autonomous Improvement
Figure 3 for Scalable Multi-Task Imitation Learning with Autonomous Improvement
Figure 4 for Scalable Multi-Task Imitation Learning with Autonomous Improvement
Viaarxiv icon