Alert button
Picture for Stephen Tu

Stephen Tu

Alert button

Sharp Rates in Dependent Learning Theory: Avoiding Sample Size Deflation for the Square Loss

Feb 08, 2024
Ingvar Ziemann, Stephen Tu, George J. Pappas, Nikolai Matni

Viaarxiv icon

Revisiting Energy Based Models as Policies: Ranking Noise Contrastive Estimation and Interpolating Energy Models

Sep 11, 2023
Sumeet Singh, Stephen Tu, Vikas Sindhwani

Viaarxiv icon

Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners

Jul 04, 2023
Allen Z. Ren, Anushri Dixit, Alexandra Bodrova, Sumeet Singh, Stephen Tu, Noah Brown, Peng Xu, Leila Takayama, Fei Xia, Jake Varley, Zhenjia Xu, Dorsa Sadigh, Andy Zeng, Anirudha Majumdar

Figure 1 for Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners
Figure 2 for Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners
Figure 3 for Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners
Figure 4 for Robots That Ask For Help: Uncertainty Alignment for Large Language Model Planners
Viaarxiv icon

Bootstrapped Representations in Reinforcement Learning

Jun 16, 2023
Charline Le Lan, Stephen Tu, Mark Rowland, Anna Harutyunyan, Rishabh Agarwal, Marc G. Bellemare, Will Dabney

Figure 1 for Bootstrapped Representations in Reinforcement Learning
Figure 2 for Bootstrapped Representations in Reinforcement Learning
Figure 3 for Bootstrapped Representations in Reinforcement Learning
Figure 4 for Bootstrapped Representations in Reinforcement Learning
Viaarxiv icon

Agile Catching with Whole-Body MPC and Blackbox Policy Learning

Jun 14, 2023
Saminda Abeyruwan, Alex Bewley, Nicholas M. Boffi, Krzysztof Choromanski, David D'Ambrosio, Deepali Jain, Pannag Sanketi, Anish Shankar, Vikas Sindhwani, Sumeet Singh, Jean-Jacques Slotine, Stephen Tu

Figure 1 for Agile Catching with Whole-Body MPC and Blackbox Policy Learning
Figure 2 for Agile Catching with Whole-Body MPC and Blackbox Policy Learning
Figure 3 for Agile Catching with Whole-Body MPC and Blackbox Policy Learning
Viaarxiv icon

Safely Learning Dynamical Systems

May 20, 2023
Amir Ali Ahmadi, Abraar Chaudhry, Vikas Sindhwani, Stephen Tu

Figure 1 for Safely Learning Dynamical Systems
Figure 2 for Safely Learning Dynamical Systems
Figure 3 for Safely Learning Dynamical Systems
Figure 4 for Safely Learning Dynamical Systems
Viaarxiv icon

The noise level in linear regression with dependent data

May 18, 2023
Ingvar Ziemann, Stephen Tu, George J. Pappas, Nikolai Matni

Viaarxiv icon

The Power of Learned Locally Linear Models for Nonlinear Policy Optimization

May 16, 2023
Daniel Pfrommer, Max Simchowitz, Tyler Westenbroek, Nikolai Matni, Stephen Tu

Figure 1 for The Power of Learned Locally Linear Models for Nonlinear Policy Optimization
Viaarxiv icon

Multi-Task Imitation Learning for Linear Dynamical Systems

Dec 01, 2022
Thomas T. Zhang, Katie Kang, Bruce D. Lee, Claire Tomlin, Sergey Levine, Stephen Tu, Nikolai Matni

Figure 1 for Multi-Task Imitation Learning for Linear Dynamical Systems
Figure 2 for Multi-Task Imitation Learning for Linear Dynamical Systems
Viaarxiv icon

Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning

Oct 05, 2022
David Brandfonbrener, Stephen Tu, Avi Singh, Stefan Welker, Chad Boodoo, Nikolai Matni, Jake Varley

Figure 1 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 2 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 3 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Figure 4 for Visual Backtracking Teleoperation: A Data Collection Protocol for Offline Image-Based Reinforcement Learning
Viaarxiv icon