Picture for Haruki Nishimura

Haruki Nishimura

Beyond Binary Success: Sample-Efficient and Statistically Rigorous Robot Policy Comparison

Add code
Mar 13, 2026
Viaarxiv icon

Impact of Different Failures on a Robot's Perceived Reliability

Add code
Mar 09, 2026
Viaarxiv icon

A Systematic Study of Data Modalities and Strategies for Co-training Large Behavior Models for Robot Manipulation

Add code
Feb 01, 2026
Viaarxiv icon

CUPID: Curating Data your Robot Loves with Influence Functions

Add code
Jun 23, 2025
Figure 1 for CUPID: Curating Data your Robot Loves with Influence Functions
Figure 2 for CUPID: Curating Data your Robot Loves with Influence Functions
Figure 3 for CUPID: Curating Data your Robot Loves with Influence Functions
Figure 4 for CUPID: Curating Data your Robot Loves with Influence Functions
Viaarxiv icon

SAFE: Multitask Failure Detection for Vision-Language-Action Models

Add code
Jun 11, 2025
Viaarxiv icon

STITCH-OPE: Trajectory Stitching with Guided Diffusion for Off-Policy Evaluation

Add code
May 27, 2025
Viaarxiv icon

Is Your Imitation Learning Policy Better than Mine? Policy Comparison with Near-Optimal Stopping

Add code
Mar 14, 2025
Viaarxiv icon

Can We Detect Failures Without Failure Data? Uncertainty-Aware Runtime Failure Detection for Imitation Learning Policies

Add code
Mar 11, 2025
Viaarxiv icon

How Generalizable Is My Behavior Cloning Policy? A Statistical Approach to Trustworthy Performance Evaluation

Add code
May 08, 2024
Figure 1 for How Generalizable Is My Behavior Cloning Policy? A Statistical Approach to Trustworthy Performance Evaluation
Figure 2 for How Generalizable Is My Behavior Cloning Policy? A Statistical Approach to Trustworthy Performance Evaluation
Figure 3 for How Generalizable Is My Behavior Cloning Policy? A Statistical Approach to Trustworthy Performance Evaluation
Figure 4 for How Generalizable Is My Behavior Cloning Policy? A Statistical Approach to Trustworthy Performance Evaluation
Viaarxiv icon

Residual Q-Learning: Offline and Online Policy Customization without Value

Add code
Jun 15, 2023
Figure 1 for Residual Q-Learning: Offline and Online Policy Customization without Value
Figure 2 for Residual Q-Learning: Offline and Online Policy Customization without Value
Figure 3 for Residual Q-Learning: Offline and Online Policy Customization without Value
Figure 4 for Residual Q-Learning: Offline and Online Policy Customization without Value
Viaarxiv icon