Picture for Chengmeng Li

Chengmeng Li

ActiveUMI: Robotic Manipulation with Active Perception from Robot-Free Human Demonstrations

Add code
Oct 02, 2025
Figure 1 for ActiveUMI: Robotic Manipulation with Active Perception from Robot-Free Human Demonstrations
Figure 2 for ActiveUMI: Robotic Manipulation with Active Perception from Robot-Free Human Demonstrations
Figure 3 for ActiveUMI: Robotic Manipulation with Active Perception from Robot-Free Human Demonstrations
Figure 4 for ActiveUMI: Robotic Manipulation with Active Perception from Robot-Free Human Demonstrations
Viaarxiv icon

PointVLA: Injecting the 3D World into Vision-Language-Action Models

Add code
Mar 10, 2025
Figure 1 for PointVLA: Injecting the 3D World into Vision-Language-Action Models
Figure 2 for PointVLA: Injecting the 3D World into Vision-Language-Action Models
Figure 3 for PointVLA: Injecting the 3D World into Vision-Language-Action Models
Figure 4 for PointVLA: Injecting the 3D World into Vision-Language-Action Models
Viaarxiv icon

Improving Vision-Language-Action Models via Chain-of-Affordance

Add code
Dec 29, 2024
Figure 1 for Improving Vision-Language-Action Models via Chain-of-Affordance
Figure 2 for Improving Vision-Language-Action Models via Chain-of-Affordance
Figure 3 for Improving Vision-Language-Action Models via Chain-of-Affordance
Figure 4 for Improving Vision-Language-Action Models via Chain-of-Affordance
Viaarxiv icon

Diffusion-VLA: Scaling Robot Foundation Models via Unified Diffusion and Autoregression

Add code
Dec 04, 2024
Viaarxiv icon