Picture for Sipeng Zheng

Sipeng Zheng

OpenT2M: No-frill Motion Generation with Open-source,Large-scale, High-quality Data

Add code
Mar 19, 2026
Viaarxiv icon

Conservative Offline Robot Policy Learning via Posterior-Transition Reweighting

Add code
Mar 17, 2026
Viaarxiv icon

Joint-Aligned Latent Action: Towards Scalable VLA Pretraining in the Wild

Add code
Feb 25, 2026
Viaarxiv icon

Rethinking Visual-Language-Action Model Scaling: Alignment, Mixture, and Regularization

Add code
Feb 10, 2026
Viaarxiv icon

Being-H0.5: Scaling Human-Centric Robot Learning for Cross-Embodiment Generalization

Add code
Jan 19, 2026
Viaarxiv icon

Spatial-Aware VLA Pretraining through Visual-Physical Alignment from Human Videos

Add code
Dec 15, 2025
Viaarxiv icon

Robust Motion Generation using Part-level Reliable Data from Videos

Add code
Dec 14, 2025
Viaarxiv icon

Being-M0.5: A Real-Time Controllable Vision-Language-Motion Model

Add code
Aug 11, 2025
Viaarxiv icon

RL from Physical Feedback: Aligning Large Motion Models with Humanoid Control

Add code
Jun 15, 2025
Viaarxiv icon

EgoDTM: Towards 3D-Aware Egocentric Video-Language Pretraining

Add code
Mar 19, 2025
Figure 1 for EgoDTM: Towards 3D-Aware Egocentric Video-Language Pretraining
Figure 2 for EgoDTM: Towards 3D-Aware Egocentric Video-Language Pretraining
Figure 3 for EgoDTM: Towards 3D-Aware Egocentric Video-Language Pretraining
Figure 4 for EgoDTM: Towards 3D-Aware Egocentric Video-Language Pretraining
Viaarxiv icon