Picture for Hangjun Ye

Hangjun Ye

OneVL: One-Step Latent Reasoning and Planning with Vision-Language Explanation

Add code
Apr 20, 2026
Viaarxiv icon

XEmbodied: A Foundation Model with Enhanced Geometric and Physical Cues for Large-Scale Embodied Environments

Add code
Apr 20, 2026
Viaarxiv icon

DriveVA: Video Action Models are Zero-Shot Drivers

Add code
Apr 05, 2026
Viaarxiv icon

UniDriveVLA: Unifying Understanding, Perception, and Action Planning for Autonomous Driving

Add code
Apr 02, 2026
Viaarxiv icon

Toward Physically Consistent Driving Video World Models under Challenging Trajectories

Add code
Mar 25, 2026
Viaarxiv icon

Learning from Mistakes: Post-Training for Driving VLA with Takeover Data

Add code
Mar 16, 2026
Viaarxiv icon

PerlAD: Towards Enhanced Closed-loop End-to-end Autonomous Driving with Pseudo-simulation-based Reinforcement Learning

Add code
Mar 16, 2026
Viaarxiv icon

LaST-VLA: Thinking in Latent Spatio-Temporal Space for Vision-Language-Action in Autonomous Driving

Add code
Mar 02, 2026
Viaarxiv icon

SEF-MAP: Subspace-Decomposed Expert Fusion for Robust Multimodal HD Map Prediction

Add code
Feb 25, 2026
Viaarxiv icon

From Pairs to Sequences: Track-Aware Policy Gradients for Keypoint Detection

Add code
Feb 25, 2026
Viaarxiv icon