Picture for Yuxin Huang

Yuxin Huang

Recurrent Reasoning with Vision-Language Models for Estimating Long-Horizon Embodied Task Progress

Add code
Mar 18, 2026
Viaarxiv icon

ReconDrive: Fast Feed-Forward 4D Gaussian Splatting for Autonomous Driving Scene Reconstruction

Add code
Mar 08, 2026
Viaarxiv icon

Beyond Benchmarks of IUGC: Rethinking Requirements of Deep Learning Methods for Intrapartum Ultrasound Biometry from Fetal Ultrasound Videos

Add code
Feb 13, 2026
Viaarxiv icon

Consensus-Aligned Neuron Efficient Fine-Tuning Large Language Models for Multi-Domain Machine Translation

Add code
Feb 05, 2026
Viaarxiv icon

Baseline Method of the Foundation Model Challenge for Ultrasound Image Analysis

Add code
Feb 01, 2026
Viaarxiv icon

FUGC: Benchmarking Semi-Supervised Learning Methods for Cervical Segmentation

Add code
Jan 22, 2026
Viaarxiv icon

Multilingual Generative Retrieval via Cross-lingual Semantic Compression

Add code
Oct 09, 2025
Figure 1 for Multilingual Generative Retrieval via Cross-lingual Semantic Compression
Figure 2 for Multilingual Generative Retrieval via Cross-lingual Semantic Compression
Figure 3 for Multilingual Generative Retrieval via Cross-lingual Semantic Compression
Figure 4 for Multilingual Generative Retrieval via Cross-lingual Semantic Compression
Viaarxiv icon

SageLM: A Multi-aspect and Explainable Large Language Model for Speech Judgement

Add code
Aug 28, 2025
Viaarxiv icon

ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver

Add code
Aug 14, 2025
Figure 1 for ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver
Figure 2 for ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver
Figure 3 for ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver
Figure 4 for ReconVLA: Reconstructive Vision-Language-Action Model as Effective Robot Perceiver
Viaarxiv icon

CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding

Add code
Jun 16, 2025
Figure 1 for CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding
Figure 2 for CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding
Figure 3 for CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding
Figure 4 for CEED-VLA: Consistency Vision-Language-Action Model with Early-Exit Decoding
Viaarxiv icon