Picture for Shuaihang Yuan

Shuaihang Yuan

Hierarchical Scoring with 3D Gaussian Splatting for Instance Image-Goal Navigation

Add code
Jun 09, 2025
Viaarxiv icon

MapBERT: Bitwise Masked Modeling for Real-Time Semantic Mapping Generation

Add code
Jun 09, 2025
Viaarxiv icon

H2-COMPACT: Human-Humanoid Co-Manipulation via Adaptive Contact Trajectory Policies

Add code
May 23, 2025
Viaarxiv icon

Embodied Chain of Action Reasoning with Multi-Modal Foundation Model for Humanoid Loco-manipulation

Add code
Apr 13, 2025
Viaarxiv icon

A Chain-of-Thought Subspace Meta-Learning for Few-shot Image Captioning with Large Vision and Language Models

Add code
Feb 19, 2025
Viaarxiv icon

GAMap: Zero-Shot Object Goal Navigation with Multi-Scale Geometric-Affordance Guidance

Add code
Oct 31, 2024
Figure 1 for GAMap: Zero-Shot Object Goal Navigation with Multi-Scale Geometric-Affordance Guidance
Figure 2 for GAMap: Zero-Shot Object Goal Navigation with Multi-Scale Geometric-Affordance Guidance
Figure 3 for GAMap: Zero-Shot Object Goal Navigation with Multi-Scale Geometric-Affordance Guidance
Figure 4 for GAMap: Zero-Shot Object Goal Navigation with Multi-Scale Geometric-Affordance Guidance
Viaarxiv icon

Reliable Semantic Understanding for Real World Zero-shot Object Goal Navigation

Add code
Oct 29, 2024
Figure 1 for Reliable Semantic Understanding for Real World Zero-shot Object Goal Navigation
Figure 2 for Reliable Semantic Understanding for Real World Zero-shot Object Goal Navigation
Figure 3 for Reliable Semantic Understanding for Real World Zero-shot Object Goal Navigation
Figure 4 for Reliable Semantic Understanding for Real World Zero-shot Object Goal Navigation
Viaarxiv icon

Exploring the Reliability of Foundation Model-Based Frontier Selection in Zero-Shot Object Goal Navigation

Add code
Oct 28, 2024
Viaarxiv icon

Zero-shot Object Navigation with Vision-Language Models Reasoning

Add code
Oct 24, 2024
Figure 1 for Zero-shot Object Navigation with Vision-Language Models Reasoning
Figure 2 for Zero-shot Object Navigation with Vision-Language Models Reasoning
Figure 3 for Zero-shot Object Navigation with Vision-Language Models Reasoning
Figure 4 for Zero-shot Object Navigation with Vision-Language Models Reasoning
Viaarxiv icon

MultiTalk: Introspective and Extrospective Dialogue for Human-Environment-LLM Alignment

Add code
Sep 24, 2024
Figure 1 for MultiTalk: Introspective and Extrospective Dialogue for Human-Environment-LLM Alignment
Figure 2 for MultiTalk: Introspective and Extrospective Dialogue for Human-Environment-LLM Alignment
Figure 3 for MultiTalk: Introspective and Extrospective Dialogue for Human-Environment-LLM Alignment
Figure 4 for MultiTalk: Introspective and Extrospective Dialogue for Human-Environment-LLM Alignment
Viaarxiv icon