Picture for Xinyu Sun

Xinyu Sun

Shapley-Guided Neural Repair Approach via Derivative-Free Optimization

Add code
Apr 01, 2026
Viaarxiv icon

OneSearch-V2: The Latent Reasoning Enhanced Self-distillation Generative Search Framework

Add code
Mar 25, 2026
Viaarxiv icon

Virtual Community: An Open World for Humans, Robots, and Society

Add code
Aug 20, 2025
Figure 1 for Virtual Community: An Open World for Humans, Robots, and Society
Figure 2 for Virtual Community: An Open World for Humans, Robots, and Society
Figure 3 for Virtual Community: An Open World for Humans, Robots, and Society
Figure 4 for Virtual Community: An Open World for Humans, Robots, and Society
Viaarxiv icon

LSceneLLM: Enhancing Large 3D Scene Understanding Using Adaptive Visual Preferences

Add code
Dec 02, 2024
Figure 1 for LSceneLLM: Enhancing Large 3D Scene Understanding Using Adaptive Visual Preferences
Figure 2 for LSceneLLM: Enhancing Large 3D Scene Understanding Using Adaptive Visual Preferences
Figure 3 for LSceneLLM: Enhancing Large 3D Scene Understanding Using Adaptive Visual Preferences
Figure 4 for LSceneLLM: Enhancing Large 3D Scene Understanding Using Adaptive Visual Preferences
Viaarxiv icon

CoNav: A Benchmark for Human-Centered Collaborative Navigation

Add code
Jun 04, 2024
Viaarxiv icon

A Simple Knowledge Distillation Framework for Open-world Object Detection

Add code
Dec 14, 2023
Viaarxiv icon

Contrastive Vision-Language Alignment Makes Efficient Instruction Learner

Add code
Nov 29, 2023
Viaarxiv icon

RGM: A Robust Generalist Matching Model

Add code
Oct 19, 2023
Figure 1 for RGM: A Robust Generalist Matching Model
Figure 2 for RGM: A Robust Generalist Matching Model
Figure 3 for RGM: A Robust Generalist Matching Model
Figure 4 for RGM: A Robust Generalist Matching Model
Viaarxiv icon

FGPrompt: Fine-grained Goal Prompting for Image-goal Navigation

Add code
Oct 11, 2023
Viaarxiv icon

$A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models

Add code
Aug 15, 2023
Figure 1 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 2 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 3 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Figure 4 for $A^2$Nav: Action-Aware Zero-Shot Robot Navigation by Exploiting Vision-and-Language Ability of Foundation Models
Viaarxiv icon