Picture for Yuankai Qi

Yuankai Qi

DGRNet: Disagreement-Guided Refinement for Uncertainty-Aware Brain Tumor Segmentation

Add code
Mar 22, 2026
Viaarxiv icon

Hierarchical Text-Guided Brain Tumor Segmentation via Sub-Region-Aware Prompts

Add code
Mar 22, 2026
Viaarxiv icon

Question-guided Visual Compression with Memory Feedback for Long-Term Video Understanding

Add code
Mar 16, 2026
Viaarxiv icon

Unlocking Prototype Potential: An Efficient Tuning Framework for Few-Shot Class-Incremental Learning

Add code
Feb 05, 2026
Viaarxiv icon

Exploring the Temporal Consistency for Point-Level Weakly-Supervised Temporal Action Localization

Add code
Feb 05, 2026
Viaarxiv icon

Boosting Point-supervised Temporal Action Localization via Text Refinement and Alignment

Add code
Feb 01, 2026
Viaarxiv icon

Multimodal Visual Surrogate Compression for Alzheimer's Disease Classification

Add code
Jan 29, 2026
Viaarxiv icon

Visual Marker Search for Autonomous Drone Landing in Diverse Urban Environments

Add code
Jan 16, 2026
Viaarxiv icon

Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning

Add code
Nov 15, 2025
Figure 1 for Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning
Figure 2 for Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning
Figure 3 for Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning
Figure 4 for Teaching Prompts to Coordinate: Hierarchical Layer-Grouped Prompt Tuning for Continual Learning
Viaarxiv icon

Experiences from Benchmarking Vision-Language-Action Models for Robotic Manipulation

Add code
Nov 14, 2025
Figure 1 for Experiences from Benchmarking Vision-Language-Action Models for Robotic Manipulation
Figure 2 for Experiences from Benchmarking Vision-Language-Action Models for Robotic Manipulation
Figure 3 for Experiences from Benchmarking Vision-Language-Action Models for Robotic Manipulation
Figure 4 for Experiences from Benchmarking Vision-Language-Action Models for Robotic Manipulation
Viaarxiv icon