Picture for Nan Zhou

Nan Zhou

CoVFT: Context-aware Visual Fine-tuning for Multimodal Large Language Models

Add code
Mar 22, 2026
Viaarxiv icon

Code-A1: Adversarial Evolving of Code LLM and Test LLM via Reinforcement Learning

Add code
Mar 16, 2026
Viaarxiv icon

A multimodal vision foundation model for generalizable knee pathology

Add code
Jan 26, 2026
Viaarxiv icon

The Illusion of Clinical Reasoning: A Benchmark Reveals the Pervasive Gap in Vision-Language Models for Clinical Competency

Add code
Dec 25, 2025
Viaarxiv icon

PhysFire-WM: A Physics-Informed World Model for Emulating Fire Spread Dynamics

Add code
Dec 19, 2025
Figure 1 for PhysFire-WM: A Physics-Informed World Model for Emulating Fire Spread Dynamics
Figure 2 for PhysFire-WM: A Physics-Informed World Model for Emulating Fire Spread Dynamics
Figure 3 for PhysFire-WM: A Physics-Informed World Model for Emulating Fire Spread Dynamics
Figure 4 for PhysFire-WM: A Physics-Informed World Model for Emulating Fire Spread Dynamics
Viaarxiv icon

QUIDS: Quality-informed Incentive-driven Multi-agent Dispatching System for Mobile Crowdsensing

Add code
Dec 18, 2025
Figure 1 for QUIDS: Quality-informed Incentive-driven Multi-agent Dispatching System for Mobile Crowdsensing
Figure 2 for QUIDS: Quality-informed Incentive-driven Multi-agent Dispatching System for Mobile Crowdsensing
Figure 3 for QUIDS: Quality-informed Incentive-driven Multi-agent Dispatching System for Mobile Crowdsensing
Figure 4 for QUIDS: Quality-informed Incentive-driven Multi-agent Dispatching System for Mobile Crowdsensing
Viaarxiv icon

MedSAM-U: Uncertainty-Guided Auto Multi-Prompt Adaptation for Reliable MedSAM

Add code
Sep 02, 2024
Figure 1 for MedSAM-U: Uncertainty-Guided Auto Multi-Prompt Adaptation for Reliable MedSAM
Figure 2 for MedSAM-U: Uncertainty-Guided Auto Multi-Prompt Adaptation for Reliable MedSAM
Figure 3 for MedSAM-U: Uncertainty-Guided Auto Multi-Prompt Adaptation for Reliable MedSAM
Figure 4 for MedSAM-U: Uncertainty-Guided Auto Multi-Prompt Adaptation for Reliable MedSAM
Viaarxiv icon

Crowd-SAM: SAM as a Smart Annotator for Object Detection in Crowded Scenes

Add code
Jul 16, 2024
Figure 1 for Crowd-SAM: SAM as a Smart Annotator for Object Detection in Crowded Scenes
Figure 2 for Crowd-SAM: SAM as a Smart Annotator for Object Detection in Crowded Scenes
Figure 3 for Crowd-SAM: SAM as a Smart Annotator for Object Detection in Crowded Scenes
Figure 4 for Crowd-SAM: SAM as a Smart Annotator for Object Detection in Crowded Scenes
Viaarxiv icon

PLDNet: PLD-Guided Lightweight Deep Network Boosted by Efficient Attention for Handheld Dual-Microphone Speech Enhancement

Add code
Jun 06, 2024
Figure 1 for PLDNet: PLD-Guided Lightweight Deep Network Boosted by Efficient Attention for Handheld Dual-Microphone Speech Enhancement
Figure 2 for PLDNet: PLD-Guided Lightweight Deep Network Boosted by Efficient Attention for Handheld Dual-Microphone Speech Enhancement
Figure 3 for PLDNet: PLD-Guided Lightweight Deep Network Boosted by Efficient Attention for Handheld Dual-Microphone Speech Enhancement
Figure 4 for PLDNet: PLD-Guided Lightweight Deep Network Boosted by Efficient Attention for Handheld Dual-Microphone Speech Enhancement
Viaarxiv icon

iVPT: Improving Task-relevant Information Sharing in Visual Prompt Tuning by Cross-layer Dynamic Connection

Add code
Apr 08, 2024
Figure 1 for iVPT: Improving Task-relevant Information Sharing in Visual Prompt Tuning by Cross-layer Dynamic Connection
Figure 2 for iVPT: Improving Task-relevant Information Sharing in Visual Prompt Tuning by Cross-layer Dynamic Connection
Figure 3 for iVPT: Improving Task-relevant Information Sharing in Visual Prompt Tuning by Cross-layer Dynamic Connection
Figure 4 for iVPT: Improving Task-relevant Information Sharing in Visual Prompt Tuning by Cross-layer Dynamic Connection
Viaarxiv icon