Picture for Jian Liang

Jian Liang

To Trust Or Not To Trust Your Vision-Language Model's Prediction

Add code
May 29, 2025
Viaarxiv icon

Test-Time Immunization: A Universal Defense Framework Against Jailbreaks for (Multimodal) Large Language Models

Add code
May 28, 2025
Viaarxiv icon

ThanoRA: Task Heterogeneity-Aware Multi-Task Low-Rank Adaptation

Add code
May 24, 2025
Viaarxiv icon

Backdoor Cleaning without External Guidance in MLLM Fine-tuning

Add code
May 22, 2025
Viaarxiv icon

From Local Details to Global Context: Advancing Vision-Language Models with Attention-Based Selection

Add code
May 19, 2025
Viaarxiv icon

R-TPT: Improving Adversarial Robustness of Vision-Language Models through Test-Time Prompt Tuning

Add code
Apr 15, 2025
Viaarxiv icon

Do We Really Need Curated Malicious Data for Safety Alignment in Multi-modal Large Language Models?

Add code
Apr 14, 2025
Viaarxiv icon

LoRASculpt: Sculpting LoRA for Harmonizing General and Specialized Knowledge in Multimodal Large Language Models

Add code
Mar 21, 2025
Viaarxiv icon

Keeping Yourself is Important in Downstream Tuning Multimodal Large Language Model

Add code
Mar 06, 2025
Viaarxiv icon

SayAnything: Audio-Driven Lip Synchronization with Conditional Video Diffusion

Add code
Feb 17, 2025
Viaarxiv icon