Picture for Ranjie Duan

Ranjie Duan

Knowledge-Guided Adversarial Training for Infrared Object Detection via Thermal Radiation Modeling

Add code
Mar 26, 2026
Viaarxiv icon

Improving Safety Alignment via Balanced Direct Preference Optimization

Add code
Mar 24, 2026
Viaarxiv icon

Obscure but Effective: Classical Chinese Jailbreak Prompt Optimization via Bio-Inspired Search

Add code
Feb 26, 2026
Viaarxiv icon

Pruning as a Cooperative Game: Surrogate-Assisted Layer Contribution Estimation for Large Language Models

Add code
Feb 08, 2026
Viaarxiv icon

YuFeng-XGuard: A Reasoning-Centric, Interpretable, and Flexible Guardrail Model for Large Language Models

Add code
Jan 22, 2026
Viaarxiv icon

A Safety Report on GPT-5.2, Gemini 3 Pro, Qwen3-VL, Grok 4.1 Fast, Nano Banana Pro, and Seedream 4.5

Add code
Jan 16, 2026
Viaarxiv icon

Towards Class-wise Fair Adversarial Training via Anti-Bias Soft Label Distillation

Add code
Jun 10, 2025
Viaarxiv icon

The Eye of Sherlock Holmes: Uncovering User Private Attribute Profiling via Vision-Language Model Agentic Framework

Add code
May 25, 2025
Figure 1 for The Eye of Sherlock Holmes: Uncovering User Private Attribute Profiling via Vision-Language Model Agentic Framework
Figure 2 for The Eye of Sherlock Holmes: Uncovering User Private Attribute Profiling via Vision-Language Model Agentic Framework
Figure 3 for The Eye of Sherlock Holmes: Uncovering User Private Attribute Profiling via Vision-Language Model Agentic Framework
Figure 4 for The Eye of Sherlock Holmes: Uncovering User Private Attribute Profiling via Vision-Language Model Agentic Framework
Viaarxiv icon

Enhancing Adversarial Robustness of Vision Language Models via Adversarial Mixture Prompt Tuning

Add code
May 23, 2025
Viaarxiv icon

DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models

Add code
Apr 25, 2025
Figure 1 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 2 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 3 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Figure 4 for DREAM: Disentangling Risks to Enhance Safety Alignment in Multimodal Large Language Models
Viaarxiv icon