Picture for Xingxing Wei

Xingxing Wei

Beijing Key Laboratory of Digital Media, School of Computer Science and Engineering, Beihang University, Beijing, China

From reactive to cognitive: brain-inspired spatial intelligence for embodied agents

Add code
Aug 24, 2025
Viaarxiv icon

NS-FPN: Improving Infrared Small Target Detection and Segmentation from Noise Suppression Perspective

Add code
Aug 09, 2025
Viaarxiv icon

Towards Class-wise Fair Adversarial Training via Anti-Bias Soft Label Distillation

Add code
Jun 10, 2025
Viaarxiv icon

Breaking the Ceiling: Exploring the Potential of Jailbreak Attacks through Expanding Strategy Space

Add code
May 28, 2025
Viaarxiv icon

Mitigating Overthinking in Large Reasoning Models via Manifold Steering

Add code
May 28, 2025
Viaarxiv icon

Enhancing Adversarial Robustness of Vision Language Models via Adversarial Mixture Prompt Tuning

Add code
May 23, 2025
Viaarxiv icon

Towards NSFW-Free Text-to-Image Generation via Safety-Constraint Direct Preference Optimization

Add code
Apr 19, 2025
Viaarxiv icon

Rethinking Multi-modal Object Detection from the Perspective of Mono-Modality Feature Learning

Add code
Mar 14, 2025
Viaarxiv icon

When Lighting Deceives: Exposing Vision-Language Models' Illumination Vulnerability Through Illumination Transformation Attack

Add code
Mar 10, 2025
Figure 1 for When Lighting Deceives: Exposing Vision-Language Models' Illumination Vulnerability Through Illumination Transformation Attack
Figure 2 for When Lighting Deceives: Exposing Vision-Language Models' Illumination Vulnerability Through Illumination Transformation Attack
Figure 3 for When Lighting Deceives: Exposing Vision-Language Models' Illumination Vulnerability Through Illumination Transformation Attack
Figure 4 for When Lighting Deceives: Exposing Vision-Language Models' Illumination Vulnerability Through Illumination Transformation Attack
Viaarxiv icon

Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency

Add code
Jan 09, 2025
Figure 1 for Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency
Figure 2 for Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency
Figure 3 for Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency
Figure 4 for Jailbreaking Multimodal Large Language Models via Shuffle Inconsistency
Viaarxiv icon