Picture for Xiaogeng Liu

Xiaogeng Liu

MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding

Add code
Jun 13, 2024
Figure 1 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 2 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 3 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 4 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Viaarxiv icon

Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Characte

Add code
May 25, 2024
Viaarxiv icon

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks

Add code
Apr 03, 2024
Viaarxiv icon

Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models

Add code
Mar 26, 2024
Figure 1 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 2 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 3 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 4 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Viaarxiv icon

AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting

Add code
Mar 14, 2024
Figure 1 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 2 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 3 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 4 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Viaarxiv icon

Automatic and Universal Prompt Injection Attacks against Large Language Models

Add code
Mar 07, 2024
Figure 1 for Automatic and Universal Prompt Injection Attacks against Large Language Models
Figure 2 for Automatic and Universal Prompt Injection Attacks against Large Language Models
Figure 3 for Automatic and Universal Prompt Injection Attacks against Large Language Models
Figure 4 for Automatic and Universal Prompt Injection Attacks against Large Language Models
Viaarxiv icon

DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions

Add code
Dec 12, 2023
Figure 1 for DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions
Figure 2 for DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions
Figure 3 for DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions
Figure 4 for DeceptPrompt: Exploiting LLM-driven Code Generation via Adversarial Natural Language Instructions
Viaarxiv icon

AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models

Add code
Oct 03, 2023
Figure 1 for AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models
Figure 2 for AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models
Figure 3 for AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models
Figure 4 for AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models
Viaarxiv icon

Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training

Add code
Jul 19, 2023
Figure 1 for Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training
Figure 2 for Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training
Figure 3 for Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training
Figure 4 for Why Does Little Robustness Help? Understanding Adversarial Transferability From Surrogate Training
Viaarxiv icon

Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency

Add code
Mar 27, 2023
Figure 1 for Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency
Figure 2 for Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency
Figure 3 for Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency
Figure 4 for Detecting Backdoors During the Inference Stage Based on Corruption Robustness Consistency
Viaarxiv icon