Picture for Chaowei Xiao

Chaowei Xiao

AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases

Add code
Jul 17, 2024
Figure 1 for AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases
Figure 2 for AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases
Figure 3 for AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases
Figure 4 for AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge Bases
Viaarxiv icon

Consistency Purification: Effective and Efficient Diffusion Purification towards Certified Robustness

Add code
Jun 30, 2024
Viaarxiv icon

UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models

Add code
Jun 27, 2024
Figure 1 for UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models
Figure 2 for UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models
Figure 3 for UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models
Figure 4 for UniGen: A Unified Framework for Textual Dataset Generation Using Large Language Models
Viaarxiv icon

MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding

Add code
Jun 13, 2024
Figure 1 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 2 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 3 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Figure 4 for MuirBench: A Comprehensive Benchmark for Robust Multi-image Understanding
Viaarxiv icon

AI Risk Management Should Incorporate Both Safety and Security

Add code
May 29, 2024
Figure 1 for AI Risk Management Should Incorporate Both Safety and Security
Viaarxiv icon

Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Characte

Add code
May 25, 2024
Viaarxiv icon

Safeguarding Vision-Language Models Against Patched Visual Prompt Injectors

Add code
May 17, 2024
Viaarxiv icon

JailBreakV-28K: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak Attacks

Add code
Apr 03, 2024
Viaarxiv icon

Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models

Add code
Mar 26, 2024
Figure 1 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 2 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 3 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Figure 4 for Don't Listen To Me: Understanding and Exploring Jailbreak Prompts of Large Language Models
Viaarxiv icon

AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting

Add code
Mar 14, 2024
Figure 1 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 2 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 3 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Figure 4 for AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting
Viaarxiv icon