Picture for Daoyuan Wu

Daoyuan Wu

SoK: Evaluating Jailbreak Guardrails for Large Language Models

Add code
Jun 12, 2025
Viaarxiv icon

IP Leakage Attacks Targeting LLM-Based Multi-Agent Systems

Add code
May 18, 2025
Viaarxiv icon

BadMoE: Backdooring Mixture-of-Experts LLMs via Optimizing Routing Triggers and Infecting Dormant Experts

Add code
Apr 29, 2025
Viaarxiv icon

GuidedBench: Equipping Jailbreak Evaluation with Guidelines

Add code
Feb 24, 2025
Viaarxiv icon

DeFiScope: Detecting Various DeFi Price Manipulations with LLM Reasoning

Add code
Feb 17, 2025
Viaarxiv icon

API-guided Dataset Synthesis to Finetune Large Code Models

Add code
Aug 15, 2024
Figure 1 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 2 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 3 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 4 for API-guided Dataset Synthesis to Finetune Large Code Models
Viaarxiv icon

SelfDefend: LLMs Can Defend Themselves against Jailbreaking in a Practical Manner

Add code
Jun 08, 2024
Viaarxiv icon

PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation

Add code
May 04, 2024
Figure 1 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 2 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 3 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 4 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Viaarxiv icon

Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs

Add code
Apr 27, 2024
Figure 1 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 2 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 3 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 4 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Viaarxiv icon

LLMs Can Defend Themselves Against Jailbreaking in a Practical Manner: A Vision Paper

Add code
Mar 04, 2024
Viaarxiv icon