Picture for Guanhong Tao

Guanhong Tao

UNIT: Backdoor Mitigation via Automated Neural Distribution Tightening

Add code
Jul 16, 2024
Figure 1 for UNIT: Backdoor Mitigation via Automated Neural Distribution Tightening
Figure 2 for UNIT: Backdoor Mitigation via Automated Neural Distribution Tightening
Figure 3 for UNIT: Backdoor Mitigation via Automated Neural Distribution Tightening
Figure 4 for UNIT: Backdoor Mitigation via Automated Neural Distribution Tightening
Viaarxiv icon

Threat Behavior Textual Search by Attention Graph Isomorphism

Add code
Apr 18, 2024
Viaarxiv icon

LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning

Add code
Mar 25, 2024
Figure 1 for LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
Figure 2 for LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
Figure 3 for LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
Figure 4 for LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning
Viaarxiv icon

Rapid Optimization for Jailbreaking LLMs via Subconscious Exploitation and Echopraxia

Add code
Feb 08, 2024
Viaarxiv icon

Make Them Spill the Beans! Coercive Knowledge Extraction from LLMs

Add code
Dec 08, 2023
Figure 1 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 2 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 3 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 4 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Viaarxiv icon

Elijah: Eliminating Backdoors Injected in Diffusion Models via Distribution Shift

Add code
Nov 27, 2023
Viaarxiv icon

ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP

Add code
Aug 04, 2023
Figure 1 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 2 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 3 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 4 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Viaarxiv icon

Backdooring Neural Code Search

Add code
Jun 12, 2023
Figure 1 for Backdooring Neural Code Search
Figure 2 for Backdooring Neural Code Search
Figure 3 for Backdooring Neural Code Search
Figure 4 for Backdooring Neural Code Search
Viaarxiv icon

Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving

Add code
Apr 28, 2023
Figure 1 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 2 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 3 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 4 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Viaarxiv icon

Detecting Backdoors in Pre-trained Encoders

Add code
Mar 23, 2023
Figure 1 for Detecting Backdoors in Pre-trained Encoders
Figure 2 for Detecting Backdoors in Pre-trained Encoders
Figure 3 for Detecting Backdoors in Pre-trained Encoders
Figure 4 for Detecting Backdoors in Pre-trained Encoders
Viaarxiv icon