Alert button
Picture for Guanhong Tao

Guanhong Tao

Alert button

Threat Behavior Textual Search by Attention Graph Isomorphism

Add code
Bookmark button
Alert button
Apr 18, 2024
Chanwoo Bae, Guanhong Tao, Zhuo Zhang, Xiangyu Zhang

Viaarxiv icon

LOTUS: Evasive and Resilient Backdoor Attacks through Sub-Partitioning

Add code
Bookmark button
Alert button
Mar 25, 2024
Siyuan Cheng, Guanhong Tao, Yingqi Liu, Guangyu Shen, Shengwei An, Shiwei Feng, Xiangzhe Xu, Kaiyuan Zhang, Shiqing Ma, Xiangyu Zhang

Viaarxiv icon

Rapid Optimization for Jailbreaking LLMs via Subconscious Exploitation and Echopraxia

Add code
Bookmark button
Alert button
Feb 08, 2024
Guangyu Shen, Siyuan Cheng, Kaiyuan Zhang, Guanhong Tao, Shengwei An, Lu Yan, Zhuo Zhang, Shiqing Ma, Xiangyu Zhang

Viaarxiv icon

Make Them Spill the Beans! Coercive Knowledge Extraction from (Production) LLMs

Add code
Bookmark button
Alert button
Dec 08, 2023
Zhuo Zhang, Guangyu Shen, Guanhong Tao, Siyuan Cheng, Xiangyu Zhang

Viaarxiv icon

Elijah: Eliminating Backdoors Injected in Diffusion Models via Distribution Shift

Add code
Bookmark button
Alert button
Nov 27, 2023
Shengwei An, Sheng-Yen Chou, Kaiyuan Zhang, Qiuling Xu, Guanhong Tao, Guangyu Shen, Siyuan Cheng, Shiqing Ma, Pin-Yu Chen, Tsung-Yi Ho, Xiangyu Zhang

Viaarxiv icon

ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP

Add code
Bookmark button
Alert button
Aug 04, 2023
Lu Yan, Zhuo Zhang, Guanhong Tao, Kaiyuan Zhang, Xuan Chen, Guangyu Shen, Xiangyu Zhang

Figure 1 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 2 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 3 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 4 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Viaarxiv icon

Backdooring Neural Code Search

Add code
Bookmark button
Alert button
Jun 12, 2023
Weisong Sun, Yuchen Chen, Guanhong Tao, Chunrong Fang, Xiangyu Zhang, Quanjun Zhang, Bin Luo

Figure 1 for Backdooring Neural Code Search
Figure 2 for Backdooring Neural Code Search
Figure 3 for Backdooring Neural Code Search
Figure 4 for Backdooring Neural Code Search
Viaarxiv icon

Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving

Add code
Bookmark button
Alert button
Apr 28, 2023
Zhiyuan Cheng, Hongjun Choi, James Liang, Shiwei Feng, Guanhong Tao, Dongfang Liu, Michael Zuzak, Xiangyu Zhang

Figure 1 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 2 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 3 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Figure 4 for Fusion is Not Enough: Single-Modal Attacks to Compromise Fusion Models in Autonomous Driving
Viaarxiv icon

Detecting Backdoors in Pre-trained Encoders

Add code
Bookmark button
Alert button
Mar 23, 2023
Shiwei Feng, Guanhong Tao, Siyuan Cheng, Guangyu Shen, Xiangzhe Xu, Yingqi Liu, Kaiyuan Zhang, Shiqing Ma, Xiangyu Zhang

Figure 1 for Detecting Backdoors in Pre-trained Encoders
Figure 2 for Detecting Backdoors in Pre-trained Encoders
Figure 3 for Detecting Backdoors in Pre-trained Encoders
Figure 4 for Detecting Backdoors in Pre-trained Encoders
Viaarxiv icon

Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks

Add code
Bookmark button
Alert button
Feb 08, 2023
Zhiyuan Cheng, James Liang, Guanhong Tao, Dongfang Liu, Xiangyu Zhang

Figure 1 for Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks
Figure 2 for Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks
Figure 3 for Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks
Figure 4 for Adversarial Training of Self-supervised Monocular Depth Estimation against Physical-World Attacks
Viaarxiv icon