Picture for Jiazhao Li

Jiazhao Li

Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment

Add code
Feb 27, 2024
Figure 1 for Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Figure 2 for Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Figure 3 for Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Figure 4 for Mitigating Fine-tuning Jailbreak Attack with Backdoor Enhanced Alignment
Viaarxiv icon

Defending against Insertion-based Textual Backdoor Attacks via Attribution

Add code
May 03, 2023
Figure 1 for Defending against Insertion-based Textual Backdoor Attacks via Attribution
Figure 2 for Defending against Insertion-based Textual Backdoor Attacks via Attribution
Figure 3 for Defending against Insertion-based Textual Backdoor Attacks via Attribution
Figure 4 for Defending against Insertion-based Textual Backdoor Attacks via Attribution
Viaarxiv icon

ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger

Add code
Apr 27, 2023
Figure 1 for ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger
Figure 2 for ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger
Figure 3 for ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger
Figure 4 for ChatGPT as an Attack Tool: Stealthy Textual Backdoor Attack via Blackbox Generative Model Trigger
Viaarxiv icon

PharmMT: A Neural Machine Translation Approach to Simplify Prescription Directions

Add code
Apr 08, 2022
Figure 1 for PharmMT: A Neural Machine Translation Approach to Simplify Prescription Directions
Figure 2 for PharmMT: A Neural Machine Translation Approach to Simplify Prescription Directions
Figure 3 for PharmMT: A Neural Machine Translation Approach to Simplify Prescription Directions
Figure 4 for PharmMT: A Neural Machine Translation Approach to Simplify Prescription Directions
Viaarxiv icon