Picture for Zhuo Zhang

Zhuo Zhang

Online Self-Preferring Language Models

Add code
May 23, 2024
Viaarxiv icon

Threat Behavior Textual Search by Attention Graph Isomorphism

Add code
Apr 18, 2024
Viaarxiv icon

FedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning

Add code
Mar 10, 2024
Figure 1 for FedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning
Figure 2 for FedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning
Figure 3 for FedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning
Figure 4 for FedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning
Viaarxiv icon

CodeArt: Better Code Models by Attention Regularization When Symbols Are Lacking

Add code
Feb 19, 2024
Figure 1 for CodeArt: Better Code Models by Attention Regularization When Symbols Are Lacking
Figure 2 for CodeArt: Better Code Models by Attention Regularization When Symbols Are Lacking
Figure 3 for CodeArt: Better Code Models by Attention Regularization When Symbols Are Lacking
Figure 4 for CodeArt: Better Code Models by Attention Regularization When Symbols Are Lacking
Viaarxiv icon

Rapid Optimization for Jailbreaking LLMs via Subconscious Exploitation and Echopraxia

Add code
Feb 08, 2024
Viaarxiv icon

MULTIVERSE: Exposing Large Language Model Alignment Problems in Diverse Worlds

Add code
Jan 25, 2024
Figure 1 for MULTIVERSE: Exposing Large Language Model Alignment Problems in Diverse Worlds
Figure 2 for MULTIVERSE: Exposing Large Language Model Alignment Problems in Diverse Worlds
Figure 3 for MULTIVERSE: Exposing Large Language Model Alignment Problems in Diverse Worlds
Figure 4 for MULTIVERSE: Exposing Large Language Model Alignment Problems in Diverse Worlds
Viaarxiv icon

Make Them Spill the Beans! Coercive Knowledge Extraction from LLMs

Add code
Dec 08, 2023
Figure 1 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 2 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 3 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Figure 4 for Make Them Spill the Beans! Coercive Knowledge Extraction from  LLMs
Viaarxiv icon

ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP

Add code
Aug 04, 2023
Figure 1 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 2 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 3 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Figure 4 for ParaFuzz: An Interpretability-Driven Technique for Detecting Poisoned Samples in NLP
Viaarxiv icon

When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods

Add code
Dec 20, 2022
Figure 1 for When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods
Figure 2 for When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods
Figure 3 for When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods
Figure 4 for When Federated Learning Meets Pre-trained Language Models' Parameter-Efficient Tuning Methods
Viaarxiv icon

Backdoor Vulnerabilities in Normally Trained Deep Learning Models

Add code
Nov 29, 2022
Figure 1 for Backdoor Vulnerabilities in Normally Trained Deep Learning Models
Figure 2 for Backdoor Vulnerabilities in Normally Trained Deep Learning Models
Figure 3 for Backdoor Vulnerabilities in Normally Trained Deep Learning Models
Figure 4 for Backdoor Vulnerabilities in Normally Trained Deep Learning Models
Viaarxiv icon