Picture for Yuning Mao

Yuning Mao

Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts

Add code
Feb 26, 2024
Figure 1 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 2 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 3 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 4 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Viaarxiv icon

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

Add code
Dec 07, 2023
Figure 1 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 2 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 3 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 4 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Viaarxiv icon

Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations

Add code
Dec 07, 2023
Viaarxiv icon

MART: Improving LLM Safety with Multi-round Automatic Red-Teaming

Add code
Nov 13, 2023
Figure 1 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 2 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 3 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 4 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Viaarxiv icon

Llama 2: Open Foundation and Fine-Tuned Chat Models

Add code
Jul 19, 2023
Figure 1 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 2 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 3 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 4 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Viaarxiv icon

LIMA: Less Is More for Alignment

Add code
May 18, 2023
Figure 1 for LIMA: Less Is More for Alignment
Figure 2 for LIMA: Less Is More for Alignment
Figure 3 for LIMA: Less Is More for Alignment
Figure 4 for LIMA: Less Is More for Alignment
Viaarxiv icon

Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization

Add code
May 06, 2023
Figure 1 for Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization
Figure 2 for Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization
Figure 3 for Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization
Figure 4 for Residual Prompt Tuning: Improving Prompt Tuning with Residual Reparameterization
Viaarxiv icon

Representation Deficiency in Masked Language Modeling

Add code
Feb 04, 2023
Figure 1 for Representation Deficiency in Masked Language Modeling
Figure 2 for Representation Deficiency in Masked Language Modeling
Figure 3 for Representation Deficiency in Masked Language Modeling
Figure 4 for Representation Deficiency in Masked Language Modeling
Viaarxiv icon

Progressive Prompts: Continual Learning for Language Models

Add code
Jan 29, 2023
Figure 1 for Progressive Prompts: Continual Learning for Language Models
Figure 2 for Progressive Prompts: Continual Learning for Language Models
Figure 3 for Progressive Prompts: Continual Learning for Language Models
Figure 4 for Progressive Prompts: Continual Learning for Language Models
Viaarxiv icon

XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models

Add code
Jan 25, 2023
Figure 1 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 2 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 3 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Figure 4 for XLM-V: Overcoming the Vocabulary Bottleneck in Multilingual Masked Language Models
Viaarxiv icon