Picture for Aniruddha Saha

Aniruddha Saha

Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion

Add code
Mar 25, 2024
Figure 1 for Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion
Figure 2 for Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion
Figure 3 for Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion
Figure 4 for Generating Potent Poisons and Backdoors from Scratch with Guided Diffusion
Viaarxiv icon

Spotting LLMs With Binoculars: Zero-Shot Detection of Machine-Generated Text

Add code
Jan 22, 2024
Viaarxiv icon

NEFTune: Noisy Embeddings Improve Instruction Finetuning

Add code
Oct 10, 2023
Figure 1 for NEFTune: Noisy Embeddings Improve Instruction Finetuning
Figure 2 for NEFTune: Noisy Embeddings Improve Instruction Finetuning
Figure 3 for NEFTune: Noisy Embeddings Improve Instruction Finetuning
Figure 4 for NEFTune: Noisy Embeddings Improve Instruction Finetuning
Viaarxiv icon

Baseline Defenses for Adversarial Attacks Against Aligned Language Models

Add code
Sep 04, 2023
Figure 1 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 2 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 3 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Figure 4 for Baseline Defenses for Adversarial Attacks Against Aligned Language Models
Viaarxiv icon

On the Reliability of Watermarks for Large Language Models

Add code
Jun 30, 2023
Figure 1 for On the Reliability of Watermarks for Large Language Models
Figure 2 for On the Reliability of Watermarks for Large Language Models
Figure 3 for On the Reliability of Watermarks for Large Language Models
Figure 4 for On the Reliability of Watermarks for Large Language Models
Viaarxiv icon

Bring Your Own Data! Self-Supervised Evaluation for Large Language Models

Add code
Jun 29, 2023
Figure 1 for Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Figure 2 for Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Figure 3 for Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Figure 4 for Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Viaarxiv icon

Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches

Add code
Jun 22, 2023
Figure 1 for Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches
Figure 2 for Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches
Figure 3 for Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches
Figure 4 for Revisiting Image Classifier Training for Improved Certified Robust Defense against Adversarial Patches
Viaarxiv icon

Backdoor Attacks on Vision Transformers

Add code
Jun 16, 2022
Figure 1 for Backdoor Attacks on Vision Transformers
Figure 2 for Backdoor Attacks on Vision Transformers
Figure 3 for Backdoor Attacks on Vision Transformers
Figure 4 for Backdoor Attacks on Vision Transformers
Viaarxiv icon

Backdoor Attacks on Self-Supervised Learning

Add code
May 21, 2021
Figure 1 for Backdoor Attacks on Self-Supervised Learning
Figure 2 for Backdoor Attacks on Self-Supervised Learning
Figure 3 for Backdoor Attacks on Self-Supervised Learning
Figure 4 for Backdoor Attacks on Self-Supervised Learning
Viaarxiv icon

Adversarial Patches Exploiting Contextual Reasoning in Object Detection

Add code
Sep 30, 2019
Figure 1 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 2 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 3 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Figure 4 for Adversarial Patches Exploiting Contextual Reasoning in Object Detection
Viaarxiv icon