Alert button
Picture for Vasu Singla

Vasu Singla

Alert button

Understanding and Mitigating Copying in Diffusion Models

May 31, 2023
Gowthami Somepalli, Vasu Singla, Micah Goldblum, Jonas Geiping, Tom Goldstein

Figure 1 for Understanding and Mitigating Copying in Diffusion Models
Figure 2 for Understanding and Mitigating Copying in Diffusion Models
Figure 3 for Understanding and Mitigating Copying in Diffusion Models
Figure 4 for Understanding and Mitigating Copying in Diffusion Models

Images generated by diffusion models like Stable Diffusion are increasingly widespread. Recent works and even lawsuits have shown that these models are prone to replicating their training data, unbeknownst to the user. In this paper, we first analyze this memorization problem in text-to-image diffusion models. While it is widely believed that duplicated images in the training set are responsible for content replication at inference time, we observe that the text conditioning of the model plays a similarly important role. In fact, we see in our experiments that data replication often does not happen for unconditional models, while it is common in the text-conditional case. Motivated by our findings, we then propose several techniques for reducing data replication at both training and inference time by randomizing and augmenting image captions in the training set.

* 17 pages, preprint. Code is available at https://github.com/somepago/DCR 
Viaarxiv icon

What Can We Learn from Unlearnable Datasets?

May 30, 2023
Pedro Sandoval-Segura, Vasu Singla, Jonas Geiping, Micah Goldblum, Tom Goldstein

Figure 1 for What Can We Learn from Unlearnable Datasets?
Figure 2 for What Can We Learn from Unlearnable Datasets?
Figure 3 for What Can We Learn from Unlearnable Datasets?
Figure 4 for What Can We Learn from Unlearnable Datasets?

In an era of widespread web scraping, unlearnable dataset methods have the potential to protect data privacy by preventing deep neural networks from generalizing. But in addition to a number of practical limitations that make their use unlikely, we make a number of findings that call into question their ability to safeguard data. First, it is widely believed that neural networks trained on unlearnable datasets only learn shortcuts, simpler rules that are not useful for generalization. In contrast, we find that networks actually can learn useful features that can be reweighed for high test performance, suggesting that image privacy is not preserved. Unlearnable datasets are also believed to induce learning shortcuts through linear separability of added perturbations. We provide a counterexample, demonstrating that linear separability of perturbations is not a necessary condition. To emphasize why linearly separable perturbations should not be relied upon, we propose an orthogonal projection attack which allows learning from unlearnable datasets published in ICML 2021 and ICLR 2023. Our proposed attack is significantly less complex than recently proposed techniques.

* 17 pages, 9 figures 
Viaarxiv icon

Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models

Dec 12, 2022
Gowthami Somepalli, Vasu Singla, Micah Goldblum, Jonas Geiping, Tom Goldstein

Figure 1 for Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models
Figure 2 for Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models
Figure 3 for Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models
Figure 4 for Diffusion Art or Digital Forgery? Investigating Data Replication in Diffusion Models

Cutting-edge diffusion models produce images with high quality and customizability, enabling them to be used for commercial art and graphic design purposes. But do diffusion models create unique works of art, or are they replicating content directly from their training sets? In this work, we study image retrieval frameworks that enable us to compare generated images with training samples and detect when content has been replicated. Applying our frameworks to diffusion models trained on multiple datasets including Oxford flowers, Celeb-A, ImageNet, and LAION, we discuss how factors such as training set size impact rates of content replication. We also identify cases where diffusion models, including the popular Stable Diffusion model, blatantly copy from their training data.

* Updated draft with the following changes (1) Clarified the LAION Aesthetics versions everywhere (2) Correction on which LAION Aesthetics version SD - 1.4 is finetuned on and updated figure 12 based on this (3) A section on possible causes of replication 
Viaarxiv icon

Autoregressive Perturbations for Data Poisoning

Jun 15, 2022
Pedro Sandoval-Segura, Vasu Singla, Jonas Geiping, Micah Goldblum, Tom Goldstein, David W. Jacobs

Figure 1 for Autoregressive Perturbations for Data Poisoning
Figure 2 for Autoregressive Perturbations for Data Poisoning
Figure 3 for Autoregressive Perturbations for Data Poisoning
Figure 4 for Autoregressive Perturbations for Data Poisoning

The prevalence of data scraping from social media as a means to obtain datasets has led to growing concerns regarding unauthorized use of data. Data poisoning attacks have been proposed as a bulwark against scraping, as they make data "unlearnable" by adding small, imperceptible perturbations. Unfortunately, existing methods require knowledge of both the target architecture and the complete dataset so that a surrogate network can be trained, the parameters of which are used to generate the attack. In this work, we introduce autoregressive (AR) poisoning, a method that can generate poisoned data without access to the broader dataset. The proposed AR perturbations are generic, can be applied across different datasets, and can poison different architectures. Compared to existing unlearnable methods, our AR poisons are more resistant against common defenses such as adversarial training and strong data augmentations. Our analysis further provides insight into what makes an effective data poison.

* 22 pages, 13 figures. Code available at https://github.com/psandovalsegura/autoregressive-poisoning 
Viaarxiv icon

Poisons that are learned faster are more effective

Apr 19, 2022
Pedro Sandoval-Segura, Vasu Singla, Liam Fowl, Jonas Geiping, Micah Goldblum, David Jacobs, Tom Goldstein

Figure 1 for Poisons that are learned faster are more effective
Figure 2 for Poisons that are learned faster are more effective
Figure 3 for Poisons that are learned faster are more effective
Figure 4 for Poisons that are learned faster are more effective

Imperceptible poisoning attacks on entire datasets have recently been touted as methods for protecting data privacy. However, among a number of defenses preventing the practical use of these techniques, early-stopping stands out as a simple, yet effective defense. To gauge poisons' vulnerability to early-stopping, we benchmark error-minimizing, error-maximizing, and synthetic poisons in terms of peak test accuracy over 100 epochs and make a number of surprising observations. First, we find that poisons that reach a low training loss faster have lower peak test accuracy. Second, we find that a current state-of-the-art error-maximizing poison is 7 times less effective when poison training is stopped at epoch 8. Third, we find that stronger, more transferable adversarial attacks do not make stronger poisons. We advocate for evaluating poisons in terms of peak test accuracy.

* 8 pages, 4 figures. Accepted to CVPR 2022 Art of Robustness Workshop 
Viaarxiv icon

Shift Invariance Can Reduce Adversarial Robustness

Mar 03, 2021
Songwei Ge, Vasu Singla, Ronen Basri, David Jacobs

Figure 1 for Shift Invariance Can Reduce Adversarial Robustness
Figure 2 for Shift Invariance Can Reduce Adversarial Robustness
Figure 3 for Shift Invariance Can Reduce Adversarial Robustness
Figure 4 for Shift Invariance Can Reduce Adversarial Robustness

Shift invariance is a critical property of CNNs that improves performance on classification. However, we show that invariance to circular shifts can also lead to greater sensitivity to adversarial attacks. We first characterize the margin between classes when a shift-invariant linear classifier is used. We show that the margin can only depend on the DC component of the signals. Then, using results about infinitely wide networks, we show that in some simple cases, fully connected and shift-invariant neural networks produce linear decision boundaries. Using this, we prove that shift invariance in neural networks produces adversarial examples for the simple case of two classes, each consisting of a single image with a black or white dot on a gray background. This is more than a curiosity; we show empirically that with real datasets and realistic architectures, shift invariance reduces adversarial robustness. Finally, we describe initial experiments using synthetic data to probe the source of this connection.

Viaarxiv icon

Low Curvature Activations Reduce Overfitting in Adversarial Training

Feb 15, 2021
Vasu Singla, Sahil Singla, David Jacobs, Soheil Feizi

Figure 1 for Low Curvature Activations Reduce Overfitting in Adversarial Training
Figure 2 for Low Curvature Activations Reduce Overfitting in Adversarial Training
Figure 3 for Low Curvature Activations Reduce Overfitting in Adversarial Training
Figure 4 for Low Curvature Activations Reduce Overfitting in Adversarial Training

Adversarial training is one of the most effective defenses against adversarial attacks. Previous works suggest that overfitting is a dominant phenomenon in adversarial training leading to a large generalization gap between test and train accuracy in neural networks. In this work, we show that the observed generalization gap is closely related to the choice of the activation function. In particular, we show that using activation functions with low (exact or approximate) curvature values has a regularization effect that significantly reduces both the standard and robust generalization gaps in adversarial training. We observe this effect for both differentiable/smooth activations such as Swish as well as non-differentiable/non-smooth activations such as LeakyReLU. In the latter case, the approximate curvature of the activation is low. Finally, we show that for activation functions with low curvature, the double descent phenomenon for adversarially trained models does not occur.

Viaarxiv icon

ASAP-NMS: Accelerating Non-Maximum Suppression Using Spatially Aware Priors

Aug 21, 2020
Rohun Tripathi, Vasu Singla, Mahyar Najibi, Bharat Singh, Abhishek Sharma, Larry Davis

Figure 1 for ASAP-NMS: Accelerating Non-Maximum Suppression Using Spatially Aware Priors
Figure 2 for ASAP-NMS: Accelerating Non-Maximum Suppression Using Spatially Aware Priors
Figure 3 for ASAP-NMS: Accelerating Non-Maximum Suppression Using Spatially Aware Priors
Figure 4 for ASAP-NMS: Accelerating Non-Maximum Suppression Using Spatially Aware Priors

The widely adopted sequential variant of Non Maximum Suppression (or Greedy-NMS) is a crucial module for object-detection pipelines. Unfortunately, for the region proposal stage of two/multi-stage detectors, NMS is turning out to be a latency bottleneck due to its sequential nature. In this article, we carefully profile Greedy-NMS iterations to find that a major chunk of computation is wasted in comparing proposals that are already far-away and have a small chance of suppressing each other. We address this issue by comparing only those proposals that are generated from nearby anchors. The translation-invariant property of the anchor lattice affords generation of a lookup table, which provides an efficient access to nearby proposals, during NMS. This leads to an Accelerated NMS algorithm which leverages Spatially Aware Priors, or ASAP-NMS, and improves the latency of the NMS step from 13.6ms to 1.2 ms on a CPU without sacrificing the accuracy of a state-of-the-art two-stage detector on COCO and VOC datasets. Importantly, ASAP-NMS is agnostic to image resolution and can be used as a simple drop-in module during inference. Using ASAP-NMS at run-time only, we obtain an mAP of 44.2\%@25Hz on the COCO dataset with a V100 GPU.

* Under Review at CVIU 
Viaarxiv icon