<?xml-model href='http://www.tei-c.org/release/xml/tei/custom/schema/relaxng/tei_all.rng' schematypens='http://relaxng.org/ns/structure/1.0'?><TEI xmlns="http://www.tei-c.org/ns/1.0">
	<teiHeader>
		<fileDesc>
			<titleStmt><title level='a'>CUDA: Convolution-based Unlearnable Datasets</title></titleStmt>
			<publicationStmt>
				<publisher>CVPR 2023</publisher>
				<date>01/01/2023</date>
			</publicationStmt>
			<sourceDesc>
				<bibl> 
					<idno type="par_id">10493818</idno>
					<idno type="doi"></idno>
					
					<author>Vinu Sadasivan</author><author>Mahdi Soltanolkotabi</author><author>Soheil Feizi</author>
				</bibl>
			</sourceDesc>
		</fileDesc>
		<profileDesc>
			<abstract><ab><![CDATA[Large-scale training of modern deep learning models heavily relies on publicly available data on the web. This potentially unauthorized usage of online data leads to concerns regarding data privacy. Recent works aim to make unlearnable data for deep learning models by adding small, specially designed noises to tackle this issue. However, these methods are vulnerable to adversarial training (AT) and/or are computationally heavy. In this work, we propose a novel, model-free, Convolution-based Unlearnable DAtaset (CUDA) generation technique. CUDA is generated using controlled class-wise convolutions with filters that are randomly generated via a private key. CUDA encourages the network to learn the relation between filters and labels rather than informative features for classifying the clean data. We develop some theoretical analysis demonstrating that CUDA can successfully poison Gaussian mixture data by reducing the clean data performance of the optimal Bayes classifier. We also empirically demonstrate the effectiveness of CUDA with various datasets (CIFAR-10, CIFAR-100, ImageNet-100, and Tiny-ImageNet), and architectures   DeIT, and MobileNetV2). Our experiments show that CUDA is robust to various data augmentations and training approaches such as smoothing, AT with different budgets, transfer learning, and fine-tuning. For instance, training a ResNet-18 on ImageNet-100 CUDA achieves only 8.96%, 40.08%, and 20.58% clean test accuracies with empirical risk minimization (ERM), L ∞ AT, and L 2 AT, respectively. Here, ERM on the clean training data achieves a clean test accuracy of 80.66%. CUDA exhibits unlearnability effect with ERM even when only a fraction of the training dataset is perturbed. Furthermore, we also show that CUDA is robust to adaptive defenses designed specifically to break it.]]></ab></abstract>
		</profileDesc>
	</teiHeader>
	<text><body xmlns="http://www.tei-c.org/ns/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Modern deep learning training frameworks heavily depend on large-scale datasets for achieving high accuracy.</p><p>This encourages deep learning practitioners to scrape data from the web for data collection <ref type="bibr">[8,</ref><ref type="bibr">31,</ref><ref type="bibr">39,</ref><ref type="bibr">42]</ref>. Since a lot of the data is publicly available online, sometime this scrapping of data is unauthorized. For instance, a recent article <ref type="bibr">[15]</ref> discloses that a private company trained a commercial face recognition system using over three billion facial images collected from the internet without any user consent. Although such massive data can significantly boost the performance of deep learning models, it raises serious concerns about data privacy and security.</p><p>To prevent the unauthorized usage of personal data, a series of recent papers <ref type="bibr">[10,</ref><ref type="bibr">17,</ref><ref type="bibr">55]</ref> propose to poison data with additive noise. The idea is to make datasets unlearnable for deep learning models by ensuring that they learn the correspondence between noises and labels. Thereby, they do not learn much useful information about the clean data, significantly degrading their clean test accuracy. However, in recent works <ref type="bibr">[11,</ref><ref type="bibr">17,</ref><ref type="bibr">49]</ref>, these unlearnability methods are shown to be vulnerable to adversarial training (AT) frameworks <ref type="bibr">[34]</ref>. Motivated by this problem, Fu et al. <ref type="bibr">[11]</ref> developed Robust Error-Minimization (REM) noises to make unlearnable data that is protected from AT. While the authors show the effectiveness of REM in multiple scenarios, we demonstrate that these methods are still not robust against different data augmentations or training settings (see Section 3.2). Furthermore, current unlearnability frameworks <ref type="bibr">[10,</ref><ref type="bibr">11,</ref><ref type="bibr">17,</ref><ref type="bibr">55]</ref> are model-dependent and require expensive optimization steps on deep learning models to obtain the additive noises. They also need to train the deep learning models from scratch to obtain noises for each new data set.</p><p>In this paper, we propose a novel Convolution-based Unlearnable DAtaset (CUDA) generation technique. We address limitations of existing unlearnable data generation techniques in Section 3.2 and motivate our CUDA technique in Section 3.3. For generating CUDA, an attacker randomly generates different convolutional filters for each class in the dataset using a private key or seed value. These filters are used to perform controlled class-wise convolutions on the clean training dataset to obtain CUDA. As we describe in Sections 3.3 and 5.2, CUDA generation performs controlled convolutions using a blur parameter p b to ensure that the semantics of the dataset are preserved (see Figure <ref type="figure">1</ref>). CUDA generation with a lower blurring parameter p b adds less perceptible noises to clean samples. A network trained by a defender on CUDA is encouraged to learn the shortcut relation between class-wise convolutional filters and labels rather than useful features for classifying the clean data. Since the seed value for generating the filters are private, its not possible for the defender to obtain clean data from CUDA alone. Additionally, CUDA exhibits unlearnability effect with ERM even when only a fraction of the training dataset is perturbed (see <ref type="bibr">Section 5)</ref>. While the existing unlearnability works use additive noises, CUDA generation technique enjoys the advantage of introducing multiplicative noises in the Fourier domain due to the convolution theorem (since convolution of signals is the same as element-wise multiplication in the Fourier domain). This lets CUDA generation add higher amounts of noise in the image space, specifically along the edges in images, and makes it resilient to AT with small additive noise budgets. In Figure <ref type="figure">2</ref>, with the help of t-SNE plots <ref type="bibr">[51]</ref>, we also find that the noises added by CUDA generation is not linearly separable while they are linearly separable for the existing works on unlearnability <ref type="bibr">[54]</ref>.</p><p>In Section 4, we theoretically show that CUDA generation can successfully poison Gaussian mixture data by degrading the clean data accuracy of the optimal Bayes classifier. We state our result informally below while the formal version is presented in Theorem 2.</p><p>Theorem 1 (Informal) Let D denote a Gaussian mixture data with two modes, P D denote the optimal Bayes classifier trained on D, and &#964; D (P D ) denote the accuracy of the classifier P D on D. Then, under some assumptions, for every clean data D, there is a CUDA D such that</p><p>Furthermore, our empirical experiments in Section 5 demonstrate the effectiveness of CUDA under various training scenarios such as ERM with various augmentations and regularizations, AT with different budgets, randomized smoothing <ref type="bibr">[6,</ref><ref type="bibr">22,</ref><ref type="bibr">27]</ref>, transfer learning, and fine-tuning. For instance, training a ResNet-18 on CIFAR-10 CUDA achieves only 18.48%, 44.4%, and 51.14% clean test accuracies with ERM, L &#8734; AT, and L 2 AT, respectively (see Figure <ref type="figure">2</ref>). Here, ERM on the clean training data achieves a clean test accuracy of 94.66%. In addition, we also design adaptive defenses to investigate if CUDA breaks with random or adversarial defense mechanisms. We find that CUDA is robust to the adaptive defenses that we specifically design to break it.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>Our CUDA generation technique is intimately related with adversarial and poisoning attacks. We first discuss some of this literature and then explain their relation with CUDA generation.</p><p>Adversarial attacks. Adversarial examples are specially designed examples that can fool deep learning models at test time <ref type="bibr">[4,</ref><ref type="bibr">12,</ref><ref type="bibr">23,</ref><ref type="bibr">24,</ref><ref type="bibr">47]</ref>. The adversary crafts these examples by adding error-maximizing noises to the clean data. Even slightly perturbed data can serve as adversarial examples. AT is a training framework proposed to make deep learning models robust to adversarial examples <ref type="bibr">[19,</ref><ref type="bibr">25,</ref><ref type="bibr">34,</ref><ref type="bibr">53,</ref><ref type="bibr">59]</ref>. AT is a min-max optimization problem where the model is trained to minimize loss on adversarial examples that have the maximum loss.</p><p>Poisoning attacks. In data poisoning, an attacker aims to hurt the deep learning model's performance by perturbing the training data <ref type="bibr">[2,</ref><ref type="bibr">20,</ref><ref type="bibr">28,</ref><ref type="bibr">32,</ref><ref type="bibr">43,</ref><ref type="bibr">52]</ref>. The backdoor attack is a special type of poisoning attack where a trigger pattern is injected into clean data at training time <ref type="bibr">[5,</ref><ref type="bibr">29,</ref><ref type="bibr">33,</ref><ref type="bibr">36]</ref>. The model trained on this data would misclassify an image with a trigger pattern at test time. Gu et al. <ref type="bibr">[13]</ref> and Li et al. <ref type="bibr">[30]</ref> use perceptible amounts of noises similar to CUDA for data poisoning. However, backdoor attacks do not affect the performance of the model on clean data <ref type="bibr">[1,</ref><ref type="bibr">5,</ref><ref type="bibr">43]</ref>.</p><p>Recent literature utilize data poisoning to protect data from being used for model training without authorization. Yuan and Wu <ref type="bibr">[55]</ref> use neural tangent kernels <ref type="bibr">[18]</ref> to generate clean label attacks that can hurt the generalization of deep learning models. Huang et al. <ref type="bibr">[17]</ref> show that errorminimizing noise addition can serve as a poisoning technique. Fowl et al. <ref type="bibr">[10]</ref> show that error-maximizing noises as well can make strong poison attacks. However, all these poisoning techniques do not offer data protection with AT <ref type="bibr">[49,</ref><ref type="bibr">54]</ref>. Fu et al. <ref type="bibr">[11]</ref> proposes a min-min-max optimization technique to generate poisoned data that offers better unlearnability effects with AT.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Convolution-based Ulearnable DAtaset (CUDA)</head><p>In this section, first we give some preliminaries about unlearnability. Then we discuss the limitations of the existing unlearnability methods. Finally, we propose our CUDA generation technique.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Preliminaries</head><p>Let {(x i , y i )} n i=1 &#8764; D n be the clean training dataset where D is the clean data distribution, x i &#8712; X &#8834; R d are the samples, and y i &#8712; Y are the corresponding labels. Suppose a network is given as f &#952; : X &#8594; Y where &#952; &#8712; &#920; is the network parameter and : Y &#215; Y &#8594; R is the loss function. ERM trains a network using a minimization problem of the form: min &#952;</p><p>where &#961; a is the adversarial perturbation radius.</p><p>We will use &#964; D (&#952;), with &#952; the clean model parameter, to denote the clean test accuracy of a model trained on the clean training dataset (i.e., clean model accuracy). In unlearnable dataset generation, an attacker uses an algorithm A : X &#8594; X to generate an unlearnable dataset {( xi = A(x i ), y i )} n i=1 &#8764; Dn from the clean training data. Here, the attacker assumes access to the full clean training datatset. Moreover, the attacker cannot modify the unlearnable dataset once it is released publicly. A defender trains using the unlearnable dataset to obtain a network f &#952; . The objective of the attacker is to design an unlearnable dataset such that the defender's model trained on the unlearnable data achieves a clean test accuracy (i.e., unlearnable model accuracy) worse than the clean model accuracy i.e. &#964; D ( &#952;) &#964; D (&#952;).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Limitations of existing works</head><p>Fu et al. <ref type="bibr">[11]</ref> show that the previous unlearnability methods including Error-Minimization (EM) <ref type="bibr">[17]</ref>, Targeted Adversarial Poisoning (TAP) <ref type="bibr">[10]</ref>, and Neural Tangent Generalization Attack (NTGA) <ref type="bibr">[55]</ref> are vulnerable to AT. Hence, they propose a Robust Error-Minimization (REM) <ref type="bibr">[11]</ref> method that exploits a min-min-max optimization procedure to generate unlearnable noises. REM first trains a noise generator f &#952; on data points {(x i , y i )} n i=1 over a loss func- </p><p>Here, T is a distribution over a set of transformations {t : X &#8594; X }, &#961; u is the defensive perturbation radius, and &#961; a controls the protection level of REM against AT. After training the noise generator, an unlearnable example (x, y) is generated via</p><p>First, note that REM is computationally expensive since it needs to generate unlearnability noises through solving optimization equation 2. Moreover, the existing techniques are model-dependent and they require gradient-based training with a neural network to generate unlearnable data. They also require neural network training from scratch for every dataset that is to be made unlearnable. Table <ref type="table">1</ref> shows the amount of time required to generate various unlearnable datasets using NVIDIA &#174; Tesla V100 GPU and 10 CPU cores. CUDA generation is significantly faster than the existing methods since it uses a model-free approach (no training required). Furthermore, REM is sensitive to hyperparameters and norm-budgets of AT since they generate noises with fixed L &#8734; norm budgets. For instance, a ResNet-18 trained on clean CIFAR-10 dataset achieves a clean test data accuracy of 94.66%. L &#8734; AT with perturbation radius &#961; a = 4/255 on REM CIFAR-10 data (generated using &#961; u = 8/255 and &#961; a = 4/255) achieves only a clean test accuracy of 48.16%. However, L &#8734; AT with perturbation radius &#961; a = 8/255 and L 2 AT with perturbation radius &#961; a = 0.75 can achieve a clean test accuracy of 78.71% and 79.65%, respectively, on the same REM data. We also find that ERM with a ResNet-18 on grayscaled REM CIFAR-10 images can achieve a high test accuracy of 70.76% on the grayscaled CIFAR-10 test data. This shows that REM relies upon the color space for poisoning clean data. Fu et al. <ref type="bibr">[11]</ref> also show that REM noise generated using ResNet-18 is not transferable to DenseNet-121.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Our method: CUDA</head><p>The major limitations of the previous works are that they are vulnerable to AT, and computationally expensive. We think the major reason for the former limitation is the usage of small additive noises for unlearnability. AT is designed to train in the presence of such additive noises. Increasing the budget of the amount of additive noises for unlearnability might destroy the semantics of the images while perturbing them. The latter limitation arises from the fact that these methods are model-dependent and they require multilevel optimizations. Hence, we are motivated to design a compute-efficient unlearnability method that is robust to AT. CUDA technique can perform convolutions to add larger amounts of noises to clean images without destroying its semantics. This can help CUDA to be robust against AT. CUDA uses randomly generated convolution filters for blurring images from each class. This makes a model trained on CUDA to learn shortcut relations between filters and labels. We empirically support these claims in Section 5.2. Additionally, randomly generating the filters makes our technique model-free. Since the keys for generating the filters are private, it is not possible to reverse the blurring effect in CUDA without having access to the corresponding clean images. Hence, we assume that the data publisher deletes the clean images after perturbing them. Moreover, CUDA technique is a novel class of non-additive noise based poisoning attack that needs to be studied.</p><p>CUDA uses convolutional filters s i &#8712; R k&#215;k for each class i &#8712; [1, K]. A random parameter, out of the k 2 parameters, in each of the filters is set to have a value of 1. The rest of the filter parameters are randomly initialized from a uniform distribution U (0, p b ) using a private seed where p b is the blur parameter. Blur parameter controls the level of blurring that occurs when an image x &#8712; [0, 1] d1&#215;d2&#215;d3 is convolved with a filter s i . Here, d 1 , d 2 , and d 3 are the height, width, and number of channels of the image, respectively. For example, a CIFAR-10 image has a dimension of 32 &#215; 32 &#215; 3. The higher p b is, the higher the blurring effect is. Let x = x * s i where x belongs to class i. The CUDA data point for x is given by x = x/MAX( x). Rescaling is performed to make sure that the CUDA image pixels lie between 0 and 1. We find that the unlearnability effect gets stronger with larger p b and k values (see supplementary material for details).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Theory for CUDA</head><p>In this section, we define a binary classification setup similar to <ref type="bibr">[19,</ref><ref type="bibr">35]</ref> to theoretically analyze CUDA. Let D be a clean dataset modelled by an isotropic Gaussian mixture model given by N (y&#181;, I), where y &#8712; {&#177;1} is the class label, &#181; &#8712; R d , and I &#8712; R d&#215;d is the identity matrix. We defer the proofs for all the lemmas and theorem to the sup- To characterize the decision boundary for CUDA, we need to use some properties of Toeplitz matrices from Noschese et al. <ref type="bibr">[37]</ref> given in the following Lemma 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Lemma 3 Any tri-diagonal Toeplitz matrix</head><p>Next we use Lemma 3 to show that the Bayes optimal decision boundary for classifying D is a quadratic plane.</p><p>). The Bayes optimal decision boundary for classifying D is given by P (x) &#8801; x Ax + b x + c = 0, where</p><p>Now we state a lemma regarding the tail of Gaussian quadratic forms which plays a crucial role in our main result.</p><p>Lemma 5 Let &#8226; denote the operator norm, &#8226; 2 denote the vector 2-norm, z &#8764; N (0, I), and Z = z Az +z b+c where A = Q&#923;Q . Using Chernoff bound, for any t &#8805; 0 and &#947; &#8712; R,</p><p>.</p><p>Lemma 5 allows us to provide an upper bound for the accuracy of the unlearnable decision boundary P on the clean dataset D, given as &#964; D ( P ), in Theorem 2 below.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Theorem 2 (Main result)</head><p>For any non-negative constants t 1 and t 2 , the accuracy of the unlearnable decision boundary P on the clean dataset D can be upper-bounded as</p><p>Moreover, for any &#181; = 0 and a -1 &#8712; [0, 0.5], &#8707;a 1 such that &#964; D ( P ) &lt; &#964; D (P ).</p><p>Poisoning is effective only if the accuracy of the unlearnable model P is less than that of the clean model P on the clean dataset D, that is, &#964; D ( P ) &lt; &#964; D (P ). To satisfy this condition, we need to carefully select a y 's. In Theorem 2, we formally state this condition <ref type="foot">1</ref> . Theorem 2 shows that CUDA can effectively poison when there are two distinct modes in the clean Gaussian mixture data model. We validate our theoretical claim through empirical analysis as well (see Figure <ref type="figure">3</ref>). Details for the analysis is given in supplementary material. We find that our upper bound for the clean test accuracy of CUDA classifiers are consistent with our empirical analysis. In our experiments, we also find that the unlearnability effect is stronger with a larger blur parameter. This effect is evident from Figure <ref type="figure">3</ref> where we find that it is likely to get a lower &#964; D ( P ) with higher a y values. These results are consistent with our experimental results in Section 5 with CIFAR-10, CIFAR-100, and ImageNet-100 datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiments</head><p>In this section, we first discuss our experimental setup. More details on the setup is deferred to supplementary material. We then show the robustness of CUDA generation with various datasets and architectures. We also run various experiments to analyze the effectiveness of CUDA under different training techniques (ERM, AT with varying budgets, randomized smoothing, transfer learning, and finetuning) and augmentation techniques (mixup [58], grayscaling, random blurring, cutout <ref type="bibr">[9]</ref>, cutmix <ref type="bibr">[56]</ref>, autoaugment <ref type="bibr">[7]</ref>, and orthogonal regularization <ref type="bibr">[3]</ref>). Finally, we also design adaptive defenses to test the robustness of CUDA. One might think that CUDA filters can be obtained by adversarially training them with the data. We show that CUDA is robust to such adaptive defenses that we design. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Experimental setup</head><p>Datasets. We use three image classification datasets -CIFAR-10, CIFAR-100, <ref type="bibr">[21]</ref> and ImageNet-100 (a subset of ImageNet made of the first 100 classes) <ref type="bibr">[40]</ref>. We use the data augmentation techniques such as random flipping, cropping, and resizing <ref type="bibr">[44]</ref>.</p><p>Architectures. We use ResNet-18 <ref type="bibr">[14]</ref>, VGG-16 <ref type="bibr">[45]</ref>, Wide ResNet-34-10 [57], and DenseNet-121 <ref type="bibr">[16]</ref>. We train the networks with hyperparameters used in Fu et al. <ref type="bibr">[11]</ref>. Previous works mainly employ ResNet-18 for most of their evaluations. Additional experiments on Tiny-ImageNet <ref type="bibr">[26]</ref>, DeIT <ref type="bibr">[50]</ref>, EfficientNetV2-S <ref type="bibr">[48]</ref>, and MobileNetV2 <ref type="bibr">[41]</ref> are provided in the supplementary material.</p><p>CUDA generation. We use filters of size k = 3 and blur parameter p b = 0.3 for both CIFAR-10 and CIFAR-100 datsets. ImageNet-100 is a higher dimensional 224&#215;224&#215;3 image dataset when compared to the 32&#215;32&#215;3 dimensional CIFAR datasets. Hence, we use larger filters of size k = 9 with p b = 0.06 for ImageNet-100. These hyperparameters are chosen such that the CUDA images are not perceptibly highly perturbed and give good unlearnability effect (see plot in Figure <ref type="figure">1</ref>). In supplementary matrial, we show the results of training on CUDA with different hyperparameters for data generation.</p><p>Baselines. We compare CUDA generation technique with four state-of-the-art unlearnability methods -REM <ref type="bibr">[11]</ref>, EM <ref type="bibr">[17]</ref>, TAP <ref type="bibr">[10]</ref>, and NTGA <ref type="bibr">[55]</ref>. We adopt the results reported in <ref type="bibr">[11]</ref> since we use the same hyperparameters for training. For REM, we select hyperparameters &#961; u = 8/255 and &#961; a = 4/255, the highest radii values in <ref type="bibr">[11]</ref>. For comparing unlearnable methods, we look at the clean test accuracy. The lower the test accuracy, the better the unlearnability method is. As mentioned in the supplementary material, we use publicly released official codebases for reproducing the baselines using their default hyperparameters.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Effectiveness of CUDA</head><p>Why does CUDA work? In order to measure how much CUDA technique's blurring affects the dataset's quality, we compare class-wise blurring (CUDA technique) against uni- versal blurring where a single convolutional filter is used for blurring all the images. We keep the filter generation parameters fixed (p b = 0.3 and k = 3) for both CUDA classwise blurring and universal blurring. ResNet-18 trained with clean CIFAR-10, universally blurred CIFAR-10, and CIFAR-10 CUDA achieve clean test accuracies of 94.66%, 90.47%, and 18.48%, respectively. This suggests that our controlled blurring does not obscure the semantics of the dataset. Hence, the significant drop in the clean test accuracy introduced by CUDA is most likely due to the usage of class-wise filters. This suggests that a model trained on CUDA learns the relation between the class-wise convolution filters and their corresponding labels. Therefore, during test time when this convolution effect is absent, the CUDA trained model fails to make correct classifications. Furthermore, the model trained on CUDA achieves an accuracy of 99.91% on the CIFAR-10 CUDA testset. This strongly supports our claim that the CUDA model learns to classify images based on the convolutional filters used to blur them.</p><p>In addition, if we permute the class-wise filters for blurring the test set (i.e., blurring class 1 images with class 2 filters, class 2 images with class 3 filters, and so on), we get a very low accuracy of 2.53% on this test set. Further details are provided in the supplementary material. Finally, we note that real-world datasets might also contain blurred images due to various factors such as motion blurring, weather conditions, issues with the camera, etc. Hence, detecting if a blurred image is poisoned might not always be possible. However, one might argue that it is possible to detect if an entire dataset is blurred. Interestingly, later in this section we show that CUDA technique exhibits unlearnability effect even when only a fraction of the training dataset is poisoned. Different datasets. We first compare the effectiveness of CUDA with different datasets using ERM and L &#8734; AT with &#961; a = 4/255. We use ResNet-18 for the experiments. The results are shown in Table <ref type="table">2</ref>. These results show that EM, TAP, and NTGA are not robust to AT. However, both CUDA and REM are successful. Here, our method CUDA outperforms REM with CIFAR-10 and ImageNet-100 datasets. Smartly designed additive noise in AT helps in achieving better generalization than ERM on the unlearnable datasets. This experiment thus demonstrates that ERM and AT are not good choices for training with CUDA and REM dataset.</p><p>Different models. Next we compare the effectiveness of CUDA using various deep learning architectures with L &#8734; AT (&#961; a = 4/255). We use CIFAR-10 for the experiments. The results are shown in Table <ref type="table">3</ref>. As we see in the table, CUDA is effective with all the five network architectures. However, REM is not seen to be transferrable with DenseNet-121.</p><p>Robustness to different training settings. In Section 3.2, we show that REM is sensitive to the training settings. REM generated using L &#8734; radii budgets of &#961; u = 8/255 and &#961; a = 4/255 for CIFAR-10 breaks with L &#8734; AT (&#961; a = 8/255) and L 2 AT (&#961; a = 0.75) to get test accuracy of 78.71% and 79.65%, respectively. Hence, we run experiments to check the robustness of CUDA with various AT norm budgets. The results are shown in Table <ref type="table">4</ref>. As we see in the table, CUDA is robust to ERM, L &#8734; , and L 2 AT settings with varying training budgets. Impressively, the highest test accuracy achieved with training on CIFAR-10 CUDA, CIFAR-100 CUDA, and ImageNet-100 CUDA are as low as 51.19%, 36.90%, and 40.08%, respectively. We also find that using a pre-trained ResNet-18 with CIFAR-10 CUDA only achieves clean test accuracy of 42.42% and 48.22% with fine-tuning the full network and a newly trained final layer, respectively (details are deferred to the supplementary material).</p><p>Different protection percentages. In Section 3.1, we assume that the attacker has access to the full clean training data. However, in real life settings, this might not be always possible. Hence, we train ResNet-18 on a mix of CIFAR-10 CUDA and clean CIFAR-10 training datasets to evaluate the effectiveness of poisoning with varying data protection percentages. Protection percentage denotes the percentage of the training data that is poisoned.</p><p>We show the results in Table <ref type="table">5</ref>. In the table, the "Mixed" column denotes the clean test accuracy of a model trained One may think that CUDA technique can be broken by learning the private filters from the data. We test this idea by training deconvolution filters to find if we can reverse the blurring effect in CUDA with adversarially trained filters. We use a novel Deconvolution-based Adversarial Training (DAT) technique that is similar to AT (check supplementary material for details). While the adversarial step in AT learns sample-wise error-maximizing additive noises, the adversarial step in DAT learns class-wise error-maximizing deconvolution filters. We train DAT with filters of varying sizes (3, 5, and 7) on CIFAR-10 CUDA using ResNet-18. The filter parameters are constrained within a finite range to make sure that the images do not get distorted with the adversarial step similar to projection in projected gradient descent. We find that CUDA is robust to DAT. DAT using filters of size 3, 5, and 7 with CUDA achieves only test accuracy of 39.05%, 46.21%, and 38.48%, respectively. DAT is not successful against CUDA since we can not invert convolutions without the knowledge of the private filters or clean images corresponding to CUDA.</p><p>Limitations and future directions. The unlearnability effect of CUDA can be defended if some fraction of the clean data and its corresponding CUDA images are leaked. The defender must also be able to detect if all samples in the dataset are poisoned. However, our work assumes a setup where the filters remain private. For example, a data publisher could simply publish their CUDA images and delete the clean images permanently to prevent this scenario. As discussed in this section, CUDA as well as other prior works do not perform well with different protection percentages with AT. Improving this can be an interesting research direction. We believe that extending CUDA technique to other domains such as tabular and text data is also an interesting future direction. It would also be interesting to see theoretical analysis of CUDA considering more complex setups.</p><p>[58] Hongyi Zhang, Moustapha Ciss&#233;, Yann N. <ref type="bibr">Dauphin</ref> At the optimal decision boundary the probabilities of any point x &#8712; R d belonging to class y = -1 and y = 1 modeled by D are the same. Here, &#181; = &#181; 1 = -&#181; -1 and</p><p>Now, the accuracy of the clean model P is to be computed. Note that if P (x) &lt; 0 the Bayes optimal classification is class -1, else the classification is class 1. Let z &#8764; N (0, I), and Z &#8764; N (0, 1), and sgn(.) be the signum function.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.2. Proof for Lemma 2</head><p>Let D 1 = N (&#181;, I). For every data point (x, y) &#8764; D 1 , let the perturbed data (A 1 x, y) be modelled by a distribution D1 . We prove that D1 = N (A 1 &#181;, A 1 A 1 ).</p><p>Tri-diagonal Toeplitz matrices A y = T (d; a y , 1, a y ) are symmetric. Hence, D = N (yA y &#181;, A 2 y ).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.3. Remarks on Lemma 3</head><p>A tri-diagonal Toeplitz matrix T (d; a 1 , a 2 , a 3 ) is represented as</p><p>The class of matrices A y = T (d; a y , 1, a y ) are symmetric and can be diagonalized as QDQ .</p><p>1/2 sin ij&#960; d+1 i,j is symmetric and it is the common eigenvector matrix to all A y matrices. As shown in Lemma 3, Q and D can be represented using trigonometric functions. Also, we have</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.4. Proof for Lemma 4</head><p>At the optimal decision boundary the probabilities of any point x &#8712; R d belonging to class y = -1 and y = 1 modeled by D are the same. Here, &#181; = &#181; 1 = -&#181; -1 and A y 's are symmetric.</p><p>2 )</p><p>Note that here if P (x) &lt; 0, the Bayes optimal classification is class -1, else the classification is class 1. Here, for shorthand notations we denote</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.5. Proof for Lemma 5</head><p>Let Z = z Az +z b+c and z &#8764; N (0, I) &#8834; R d where A = Q&#923;Q . Also,</p><p>For any t &#8805; 0 and x &#8764; N (0, I), we write the moment generating function for a quadratic random variable Y = x Ax as 2</p><p>.</p><p>Using the Chernoff bound and E z Az = Tr(AE[zz ]) = Tr(A), for some &#947;,</p><p>since &#923; is a diagonal matrix. Using Woodbury matrix identity, we get (I -2t&#923;) -1 = I -(I -1 2t &#923; -1 ) -1 . This gives us</p><p>.</p><p>A.6. Proof for Theorem 2</p><p>Note that if P (x) &lt; 0, the classifier predicts a label for class -1, else the predicted label would be 1. Here, x = y&#181; + z where z &#8764; N (0, I) and y &#8712; {&#177;1} since (x, y) &#8764; D. </p><p>We can see that</p><p>Using Lemma 5, with &#947; = &#947; 1 , t = t 1 for computing p 1 and &#947; = &#947; 2 , t = t 2 for computing p 2 where t 1 , t 2 are some non-negative constants, we get</p><p>, and</p><p>This gives us the upper bound for &#964; D ( P ). However, to make sure that this upper bound is smaller than 1, we need to assert more conditions. p 1 and p 2 become smaller as &#947; 1 and &#947; 2 are larger positive numbers. However,</p><p>) 0. Hence, we look at separately at cases when either &#947; 1 &gt; 0 or &#947; 2 &gt; 0.</p><p>If &#947; 1 &gt; 0, then &#964; D ( P ) = 1 2 (p 1 + 1) &lt; 1. Else, if &#947; 2 &gt; 0, then &#964; D ( P ) = 1 2 (p 2 + 1) &lt; 1. We know that for &#181; = 0, &#964; D (P ) = &#966;(&#181;) &gt; 1 2 . Moreover, for any a -1 &#8712; [0, 0.5], &#8707;a 1 such that &#964; D ( P ) &lt; &#964; D (P ). This can be satisfied by picking a 1 such that either &#947; 1 or &#947; 2 is very large, i.e., 1  2 &lt; &#964; D ( P ) = A.7. Details on generating Figure <ref type="figure">3</ref> We use &#181; &#8712; R d , d = 100 to generate clean dataset with 1000 data points. They are randomly split into training and testing partitions of equal size. All the assumptions are consistent with the details provided in the main body. We use 30 &#215; 30 mesh-grid to plot the contour plots. While plotting the theoretical upper bounds, we choose the best t 1 , t 2 with grid search from a search space</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.8. Experimental details</head><p>This subsection provides the details for experiments in Section 5.</p><p>Hardware. We use NVIDIA &#174; RTX A4000 GPU with 16GB memory with 16 AMD &#174; EPYC 7302P CPU cores.</p><p>Data augmentations. For CIFAR-10 and CIFAR-100, we use random flipping, 4 pixel padding, and random 32 &#215; 32 size cropping. For ImageNet-100, we use random flipping and random cropping with resizing to 224 &#215; 224 size. All the images are rescaled to have pixel values in the range [0, 1].</p><p>Baselines. We compare CUDA against error-minimizing noise <ref type="bibr">[17]</ref>, targeted adversarial poisoning <ref type="bibr">[10]</ref>, neural tangent generalization attack <ref type="bibr">[55]</ref>, and robust error-minimizing noise <ref type="bibr">[11]</ref>. We use the experimental outputs reported in <ref type="bibr">[11]</ref> for our comparisons. For REM we choose &#961; u = 8/255 and &#961; a = 4/255 since REM works the best when &#961; u = 2&#961; a <ref type="bibr">[11]</ref>. We perform experiments on REM not present in their work using their code available publicly on GitHub 3 (MIT License).</p><p>Networks. For consistency, we use the same architectures used in <ref type="bibr">[11]</ref>. We use their GitHub script 4 for this purpose.</p><p>Training. We train all the networks for 100 epochs. The initial learning rate is 0.1. Learning rate decays to 0.01 at epoch 40 and to 0.001 at epoch 80. We use a stochastic gradient descent optimizer with a momentum factor of 0.9, weight decay factor of 0.0005, and batch size of 128. For adversarial training, we follow the procedure in <ref type="bibr">[34]</ref>. We use 10 steps of projected gradient descent with a step size of 0.15&#961; a .</p><p>3 <ref type="url">https://github.com/fshp971/robust-unlearnableexamples</ref> 4 <ref type="url">https://github.com/fshp971/robust-unlearnableexamples/tree/main/models</ref> Analysis of CUDA. For grayscaling experiments, we use images with their average channel values as the input to the network. Test accuracy is computed on the grayscaled test datasets. For smoothing, we use the GitHub codes<ref type="foot">foot_2</ref> from <ref type="bibr">[6]</ref> (MIT License). For mixup [58], we use the default value of &#945; = 1.0.</p><p>Deconvolution-based adversarial training (DAT). We experiment with various filter sizes of 3,5, and 7 for the transpose convolution filters. For each batch of data, we use 10 steps of projected gradient descent with a learning rate of 0.1 to learn transpose convolution filters for each class. The weights and biases of the transpose convolution filters are constrained to be within [-C, C]. We choose C = 5. After 10 steps of inner maximizing optimization, the resulting image is rescaled such that the pixel values lie in [0, 1]. See Figure <ref type="figure">4</ref> for clean test accuracy vs. epochs plot for DAT with varying transpose filter sizes. As seen in Figure <ref type="figure">1</ref>, DAT can break CUDA CIFAR-10 with a low blur parameter value of p b = 0.1 to get a clean test accuracy &#8764;78%. However, with higher p b values DAT can not achieve more than 50% clean test accuracy. DAT solves the following optimization problem:</p><p>where denotes the transpose convolution operator, s yi denotes the transpose convolution filter for class y i , and is the soft-max cross-entropy loss function.</p><p>CUDA with augmentations. We use mixup with the default &#945; = 1.0 [58]. See Figure <ref type="figure">5</ref> for the training curve. For random blurring augmentations, we use p b = 0.1, 0.3 and k = 3. With both these parameters, CUDA is seen to be effective. See Figure <ref type="figure">5</ref> for the training curve with p b = 0.3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.9. More experimental results</head><p>Figure <ref type="figure">6</ref> shows the CUDA CIFAR-10 data generated using k = 3 and different p b blur parameters. Figure <ref type="figure">7</ref> shows the CUDA CIFAR-100 and CUDA ImageNet-100 data generated using k = 3, p b = 0.3 and k = 9, p b = 0.06, respectively. Figure <ref type="figure">8</ref> shows the clean test accuracy of ResNet-18 with CUDA CIFAR-10 generated using different blur parameters. As we see in the plots, higher the blur parameter, better the effectiveness of CUDA is. However, we choose p b = 0.3 for our experiments since the the images generated using this hyperparameter look perceptibly more similar to the clean images (when compared to p b = 0.5) while giving a very low clean test accuracy. A lower value of p b = 0.1 gives better unlearnability. However, CUDA CIFAR-10 generated using p b = 0.1 is not robust with   We show the effectiveness of CUDA with Tiny-ImageNet <ref type="bibr">[26]</ref>, DeIT <ref type="bibr">[50]</ref>, EfficientNetV2 <ref type="bibr">[48]</ref>, and Mo-bileNetV2 <ref type="bibr">[41]</ref> below.    (see Figure <ref type="figure">11</ref>). This experiment also demonstrates that the blurring we perform does not make the dataset useless or destroy its semantics. For this experiment, we use models with fixed initialization and random seeds.</p><p>A.11. Why does CUDA work?</p><p>In this section, we perform experiments that show that a model trained on CUDA dataset learns the relation between the class-wise filters and the labels. We train ResNet-18 using the CUDA CIFAR-10 dataset for the experiments. We perform three independent trials for each of the experiments and report the mean performance scores. Trained models achieve a mean clean test accuracy of 21.34%. Now, we use the class-wise filters to perturb the images in the test set based on their corresponding labels. Trained models achieve a very high mean accuracy of 99.91% on this perturbed test set. This shows that the trained models learned the relation between the filters and their corresponding labels. Next, we permute the filters to perturb the test set such that test set images with label i are perturbed with the filters of class (i + 1)%10. Trained models achieve a very low mean accuracy of 2.53% on this perturbed test set. This is evidence that CUDA can also be used for backdoor attacks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.12. Effect of transfer learning</head><p>In this section, we experiment the effect of using a pretrained ResNet-18 with PyTorch <ref type="bibr">[38]</ref>. We train it on the CUDA CIFAR-10 dataset in two different ways. First, we fine-tune the whole network on the CUDA dataset with a learning rate of 0.001 for 15 epochs. This achieves a clean test accuracy of 42.42%. Fine-tuning the network with clean training data gives 94.19% clean test accuracy. Next, we freeze all the layers except the final layer to train a linear classifier with the pre-trained weights using the CUDA  CIFAR-10 dataset. We call this "Freeze and learn". We use a SGD optimizer to train the linear layer for 15 epochs with an initial learning rate of 0.1. The learning rate is decayed by a factor of 10 after every 5 epochs. This achieves a clean test accuracy of 48.22%. The results are shown in Figure <ref type="figure">12</ref>. This experiment shows that pre-trained network with CUDA data training does not help achieve good generalization on the clean data distribution.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.13. Effect of CUDA with regularization techniques</head><p>In this section, we study the effect of training a ResNet-18 with CUDA CIFAR-10 dataset using various regularization techniques such as mixup [58], cutout <ref type="bibr">[9]</ref>, cutmix <ref type="bibr">[56]</ref>, autoaugment <ref type="bibr">[7]</ref>, and orthogonal regularization <ref type="bibr">[3]</ref>. We perform mixup, cutout, cutmix, autoaugment, and orthogonal regularization to achieve 25.53%, 25.80%, 26.93%, 34.09%, and 50.72%. Even though these regularizations help in improving the vanilla ERM training, these networks still do not achieve good generalization on the clean data distribution. We use cutout using GitHub codes <ref type="foot">6</ref>with length=16 and n holes=1, cutmix using GitHub codes <ref type="foot">7</ref> with &#945; = 1, autoaugment using PyTorch <ref type="bibr">[38]</ref>, mixup using GitHub codes<ref type="foot">foot_5</ref> with &#945; = 1, and orthogonal regularization using GitHub codes<ref type="foot">foot_6</ref> with reg=1e-6 (all MIT licenses).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.14. Network parameter distribution</head><p>In this section, we compare the network parameter distributions of ResNet-18 trained on clean and CUDA CIFAR-10 datasets (see Figure <ref type="figure">13</ref>). Both the distributions are similar to normal distributions with a mean of 0. However, the parameter distribution of the clean model has a higher standard deviation than the CUDA-based model's parameter distribution.</p></div><note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0"><p>We note that the conditions &#181; A&#181; + b &#181; + c + Tr(A) +</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_1"><p>2A&#181; + b 2 &lt; 0 or -&#181; A&#181; + b &#181; -c -Tr(A) + 2A&#181; -b 2 &lt; 0 can always be satisfied by picking a sufficiently large &#181; in the direction of an eigenvector corresponding to a negative or positive eigenvalue of A (note that A has negative and positive eigenvalues).</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2"><p>https : / / github . com / Hadisalman / smoothingadversarial</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3"><p>https://github.com/uoguelph-mlrg/Cutout/blob/ master/util/cutout.py</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4"><p>https://github.com/hysts/pytorch_cutmix/blob/ master/cutmix.py</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_5"><p>https : / / github . com / facebookresearch / mixup -cifar10/blob/main/train.py</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6"><p>https://github.com/kevinzakka/pytorch-goodies</p></note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_7"><p>(a) Test accuracy (b) Training loss</p></note>
		</body>
		</text>
</TEI>
