<?xml-model href='http://www.tei-c.org/release/xml/tei/custom/schema/relaxng/tei_all.rng' schematypens='http://relaxng.org/ns/structure/1.0'?><TEI xmlns="http://www.tei-c.org/ns/1.0">
	<teiHeader>
		<fileDesc>
			<titleStmt><title level='a'>SF-PATE: Scalable, Fair, and Private Aggregation of Teacher Ensembles</title></titleStmt>
			<publicationStmt>
				<publisher></publisher>
				<date>2023</date>
			</publicationStmt>
			<sourceDesc>
				<bibl> 
					<idno type="par_id">10437846</idno>
					<idno type="doi">10.24963/ijcai.2023/56</idno>
					<title level='j'>International Joint Conference on Artificial Intelligence (IJCAI)</title>
<idno></idno>
<biblScope unit="volume"></biblScope>
<biblScope unit="issue"></biblScope>					

					<author>Cuong Tran</author><author>Keyu Zhu</author><author>Ferdinando Fioretto</author><author>Pascal Van Hentenryck</author>
				</bibl>
			</sourceDesc>
		</fileDesc>
		<profileDesc>
			<abstract><ab><![CDATA[A critical concern in data-driven processes is to build models whose outcomes do not discriminate against some protected groups. In learning tasks, knowledge of the group attributes is essential to ensure non-discrimination, but in practice, these attributes may not be available due to legal and ethical requirements. To address this challenge, this paper studies a model that protects the privacy of individuals' sensitive information while also allowing it to learn non-discriminatory predictors. A key feature of the proposed model is to enable the use of off-the-shelves and non-private fair models to create a privacy-preserving and fair model. The paper analyzes the relation between accuracy, privacy, and fairness, and assesses the benefits of the proposed models on several prediction tasks. In particular, this proposal allows both scalable and accurate training of private and fair models for very large neural networks.]]></ab></abstract>
		</profileDesc>
	</teiHeader>
	<text><body xmlns="http://www.tei-c.org/ns/1.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink">
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>A number of decision processes with high societal impact, such as criminal assessment, lending, and hiring procedures, are increasingly being supported by machine-learning systems. A critical concern is that the learned models may report outcomes that are discriminatory against some demographic group, including gender, ethnicity, or age.</p><p>To ensure non-discrimination in learning tasks, knowledge of the sensitive attributes is essential. At the same time, legal and ethical requirements often prevent the use of this sensitive data. For example, U.S. law prevents using racial identifiers in the development of models for consumer lending or credit scoring, and the GDPR in the E.U. prevents the collection of protected user attributes. In this scenario, an important tension arise between (1) the demand for models to be non-discriminatory, (2) the requirement for such model to use the protected attribute during training, as adopted by common fairness models, and (3) the restriction on the data or protected attributes that can be used. There is thus a need to develop learning models that can both guarantee nondiscriminatory decisions and protect the privacy of the individuals' groups attributes.</p><p>To this end, this paper introduces a novel differentially private learning framework that satisfy group fairness while providing privacy of the protected attributes. The proposed framework, called Scalable, Fair, and Private Aggregation of Teacher Ensemble (SF-PATE) is inspired by the success of private teachers ensemble learning <ref type="bibr">[Papernot et al., 2018]</ref>. These frameworks transfer the classification knowledge learned from a pretrained ensemble of models (called teachers) to a target model (called student) via a privacypreserving aggregation process. This paper exploits this key idea, but rather than transferring the classification capability of the models, it seeks to answer an important and unanswered question: can fairness properties of a model ensemble be transferred in a privacy-perserving way to a target model?</p><p>In addition to providing an affirmative answer to the question above, this paper makes four key contributions: (1) It proposes two flavors of SF-PATE that enforce fairness properties while protecting the demographic group attributes. (2) In addition to guaranteeing differential privacy, the paper provides an analysis on the fairness properties of SF-PATE and shows that unfairness can be bounded in many practical settings. (3) Importantly, SF-PATE decouples the implementation of fairness and privacy requirements and it can be systematically built on top of (non-private) fair models which are viewed as black-box. This engineering benefit is truly unique: it simplifies the development of private and fair models, and facilitates the adoption of new fairness metrics in privacypreserving ML, as the paper shows in the evaluation. (4) Evaluations on both tabular and image datasets show that SF-PATE not only achieves better accuracy, privacy, and fairness tradeoffs with respect to the current state of the art, but it is also significantly faster. This aspect is important because the added computational cost of considering privacy and fairness in already computationally heavy models (e.g., large vision or language models) can make deployment more expensive and may discourage the release of ethical and trustworthy models. Supplemental material. <ref type="bibr">[Tran et al., 2022]</ref> contains an extended version of this paper including proofs of all theorems, and additional experiments. All references to the appendix in this paper refer to such an extended version.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>The interconnection between privacy and fairness is receiving increasing attention and the reader is referred to <ref type="bibr">[Fioretto et al., 2022]</ref> for an overview about the state of the field. Within this literature, a recent line of work observed that private models may have a negative impact towards fairness <ref type="bibr">[Bagdasaryan et al., 2019;</ref><ref type="bibr">Pujol et al., 2020;</ref><ref type="bibr">Tran et al., 2021c;</ref><ref type="bibr">Tran et al., 2021a;</ref><ref type="bibr">Tran and Fioretto, 2023]</ref>. Building from these observations, <ref type="bibr">[Ekstrand et al., 2018]</ref> raise questions about the tradeoff between privacy and fairness and, <ref type="bibr">[Jagielski et al., 2019]</ref> and <ref type="bibr">[Mozannar et al., 2020]</ref> proposed two simple, yet effective algorithms that satisfy (&#9999;, ) and &#9999;differential privacy, respectively, for equalized odds. In particular, state of the art model M, proposed in <ref type="bibr">[Mozannar et al., 2020]</ref>, adds calibrated noise to the group attributes using randomized response prior to use them as input to a fair classifier. <ref type="bibr">[Xu et al., 2019]</ref> proposed a private and fair logistic regression model making use of the functional mechanism <ref type="bibr">[Chaudhuri et al., 2011]</ref>. Finally, <ref type="bibr">[Tran et al., 2021b]</ref> proposed PF-LD, a method to train a private classifier under differential privacy stochastic gradient descent <ref type="bibr">[Abadi et al., 2016]</ref> while imposing fairness constraints. These constraints are imposed using a privacy-preserving extension of the Dual Lagrangian framework of <ref type="bibr">[Fioretto et al., 2020]</ref>. Although this method was shown effective to balance privacy and fairness, it is computationally expensive.</p><p>In contrast, this work introduces a semi-supervised framework that relies on transferring privacy and fairness from a model ensemble to construct high-quality private and fair classifiers, while also being practical and scalable.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Problem Settings and Goals</head><p>The paper considers datasets D consisting of n individual data points (X i , A i , Y i ), with i 2 [n] drawn i.i.d. from an unknown distribution &#8679;. Therein, X i 2 X is a non-sensitive feature vector, A i 2 A, where A = [m] (for some finite m) is a demographic group attribute and Y i 2 Y is a class label. The goal is to learn a classifier M &#10003; : X ! Y, where &#10003; is a vector of real-valued parameters, that ensures a specified non-discriminatory notion with respect to A while guaranteeing the privacy of the group attribute A. The model quality is measured in terms of a non-negative, and assumed differentiable, loss function L : Y &#8677; Y ! R + , and the problem is that of minimizing the empirical risk function (ERM):</p><p>(P)</p><p>The paper focuses on learning general classifiers, such as neural networks, that satisfy group fairness (as defined next) and protect the disclosure of the group attributes using the notion of differential privacy. Importantly, the paper assumes that the attribute A is not part of the model input during inference. This is crucial in the application of interest to this work as the protected attributes cannot be disclosed.</p><p>Fairness. This work constrains a classifier M to satisfy a given group fairness notion under a distribution over (X, A, Y ) for the protected attribute A as defined below.</p><p>Definition 1 (&#8629;-fairness). A model M &#10003; is &#8629;-fair w.r.t. a joint distribution (X, A, Y ) and fairness function h(&#8226;) iff:</p><p>(1) wehre &#8672;(D, h, &#10003;) is referred to as fairness violation.</p><p>The above compares a property for a group of individuals with respect to the whole population and quantifies its difference. Therein, h(M &#10003; (X), Y ) 2 R, referred here as fairness function, defines a target group fairness notion while parameter &#8629; 2 R + represents the fairness violation.</p><p>The above capture an example class of group fairness notions, including Demographic parity, Equalized odds, and Accuracy parity. A comprehensive review of these fairness definitions is provided in Appendix A.</p><p>Differential Privacy. Differential Privacy (DP) <ref type="bibr">[Dwork et al., 2006</ref>] is a strong privacy notion used to quantify and bound the privacy loss of an individual participation to a computation. Similarly to <ref type="bibr">[Jagielski et al., 2019;</ref><ref type="bibr">Tran et al., 2021b]</ref>, this work focuses on the instance where the protection is restricted to the group attributes only. A dataset D 2 D = (X &#8677;A&#8677;Y) of size n can be described as a pair (D P , D S ) where D P 2 (X &#8677; Y) n describes the public attributes and D S 2 A n the group attributes. The privacy goal is to ensure that the output of the learning model does not significantly change when a single group attribute is changed.</p><p>The action of changing a single attribute from a dataset D S , resulting in a new dataset D 0 S , defines the notion of dataset adjacency. Two dataset D S and D 0 S 2 A n are said adjacent, denoted D S &#8672; D 0 S , if they differ in at most a single entry (e.g., in one individual's group membership).</p><p>Definition 2 (Differential Privacy). A randomized mechanism M : D ! R with domain D and range R is (&#9999;, )differentially private w.r.t. attribute A, if, for any dataset D P 2 (X &#8677; Y) n , any two adjacent inputs D S , D 0 S 2 A n , and any subset of output responses R &#10003; R:</p><p>S ) 2 R)+ . When = 0 the algorithm is said to satisfy &#9999;-DP. Parameter &#9999; &gt; 0 describes the privacy loss of the algorithm, with values close to 0 denoting strong privacy, while parameter 2 [0, 1] captures the probability of failure of the algorithm to satisfy &#9999;-DP. The global sensitivity f of a realvalued function f : D ! R k is defined as the maximum amount by which f changes in two adjacent inputs D and D 0 :</p><p>In particular, the Gaussian mechanism, defined by</p><p>) is the Gaussian distribution with 0 mean and standard deviation f , satisfies (&#9999;, )-DP for &gt; </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Private &amp; Fair Learning: Challenges</head><p>Constructing models enforcing both privacy and fairness brings with it three fundamental challenges:</p><p>&#8226; Scalability: When interpreted as constraints of the form (1), fairness properties can be explicitly imposed to problem (P). The resulting problem is typically non-convex, The approach proposed in this paper avoids these difficulties by providing a teachers ensemble model that (1) provides state-of-the-art accuracy, privacy, and fairness tradeoffs, (2) adds negligible computational cost to standard (non-private) training, and (3) directly exploits the existence of fair (nonprivate) algorithms. This last property significantly simplifies the engineering of fair and private models, facilitates the adoption of new fairness notions, and avoids the complications introduced by fairness constraints in the DP analysis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">The SF-PATE Framework</head><p>This section discusses two algorithms for transferring fairness considerations during the private learning process. Both algorithms rely on the presence of an ensemble of teacher models T = {M (i) } K i=1 , with each M (i) trained on a nonoverlapping portion D i of the dataset D. This ensemble is used to transfer knowledge to a student model M&#10003; : X ! Y.</p><p>The student model M is trained using a subset D &#10003; D P whose samples are randomly selected from the original training dataset but whose protected group attributes are unrevealed. As characteristic of ensemble models, the student's queries (data entries X 2 D) are processed by the teacher ensemble to predict the associated label Y through a voting process (which is elaborated below). This process allows the teachers knowledge to be transferred to the student model. The framework, and the two variants introduced next, are depicted schematically in Figure <ref type="figure">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Transfer Learning with Fair Student</head><p>The first algorithm presented, called SF S -PATE, trains a student model with privacy-preserving group attributes chosen by an ensemble of teachers. Subscript S in the algorithm's name stands for "student" to emphasize that it is the student that enforces the fairness constraints during training.</p><p>The teachers ensemble T = {M (i) } K i=1 is comprised of models M (i) : X ! A. Its goal is to predict the group attributes from a sample features, but not the labels as classically done in classification tasks. Note that, importantly, the teacher models are standard classifiers: they are neither private nor fair. Their role is to transfer the information of the group attributes associated with the samples (X i , Y i ) 2 D provided by the student. The student model M&#10003; : X ! Y solves the following regularized constrained minimization:</p><p>where &#7805;A : A K ! A is a private voting scheme returning the group attribute &#195; chosen by the teachers ensemble:</p><p>&#10003; (X) = a} associated to group a 2 A with Gaussian noise of zero mean and standard deviation .</p><p>Problem (2) minimizes the standard empirical risk function (first component of (2a)), while encouraging the student model parameters &#10003; to be close to the optimal, non-fair, parameters &#10003; &#8676; (second component of (2a)). Note that &#10003; &#8676; can be considered as the most accurate model since it solves problem (P) using all training data D. This model will not leak any private information w.r.t. the protected attributes A because it is trained using only features X and labels Y . The regularization parameter &gt; 0 controls the trade-off between accuracy (when is large, the model is pushed towards &#10003; &#8676; ) and fairness (when is small, the model prioritizes fairness violations). The fairness constraints expressed in Equation (2b) can be enforced through the adoption of off-the-shelf techniques and this paper relies on the use of the Lagrangian dual deep learning framework of <ref type="bibr">[Fioretto et al., 2020]</ref>. This choice also allows us to make a fair comparison with state-ofthe-art PF-LD <ref type="bibr">[Tran et al., 2021b]</ref> in our experiments.</p><p>SF S -PATE achieves (&#9999;, )-DP by introducing calibrated noise on the counts of the protected attributes predicted by the teachers ensemble, as illustrated in Equation ( <ref type="formula">3</ref>). The privacy analysis follows from <ref type="bibr">[Papernot et al., 2018]</ref> with a few additional considerations and is provided in Appendix C.</p><p>Fairness analysis. The rest of this section discusses the fairness guarantees provided by SF S -PATE. The proofs of all theorems are reported in the Appendix B. Note that SF S -PATE relies on a (non-private) fair classifier to train the student model. The performance of many such fair classifiers has been established in the literature. The key question however is to assess the impact of the privacy-preserving voting scheme on the fairness of the student model. The following theorem bounds the fairness violation of the student M&#10003; w.r.t. (X, A, Y )-i.e., the original, non-noisy data-when the group attribute A and its privacy-preserving counterpart &#195; are close enough to each other statistically, i.e.,</p><p>for events (X = x, Y = x) and value a. Note that, above, &#8984; A characterizes the confidence of the noisy voting process. Small &#8984; A values correspond to ensembles with high prediction agreement (e.g., many teachers predict the same output &#195;). Conversely, large &#8984; A values correspond to ensembles with low agreements, rendering the noisy votes less tolerant to noise. Additionally, this property specifies that the group attributes can be inferred from the sample features and labels, which is also what allows a notion of disparity to arise.</p><p>Theorem 1. Let M&#10003; be &#8629;-fair w.r.t. (X, &#195;, Y ) and h(&#8226;). Then, M&#10003; is &#8629; 0 -fair w.r.t. (X, A, Y ) and h(&#8226;) with:</p><p>where B is the supremum of the fairness function h(&#8226;).</p><p>Thus, this result help us quantifying the fairness guarantees attained on the original data (X, A, Y ) when only private noisy data (X, &#195;, Y ) (as used by SF S -PATE) is used. Quantity Pr( &#195; = a), however, may be difficult to compute, due to the noisy voting process. The next result allows us to compute &#8629; 0 as a function of quantity Pr(A = a) only, which is independent from the noisy process and can be computed empirically from the data, in non-restrictive settings.</p><p>Corollary 1. When the probability of A belonging to any class a 2 [m] is at least &#8984; A (i.e., Pr (A = a) &gt; &#8984; A , 8 a 2 [m]), the fairness bound &#8629; 0 can then be given by</p><p>The appendix in <ref type="bibr">[Tran et al., 2022</ref>] also provides a result (described in Theorem 3) to compare the fairness guarantees achieved when the protected attributes &#195; are computed using randomized response, which is the core process used by the baseline and state of the art model M <ref type="bibr">[Mozannar et al., 2020]</ref>. This result shows that, when compared with the baseline model M, SF S -PATE makes the privacy-preserving group attributes &#195; much closer to their original counterparts A under the same privacy budgets. This aspect is also further discussed in the experiments (as emphasized in Figure <ref type="figure">3</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Transfer Learning with Fair Teachers</head><p>SF S -PATE transfers privacy-preserving group attributes from an ensemble of teachers to a fair student model. This section introduces an SF-PATE variant that transfer (non-private) fairness through the semi-supervised learning scheme. The proposed algorithm, called SF T -PATE, trains a student model from an ensemble of fair teacher models. Subscript T in the algorithm's name stands for "teachers" and emphasizes that the fairness constraints are enforced by the teachers and transferred to the student during training.</p><p>The teachers transfer fairness properties via model prediction to the student. The teachers ensemble T = {M (i) } K i=1 is composed of pre-trained classifiers M (i) : X ! Y that are non-private but fair over their training data D i . Each SF T -PATE teacher solves an empirical risk minimization problem subject to fairness constraint:</p><p>Similar to SF S -PATE, the implementation uses the Lagrangian dual method <ref type="bibr">[Fioretto et al., 2020]</ref> to achieve this goal. This is again an important aspect of the SF-PATE framework since, by relying on black-box fair (and not private) algorithms, it decouples the dependency of developing joint private and fair analysis. The student model M&#10003; : X ! Y solves a standard empirical risk minimization:</p><p>where &#10003; &#8676; represents the student model parameters obtained solving the standard classification task (without fairness considerations) (P) on D and the private voting scheme &#7805;Y : Y K !Y reports the label &#7928; selected by the teachers:</p><p>SF T -PATE protects the privacy of the sensitive group information A and of the labels Y when = 0. The reason comes from the fact that when = 0, the student only utilize the private class label &#7805;Y (T (X) but not the true label Y , implicitly encoded in &#10003; &#8676; , to train the model. Thus, SF T -PATE can be adopted in contexts where both the protected group and the labels are sensitive information. Finally, note that the voting scheme above, emulates the GNMAX framework proposed in <ref type="bibr">[Papernot et al., 2018]</ref>, but the two have fundamentally different goals: GNMAX aims at protecting the participation of individuals into the training data, while SF T -PATE aims at privately transferring the fairness properties of the teachers ensemble to a student model.</p><p>SF T -PATE achieves (&#9999;, )-DP by introducing calibrated noise on the counts of the predicted labels reported by the teachers ensemble (see Appendix C).</p><p>Fairness analysis. The next results considers how well the fairness properties of the ensemble are transferred to the student model. They provides theoretical insights on the potential fairness violations induced by the voting mechanism &#7805;Y . For notational convenience, let B := sup y2Y,y 0 2Y h(y, y 0 ) denote the supremum of the fairness function h(&#8226;). Let Z be the random vector (M </p><p>The following theorem bounds the fairness violation of the voting mechanism &#7805;Y (T ) when the joint distributions of the teachers ensemble and labels Y are roughly similar across the different group classes. That is, for each group attribute a 2 [m], the total variation distance between Z and Z a is bounded from the above by &#8984; Y &gt; 0, i.e.,</p><p>where &#8998; is used to denote the shared probability space.</p><p>Theorem 2. The voting mechanism &#7805;Y (T ) is &#8629; 0 fair w.r.t.</p><p>The result above shed light on the (non-restrictive) conditions required for the ensemble votes to transfer fairness knowledge accurately. The next corollary is a direct consequence of Theorem 2 and provides a sufficient condition for perfect fairness of the voting mechanism &#7805;Y (T ).</p><p>Corollary 2. Suppose that the random vector Z is independent of A, i.e., {Z a } a2[m] and Z are identically distributed. Then, the voting mechanism &#7805;Y (T ) is perfectly fair (i.e., fairness violation &#8629; = 0) w.r.t. (X, A, Y ) and h(&#8226;).</p><p>These results are reassuring: informally speaking, they show that SF T -PATE provides fairness guarantees that are "inversely propertional" to the predictive impact of the pro-tected attributes. The experimental results presented subsequently confirm that SF T -PATE achieves state-of-the-art tradeoffs among accuracy, privacy, and fairness.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Experiments</head><p>This section evaluates the performance of the SF-PATE algorithms against the prior approaches of <ref type="bibr">[Tran et al., 2021b]</ref> and <ref type="bibr">[Mozannar et al., 2020]</ref>, denoted by PF-LD and M, respectively. They represents the state-of-the-art for learning private and fair classifiers in the context studied in this paper. In addition to assess the competitiveness of the proposed solutions, the evaluation focuses on two key aspects that set the proposed framework apart from existing methods. (1) It shows that SF-PATE can naturally handle new fairness notions, even when no viable privacy analysis exists about these notions.</p><p>(2) It shows that SF-PATE has a low computational overhead compared to classical (non-private, non-fair) classifiers, rendering it a practical choice for the training of very large models. These two properties are unique to SF-PATE and make it to applicable to a broad class of challenging decision tasks.</p><p>Datasets and settings. The evaluation is conducted using four UCI tabular datasets: Bank, Parkinson, Income and Credit Card <ref type="bibr">[Blake, 1998]</ref>, and UTKFace <ref type="bibr">[Hwang et al., 2020]</ref>, a vision dataset. The latter is used to demonstrate the scalability of SF-PATE when trained on very large models. All experiments are repeated using 100 random seeds.</p><p>Models and hyperparameters. To ensure a fair comparison, the experimental analysis uses the same architectures, model initialization &#10003;, and parameters for all models (including the baselines models PF-LD and M). For tabular datasets, the underlying classifier is a feedforward neural network with two hidden layers and nonlinear ReLU activations. The fair, non-private, classifiers adopted by the two SF-PATE variants implement a Lagrangian dual scheme <ref type="bibr">[Fioretto et al., 2020]</ref>, which is also the underlying scheme adopted by baseline models. For vision tasks on the UTK-Face dataset, the evaluation uses a Resnet 50 classifier. A more detailed description of these approaches, the settings adopted, hyperparameters optimization, and the datasets is deferred to Appendix D.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Accuracy, Privacy, and Fairness Trade-off</head><p>We now compare the accuracy, fairness, and privacy tradeoffs of the proposed model variants SF S -and SF T -PATE against the baseline models PF-LD and M on the tabular datasets. Figure <ref type="figure">2</ref>(a) illustrates the accuracy (top subplots) and fairness violations &#8672;(&#10003;, h, D) (bottom subplots) when varying the privacy loss &#9999; (x-axis). The fairness notion adopted is demographic parity and additional results on other fairness metrics and datasets are deferred to the Appendix (in <ref type="bibr">[Tran et al., 2022]</ref>), showing similar trends. The figures clearly illustrate that both SF-PATE variants achieve better accuracy/fairness tradeoffs for various privacy parameters. The property of retaining high accuracy with low fairness violation is especially relevant in the tight privacy regime adopted (&#9999; &lt; 2). Observe that the figures y-axes have different scales.</p><p>First, notice that, consistently with previous work showing that teachers ensemble models can outperform DP-SGD  based models in terms of accuracy <ref type="bibr">[Uniyal et al., 2021]</ref>, all SF-PATE models typically outperform the FP-LD model based on DP-SGD. Remarkably, both SF-PATE variants also report lower fairness violations, for the fairness notions analyzed, indicating the strength of this approach. Additionally, recall that both SF S -PATE and M generate privacy-preserving group features and input them to a fair model. However, in contrast to M, the model ensemble used in SF S -PATE exploits the relationship between the sample features X and its associated group information A to derive more accurate private group information &#195;. This results in student models which are often more accurate, and fairer, than the baseline M. This result is further highlighted in Figure <ref type="figure">3</ref>, which reports the expected accuracy of the private group attributes Pr( &#195;i = A i ) produced by SF S -PATE for different sizes K of the ensemble and compares it with that obtained by model M. Note the distinctive ability of the teachers ensemble to generate high-quality privacy-preserving group attributes &#195; = &#7805;A (T (X)), especially under tight privacy regimes. Additionally, increasing the ensemble size also enables the teachers to transfer higher quality private group attributes, which is a fundamental aspect to ensure fairness.</p><p>The second SF-PATE variant, SF T -PATE, which operates by privately transferring the fairness knowledge from a teachers ensemble to a student, is found to always outperform both M and PF-LD on tabular datasets (see Figure <ref type="figure">2</ref>(a) and Appendix D). Finally, the analysis also shows that the average accuracy of the SF-PATE models is within 2% of their nonprivate counterpart and brings up to an order of magnitude fairness violation gains over existing methods. This is significant due to the challenging nature of the tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Handling New Group Fairness Notions</head><p>The next results consider the ability of SF-PATE to handle arbitrary fairness notions, even if a privacy analysis is missing, as long as a fair model can be derived. This feature stems from the use of black-box (non-private but fair) models in SF-PATE. This is in sharp contrast with state-of-the-art model PF-LD, that requires the development of a privacy analysis for each fairness property of interest, in order to calibrate the amount of noise to apply in both primal and dual step. To demonstrate this benefit, this section introduces a new fairness notion, which generalizes demographic parity.</p><p>Definition 3 (Generalized demographic parity). requires the distribution over the predictions of M to be statistically independent of the protected group attribute A. That is, 8a 2 A, &#8984; 2 [0, 1], Pr(M(X) &#8984;|A = a) = Pr(M(X) &#8984;). This generalizes demographic parity, which states that Pr (M &#10003; (X) 0.5 | A = a) = Pr (M &#10003; (X) 0.5). Generalized demographic parity is useful in settings where the decision threshold (e.g., 0.5 above) might not be available at the training time. Matching the distribution of score functions (e.g., credit or income scores) among different groups attributes guarantees demographic parity fairness regardless of the decision threshold adopted. Such fairness constraint can be implemented by equalizing different order statistics of the score functions between group and population level: The experiments H = 2 and the Lagrangian Dual method of <ref type="bibr">[Fioretto et al., 2020]</ref> to enforce these constraints during training. Notice that it is highly non-trivial to derive a privacy analysis for such fairness notion-the PF-LD model only does so for H = 1, and, thus, not viable in this setting.</p><p>Figure <ref type="figure">2</ref>(b) reports the accuracy and fairness violations obtained by the SF-PATE models and the baseline model M which adopts, as a post-processing step, a non-private but fair classifier. Fairness is evaluated in terms of the Wasserstein distance between the score functions of different groups. The smaller the distance the lower the fairness violation. The plots clearly illustrates the advantages of SF-PATE in terms of both accuracy and fairness when compared to model M. Remarkably, the fairness violations reported by SF-PATE are often significantly lower than those reported by model M for various privacy loss parameters &#9999;.</p><p>These results are significant from an engineering standpoint: the development of private and fair analysis is complex and SF-PATE immediately lowers the accessibility barrier for non-privacy experts to develop private and fair ML models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3">Computational Time and Scalability</head><p>The last results demonstrate another key benefits of SF-PATE: its ability to scale to large data and perform well on very deep networks. These experiments use a ResNet 50 (&gt; 23M parameters) and PF-LD was unable to train even a single epoch (in 1h) due to its use of computational expensive per-sample gradient clipping performed during training. This renders such model unusable for many realistic settings. The comparison thus focuses on SF-PATE and M.</p><p>Figure <ref type="figure">2</ref>(c) shows the accuracy and fairness trade-off for different privacy values &#9999;. The models enforce accuracy parity and the figure shows that both versions of SF-PATE significantly improve accuracy compared to model M, while also maintaining similar or even reducing fairness violations. For additional results see Appendix D.</p><p>Table <ref type="table">1</ref> shows the training time required for the algorithms to create a private model (with &#9999; = 1.0) on the benchmarks set. Notice the significant training time differences across the models with SF S -PATE being up to three times faster than M and up to two orders of magnitude faster than PF-LD.</p><p>These results show that SF-PATE can become a practical tool for private and fair ML especially for large, overparametetrized models and under stringent privacy regimes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.4">Discussion and Limitations</head><p>While the previous section highlighted the advantages of the proposed framework over existing models, this section sheds light on key elements and usage guidelines for SF S -and SF T -PATE. While both SF-PATE variants rely on the same framework, they diverge in the fairness enforcing mechanism. SF S -PATE delegates this to its student while SF T -PATE to its teachers. Notice that, by training an ensemble of fair teachers, SF T -PATE is able to transfer fairness knowledge directly which, as observed in the experiments, often results in smaller fairness violations than those attained by SF S -PATE. This is notable, especially in the case of "complex" fairness constraints; i.e., when multiple concurrent constraints are to be enforced, as in the case of generalized demographic parity which imposes multiple order moments matching. Being treated as soft penalty functions, these constraints are added to the original empirical loss function (see Problem (2)). SF S -PATE adds noise to each constraint terms (e.g., those represented by the constrains in (2b)) since the private voting scheme affects the accuracy of the constraints. In contrast, SF T -PATE adds noise only to the original loss term (e.g., the first term of (2a)) because the voting scheme acts on the labels Y and not the protected groups A and fairness is enforced by the teachers.</p><p>A shortcoming of the SF T -PATE algorithm, however, is that it has to enforce fairness in each teacher model. Thus, one has to ensure that enough data (with large enough representation from all the protected groups) is assigned to each teacher. On the other hand, by training a single fair model (the student) SF S -PATE avoids such potential issue.</p><p>It is also worth noting that a limitation of all ensemble models, including those proposed in this work, is the need to store a model for each of the K teachers. This, however, also represents an opportunity for future research to develop effective model storage and pruning techniques that minimize the loss in accuracy and fairness while retaining privacy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Conclusions</head><p>This paper proposed a framework to train deep learning models that satisfy several notions of group fairness while ensuring that the model satisfies differential privacy for the protected attributes. The proposed framework, called Scalable, Fair, and Private Aggregation of Teacher Enseble (SF-PATE) transfer fairness knowledge learned from a pretrained ensemble of models to a target model via a privacy-preserving voting process. The paper analyzes the fairness properties of SF-PATE and shows that unfairness can be bounded in many practical settings. An important property of SF-PATE is to allow the adoption of black-box (non-private) fair models during the knowledge transfer process, which may simplify the development and boosts the adoption of new fairness metrics in privacy-preserving ML.</p><p>Finally, evaluation on both tabular and image datasets shows not only that SF-PATE achieves better accuracy, privacy, and fairness tradeoffs with respect to the current stateof-the-art, but it is also significantly faster. These properties render SF-PATE amenable to train large, overparameterized, models, that ensure privacy, accuracy, and fairness simultaneously, showing that it may become a practical tool for privacy-preserving and fair decision making.</p></div></body>
		</text>
</TEI>
