<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Measuring and Mitigating Bias for Tabular Datasets with Multiple Protected Attributes ⋆</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Manh</forename><forename type="middle">Khoi</forename><surname>Duong</surname></persName>
							<email>manh.khoi.duong@hhu.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Heinrich Heine University</orgName>
								<address>
									<addrLine>Universitätsstraße 1</addrLine>
									<postCode>40225</postCode>
									<settlement>Düsseldorf</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Stefan</forename><surname>Conrad</surname></persName>
							<email>stefan.conrad@hhu.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Heinrich Heine University</orgName>
								<address>
									<addrLine>Universitätsstraße 1</addrLine>
									<postCode>40225</postCode>
									<settlement>Düsseldorf</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Measuring and Mitigating Bias for Tabular Datasets with Multiple Protected Attributes ⋆</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">9A821F217827FA2D862BBA8BB6EF823E</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:37+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Machine Learning</term>
					<term>Bias Mitigation</term>
					<term>Intersectional Discrimination</term>
					<term>Fairness</term>
					<term>AI Act</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Motivated by the recital (67) of the current corrigendum of the AI Act in the European Union, we propose and present measures and mitigation strategies for discrimination in tabular datasets. We specifically focus on datasets that contain multiple protected attributes, such as nationality, age, and sex. This makes measuring and mitigating bias more challenging, as many existing methods are designed for a single protected attribute. This paper comes with a twofold contribution: Firstly, new discrimination measures are introduced. These measures are categorized in our framework along with existing ones, guiding researchers and practitioners in choosing the right measure to assess the fairness of the underlying dataset. Secondly, a novel application of an existing bias mitigation method, FairDo, is presented. We show that this strategy can mitigate any type of discrimination, including intersectional discrimination, by transforming the dataset. By conducting experiments on real-world datasets (Adult, Bank, COMPAS), we demonstrate that de-biasing datasets with multiple protected attributes is possible. All transformed datasets show a reduction in discrimination, on average by 28%. Further, these datasets do not compromise any of the tested machine learning models' performances significantly compared to the original datasets. Conclusively, this study demonstrates the effectiveness of the mitigation strategy used and contributes to the ongoing discussion on the implementation of the European Union's AI Act.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Discrimination in artificial intelligence (AI) applications is a growing concern since the adoption of the AI Act by the European Parliament on March 13, 2024 <ref type="bibr" target="#b0">[1]</ref>. It still remains a significant challenge across numerous domains <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5]</ref>. To prevent biased outcomes, pre-processing methods are often used to mitigate biases in datasets before training machine learning models <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8,</ref><ref type="bibr" target="#b8">9]</ref>. The current corrigendum of the AI Act <ref type="bibr" target="#b0">[1]</ref> emphasizes this in Recital (67): " <ref type="bibr">[...]</ref> The data sets should also have the appropriate statistical properties, including as regards the persons or groups of persons in relation to whom the high-risk AI system is intended to be used, with specific attention to the mitigation of possible biases in the data sets <ref type="bibr">[..</ref></p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>.]"</head><p>Since datasets often consist of multiple protected attributes, pre-processing methods should be able to handle these cases. However, only a few works have addressed this issue <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b9">10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13</ref>] and de-biasing such datasets is still an ongoing research topic. In addition, there is no straightforward approach to managing multiple protected attributes, as shown in Figure <ref type="figure">1</ref>.</p><p>Our paper mainly focuses on how to measure and mitigate discrimination in datasets where multiple protected attributes are present. In our first contribution, we provide a comprehensive categorization of discrimination measuring methods. Besides introducing new measures for some of these cases, we also categorize existing measures from the literature. Some of the listed measures specifically address intersectional discrimination and non-binary groups. The second contribution deals with bias mitigation. For this, we use our published pre-processing framework, FairDo <ref type="bibr" target="#b8">[9]</ref>, that is fairness-agnostic. The fairness-agnostic property makes it possible to define any discrimination measure that should be Color Shape</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Intersectional</head><p>Non-intersectional</p><note type="other">Shape Color</note><p>Figure <ref type="figure">1</ref>: Stick figures can be differentiated by their color and shape. In intersectional discrimination, attributes are intersected, which leads to new subgroups. In non-intersectional, each attribute is treated independently, i.e., colors and shapes are not intersecting in this case.</p><p>minimized. By implementing the introduced measures, we can therefore mitigate biases for multiple protected attributes. Another advantage of FairDo is that it preserves data integrity and does not modify the features of individuals during the optimization process, unlike other methods <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b6">7]</ref>. We evaluated our methodology on popular tabular datasets with fairness concerns, such as Adult <ref type="bibr" target="#b14">[15]</ref>, Bank <ref type="bibr" target="#b15">[16]</ref>, and COMPAS <ref type="bibr" target="#b16">[17]</ref>. We used different discrimination measures to evaluate the effectiveness of the bias mitigation process. Because a successful mitigation process does not guarantee that the outcomes of machine learning models are fair, we trained machine learning models on the transformed datasets and evaluated their predictions regarding fairness and performance. The code for the experiments can be found in the accompanying repository: https://github.com/mkduong-ai/fairdo/evaluation.</p><p>The results of the bias mitigation process as well as the performance of the machine learning models are promising. They indicate that achieving fairness in datasets with multiple protected attributes is possible, and FairDo is a proper framework for this task. Overall, our work contributes technical solutions for stakeholders to enhance the fairness of datasets and machine learning models, aiming for compliance with the AI Act <ref type="bibr" target="#b0">[1]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Preliminaries</head><p>To handle multiple protected attributes, we define 𝒵 = {𝑍 1 , . . . , 𝑍 𝑝 } as a set of protected attributes. It can represent the set of sociodemographic features such as age, gender, and ethnicity. These factors may make individuals vulnerable to discrimination. Each protected attribute 𝑍 𝑘 ∈ 𝒵 is formally a discrete random variable that can take on values from the sample space 𝑔 𝑘 . In this context, we refer 𝑔 𝑘 to groups that describe distinct social categories of a protected attribute. For example, let 𝑍 𝑘 represent gender; then 𝑔 𝑘 is a set containing the genders male, female, and non-binary. To avoid limitations to a particular group fairness notion, we introduce a generalized notation based on the works of Žliobaitė <ref type="bibr" target="#b1">[2]</ref>, Duong and Conrad <ref type="bibr" target="#b8">[9]</ref> in the following.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Definition 2.1 (Treatment).</head><p>Let 𝐸 1 , 𝐸 2 be events and 𝑍 𝑘 be a random variable that can take on values from 𝑔 𝑘 , then we call the conditional probability</p><formula xml:id="formula_0">𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖)</formula><p>treatment, where 𝑖 ∈ 𝑔 𝑘 . 𝐸 1 describes some favorable outcome, such as getting accepted for a job, while 𝐸 2 often represents some additional information about the individual, such as their qualifications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Definition 2.2 (Fairness Criteria).</head><p>With the definition of treatment, we can define fairness criteria that demand equal treatment for different groups. Let 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖) and 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑗) be treatments, then we call the following equation:</p><formula xml:id="formula_1">𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖) = 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑗)</formula><p>a fairness criterion, for all 𝑖, 𝑗 ∈ 𝑔 𝑘 . Definition 2.2 allows us to define various group fairness criteria, including statistical parity <ref type="bibr" target="#b17">[18]</ref>, predictive parity <ref type="bibr" target="#b2">[3]</ref>, equality of opportunity <ref type="bibr" target="#b18">[19]</ref>, etc. They all demand some sort of equal outcome for different groups and can be defined by configuring the events 𝐸 1 , 𝐸 2 . For instance, statistical parity <ref type="bibr" target="#b17">[18]</ref> requires that two different groups have an equal probability of receiving a favorable outcome (𝑌 = 1).</p><p>Example 2.1 (Statistical Parity <ref type="bibr" target="#b17">[18]</ref>). To define statistical parity for the attribute 𝑍 𝑘 using our notation, we set 𝐸 1 := (𝑌 = 1) and 𝐸 2 := Ω. By setting 𝐸 2 to the sample space Ω, we compare the probabilities of the event 𝑌 = 1 across different groups without conditioning on any additional event:</p><formula xml:id="formula_2">𝑃 (𝑌 = 1 | Ω, 𝑍 𝑘 = 𝑖) = 𝑃 (𝑌 = 1 | Ω, 𝑍 𝑘 = 𝑗) ⇐⇒ 𝑃 (𝑌 = 1 | 𝑍 𝑘 = 𝑖) = 𝑃 (𝑌 = 1 | 𝑍 𝑘 = 𝑗),</formula><p>where 𝑖, 𝑗 ∈ 𝑔 represent different groups.</p><p>In real-world applications, achieving equal probabilities for certain outcomes is not always possible. Due to variations in sample sizes in the groups, it is common to yield unequal treatments, even when they are similar. Thus, existing literature <ref type="bibr" target="#b1">[2]</ref> uses the absolute difference to quantify the strength of discrimination.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Definition 2.3 (Disparity</head><formula xml:id="formula_3">). Let 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖) and 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑗) be two treatments, then we refer to 𝛿 𝑍 𝑘 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 ) = |𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖) − 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑗)|</formula><p>as the disparity, for all 𝑖, 𝑗 ∈ 𝑔 𝑘 . Trivially, 𝛿 𝑍 𝑘 is commutative regarding 𝑖, 𝑗. In practice, it prevents reverse discrimination due to the absolute value.</p><p>Definition 2.4 (Discrimination). We use 𝜓 : D → R to denote some discrimination measure that quantifies the discrimination inherent in any dataset 𝒟 ∈ D. A dataset 𝒟 consists of features, protected attributes, and labels for each individual. The explicit form of 𝜓 depends on the cases introduced in Section 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Measuring Discrimination for Multiple Attributes</head><p>We found that numerous scenarios arise when dealing with multiple protected attributes. We categorize these scenarios based on the number of groups, denoted as |𝑔|, and the number of protected attributes, denoted as |𝒵|. By going through all cases, we present possible approaches from the literature as well as our own suggestions to measure discrimination.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Single Protected Attribute (|𝒵| = 1)</head><p>In the case of having only one protected attribute, i.e., |𝒵| = |{𝑍 1 }| = 1, we distinguish between cases by the number of available groups |𝑔| in the dataset. We categorize the cases by |𝑔| = 0, 1, 2, and |𝑔| &gt; 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">No Groups (|𝑔| = 0)</head><p>When there are no groups, the measurement of discrimination is impossible if no assumptions are being made. Discrimination can be assessed through proxy variables <ref type="bibr" target="#b19">[20]</ref>; however, this approach can be imprecise and may introduce new biases. This case is equivalent to having no protected attribute, i.e., |𝒵| = 0.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Single Group (|𝑔| = 1)</head><p>Similarly to the case of having no groups, discrimination cannot be measured when having only one group. For this, we propose practices where prior information can be incorporated:</p><p>1. No discrimination: As no difference towards any other group can be measured, returning a discrimination score of 0 is one viable option.</p><formula xml:id="formula_4">𝜓(𝒟) = 0.<label>(1)</label></formula><p>2. Difference to optimal treatment: Another way is to return the absolute difference of the group's outcome to the optimal treatment. For example, group 𝑖 has an 80% chance of receiving the favorable treatment. Ideally, having a 100% chance would represent the optimal scenario. Therefore, the discrimination score is 20% in this case. It is given by:</p><formula xml:id="formula_5">𝜓(𝒟) = |𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑖) − 1|.<label>(2)</label></formula><p>3. Difference to expected treatment: We can use the expected treatment as a reference point. For example, we know that a company has a 50% acceptance rate for job applications. Now a machine learning classifier is trained to predict whether an applicant will be accepted and the model's predictions result in a 60% acceptance rate for group 𝑖. Hence, the model is positively biased towards group 𝑖 by 10%. This can be formulated as:</p><formula xml:id="formula_6">𝜓(𝒟) = |𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑖) − 𝑝 expect. |,<label>(3)</label></formula><p>where 𝑝 expect. is the expected treatment. It can describe the average treatment across all groups <ref type="bibr" target="#b20">[21]</ref> or some other prior information that is not included in the dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.3.">Binary Groups (|𝑔| = 2)</head><p>Without using any prior information, we can calculate the discrimination score by taking the absolute difference between the treatments of the two groups, as advised by Žliobaitė <ref type="bibr" target="#b1">[2]</ref>. The discrimination measure 𝜓 is then simply given by the disparity as mentioned in Definition 2.3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.4.">Non-binary Groups (|𝑔| &gt; 2)</head><p>While the case for binary attributes is straightforward, it becomes non-trivial for non-binary attributes that arise naturally in real-world data. We can fall back to |𝑔| = 2 by calculating the absolute difference between every distinct group 𝑖, 𝑗 ∈ 𝑔. Because the discrimination between 𝑖 and 𝑗 is the same as between 𝑗 and 𝑖, only</p><formula xml:id="formula_7">(︀ |𝑔|<label>2</label></formula><p>)︀ pairs need to be compared and we use an aggregation function agg (1) to report the differences <ref type="bibr" target="#b1">[2]</ref>. Lum et al. <ref type="bibr" target="#b21">[22]</ref> refers to measures that aggregate or summarize discrimination scores as meta-metrics. The aggregate can be the sum or maximum function, depending on the use case. The result for a single protected attribute 𝑍 𝑘 with two or more groups can be computed as follows:</p><formula xml:id="formula_8">𝜓(𝒟) = agg (1) 𝑖,𝑗∈𝑔 𝑘 ,𝑖&lt;𝑗 𝛿 𝑍 𝑘 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 ),<label>(4)</label></formula><p>where 𝛿 𝑍 𝑘 is the disparity as defined in Definition 2.3 and 𝑖 &lt; 𝑗 ensures that each pair is considered only once (assuming label-encoded groups). According to Žliobaitė <ref type="bibr" target="#b1">[2]</ref> and her personal discussions with legal experts, she advocates using the maximum function, i.e.,</p><formula xml:id="formula_9">𝜓(𝒟) = max 𝑖,𝑗∈𝑔 𝑘 ,𝑖&lt;𝑗 𝛿 𝑍 𝑘 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 )<label>(5)</label></formula><p>= max</p><formula xml:id="formula_10">𝑖∈𝑔 𝑘 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑖) − min 𝑗∈𝑔 𝑘 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 𝑘 = 𝑗).<label>(6)</label></formula><p>Equation ( <ref type="formula" target="#formula_9">5</ref>) describes the maximum discrimination obtainable between two groups. An alternative and equivalent formulation is given in Equation ( <ref type="formula" target="#formula_10">6</ref>) <ref type="bibr" target="#b6">[7]</ref>. The latter is computationally more efficient as it requires 𝒪(2|𝑔|) operations compared to 𝒪(|𝑔| 2 ) operations for the former.</p><p>A more general approach to measuring discrimination is to calculate some form of correlation coefficient between the protected attribute and the outcome. The correlation coefficient can be calculated using Pearson's correlation <ref type="bibr" target="#b22">[23]</ref>, Spearman or Kendall's rank correlation <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b24">25]</ref>. The discrimination measure can then be defined as the absolute value of the correlation coefficient:</p><formula xml:id="formula_11">𝜓(𝒟) = |Corr(𝐸 1 , 𝑍 𝑘 )|.<label>(7)</label></formula><p>This approach can be applied to any number of groups. Fairlearn provides a pre-processing method that removes the correlation between the protected attribute and the outcome by transforming the data <ref type="bibr" target="#b6">[7]</ref>. However, the given approach violates data integrity constraints as categorical attributes are transformed into continuous values. Moreover, zero correlation does not imply independence between two variables.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Multiple Protected Attributes (|𝒵| &gt; 1)</head><p>There are several ways to measure discrimination for multiple protected attributes (|𝒵| &gt; 1). Based on the works of Kearns et al. <ref type="bibr" target="#b20">[21]</ref>, Yang et al. <ref type="bibr" target="#b10">[11]</ref> and Kang et al. <ref type="bibr" target="#b12">[13]</ref>, we categorize them into two approaches: intersectional and non-intersectional (see Figure <ref type="figure">1</ref>). Intersectional approaches consider the intersection of identities. The overlapping of such identities forms subgroups <ref type="bibr" target="#b20">[21]</ref>. Non-intersectional approaches treat each protected attribute independently <ref type="bibr" target="#b10">[11]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Intersectional Discrimination</head><p>The central idea of intersectionality is that individuals experience overlapping forms of oppression or privilege based on the combination of multiple social categories they belong to. In the following, we will introduce definitions to formulate intersectional discrimination, which is based on the work of Kearns et al. <ref type="bibr" target="#b20">[21]</ref>. </p><formula xml:id="formula_12">𝛿 ^𝒵 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 ) = |𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑖 1 , . . . , 𝑍 𝑝 = 𝑖 𝑝 ) − 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑗 1 , . . . , 𝑍 𝑝 = 𝑗 𝑝 )|.</formula><p>Similarly to Equation (4), we can calculate the discrimination score for multiple protected attributes by aggregating disparities across all subgroups. A subgroup can be treated like a normal group. According to Definition 3.1, there are theoretically at least 2 𝑝 subgroups, where 𝑝 is the number of protected attributes. However, not all subgroups may be available in the dataset. For unavailable subgroups, the disparity cannot be calculated as the corresponding treatment is undefined. Let us denote the set of available subgroups as 𝐺 avail ⊆ 𝑔 1 × . . . × 𝑔 𝑘 . To finally capture the discrepancies across all available subgroup pairs, an aggregation function agg (1) is applied to the subgroup disparities 𝛿 ^𝒵 :</p><p>𝜓 intersect (𝒟) = agg (1)   𝑖,𝑗∈𝐺 avail 𝛿 ^𝒵 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 ). Equation ( <ref type="formula" target="#formula_13">8</ref>) represents the aggregated discrimination between all available subgroups in the dataset. When using the maximum function as the aggregator, the calculations are equivalent to Equation ( <ref type="formula" target="#formula_9">5</ref>) and Equation <ref type="bibr" target="#b5">(6)</ref>. The only difference is that the conditionals are now subgroups instead of groups:</p><formula xml:id="formula_14">𝜓 intersect (𝒟) = max 𝑖,𝑗∈𝐺 avail 𝛿 ^𝑍𝑘 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 )<label>(9)</label></formula><p>= max</p><formula xml:id="formula_15">𝑖∈𝐺 avail 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑖 1 , . . . , 𝑍 𝑝 = 𝑖 𝑝 ) − min 𝑗∈𝐺 avail 𝑃 (𝐸 1 | 𝐸 2 , 𝑍 1 = 𝑗 1 , . . . , 𝑍 𝑝 = 𝑗 𝑝 ).</formula><p>Kang et al. <ref type="bibr" target="#b12">[13]</ref> also dealt with intersectional discrimination in their work by introducing a multivariate random variable 𝑍 where each dimension represents a protected attribute. Their fairness objective is to minimize the mutual information between the outcome and the multivariate random variable. By minimizing the mutual information, the outcome is independent of the protected attributes, which is a desirable property for fairness <ref type="bibr" target="#b13">[14,</ref><ref type="bibr" target="#b25">26]</ref>. In this context, zero mutual information implies the absence of intersectional discrimination <ref type="bibr" target="#b12">[13]</ref>. However, this approach relies on expensive techniques to approximate the mutual information. Using our notation, their formulation can be written as <ref type="bibr" target="#b12">[13]</ref>:</p><formula xml:id="formula_16">𝜓 MI (𝒟) = MI(𝐸 1 , 𝑍),<label>(10)</label></formula><p>where MI denotes the mutual information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Non-intersectional Discrimination</head><p>The problem with measuring discrimination for intersectional groups is that it has an upward bias when using meta-metrics <ref type="bibr" target="#b21">[22]</ref>. This is because the number of subgroups grows exponentially with the number of protected attributes. This leads to many subgroups where the number of samples in each subgroup is possibly small, resulting in larger noise in the treatment estimates <ref type="bibr" target="#b21">[22]</ref>. Besides intersectional groups, Yang et al. <ref type="bibr" target="#b10">[11]</ref> listed a non-intersectional definition of groups, called independent groups. Building on the definition of independent groups, we propose an appropriate approach to measure discrimination for this type of groups. It is more suitable when dealing with a large number of subgroups or when intersectional discrimination is not deemed important. Our nonintersectional approach treats each protected attribute independently and aggregates the discrimination scores across all protected attributes. For this, a second aggregate function with agg (2) is introduced, yielding the following equation: (2)   𝑍 𝑘 ∈𝒵</p><formula xml:id="formula_17">𝜓 indep (𝒟) = agg</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>{︃</head><p>agg (1)   𝑖,𝑗∈𝑔 𝑘 ,𝑖&lt;𝑗</p><formula xml:id="formula_18">𝛿 𝑍 𝑘 (𝑖, 𝑗, 𝐸 1 , 𝐸 2 ) }︃ . (<label>11</label></formula><formula xml:id="formula_19">)</formula><p>The first-level aggregator agg (1) aggregates disparities within a protected attribute, considering unique pairs of groups 𝑖 and 𝑗. The second-level aggregator agg (2) then combines the results across all protected attributes. By applying both operators, we obtain a discrimination measure that captures disparities between groups across multiple attributes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Example</head><p>Let us consider a dataset with two protected attributes, age and sex (see Table <ref type="table" target="#tab_1">1</ref>). The set of protected attributes is 𝒵 = {𝑍 1 , 𝑍 2 } = {Age, Sex} and the set of available subgroups in the dataset is 𝐺 avail = {Old, Young} × {Male, Female}. We measure discrimination using statistical disparity. For simplicity, all aggregation functions are set to the maximum function. The intersectional approach yields the following discrimination score:</p><formula xml:id="formula_20">𝜓 intersect (𝒟) = max 𝑖,𝑗∈𝐺 avail 𝛿 ^𝒵 (𝑖, 𝑗, (𝑌 = 1), Ω)<label>(12)</label></formula><p>= max 𝑖,𝑗∈𝐺 avail 𝛿 ^{Age, Sex} (𝑖, 𝑗, (𝑌 = 1), Ω)</p><p>= max</p><formula xml:id="formula_21">𝑖∈𝐺 avail 𝑃 (𝑌 = 1 | 𝑍 1 = 𝑖 1 , 𝑍 2 = 𝑖 2 ) − min 𝑗∈𝐺 avail 𝑃 (𝑌 = 1 | 𝑍 1 = 𝑗 1 , 𝑍 2 = 𝑗 2 ) = |𝑃 (𝑌 = 1 | Age = Old, Sex = Male) − 𝑃 (𝑌 = 1 | Age = Young, Sex = Male)| = 1,</formula><p>while the discrimination score for the non-intersectional approach is given by:</p><formula xml:id="formula_22">𝜓 indep (𝒟) = max 𝑍 𝑘 ∈𝑍 {︂ max 𝑖,𝑗∈𝑔 𝑘 ,𝑖&lt;𝑗</formula><p>𝛿 𝑍 𝑘 (𝑖, 𝑗, (𝑌 = 1), Ω)</p><formula xml:id="formula_23">}︂ (<label>13</label></formula><formula xml:id="formula_24">)</formula><p>= max {︀ 𝛿 Age (Old, Young, (𝑌 = 1), Ω), 𝛿 Sex (Male, Female, (𝑌 = 1), Ω) }︀ = max{|0.5 − 0.5|, |0.5 − 0.5|} = max{0, 0} = 0.</p><p>The non-intersectional approach yields a discrimination score of 0 because the disparities for both protected attributes are 0. This is quite different from the intersectional approach, which reports a discrimination score of 1. As seen, the results can differ depending on the approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head><p>Our experimentation follows a pipeline consisting of data pre-processing, bias mitigation, model training, and evaluation. To mitigate bias in tabular datasets with multiple protected attributes, we used the sampling method, FairDo <ref type="bibr" target="#b8">[9]</ref>, that constructs fair datasets by selectively sampling data points. The method is very flexible and only requires the user to define the discrimination measure that should be minimized. In our case, we are interested in a dataset that has minimal bias across multiple protected attributes. The experiments revolve around the following research questions:</p><p>• RQ1 Is it possible to yield a fair dataset with FairDo, where bias for multiple protected attributes is reduced? • RQ2 Are machine learning models trained on fair datasets more fair in their predictions than those trained on original datasets?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Experimental Setup</head><p>Datasets and Pre-processing The tabular datasets employed in our experiments include the Adult <ref type="bibr" target="#b14">[15]</ref>, Bank <ref type="bibr" target="#b15">[16]</ref>, and COMPAS <ref type="bibr" target="#b16">[17]</ref> datasets. They are known for their use in fairness research and contain multiple protected attributes. We pre-processed the datasets by applying one-hot encoding to categorical variables and label encoding to protected attributes. Table <ref type="table" target="#tab_2">2</ref> shows important characteristics of the datasets after pre-processing. Each dataset was divided into training and testing sets using an 80/20 split, respectively. We ensured that the split was stratified (if possible) based on protected attributes to maintain representativeness across different groups in both sets. Bias Mitigation Applying the bias mitigation method FairDo <ref type="bibr" target="#b8">[9]</ref> to the datasets can be regarded as a pre-processing step, too. This is because the method simply returns a dataset that is fair with respect to the given discrimination measure. FairDo <ref type="bibr" target="#b8">[9]</ref> offers a variety of options to mitigate bias, and we chose the undersampling method that removes samples. In this option, the optimization objective is stated as <ref type="bibr" target="#b8">[9]</ref>: min</p><formula xml:id="formula_25">𝒟 fair ⊆𝒟 𝜓(𝒟 fair ),<label>(14)</label></formula><p>where 𝒟 is the training set of Adult, Bank, or COMPAS, and 𝜓 is the fairness objective function. We experimented with both 𝜓 intersect and 𝜓 indep as objectives functions. Bias mitigation is only applied to the training set and the testing set remains unchanged. FairDo internally uses genetic algorithms to select a subset of the training set that minimizes the objective function. We used the same settings and operators as provided in the package and only adjusted the population size (200) and the number of generations (400).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model Training</head><p>We utilized the scikit-learn library <ref type="bibr" target="#b26">[27]</ref> to train various machine learning classifiers, namely Logistic Regression (LR), Support Vector Machine (SVM), Random Forest (RF), and Artificial Neural Network (ANN). These classifiers were trained on both the original and fair datasets. Classifiers trained on the original datasets serve as a baseline for comparison. We used the default hyperparameters given by scikit-learn package for each classifier.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Evaluation Metrics</head><p>We evaluated the models' predictions on fairness and performance using the test set. For fairness, we assessed 𝜓 intersect and 𝜓 indep . For the classifiers' performances, we report the area under the receiver operating characteristic curve (AUROC) <ref type="bibr" target="#b27">[28]</ref>, where higher values indicate better performances. Because removing data points can compromise the overall quality of the data, we also report the number of subgroups before and after bias mitigation to check for representativeness.</p><p>Trials For each dataset and discrimination measure combination, the bias mitigation process was repeated 10 times. The results were averaged over the trials to obtain a more robust evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head><p>Fair Dataset Generation Table <ref type="table" target="#tab_3">3</ref> shows the average discrimination before and after mitigating bias in the training sets. On all datasets, discrimination was reduced after applying FairDo. Without considering group intersections, discrimination was reduced by 7%, 19%, and 25% for Adult, Bank, and COMPAS, respectively. When considering intersectionality, the discrimination was reduced by 15%, 18%, and 83%. Hence, discrimination was reduced by 28% on average across all datasets, thus answering RQ1 positively. When comparing the discrimination scores, it can be observed that the intersectional discrimination scores are generally higher. This is because in the intersectional setting, more subgroups are considered, which potentially leads to larger differences between them <ref type="bibr" target="#b20">[21]</ref>.</p><p>We also report the number of subgroups before and after bias mitigation to assess the impact of the undersampling method on the dataset. The removal of subgroups can only be observed in the intersectional setting. In the COMPAS dataset 5.2 out of 34 subgroups were removed on average, indicating the largest amount of subgroups removed across all datasets. While the Bank dataset consists of 48 subgroups, only 1.8 subgroups were removed on average. Because the COMPAS dataset's initial intersectional discrimination score is 100%, removing more subgroups seems inevitable to reduce bias.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model Performance and Fairness</head><p>Figure <ref type="figure" target="#fig_0">2</ref> shows the results of the classifiers' performances on the test set. The classifiers' performances are displayed on the y-axis, while the discrimination values are shown on the x-axis. We note that the axes do not share the same scale across the subfigures for analytical purposes.</p><p>Classifiers trained on fair datasets did not suffer a significant decline in performance compared to those trained on original datasets. In all cases, only a slight decrease of 1%-3% in performance can be noted. This indicates that the bias mitigation process does not compromise the dataset's fidelity and, therefore, the classifiers' performances. Regarding discrimination, a significant reduction is evident. The x-axis scales are much larger than the y-axis scales, suggesting that changes in discrimination are larger than changes in performance. For example, the RF classifier trained on the Bank dataset (Figure <ref type="figure" target="#fig_0">2g</ref>) shows a decrease in intersectional discrimination from 38% to 15%, while the performance only decreases by 2%. Similar results can be observed for the other classifiers and datasets as well, successfully addressing RQ2. The results suggest that FairDo can be reliably used to mitigate bias in tabular datasets for various measures that consider multiple protected attributes. Still, we advise users to carefully perform similar analyses when applying the method to their datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussion</head><p>The results of our experiments show that the presented measures detect discrimination in datasets with multiple protected attributes differently. When using the intersectional discrimination measure, more groups are identified and compared to each other. While subgroups are not ignored by this measure, measuring higher discrimination scores by random chance becomes more likely <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref>. In contrast, treating each protected attribute separately prevents this issue but may lead to overlooking discrimination. The choice of measure is up to the stakeholders and depends on the context of the dataset and the regulations that apply to the AI system. We generally recommend using the intersectional discrimination measure if the number of individuals in each subgroup is large enough to draw statistically significant conclusions. Otherwise, treating each protected attribute separately is more suitable.</p><p>By using the mitigation strategy FairDo <ref type="bibr" target="#b8">[9]</ref>, the resulting datasets in the experiments have improved statistical properties regarding fairness. Whether intersectionality was considered or not, reducing discrimination in datasets was possible. At the current state, the AI Act <ref type="bibr" target="#b0">[1]</ref> does not explicitly mention intersectional discrimination nor how to deal with multiple protected attributes generally. While recital (67) states that datasets "should [...] have the appropriate statistical properties", it does not specify what these properties are. Hence, our work serves as an initial guideline for what these properties could be and how to achieve them in practice.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>Datasets often come with multiple protected attributes, which makes measuring and mitigating discrimination more challenging. Most existing studies only deal with a single protected attribute, and works that consider multiple protected attributes often focus on intersectionality. In opposition to this, we proposed a new non-intersectional measure that treats each protected attribute separately. This is more suitable when the number of subgroups is too large or the number of individuals in each subgroup is small. We used both intersectional and non-intersectional measures as objectives and applied the FairDo framework to mitigate discrimination in multiple datasets. The experiments show that discrimination was reduced in all datasets and on average by 28%. Machine learning models trained on the bias-mitigated datasets also improved their fairness while maintaining performance compared to models trained on the original datasets.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Results on the test set. The x-axis represents the discrimination values (legend indicates used measure) and the y-axis represents the classifiers' performances. We compare the pre-processed (fair) data with the original data. The points/stars represent averages, and the error bars display the standard deviations of the AUROC and discrimination values over 10 trials.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Example dataset of individuals receiving a favorable (𝑌 = 1) or unfavorable (𝑌 = 0) outcome. The dataset shows four individuals with their respective age group and sex.</figDesc><table><row><cell cols="2">Individual Age</cell><cell>Sex</cell><cell>Outcome (𝑌 )</cell></row><row><cell>1</cell><cell>Old</cell><cell>Male</cell><cell>1</cell></row><row><cell>2</cell><cell>Old</cell><cell>Female</cell><cell>0</cell></row><row><cell>3</cell><cell cols="2">Young Male</cell><cell>0</cell></row><row><cell>4</cell><cell cols="2">Young Female</cell><cell>1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Overview of Datasets</figDesc><table><row><cell>Dataset</cell><cell cols="2">Samples Feats. Label</cell><cell cols="3">Protected Attributes</cell><cell>Description</cell></row><row><cell>Adult [15]</cell><cell>32 561</cell><cell>21 Income</cell><cell cols="3">Race: White, Black, Asian-</cell><cell>Indicates individuals</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Pacific-Islander,</cell><cell>American-</cell><cell>earning over $50,000</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="3">Indian-Eskimo, Other</cell><cell>annually</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Sex: Male, Female</cell><cell></cell></row><row><cell>Bank [16]</cell><cell>41 188</cell><cell>50 Term</cell><cell>Job:</cell><cell cols="2">Admin, Blue-Collar,</cell><cell>Shows whether the</cell></row><row><cell></cell><cell></cell><cell>deposit</cell><cell cols="3">Technician, Services, Manage-</cell><cell>client has subscribed</cell></row><row><cell></cell><cell></cell><cell>subscription</cell><cell cols="3">ment, Retired, Entrepreneur,</cell><cell>to a term deposit.</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Self-Employed,</cell><cell>Housemaid,</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="3">Unemployed, Student, Unknown</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Marital Status:</cell><cell>Divorced,</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="3">Married, Single, Unknown</cell></row><row><cell>COMPAS [17]</cell><cell>7 214</cell><cell>13 2-year</cell><cell cols="3">Race: African-American, Cau-</cell><cell>Displays individuals</cell></row><row><cell></cell><cell></cell><cell>recidivism</cell><cell cols="3">casian, Hispanic, Other, Asian,</cell><cell>that were rearrested</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Native American</cell><cell></cell><cell>for a new crime</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="2">Sex: Male, Female</cell><cell></cell><cell>within 2 years after</cell></row><row><cell></cell><cell></cell><cell></cell><cell cols="3">Age Category: &lt;25, 25-45, &gt;45</cell><cell>initial arrest.</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Average discrimination and number of subgroups before and after pre-processing the training sets with FairDo.</figDesc><table><row><cell>Dataset</cell><cell>Metric</cell><cell cols="4">Disc. Before Disc. After Subgroups Before Subgroups After</cell></row><row><cell>Adult</cell><cell>𝜓 indep</cell><cell>20%</cell><cell>13%</cell><cell>10</cell><cell>10</cell></row><row><cell></cell><cell>𝜓 intersect</cell><cell>31%</cell><cell>16%</cell><cell>10</cell><cell>10</cell></row><row><cell>Bank</cell><cell>𝜓 indep</cell><cell>24%</cell><cell>5%</cell><cell>48</cell><cell>48</cell></row><row><cell></cell><cell>𝜓 intersect</cell><cell>33%</cell><cell>15%</cell><cell>48</cell><cell>46.2</cell></row><row><cell cols="2">COMPAS 𝜓 indep</cell><cell>30%</cell><cell>5%</cell><cell>34</cell><cell>34</cell></row><row><cell></cell><cell>𝜓 intersect</cell><cell>100%</cell><cell>17%</cell><cell>34</cell><cell>28.8</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Original ψ multi.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fair ψ multi.</head><p>Original ψ intersect. Fair ψ intersect. </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<ptr target="Ac-cessed:17" />
		<title level="m">Artificial Intelligence Act, Corrigendum</title>
				<imprint>
			<date type="published" when="2024-04-19">19 April 2024. 2024. May 2024</date>
		</imprint>
	</monogr>
	<note>European Commission</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Measuring discrimination in algorithmic decision making</title>
		<author>
			<persName><forename type="first">I</forename><surname>Žliobaitė</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Data Mining and Knowledge Discovery</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="1060" to="1089" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Fairness beyond disparate treatment &amp; disparate impact: Learning classification without disparate mistreatment</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">B</forename><surname>Zafar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Valera</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Gomez</forename><surname>Rodriguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">P</forename><surname>Gummadi</surname></persName>
		</author>
		<idno type="DOI">10.1145/3038912.3052660</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 26th International Conference on World Wide Web</title>
				<meeting>the 26th International Conference on World Wide Web</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Algorithmic decision making and the cost of fairness</title>
		<author>
			<persName><forename type="first">S</forename><surname>Corbett-Davies</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pierson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Feller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Goel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Huq</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="797" to="806" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Fairness and Machine Learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Barocas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Narayanan</surname></persName>
		</author>
		<ptr target="org" />
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>fairmlbook</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Certifying and removing disparate impact</title>
		<author>
			<persName><forename type="first">M</forename><surname>Feldman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Friedler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Moeller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Scheidegger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Venkatasubramanian</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining</title>
				<meeting>the 21th ACM SIGKDD international conference on knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="259" to="268" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Fairlearn: A toolkit for assessing and improving fairness in AI</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bird</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dudík</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Edgar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Horn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Milan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sameki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wallach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Walker</surname></persName>
		</author>
		<idno>MSR- TR-2020-32</idno>
		<ptr target="https://www.microsoft.com/en-us/research/publication/fairlearn-a-toolkit-for-assessing-and-improving-fairness-in-ai/" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
		<respStmt>
			<orgName>Microsoft</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A reductions approach to fair classification</title>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Beygelzimer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dudík</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Langford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wallach</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="60" to="69" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Towards fairness and privacy: A novel data pre-processing optimization framework for non-binary protected attributes</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">K</forename><surname>Duong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Conrad</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Data Science and Machine Learning</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Benavides-Prado</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Erfani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Fournier-Viger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><forename type="middle">L</forename><surname>Boo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Koh</surname></persName>
		</editor>
		<meeting><address><addrLine>Singapore, Singapore</addrLine></address></meeting>
		<imprint>
			<publisher>Springer Nature</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="105" to="120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Bayesian Modeling of Intersectional Fairness: The Variance of Bias</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Foulds</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Islam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">N</forename><surname>Keya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Pan</surname></persName>
		</author>
		<idno type="DOI">10.1137/1.9781611976236.48</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="424" to="432" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Fairness with overlapping groups</title>
		<author>
			<persName><forename type="first">F</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cisse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Koyejo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 34th International Conference on Neural Information Processing Systems, NIPS &apos;20</title>
				<meeting>the 34th International Conference on Neural Information Processing Systems, NIPS &apos;20<address><addrLine>Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Fair classification with noisy protected attributes: A framework with provable guarantees</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Celis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Keswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">K</forename><surname>Vishnoi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th International Conference on Machine Learning</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Meila</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</editor>
		<meeting>the 38th International Conference on Machine Learning<address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">139</biblScope>
			<biblScope unit="page" from="1349" to="1361" />
		</imprint>
	</monogr>
	<note>Proceedings of Machine Learning Research</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Infofair: Information-theoretic intersectional fairness</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Maciejewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Tong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Big Data (Big Data)</title>
				<imprint>
			<date type="published" when="2021">2022. 2021</date>
			<biblScope unit="page" from="1455" to="1464" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Learning fair representations</title>
		<author>
			<persName><forename type="first">R</forename><surname>Zemel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Swersky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pitassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Dwork</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<imprint>
			<publisher>PMLR</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="325" to="333" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Scaling up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid</title>
		<author>
			<persName><forename type="first">R</forename><surname>Kohavi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">KDD&apos;96</title>
				<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="1996">1996</date>
			<biblScope unit="page" from="202" to="207" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">A data-driven approach to predict the success of bank telemarketing</title>
		<author>
			<persName><forename type="first">S</forename><surname>Moro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cortez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rita</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Decision Support Systems</title>
		<imprint>
			<biblScope unit="volume">62</biblScope>
			<biblScope unit="page" from="22" to="31" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Machine bias</title>
		<author>
			<persName><forename type="first">J</forename><surname>Larson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Angwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mattu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kirchner</surname></persName>
		</author>
		<ptr target="https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing" />
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Building classifiers with independency constraints</title>
		<author>
			<persName><forename type="first">T</forename><surname>Calders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Kamiran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pechenizkiy</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICDMW.2009.83</idno>
	</analytic>
	<monogr>
		<title level="m">2009 IEEE International Conference on Data Mining Workshops</title>
				<imprint>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="13" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Equality of opportunity in supervised learning</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Price</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Srebro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A survey on bias and fairness in machine learning</title>
		<author>
			<persName><forename type="first">N</forename><surname>Mehrabi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Morstatter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Saxena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lerman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Galstyan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys (CSUR)</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="1" to="35" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Preventing fairness gerrymandering: Auditing and learning for subgroup fairness</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kearns</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Neel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">S</forename><surname>Wu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 35th International Conference on Machine Learning</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Dy</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Krause</surname></persName>
		</editor>
		<meeting>the 35th International Conference on Machine Learning<address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">80</biblScope>
			<biblScope unit="page" from="2564" to="2572" />
		</imprint>
	</monogr>
	<note>Proceedings of Machine Learning Research</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">De-biasing &quot;bias&quot; measurement</title>
		<author>
			<persName><forename type="first">K</forename><surname>Lum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bower</surname></persName>
		</author>
		<idno type="DOI">10.1145/3531146.3533105</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT &apos;22</title>
				<meeting>the 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT &apos;22<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="379" to="389" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Notes on regression and inheritance in the case of two parents</title>
		<author>
			<persName><forename type="first">K</forename><surname>Pearson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the Royal Society of London</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="240" to="242" />
			<date type="published" when="1895">1895</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">The proof and measurement of association between two things</title>
		<author>
			<persName><forename type="first">C</forename><surname>Spearman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">American Journal of Psychology</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="72" to="101" />
			<date type="published" when="1904">1904</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">A new measure of rank correlation</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">G</forename><surname>Kendall</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Biometrika</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="81" to="93" />
			<date type="published" when="1938">1938</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Fairness in supervised learning: An information theoretic approach</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ghassami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Khodadadian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kiyavash</surname></persName>
		</author>
		<idno type="DOI">10.1109/ISIT.2018.8437807</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE International Symposium on Information Theory (ISIT)</title>
				<imprint>
			<publisher>IEEE Press</publisher>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="176" to="180" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Scikit-learn: Machine learning in Python</title>
		<author>
			<persName><forename type="first">F</forename><surname>Pedregosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Varoquaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gramfort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Thirion</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Grisel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Blondel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Prettenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Weiss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dubourg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Vanderplas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Cournapeau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brucher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Perrot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Duchesnay</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="2825" to="2830" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">An introduction to ROC analysis</title>
		<author>
			<persName><forename type="first">T</forename><surname>Fawcett</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.patrec.2005.10.010</idno>
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition Letters</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="861" to="874" />
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
