<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Exploration of Potential New Benchmark for Fairness Evaluation in Europe</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Magali</forename><surname>Legast</surname></persName>
							<email>magali.legast@uclouvain.be</email>
							<affiliation key="aff0">
								<orgName type="institution">Université catholique de Louvain</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Lisa</forename><surname>Koutsoviti-Koumeri</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Hasselt University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yasaman</forename><surname>Yousefi</surname></persName>
							<email>yasaman.yousefi3@unibo.it</email>
							<affiliation key="aff2">
								<orgName type="institution">Università di Bologna</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Axel</forename><surname>Legay</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Université catholique de Louvain</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<settlement>Mainz</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Exploration of Potential New Benchmark for Fairness Evaluation in Europe</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6146658D0A42EFC5982092A02A0BD248</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Fairness</term>
					<term>Datasets</term>
					<term>Machine learning</term>
					<term>Bias mitigation</term>
					<term>Classification</term>
					<term>Fair Classification</term>
					<term>Benchmark</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>With the increase use of artificial intelligence systems and the associated concerns regarding automated discrimination, research in the field of fairness has increased in the past years. To evaluate their work in fair machine learning, researchers have often been using the same three datasets (Adult, COMPAS, and German credit) as benchmarks. However, those datasets each present serious limitations. In this work, we first explore what other datasets could potentially be used as replacement, specifically in a European context. We then use an experimental approach to compare Adult and COMPAS with a new candidate, Student Performance (a.k.a Student Alcohol Consumption). Our early results highlight the scarcity of easily accessible European datasets suitable as benchmarks for fairness evaluation of problems with positive or negative outcome, as well as the high influence dataset selection can have on experimental results.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>With the increase use of artificial intelligence systems, legitimate ethical and legal concerns have been growing, including the risk that some people may be treated more negatively than others, thus resulting in discrimination <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b20">21,</ref><ref type="bibr" target="#b21">22]</ref>. This problem is prevalent in machine learning, where the training data is of paramount importance, while usually retaining historical and social biases that are then learned by the prediction models <ref type="bibr" target="#b20">[21]</ref>.</p><p>Research in the field of fairness has been constantly growing in the past few years, with the problem of fair classification receiving the most attention <ref type="bibr" target="#b20">[21]</ref>. Many fairness metrics and bias mitigation methods have been developed in that sub-field, but less attention has been given towards quality benchmark datasets, specially regarding European data. A subset of only three datasets has been surpassing all others in term of popularity <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b16">17]</ref>, namely Adult <ref type="bibr" target="#b2">[3]</ref>, COMPAS <ref type="bibr" target="#b9">[10]</ref>, and German credit <ref type="bibr" target="#b15">[16]</ref>. Those popular datasets have nevertheless been shown to present serious limitations <ref type="bibr" target="#b11">[12]</ref>, such as, but not limited to, old age (Adult and German credit), noisy data <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b13">14]</ref>, label bias (COMPAS <ref type="bibr" target="#b0">[1]</ref>), or even coding mistakes preventing retrieval of the sensitive attribute (German credit <ref type="bibr" target="#b13">[14]</ref>).</p><p>Given these observations, the emerging consensus is that their use as benchmarks should be avoided, lest duly justified. Efforts have been made to provide alternatives (for instance a dataset suite to replace Adult <ref type="bibr" target="#b8">[9]</ref>), to disseminate good practices (like in <ref type="bibr" target="#b0">[1]</ref>), or to facilitate access to other datasets (such as a large collection of fairness datasets <ref type="bibr" target="#b11">[12]</ref> along with its search engine <ref type="bibr" target="#b12">[13]</ref>). Building on these works, we explore potential alternatives to the three popular yet flawed historical benchmarks, while focusing on the European context. We also compare, with an experimental approach, results stemming from the use of Adult and COMPAS as benchmarks with a potential replacement, Student Performance <ref type="bibr" target="#b6">[7]</ref>. More specifically, we compare the results of a bias mitigation method and several popular fairness metrics for models trained with each of those three datasets.</p><p>Our results highlight the scarcity of open access and easily accessible datasets for fairness evaluation in a European context. Further, the analysis of those three examples confirms that different datasets may lead to different results when evaluating bias mitigation methods and fairness metrics. This stresses the importance of dataset selection in experiments. We also aim to expand this work with a more thorough search for datasets and further experiments encompassing more bias mitigation methods, fairness metrics and sensitive attributes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">European Datasets</head><p>As mentioned in the introduction, the datasets Adult, COMPAS and German credit have been widely used as benchmarks in fair classification studies, even though they are not as suitable for fair machine learning research as previously thought. All three of them contain real life tabular data about individuals with one or more attribute(s) recognized as sensitive/protected and a target label that is deemed positive or negative for the individual.</p><p>Those key characteristics for fair classification should be shared by potential replacement datasets. Being open access also increased their appeal. We thus aim to find a potential new benchmark that is also easily accessible, specially considering the fast past environment of computer science conferences where there is often little time dedicated to the selection of datasets. Lastly, we focus on the European context. Indeed, the worldwide influence of the United-States and difference in data privacy culture and legislation across countries have made datasets from the USA dominant in the field, while data from other areas, including Europe, remains less accessible. This can cause further bias given that models and results from a certain place aren't necessary applicable to other geographical locations <ref type="bibr" target="#b8">[9]</ref>.</p><p>Merging the positive characteristics of the popular datasets with our context of interest, we formulate our dataset selection criteria as follows: An open access dataset with tabular data about European subjects that is no more than 25 years old and is adapted to the problem of fair classification leading either to a positive or negative outcome for the subject.</p><p>To find such dataset, we used the search engine for fairness datasets<ref type="foot" target="#foot_0">1</ref> presented in <ref type="bibr" target="#b12">[13]</ref> as it is the most complete collection of datasets for fair machine learning to the best of our knowledge. Its database comprises over 200 datasets for diverse domains and fairness tasks.</p><p>Filtering on tabular data and fair classification with a positive or negative outcome, we are left with 22 datasets (as of January 29 and February 2, 2024). Out of those, fourteen datasets contain data from the United-States, two from elsewhere in America, three have no mention of localization in their description, three are European, and only one is from Asia. There are thus no dataset referenced for Africa nor Oceania. Out of the three European datasets, only two contain data collected in this century, Dutch Census <ref type="bibr" target="#b5">[6]</ref> and Student Performance <ref type="bibr" target="#b6">[7]</ref> (a.k.a "Student" and "Student Alcohol Consumption"). Since Dutch Census is part of the IPUMS International collection <ref type="foot" target="#foot_1">2</ref> and requires approval to be accessed, only Student Performance fully fits our criteria. Presented in <ref type="bibr" target="#b7">[8]</ref>, this dataset contains social, gender and study information about students in two Portuguese schools for the core classes of Mathematics and Portuguese in secondary education (high school). It has been used in a few fairness studies and is referenced in the dataset survey <ref type="bibr" target="#b18">[19]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experiment</head><p>With this experiment <ref type="foot" target="#foot_2">3</ref> , we compare different models trained with bias mitigation on Adult, COMPAS, and Student Performance. For Student Performance, we use data related to the Portuguese subject version of the dataset since it has the most instances (649). We consider the sensitive attributes sex and age, with students who are 18 or older as the protected group, as in <ref type="bibr" target="#b18">[19]</ref>. We take the usual sensitive attributes sex and race for Adult and COMPAS. We did not include German credit, as it is impossible to retrieve its protected attribute (sex), making interpretation of results misleading.</p><p>For each of the considered datasets, we compute several classifiers using the training and in-processing bias mitigation meta-algorithm presented in <ref type="bibr" target="#b4">[5]</ref>. This algorithm considers a constrained optimization problem that is approximately solved with provable guarantees. The constraint enforces a minimal value, the fairness penalty parameter 𝜏 , for a chosen fairness metric. We use the AIF360 <ref type="bibr" target="#b3">[4]</ref> implementation with Statistical Parity ratio as the fairness constraint and gradient descent. We train different models with 𝜏 value ranging from 0 (no bias mitigation) to 1 (constraint of perfect statistical parity). We then evaluate the performance and fairness of the resulting models using different metrics to assess their evolution with different constraint levels. The fairness metrics we use are Statistical Rate difference (SR) <ref type="bibr" target="#b10">[11]</ref>, the most used metric and based on prediction only, Equality of Opportunity (Eq. Opp) and Equalized Odds (Eq. Odds) <ref type="bibr" target="#b14">[15]</ref>, the most used metrics based both on predictions and ground truth, and Consistency <ref type="bibr" target="#b23">[24]</ref>, the most used metric based on similarity <ref type="bibr" target="#b22">[23,</ref><ref type="bibr" target="#b16">17]</ref>. We evaluate each metric for each classifier, then report the average over 10 folds and the corresponding confidence interval.</p><p>You can see in Figure <ref type="figure" target="#fig_0">1</ref> the results for the different models, each represented by the fairness penalty parameter 𝜏 it was trained with. Results for Adult and COMPAS with sensitive attribute race are close to those with sex and are not presented here due to space restriction.</p><p>Let us first note that Consistency, an individual fairness metric, is not impacted by the Statistical Parity mitigation, which is based on group fairness. We thus focus only on the three  other metrics, related to group fairness, for the remainder of this section.</p><p>In Figure <ref type="figure" target="#fig_0">1a</ref>, we see that the model trained on Adult without bias mitigation (𝜏 = 0) is already very close to fairness according to all metrics. Bias mitigation still induces a general improvement for all group fairness metrics. Eq. Opp indicates that the protected group (Women) becomes the one with a slightly higher true positive rate when the fairness constraint is greater than or equal to 0.7 . We also note a significant drop in accuracy and increase of the F1-score as soon as bias mitigation is introduced, but only marginal changes after that.</p><p>In Figure <ref type="figure" target="#fig_0">1b</ref>, for models trained on COMPAS, there is first an overall decrease in fairness for all group metrics from 𝜏 = 0.1 to 𝜏 = 0.7. Fairness then increases again for higher values of 𝜏 , which is coincident with a significant drop in accuracy, while the F1-score remains steady.</p><p>In Figure <ref type="figure" target="#fig_0">1c</ref>, the unconstrained model trained on Student Performance with sex as protected attribute is near perfect fairness according to SR and Eq. Opp. The constrained models never surpass these values. Eq. Odds shows more bias than both SR and Eq. Opp for all models but one. This metric is less mitigated than for Adult, even though the original value was higher to start with.</p><p>In Figure <ref type="figure" target="#fig_0">1d</ref>, the model trained on Student Performance with age as sensitive attribute and no bias mitigation shows significant bias according to SR. This bias is efficiently mitigated by the model, even with the lowest level of fairness constraint. Bias reported by SR and Eq. Opp is very low even before bias mitigation. Except for a brief increase when 𝜏 is 0.1, their bias level is reduced, with Eq. Opp extremely close to equality.</p><p>Overall, even though the same algorithm as been used for the training and bias mitigation of all of these models, the results vary significantly for each of the different datasets studied. Additionally, for Adult and COMPAS the overall tendencies are similar for the two sensitive attributes considered (sex and race), but we see a very significant difference when considering sex or age for Student Performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Discussion</head><p>On the one hand, the search for new potential benchmarks highlights the scarcity of European datasets for use in fair classification with a positive or negative outcome. Indeed, out of the over 200 datasets referenced in the search engine used <ref type="bibr" target="#b12">[13]</ref>, only Student Performance <ref type="bibr" target="#b6">[7]</ref> fits our selection criteria. This dataset shows very little to no bias with regard to the sensitive attribute sex, at least regarding the most common fairness definitions, strongly reducing its interest when studying fairness related problems. The other most common sensitive attribute considered for this dataset is age. However, being an older student is most often a direct result of past failures, which is itself usually deemed an appropriate criterion to predict future exam results. It is thus questionable whether this attribute should be considered protected or not. Other attributes we have not explored here could also be of interest. For example, attributes related to alcohol consumption are studied in <ref type="bibr" target="#b19">[20]</ref>, which also extends the discussion to label bias.</p><p>So, despite the existing efforts to mitigate the collective data documentation debt and offer new alternatives to Adult, COMPAS and German credit, there is still a need to bring forward new European datasets, as well as data from other underrepresented continents. This may include improvement of the visibility and centralization of existing datasets or collection of new data adapted to fairness related questions.</p><p>On the other hand, our study illustrates that the same procedure applied to different datasets may lead to significantly different results, which is congruent with the results in <ref type="bibr" target="#b18">[19]</ref>. This showcases the importance the choice of dataset can have on fairness evaluation and when presenting results. We thus recommend to use several different datasets since they may lead to varying results, to look beyond open access data if needed, and to avoid making broad claims based on few examples, which echoes some of the recommendations in <ref type="bibr" target="#b0">[1]</ref> and <ref type="bibr" target="#b18">[19]</ref>. We also encourage researchers to consider the selection of data not as a minor step, but as a meaningful part of the research, and to provide justification on the choices made in that regard.</p><p>Beyond this discussion, we aim to expand this work to study more datasets, include non-open access data, as well as more bias mitigation methods, fairness metrics and sensitive attributes.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Fairness and performance results for classifiers trained with different levels of fairness constraint</figDesc></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">This search engine is available at http://fairnessdata.dei.unipd.it/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">See Harmonized International Census Data for Social Science and Health Research https://www.ipums.org/projects/ ipums-international</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">The full code of the experiment is available at https://github.com/Magalii/AIF360/tree/EWAF2024</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">It&apos;s COMPASlicated: The messy relationship between RAI datasets and algorithmic fairness benchmarks</title>
		<author>
			<persName><forename type="first">Michelle</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Angela</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Samantha</forename><surname>Zottola</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Brian</forename><surname>Brubach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sarah</forename><surname>Desmarais</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aaron</forename><surname>Horowitz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><surname>Lum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Suresh</forename><surname>Venkatasubramanian</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2106.05498" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">ArXiv</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Propublica&apos;s compas data revisited</title>
		<author>
			<persName><forename type="first">Matias</forename><surname>Barenstein</surname></persName>
		</author>
		<idno>ArXiv</idno>
		<ptr target="http://arxiv.org/abs/1906.04711" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Adult. UCI Machine Learning Repository</title>
		<author>
			<persName><forename type="first">Barry</forename><surname>Becker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ronny</forename><surname>Kohavi</surname></persName>
		</author>
		<idno type="DOI">10.24432/C5XW20</idno>
		<ptr target="https://doi.org/10.24432/C5XW20" />
		<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">E</forename><surname>Rachel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kuntal</forename><surname>Bellamy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><surname>Hind</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Samuel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stephanie</forename><surname>Hoffman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kalapriya</forename><surname>Houde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pranay</forename><surname>Kannan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jacquelyn</forename><surname>Lohia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sameep</forename><surname>Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aleksandra</forename><surname>Mehta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Seema</forename><surname>Mojsilovic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Karthikeyan</forename><surname>Nagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><surname>Natesan Ramamurthy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Diptikalyan</forename><surname>Richards</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Prasanna</forename><surname>Saha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Moninder</forename><surname>Sattigeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kush</forename><forename type="middle">R</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yunfeng</forename><surname>Varshney</surname></persName>
		</author>
		<author>
			<persName><surname>Zhang</surname></persName>
		</author>
		<idno>ArXiv</idno>
		<ptr target="https://arxiv.org/abs/1810.01943" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Classification with fairness constraints: A meta-algorithm with provable guarantees</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">Elisa</forename><surname>Celis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lingxiao</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vijay</forename><surname>Keswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nisheeth</forename><forename type="middle">K</forename><surname>Vishnoi</surname></persName>
		</author>
		<idno type="DOI">10.1145/3287560.3287586</idno>
		<ptr target="https://doi.org/10.1145/3287560.3287586" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the conference on fairness, accountability, and transparency</title>
				<meeting>the conference on fairness, accountability, and transparency</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="319" to="328" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<idno type="DOI">10.18128/D020.V6.4</idno>
		<ptr target="http://doi.org/10.18128/D020.V6.4" />
		<title level="m">Integrated public use microdata series, international: Version 6.4 2001 dutch census</title>
				<meeting><address><addrLine>Minneapolis</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
		<respStmt>
			<orgName>Minnesota Population Center ; University of Minnesota</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Student performance. UCI Machine Learning Repository</title>
		<author>
			<persName><forename type="first">Paulo</forename><surname>Cortez</surname></persName>
		</author>
		<ptr target="https://archive.ics.uci.edu/dataset/320" />
		<imprint>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Using data mining to predict secondary school student performance</title>
		<author>
			<persName><forename type="first">Paulo</forename><surname>Cortez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alice</forename><surname>Maria</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gonçalves</forename><surname>Silva</surname></persName>
		</author>
		<ptr target="https://repositorium.sdum.uminho.pt/handle/1822/8024" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of 5th Annual Future Business Technology Conference</title>
				<meeting>5th Annual Future Business Technology Conference<address><addrLine>Porto</addrLine></address></meeting>
		<imprint>
			<publisher>EUROSIS-ETI</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="5" to="12" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Retiring adult: New datasets for fair machine learning</title>
		<author>
			<persName><forename type="first">Frances</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Moritz</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ludwig</forename><surname>Schmidt</surname></persName>
		</author>
		<idno>ArXiv</idno>
		<ptr target="http://arxiv.org/abs/2108.04884" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The accuracy, fairness, and limits of predicting recidivism</title>
		<author>
			<persName><forename type="first">Julia</forename><surname>Dressel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hany</forename><surname>Farid</surname></persName>
		</author>
		<idno type="DOI">10.1126/sciadv.aao5580</idno>
		<ptr target="https://www.science.org/doi/10.1126/sciadv.aao5580" />
	</analytic>
	<monogr>
		<title level="j">Science Advances</title>
		<imprint>
			<biblScope unit="volume">4</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Fairness through awareness</title>
		<author>
			<persName><forename type="first">Cynthia</forename><surname>Dwork</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Moritz</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Toniann</forename><surname>Pitassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Omer</forename><surname>Reingold</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Richard</forename><surname>Zemel</surname></persName>
		</author>
		<idno type="DOI">10.1145/2090236.2090255</idno>
		<ptr target="https://doi.org/10.1145/2090236.2090255" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, ITCS &apos;12</title>
				<meeting>the 3rd Innovations in Theoretical Computer Science Conference, ITCS &apos;12</meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="214" to="226" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Algorithmic fairness datasets: the story so far</title>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Fabris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Stefano</forename><surname>Messina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gianmaria</forename><surname>Silvello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gian</forename><forename type="middle">Antonio</forename><surname>Susto</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10618-022-00854-z</idno>
		<ptr target="https://link.springer.com/10.1007/s10618-022-00854-z" />
	</analytic>
	<monogr>
		<title level="j">Data Mining and Knowledge Discovery</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="2074" to="2152" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">A search engine for algorithmic fairness datasets</title>
		<author>
			<persName><forename type="first">Alessandro</forename><surname>Fabris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fabio</forename><surname>Giachelle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alberto</forename><surname>Piva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gianmaria</forename><surname>Silvello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gian</forename><forename type="middle">Antonio</forename><surname>Susto</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3442/paper-08.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2nd European Workshop on Algorithmic Fairness</title>
				<meeting>the 2nd European Workshop on Algorithmic Fairness</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
	<note>CEUR Workshop Proceedings</note>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">South german credit data: Correcting a widely used data set</title>
		<author>
			<persName><forename type="first">Ulrike</forename><surname>Grömping</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<pubPlace>Berlin, Germany</pubPlace>
		</imprint>
		<respStmt>
			<orgName>Mathematics, Physics and Chemistry, Department II, Beuth University of Applied Sciences</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical report</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Equality of opportunity in supervised learning</title>
		<author>
			<persName><forename type="first">Moritz</forename><surname>Hardt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eric</forename><surname>Price</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nathan</forename><surname>Srebro</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper_files/paper/2016/file/9d2682367c3935defcb1f9e247a97c0d-Paper.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">29</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">UCI Machine Learning Repository</title>
		<author>
			<persName><forename type="first">Hans</forename><surname>Hofmann</surname></persName>
		</author>
		<idno type="DOI">10.24432/C5NC77</idno>
		<ptr target="https://doi.org/10.24432/C5NC77" />
		<imprint>
			<date type="published" when="1994">1994</date>
		</imprint>
	</monogr>
	<note>Statlog (German Credit Data)</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Bias mitigation for machine learning classifiers: A comprehensive survey</title>
		<author>
			<persName><forename type="first">Max</forename><surname>Hort</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhenpeng</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jie</forename><forename type="middle">M</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Harman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Federica</forename><surname>Sarro</surname></persName>
		</author>
		<idno type="DOI">10.1145/3631326</idno>
		<ptr target="https://doi.org/10.1145/3631326" />
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">52</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Algorithmic fairness</title>
		<author>
			<persName><forename type="first">Jon</forename><surname>Kleinberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jens</forename><surname>Ludwig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sendhil</forename><surname>Mullainathan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ashesh</forename><surname>Rambachan</surname></persName>
		</author>
		<idno type="DOI">10.1257/pandp.20181018</idno>
		<ptr target="https://pubs.aeaweb.org/doi/10.1257/pandp.20181018" />
	</analytic>
	<monogr>
		<title level="m">EAE Papers and Proceedings</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">108</biblScope>
			<biblScope unit="page" from="22" to="27" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A survey on datasets for fairness-aware machine learning</title>
		<author>
			<persName><forename type="first">Tai</forename><surname>Le Quy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arjun</forename><surname>Roy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vasileios</forename><surname>Iosifidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wenbin</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eirini</forename><surname>Ntoutsi</surname></persName>
		</author>
		<idno type="DOI">10.1002/widm.1452</idno>
		<ptr target="https://doi.org/10.1002/widm.1452" />
	</analytic>
	<monogr>
		<title level="j">WIREs Data Mining and Knowledge Discovery</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page">e1452</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Real-life performance of fairness interventionsintroducing a new benchmarking dataset for fair ML</title>
		<author>
			<persName><forename type="first">Daphne</forename><surname>Lenders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Toon</forename><surname>Calders</surname></persName>
		</author>
		<idno type="DOI">10.1145/3555776.3577634</idno>
		<ptr target="https://dl.acm.org/doi/10.1145/3555776.3577634" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing</title>
				<meeting>the 38th ACM/SIGAPP Symposium on Applied Computing</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="350" to="357" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A survey on bias and fairness in machine learning</title>
		<author>
			<persName><forename type="first">Ninareh</forename><surname>Mehrabi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Fred</forename><surname>Morstatter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nripsuta</forename><surname>Saxena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristina</forename><surname>Lerman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aram</forename><surname>Galstyan</surname></persName>
		</author>
		<idno type="DOI">10.1145/3457607</idno>
		<ptr target="https://doi.org/10.1145/3457607" />
	</analytic>
	<monogr>
		<title level="j">ACM Computing Surveys</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">35</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Algorithmic fairness: Choices, assumptions, and definitions</title>
		<author>
			<persName><forename type="first">Shira</forename><surname>Mitchell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eric</forename><surname>Potash</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Solon</forename><surname>Barocas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D'</forename><surname>Alexander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristian</forename><surname>Amour</surname></persName>
		</author>
		<author>
			<persName><surname>Lum</surname></persName>
		</author>
		<idno type="DOI">10.1146/annurev-statistics-042720-125902</idno>
		<ptr target="https://doi.org/10.1146/annurev-statistics-042720-125902" />
	</analytic>
	<monogr>
		<title level="j">Annual Review of Statistics and Its Application</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="141" to="163" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Fairness definitions explained</title>
		<author>
			<persName><forename type="first">Sahil</forename><surname>Verma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julia</forename><surname>Rubin</surname></persName>
		</author>
		<idno type="DOI">10.1145/3194770.3194776</idno>
		<ptr target="https://dl.acm.org/doi/10.1145/3194770.3194776" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Workshop on Software Fairness</title>
				<meeting>the International Workshop on Software Fairness</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Learning fair representations</title>
		<author>
			<persName><forename type="first">Richard</forename><forename type="middle">S</forename><surname>Zemel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yu</forename><surname>Ledell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kevin</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Toniann</forename><surname>Swersky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Cynthia</forename><surname>Pitassi</surname></persName>
		</author>
		<author>
			<persName><surname>Dwork</surname></persName>
		</author>
		<ptr target="https://proceedings.mlr.press/v28/zemel13.html" />
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
