<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Data Quality and Fairness: Rivals or Friends?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>(Discussion Paper)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabio Azzalini</string-name>
          <email>fabio.azzalini@polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cinzia Cappiello</string-name>
          <email>cinzia.cappiello@polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Chiara Criscuolo</string-name>
          <email>chiara.criscuolo@polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sergio Cuzzucoli</string-name>
          <email>sergio.cuzzucoli@mail.polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alessandro Dangelo</string-name>
          <email>alessandro.dangelo@mail.polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Camilla Sancricca</string-name>
          <email>camilla.sancricca@polimi.it</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Letizia Tanca</string-name>
          <email>letizia.tanca@polimi.it</email>
        </contrib>
      </contrib-group>
      <abstract>
        <p>In the last decade, data-driven decision-making is considered one of the main drivers for organizational success. Within this approach, decisions are based on insights and patterns identified through data analysis. In this scenario, input data must be reliable to guarantee the accuracy of the results: they should be correct and complete but also unbiased, i.e., both Data Quality (DQ) and Fairness should be guaranteed. However, maximizing DQ and Fairness simultaneously is not trivial, since data quality improvement techniques can negatively afect Fairness and vice versa. Understanding and thoroughly analyzing this relationship between DQ and Fairness is therefore paramount, and is this paper's goal. The results of our experiments, based on a well-known biased dataset (the Adult Census Income) provided details about this trade-of and allowed us to draw some guidelines.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Data Quality</kwd>
        <kwd>Fairness</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In the last decades, the possibility to use large amounts of data to extract information, and
gain deeper knowledge in several domains, has caused the spread of a data-driven culture,
making data collection and management extremely important. In fact, in this scenario, strategic
decisions are made on the basis of data analysis and interpretation, and relying on dependable
results becomes imperative. The performance of Machine Learning (ML) algorithms may be, for
example, seriously afected by the poor quality of the training data [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]: inaccurate, incomplete,
and inconsistent data may decrease the accuracy of the analysis results. Therefore, in addition
to the well-known storage and processing problems related to data collection, addressing Data
Quality (DQ) has become a fundamental issue [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
      </p>
      <p>
        Another key property for efective data-driven decision-making applications is the ethical
level of the data. In fact, even the most accurate application for collecting data might be afected
by ethical issues, since also high-quality data might lead to unfair outcomes. In [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] the authors
note that, for Data Science to be reliable, DQ should also include some ethical dimensions
because, in many critical fields, data can be considered of good quality only if it also conforms
to high ethical standards. Therefore the authors propose to include the most common ethical
requirements among the dimensions of quality, grouped in an Ethics Cluster:
• Fairness is defined as the lack of bias, since an algorithmic bias might result from training
a system with biased data.
• Transparency is the possibility to control the knowledge extraction process to verify the
reasons of the results.
• Diversity is the degree to which diferent kinds of objects are represented in a dataset.
• Finally, Data Protection concerns the ways to protect data, algorithms and models from
unauthorized access.
      </p>
      <p>
        Looking at this list, it is immediate to see that there may be contrasting objectives also among the
dimensions of Ethics, for instance as regards an obvious conflict between Transparency and Data
Protection. In the same way, the relationship between many well-known DQ dimensions [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and
the ethical ones is complex. In fact, maximizing both aspects simultaneously is not trivial, since
data quality techniques can negatively afect Fairness and vice versa. For example, commonly
used DQ improvement techniques, e.g., imputing missing values using the mean value, can
modify the overall distribution of values in the dataset and might lead to a reduction of Fairness;
on the other hand, some bias mitigation techniques modify real data values to remove unfairness,
thus lowering Accuracy, which is a fundamental dimension of DQ.
      </p>
      <p>With this work, we want to investigate this trade-of, especially focusing on the Completeness
and Accuracy dimensions of DQ, and on the Fairness dimension of ethics. To this aim, we
have designed experiments that take as input a dataset and perform an assessment of DQ and
Fairness before and after the application of some operations that should improve them. After
the description of the experiments we conclude the paper with some takeaway messages for
the researchers that look for the best strategy for applying changes in data to improve Fairness
or DQ according to their needs.</p>
      <p>The rest of the paper is organized as follows: Section 2 summarizes related work, while
Section 3 introduces preliminary concepts of both areas of DQ and Fairness and describes the
method we used to analyze the relationship between DQ and Fairness; Section 4 presents the
experiments we conducted on a real-world dataset, and Section 5 concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Research studies on the relationship between DQ and Fairness are in a very preliminary phase.
In this section we will first present seminal works in Fairness and then introduce two first
attempts at studying this important relationship. We do not focus on DQ systems since in this
paper we will resort to well-known and established DQ techniques.</p>
      <p>
        In the literature, one of the most notable solutions is AI Fairness 360 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], an open-source
framework aiming to measure and enforce Fairness. Its aim is to mitigate data bias, quantified
using diferent statistical measures, by exploiting pre-processing (i.e., procedures that, before the
application of a prediction algorithm, make sure that the learning data are fair), in-processing
(i.e., procedures that ensure that, during the learning phase, the algorithm does not pick up
the bias present in the data) and post-processing techniques (i.e., procedures that correct the
algorithm’s decisions with the objective of making them fair). The user can choose between
four pre-processing techniques and five statistical measures to solve bias in the dataset.
      </p>
      <p>
        Similarly, Fairlearn [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], another pre-processing, open-source, community-driven project, aims
to help data scientists improve the Fairness of their ML systems by means of statistical Fairness
metrics. These works focus on techniques that manipulate the data, seeking to make them more
fair. They do not consistently consider the impact that their techniques have on both the DQ
and Fairness dimensions.
      </p>
      <p>
        A preliminary system that considers both DQ and Fairness is the paper by Abraham et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
who proposed FairLOF, a Fairness-aware outlier detection framework. This work starts from
the fact that underrepresented groups could be identified as outliers, although they are relevant
in the dataset. Specifically, it focuses on calibrating the so-called local outlier factor, a local
outlier detection method by means of which a fairer outlier detection is possible. Though this
system actually focuses on a specific DQ problem, it can be considered as a starting point for
studying the relationship between DQ and Fairness.
      </p>
      <p>
        A similar system has been presented by Biswas et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. The authors’ goal is to investigate
the impact of data preparation pipelines on algorithmic Fairness, focusing on deep-learning
techniques. The authors conduct a detailed evaluation of several Fairness metrics applied
to diferent deep learning applications, and discover that many data preparation actions can
introduce bias in the data and, consequently, in the final prediction. However, they do not
employ any Fairness improvement technique inside their pipelines, thus considering only how
data quality techniques impact Fairness, and not vice versa.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Experiment Design</title>
      <p>This section presents the method we used to investigate the relationship between DQ and
Fairness. Before describing the work, we introduce some preliminary theoretical concepts
related to various Data Quality and Data Ethics aspects.</p>
      <p>
        Data Quality Data Quality (DQ) is defined as “fitness for use,” i.e., the ability of a data
collection to meet the user requirements [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Data Quality is a multi-dimensional concept: a
DQ model is composed of DQ dimensions representing the diferent aspects to consider (i.e.,
errors, duplicates, format errors, typos, or missing values). As already mentioned, our work
focuses on the Accuracy and Completeness dimensions:
• Accuracy is defined as the closeness between a data value v and a data value v’,
considered as the correct representation of the real-life phenomenon that the value v aims
to represent [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. It is associated with syntactic and semantic issues that might create a
discrepancy between the value stored in the dataset and the correct value.
• Completeness characterizes the extent to which a dataset represents the corresponding
real-world [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. For instance, in a relational database, Completeness is strictly related
to the presence of null values. A simple way to assess the completeness of a table is to
calculate the ratio between the number of non-null values and the number of cells in the
table.
      </p>
      <p>Fairness Fairness is one of the most important dimensions of Data Ethics. The most used
definition of Fairness is: “it is the absence of any prejudice or favoritism toward an individual
or a group based on their inherent or acquired characteristics” [10, p.100].</p>
      <p>
        Fairness is based on the idea of protected or sensitive attribute. A protected attribute is a
characteristic for which non-discrimination should be established, such as religion, race, sex,
and so on [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. A protected group is a set of individuals identified by the same value of a
protected attribute (e.g.: females, young people, Hispanic people).
      </p>
      <p>
        There is no unique definition of Fairness, but many facets exist, and a model is considered
fair if it satisfies some or all these definitions. The most used technique to identify unfairness
in datasets is to train a classification algorithm to predict the binary value of the target class
and then use Fairness metrics to understand whether the prediction of this model encompasses
discrimination: if the metrics results show unfairness, we can conclude that also the original
dataset contains unfair behaviors, since the model learned the bias from it. Specifically, we
measure the importance of protected attributes in determining the result of the model. The
following statistical metrics study how specific values of the protected attributes impact the
result of the prediction algorithm (e.g., being a woman determines that the salary is lower than
50k$/year, and being a man determines that the salary is higher than 50k$/year). Informally:
• Disparate Impact is the probability to get a positive outcome regardless of whether the
person is in the protected group [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ];
• Predictive Parity evaluates if both protected and unprotected groups have equal probability
that a group member with positive predictive value belongs to the negative class [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ];
• False Positive Ratio: evaluates if the probability of having a false positive prediction is the
same for all protected groups [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>The Method Adopted This section presents the two pipelines we defined to execute the
experiments and study the relationship between DQ and Fairness. In the first one, we inject
errors in the dataset, causing data quality issues, then apply DQ improvement techniques
and measure their impact on Fairness. In the second pipeline, since the Adult Census Income
dataset1 already contains bias w.r.t. the income of US citizens we do not need to inject further
bias to perform the experiments, therefore, we already start from a biased dataset, apply bias
mitigation techniques and measure their impact on DQ. By means of these results, we evaluate
the relationship between Fairness and DQ.</p>
      <p>The dataset used for the experiments had the following characteristics: (i) it contained bias,
so the classification algorithm would be afected by unfairness, (ii) it had been pre-processed so
that the classification algorithm could be executed in a correct manner. For example, missing
values had to be dealt with, and encoding and normalization operations were performed.</p>
      <p>
        The last operation to be performed before entering the two pipelines was applying a first
classification algorithm in order to compute the Fairness level of the dataset. As for the DQ
measure, we already knew that it was 100%. We now describe the two pipelines.
• Data Quality Improvement Pipeline: The input dataset is free of DQ problems. For this
reason, we had to inject errors in order to evaluate the impact of DQ improvement
1https://archive.ics.uci.edu/ml/datasets/adult
techniques. By injecting in a uniform manner a diferent percentage of DQ errors (from
90% to 0%, with a decreasing step of 10%) related to a specific DQ dimension, the Error
Injection phase generated ten instances of the original dataset at diferent levels of quality.
For example, for the Completeness dimension, a certain number of values are replaced
with null values. The obtained ten dirty datasets were the input of the Data Quality
Improvement phase, in which a DQ improvement technique was applied. In the case of the
Completeness dimension, an imputation technique was selected. The ten clean datasets
obtained as output were analyzed in the second Evaluation phase, in order to check the
impact of the DQ improvement on the Fairness measures. This procedure was repeated
for a number of diferent imputation methods.
• Fairness Improvement Pipeline: As regards Fairness, we did not have an error injection
phase since the considered database was already biased. The improvement phase (i.e., Bias
Mitigation phase) consisted of applying a bias mitigation technique to remove unfairness.
The repaired dataset, output of this phase, was analyzed in the second Evaluation phase
in order to assess the impact of the bias mitigation on the DQ level. This phase was
repeated for all the selected bias mitigation techniques. Since some of these techniques
act by directly substituting the data values with other (faked) values, they also allow
for partial bias repairs. For example, Correlation Remover [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], fully described in the
next section, modifies the actual values in order to minimize the correlation between the
feature attributes and the sensitive ones. If possible, ten repaired datasets (with a level of
repair from 10% to 100%) were the output of this phase.
      </p>
      <p>The last step was the analysis of the results of the two second Evaluation phases.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>In this section, we first introduce the experimental setup and then describe the results of the
experiments, both from the DQ and the Fairness perspectives.</p>
      <sec id="sec-4-1">
        <title>4.1. Experimental setup</title>
        <p>
          DQ Improvement phase In this paper, we consider two types of Data Imputation techniques:
Density-based, i.e., missing values are imputed for each feature with the same distribution of
the non-empty values, and Rare-based, i.e., the less frequent value is imputed.
Bias Mitigation phase Two bias mitigation techniques are proposed in order to remove
the unfairness from data. The former one, Correlation Remover [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], removes the negative
correlation between the protected attribute and the classification label by modifying the
nonprotected attributes that are in turn correlated to the protected one: mathematically speaking, it
poses a minimization problem of the correlation between the feature attributes and the sensitive
ones by centering the sensitive values, training a linear regressor on the non-sensitive ones
and reporting the residual. The latter, Optimized Preprocessing [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], solves an optimization
problem with the objective of minimizing the diference between the modified distribution and
the original one; specifically, it aims to reduce discrimination by mapping diferent feature
attributes and classification labels of the individuals inside a dataset while keeping the protected
attributes unchanged, to limit the dependency of the prediction on the sensitive attributes.
Evaluation Metrics To evaluate the DQ level of the dataset in the Evaluation phase (see
Section 3), the Accuracy dimension has been selected. To this aim, the distance between the
original and the final dataset has been computed. Thus, the number of values ℎ that
match between the original and the final dataset was extracted and the Accuracy dimension
has been measured as follows:
        </p>
        <p>Accuracy = ℎ

where  is the total number of cells.</p>
        <p>The three metrics selected to evaluate Fairness (see Section 3) were applied by using the
following formulas:
(1)
(2)
(3)
(4)</p>
        <p>Disparate Impact (DIR) =</p>
        <p>Predictive Parity (PPR) =
False Positive Ratio (FPR) =
 (ˆ = 1| = )
 (ˆ = 1| = )
 ( = 0|ˆ = 1,  = )
 ( = 0|ˆ = 1,  = )
 (ˆ = 1| = 0,  = )
 (ˆ = 1| = 0,  = )
where:
• : protected attribute that has two values discr (=discriminated), priv (=privileged);
• : additional data regarding the individual;
•  : actual classification result, two values (or labels) 0 or 1;
• ˆ : the algorithm predicted decision for the individual, two values of the outcome 0 or 1.
Dataset and classification algorithm The Adult Census Income dataset is obtained as an
extraction of the 1994 U.S. Census database. It is typically used to predict whether the income
of an individual exceeds 50k$ per year. It comprises 48842 tuples, described by 15 attributes,
including the targeted class. There are some sensitive attributes, such as ‘race’, ‘sex’, and ‘native
country’. In particular, for the experiments shown in this paper, we compute the Evaluation
considering the sex as a protected attribute (see Section 3). Finally, the Decision Tree Classifier
ofered by the scikit-learn2 Python library was used as classification algorithm .</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Result evaluation</title>
        <p>This section presents the main results we obtained. In Figure 1, the x-axis represents the
Completeness level; instead, in Figure 2, the x-axis shows the degree of bias mitigation. In both
Figures, the y-axis represents the level of the evaluated metrics.
Data Quality Perspective The plots shown in Figure 1 focus on the DQ perspective in which
the Evaluation results are compared for the two imputation techniques explained above (see
Section 4.1). As we can notice, the Density-based Imputation is widely better: the Accuracy
measure does not reach less than 50% for the Accuracy assessment, and the Fairness measures
increase as the percentage of injected errors increases. This is related to a vast majority of the
target class that has a value lower than 50k$/year in the dataset; since the imputation follows
the value distribution, it means that those labels have a higher probability of being assigned to
men (who are over-represented). In this way, the dataset will be balanced. We can conclude that
the application of this Imputation method improves Fairness. Instead, applying the Rare-based
Imputation, we have a big deterioration of Accuracy, and from 100% to 40% of Completeness,
the Fairness increases; instead, for Completeness values below 40%, Fairness decreases very
quickly. In this specific case, this happens because by imputing the less frequent values, the
dataset will be more balanced in favor of the protected classes. As the percentage of injected
errors increases, the rare values become too many, unbalancing the dataset. We can also observe
that the Predictive Parity (PPR) metric can assume values greater than 1. This means that the
privileged class (men) is discriminated for that specific Fairness aspect; False Positive Ratio
(FPR) always takes opposite values with respect to PPR. The two metrics are symmetrical since
they represent opposite Fairness aspects. From these results, we notice that a trade-of between
DQ and Fairness is present and that, from the DQ perspective, this trade-of can be more or less
emphasized depending on the DQ improvement technique applied.</p>
        <p>
          Fairness Perspective The plots shown in Figure 2 focus on the Fairness perspective. The
Evaluation results are compared for two diferent bias mitigation techniques: Correlation
Remover ofered by Fairlearn [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and Optimized preprocessing in AI Fairness 360 tool [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] (see
Section 4.1). In applying Correlation Remover, the Fairness metrics (DIR, FPR and PPR) slightly
increase with an important loss in Accuracy (from 1.0 to 0.6) for a partial bias mitigation between
0 and 1. This happens because the removal of correlation strongly modifies the data, greatly
decreasing the Accuracy. On the other hand, Optimized Preprocessing increases one metric
over three (FPR), and the Accuracy remains unchanged before and after the mitigation process.
This happens because data modification is at a minimum, not afecting the Accuracy. From the
Fairness perspective, a trade-of between DQ and Fairness is present, and this trade-of can be
more or less evident depending on the bias mitigation technique applied.
        </p>
        <p>Takeaway message From our experiments, we can notice that in some particular cases, the
bias mitigation technique that less afects the DQ is not the one that improves Fairness the most,
and vice-versa; for these cases, we can deduce that techniques that succeed in preserving both
DQ and Fairness do not exist. Therefore, as a takeaway message, we can afirm that the best
DQ improvement/bias mitigation techniques to apply strictly depends on the analysis goal. If
a user is more interested in preserving Fairness aspects, s/he will concentrate on a subset of
techniques at the cost of losing DQ; if the major interest is to optimize the improvement of the
DQ, the user will apply a subset of DQ improvement tasks that could deteriorate Fairness. It is
worth noting that situations may also exist in which the DQ and the Fairness aspects are not in
conflict; however, this is strictly context-dependent and could be rare to observe.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions and Future Work</title>
      <p>
        In this work, we analyzed the relationship between Data Quality (DQ) and Fairness. In particular,
through a series of experiments, we demonstrated that between DQ and Fairness, a trade-of
is present. In fact, the experiments showed us that the application of Fairness improvement
operations could lead to a deterioration of the DQ and vice-versa. Analyzing the experiments
more in detail, we can also state that the amount of DQ deterioration after Fairness
improvements depends on the bias mitigation technique, as well as the deterioration of Fairness can
depend on the selected DQ improvement technique. Future work will focus on the definition
of clear guidelines to recommend the best choice of DQ improvement/bias mitigation
techniques to be applied depending on the scope of the analysis. Moreover, we could enrich the
gathered knowledge with more datasets, DQ dimensions, Fairness metrics and bias mitigation
techniques [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ].
This research was supported by EU Horizon Framework grant agreement 101069543
(CS-AWARENEXT).
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Nagalapatti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Guttula</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mujumdar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Afzal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. Sharma</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Munigala</surname>
          </string-name>
          ,
          <article-title>Overview and importance of data quality for machine learning tasks</article-title>
          ,
          <source>in: Proceedings of the 26th ACM SIGKDD</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>3561</fpage>
          -
          <lpage>3562</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Batini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Scannapieco</surname>
          </string-name>
          ,
          <source>Data and Information Quality - Dimensions, Principles and Techniques, Data-Centric Systems and Applications</source>
          , Springer,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Sancricca</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Cappiello, Supporting the design of data preparation pipelines (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D.</given-names>
            <surname>Firmani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tanca</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Torlone</surname>
          </string-name>
          ,
          <article-title>Ethical dimensions for data quality</article-title>
          ,
          <source>JDIQ</source>
          <volume>12</volume>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R. K.</given-names>
            <surname>Bellamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hind</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Hofman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Houde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kannan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Lohia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mojsilović</surname>
          </string-name>
          , et al.,
          <source>Ai</source>
          fairness
          <volume>360</volume>
          :
          <article-title>An extensible toolkit for detecting and mitigating algorithmic bias</article-title>
          ,
          <source>IBM Journal of Research and Development</source>
          <volume>63</volume>
          (
          <year>2019</year>
          )
          <fpage>4</fpage>
          -
          <lpage>1</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bird</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dudík</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Edgar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Horn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Lutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Milan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sameki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wallach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Walker</surname>
          </string-name>
          ,
          <article-title>Fairlearn: A toolkit for assessing and improving fairness in ai</article-title>
          , Microsoft,
          <source>Tech. Rep. MSR-TR-2020-32</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Abraham</surname>
          </string-name>
          ,
          <article-title>Fairlof: fairness in outlier detection</article-title>
          ,
          <source>Data Science and Engineering</source>
          <volume>6</volume>
          (
          <year>2021</year>
          )
          <fpage>485</fpage>
          -
          <lpage>499</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Rajan</surname>
          </string-name>
          ,
          <article-title>Fair preprocessing: towards understanding compositional fairness of data transformers in machine learning pipeline</article-title>
          ,
          <source>in: Proceedings of the 29th ACM Joint Meeting on ESEC/FSE</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>981</fpage>
          -
          <lpage>993</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>R. Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. M.</given-names>
            <surname>Strong</surname>
          </string-name>
          , Beyond accuracy:
          <article-title>What data quality means to data consumers</article-title>
          ,
          <source>Journal of management information systems 12</source>
          (
          <year>1996</year>
          )
          <fpage>5</fpage>
          -
          <lpage>33</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Huang</surname>
          </string-name>
          , E. DeFilippis, G. Radanovic,
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Parkes</surname>
          </string-name>
          , Y. Liu,
          <article-title>How do fairness definitions fare? testing public attitudes towards three algorithmic definitions of fairness in loan allocations</article-title>
          ,
          <source>Artif. Intell</source>
          .
          <volume>283</volume>
          (
          <year>2020</year>
          )
          <fpage>103238</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rubin</surname>
          </string-name>
          ,
          <article-title>Fairness definitions explained</article-title>
          ,
          <source>in: Proceedings of the FairWare@ICSE</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mehrabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Morstatter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galstyan</surname>
          </string-name>
          ,
          <article-title>A survey on bias and fairness in machine learning</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>54</volume>
          (
          <year>2022</year>
          )
          <volume>115</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>115</lpage>
          :
          <fpage>35</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Azzalini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Criscuolo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Tanca</surname>
          </string-name>
          ,
          <article-title>E-fair-db: functional dependencies to discover data bias and enhance data equity</article-title>
          ,
          <source>ACM Journal of Data and Information Quality</source>
          <volume>14</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Azzalini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Criscuolo</surname>
          </string-name>
          , L. Tanca,
          <article-title>Fair-db: A system to discover unfairness in datasets</article-title>
          ,
          <source>in: 2022 IEEE 38th International Conference on Data Engineering (ICDE)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>3494</fpage>
          -
          <lpage>3497</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>