<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Transparency and Proportionality in Post-Processing Algorithmic Bias Correction</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Juliett Suárez-Ferreira</string-name>
          <email>juliettsuarez@correo.ugr.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marija Slavkovik</string-name>
          <email>Marija.Slavkovik@uib.no</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jorge Casillas</string-name>
          <email>casillas@decsai.ugr.es</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer and Telecommunications Engineering</institution>
          ,
          <addr-line>18071 Granada</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Data Science and Computational Intelligence Institute (DaSCI), University of Granada, Calle Periodista Daniel Saucedo Aranda</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Computer Science and Artificial Intelligence (DCSAI), University of Granada, Higher Technical School of</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Department of Information Science and Media Studies, University of Bergen</institution>
          ,
          <addr-line>Fosswinckels gate 6, 5007 Bergen</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <addr-line>s/n. 18071 Granada</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Algorithmic decision-making systems sometimes produce errors or skewed predictions toward a particular group, leading to unfair results. Debiasing practices, applied at diferent stages of the development of such systems (pre-processing, in-processing, post-processing), occasionally introduce new forms of unfairness or exacerbate existing inequalities. We focus on post-processing techniques that modify algorithmic predictions to achieve fairness in classification tasks, examining the unintended consequences of these interventions. To address this challenge, we develop a set of measures that quantify the disparity in the flips applied to the solution in the post-processing stage. The proposed measures will help practitioners: (1) assess the proportionality of the debiasing strategy used, (2) have transparency to explain the efects of the strategy in each group, and (3) based on those results, analyze the possibility of the use of some other approaches for bias mitigation or to solve the problem. We introduce a methodology for applying the proposed metrics during the post-processing stage and illustrate its practical application through an example. This example demonstrates how analyzing the proportionality of the debiasing strategy complements traditional fairness metrics, providing a deeper perspective to ensure fairer outcomes across all groups.</p>
      </abstract>
      <kwd-group>
        <kwd>fairness</kwd>
        <kwd>bias mitigation</kwd>
        <kwd>debias</kwd>
        <kwd>proportionality</kwd>
        <kwd>post-processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Bias, as defined by Tversky and Kahneman [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], usually signifies a systematic inclination or prejudice
that distorts judgment or decision making, causing unfair outcomes. In Artificial Intelligence (AI)
systems, bias implies the propensity of a system to consistently produce certain types of error or skewed
predictions due to flaws in the data, algorithm design, or training process, and has been recognized as
one of the risks of Algorithmic Decision Making (ADM) systems [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Debiasing refers to a range of strategies and interventions aimed at reducing or eliminating biases in
decision-making processes where the goal is to improve objectivity and ensure that decisions are more
aligned with normative standards of fairness and accuracy [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. In the development of ADM systems,
these interventions have diverse terminology, are called bias mitigation techniques [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], methods for fair
machine learning (ML) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] or fairness interventions [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], and are applied by practitioners at diferent
stages of the development of the ADM system to ameliorate the efect of bias and obtain fairer solutions.
These stages, part of the ML pipeline: pre-processing, in-processing, and post-processing, can be
observed in Figure 1.
      </p>
      <p>However, an efort to debias a decision can sometimes itself introduce new forms of unfairness or
exacerbate existing inequalities. One reason for this is that debiasing techniques may inadvertently
privilege certain groups over others in their aim to achieve a fairer result. For example, when adjusting
for bias in predictions, post-processing methods can disproportionately impact certain demographic
(J. Casillas)</p>
      <p>CEUR
Workshop</p>
      <p>
        ISSN1613-0073
groups by relabeling the outcome of historically advantaged groups to achieve a certain fairness criterion
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Furthermore, debiasing interventions may not address the root causes of biases but rather shift
them in ways that perpetuate or even amplify existing disparities [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. A critical question emerges: How
disproportionate are the results we obtain with the methods we use to debias the outcomes of
our algorithms?
      </p>
      <p>Disproportionality occurs when a debiasing method afects some demographic groups more than
others, either by changing their predictions more frequently or by imposing more harmful adjustments
(like switching favorable outcomes to unfavorable ones) on one group compared to others. We develop
a set of measures to quantify the proportionality of debiasing interventions and define a methodology
for applying the proposed metrics in the post-processing stage.</p>
      <p>We focus specifically on examining the unintended consequences of post-processing techniques that
modify algorithmic predictions to achieve fairness in classification tasks. Here, we present the study of
binary classification and binary protected attributes. However, the proposed metrics can be extended to
multi-classification scenarios.</p>
      <p>The measures we propose are intended to help practitioners (1) assess the proportionality of the
debias strategy used, (2) have transparency that allows them to explain the efects of the strategy in
each group, and (3) based on those results, analyze the possibility of using some other strategies for
bias mitigation.</p>
      <p>This work is structured as follows. Section 2 reviews some works studying sources of bias in ML and
mitigation techniques along its pipeline with a focus on post-processing methods for binary classification
tasks. We introduce what proportionality is in this context in Section 3. Section 4 introduces a set
of metrics to assess the proportionality of post-processing debiasing interventions, describing their
mathematical formulation and characteristics. Section 5 presents a methodology for applying the
proposed metrics in real-world scenarios along with a practical example of how to use it. This final
analysis demonstrates how the proposed metrics provide deeper insights into fairness, complementing
traditional fairness metrics with this proportionality analysis. Finally, the discussion, conclusions, and
future work highlights the trade-ofs identified, discusses the broader implications of proportionality
metrics, and identifies opportunities for extending the contributions of this paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        Bias exists in many forms, leading to a unfairness in some cases. In [
        <xref ref-type="bibr" rid="ref4 ref8 ref9">8, 4, 9</xref>
        ], the authors discuss various
sources of bias in ML, providing categorizations and descriptions to inspire future solutions. A variety
of bias mitigation techniques have been developed, each targeting diferent stages of the ML pipeline
where bias can manifest. They are broadly categorized into three types: pre-processing, in-processing,
and post-processing approaches [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">4, 5, 3</xref>
        ]. Each category represents diferent stages of the ML pipeline
where practitioners can apply diferent interventions to mitigate bias, and each has distinct operational
mechanisms and implications on the outcomes, see Figure 1.
      </p>
      <p>Training Data
A1 A2 … An C
1 0 … 0 1
1 1 … 0 0
… … … … …
0 1 … 1 1</p>
      <p>ML Model selection
M1 {PM1} : {Acc=A1, FM = MV1}
M2 {PM2} : {Acc=A2, FM = MV2}
M3 {PM3} : {Acc=A3, FM = MV3}
M4 {PM4} : {Acc=A4, FM = MV4}
M4 {PM4} : {Acc=A5, FM = MV5}</p>
      <p>ML Model
evaluation</p>
      <p>M2</p>
      <p>ML Model
deployment
ADM system</p>
      <p>M2</p>
      <p>Decision
pre-processing
in-processing
post-processing</p>
      <p>Problem Instance</p>
      <p>
        We focus on the results of the post-processing methods specifically designed for classification tasks
where the goal is to predict a label (y) from a set of inputs using a pre-trained model1. The debiasing
methods in these cases are applied after the model has been trained and act by modifying the prediction
of the model to ensure fairness without altering the model itself or the training data. The main
techniques include calibration [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], thresholding [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and transformation [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>y
y predicted
y corrected</p>
      <p>
        We illustrate a very simple case in Figure 2. The figure represents the outcome of 10 instances of a
classification problem ( y) with two possible labels (+ and -) belonging to two groups (light gray and dark
gray). Labels represent the outcome of the classification and groups symbolize a protected attribute.
The predicted labels for the dark gray group (y predicted) have 4 out of 5 examples in the positive class
(80%) while the light gray group has 3 out of 5 examples in the positive class (60%) this will represent a
diference of 20% in statistical parity 2 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] which could be considered unfair. Consider that applying
a post-processing debiasing method obtains the y corrected labels. Although the number of positive
outcomes in the two groups has the same proportion (2 out of 5 for a 40%), we can observe that even
when the flips occur in both groups towards the negative label, they impact more the dark gray group
compared to the light gray group. Here, we aim to evaluate the unintended consequences of debiasing
techniques that result in the alteration of the outcomes with the purpose of achieving fairness.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. What Proportionality Means</title>
      <p>Classic group fairness metrics (e.g. statistical parity or equalized odds) summarize disparities in the final
predictions produced by a classifier. They do not, however, reveal how a post-processing intervention
arrived at those predictions, nor who gained or lost during that process. Proportionality fills this gap:
it asks whether the benefits and burdens that arise when we flip labels in the post-processing stage are
distributed in a way that is normatively justified and legally defensible.</p>
      <p>
        Let a flip be any change of a predicted label induced by a post-processing rule. We distinguish:
• Balancing flips across groups. The counts and rates of flips should not be so unequal that
one group bears virtually all harmful flips or garners all beneficial ones. Our metrics (Section 4)
make this distribution explicit, extending the logic of statistical parity [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] from outcomes to
interventions.
• Harmful versus beneficial flips. Changing a positive outcome to a negative one is usually
a genuine loss for the afected individual (e.g., losing a job ofer or loan). A proportional rule
therefore seeks to minimise harmful flips overall and to avoid concentrating them on historically
marginalised groups. By separately tracking harmful and beneficial flips, we expose potential
1A model can be considered a mathematical function that map input data to output predictions. For classification tasks, they
are produced by training an algorithm with predefined data and using specific parameters.
2A fairness metric which declares that the likelihood of a positive outcome should be the same irrespective of an individual’s
group membership.
      </p>
      <p>
        levelling-down (many losses for one group, few gains elsewhere) and encourage levelling-up
strategies that improve outcomes for disadvantaged groups [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        Numerous dimensions of moral and political philosophy elucidate the significance of proportionality.
The equality of opportunities requires that people with comparable talent or qualifications face similar
chances of desirable outcomes, regardless of protected attributes [
        <xref ref-type="bibr" rid="ref15 ref16">15, 16</xref>
        ]. A post-processing rule that
places most negative flips on one group violates this principle, whereas a rule proportionate levels the
playing field without arbitrarily closing doors to the otherwise qualified.
      </p>
      <p>
        Desert-based accounts hold that benefits should align with efort or qualification [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Excessive
harmful flips against high performers signal a desert violation; a proportionate intervention corrects
bias while continuing to reward merit.
      </p>
      <p>
        Suficiency and prioritarian theories prioritize improving the situation of those who are worse of
[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Therefore, proportionality disfavors levelling down, making advantaged groups worse of without
materially helping the disadvantaged, a practice strongly criticized [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
      </p>
      <p>
        Proportionality also echoes the established equality doctrine. EU fundamental-rights law applies a
four-step proportionality test (suitability, necessity, balancing, and consistency [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]) whenever a policy
imposes diferential treatment [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. The UK Equality Act adopts a near-identical standard for justifying
indirect discrimination: the measure must be “a proportionate means of achieving a legitimate aim”
[
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. Our proposal operationalize these legal ideas: they quantify whether a debiasing strategy imposes
an excessive share of negative flips on any group and thus provide empirical evidence for (or against)
legal proportionality.
      </p>
      <p>Proportionality evaluates whether fairness corrections themselves are fair. Grounding our metrics
in normative theory and equality law serves to equip practitioners with principled diagnostics that
complement traditional group-fairness statistics and guard against well-intentioned but excessive
interventions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Assessing the Efects of Flips Produced by Post-Processing</title>
    </sec>
    <sec id="sec-5">
      <title>Debiasing Techniques</title>
      <p>In this section, we provide a characterization of the flips in the solution (Section 4.1) that ofer a general
picture of how the debiasing algorithm afects the model’s predictions, and subsequently, we extend
the analysis with the objective of evaluating whether the debiasing algorithm impacts diferent groups
equitably by introducing a series of flip proportionality metrics in Section 4.2.</p>
      <p>For each metric, we present both its mathematical definition and its interpretation. Each subsection
concludes with a summary table of the proposed metrics, clarifying their boundaries, a short description,
and edge cases.</p>
      <sec id="sec-5-1">
        <title>4.1. Characterization of Flips in a Solution</title>
        <p>In this Section we define key concepts and metrics that quantify how a debiasing algorithm modifies
the original predictions.</p>
        <p>We first start by characterizing a classification problem: given a set of features  ∈ ℝ × , where
 is the number of instances and  is the number of features. The goal of the classification task is
to learn a classifier  ∶ ℝ  → 0, 1 that predicts the binary outcome  ∈ 0, 1  ∶  ̂ =  ( ) where
 = { 1,  2, … ,   } represents the feature vectors for  ,  = { 1,  2, … ,   } represents the true binary
outcomes, and  =̂ { ̂ 1,  2̂, … ,  ̂ } represents the predicted outcomes.</p>
        <p>Let  predicted =  ∈̂ {0, 1}  be the vector of predicted labels from the classifier for the same instances
 and consider that 0 is the unfavorable outcome and 1 the favorable one. For illustrative purposes,
consider a classification problem in which possible outcomes entail either accepting or rejecting a
candidate. A favorable or positive outcome is accepting the candidate and will have the value 1 in
 predicted.
 corrected ∈ {0, 1} .</p>
        <p>Definition 1
process.</p>
        <p>predicted = { predicted,1,  predicted,2, … ,  predicted, }
After applying a debiasing algorithm, the predicted labels can be adjusted to form the corrected labels
 corrected = { corrected,1,  corrected,2, … ,  corrected, }
We will compare</p>
        <p>with  
on the classifier’s predictions (</p>
        <p>. This comparison isolates the efect of the debiasing algorithm
) measuring how much the debiasing method has adjusted the
predictions to correct potential biases. A flip occurs when the label changes from one label to another
(e.g., from 0 to 1 or from 1 to 0) as a result of a debias algorithm.</p>
        <p>(Flip). Let  predicted and  corrected be sets of the corresponding outputs after a debiasing</p>
        <sec id="sec-5-1-1">
          <title>A flip between  predicted and  corrected is defined as:</title>
          <p>Flip = {
1
0
if  predicted, ≠  corrected,
if  predicted, =  corrected,</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>Definition 2 (Number of flips) . The number of flips,  flips , represents the total count of instances in which</title>
          <p>the predicted label  predicted difers from the corrected label  corrected after applying the debiasing algorithm:
Definition 3 (Flip Rate (FR)). Is defined as the proportion of instances that experienced a flip over the
total number of instances.</p>
          <p>flips = ∑ Flip

=1
Flip Rate =
vorable outcome (0) to a favorable outcome (1), indicating an increase in the number of positive decisions.
This type of flip represents a shift towards a more favorable outcome for the instance.</p>
          <p>Favorable Flip = {
1
0
if  predicted, = 0 and  corrected, = 1
otherwise
The total number of positive flips,  favorable flips , is given by:
 favorable flips = ∑ Favorable Flip

(2)
(3)
(4)
(5)
(6)
(7)
Definition 5 (Unfavorable Flips). A negative flip or a harmful flip occurs when the predicted label changes
from 1 to 0, indicating a decrease in the number of positive decisions. This type of flip represents a shift
towards a less favorable outcome for the instance.</p>
          <p>Unfavorable Flip = {
1 if  predicted, = 1 and  corrected, = 0
0 otherwise</p>
        </sec>
        <sec id="sec-5-1-3">
          <title>The total number of negative flips,  unfavorable flips , is given by:</title>
          <p>unfavorable flips = ∑ Unfavorable Flip

=1</p>
          <p>These classifications help in understanding the nature of the changes made by the debiasing algorithm.
Analyzing the nature of the flips can provide insight into how the debiasing process impacts overall
decision making.
unfavorable flips (from 1 to 0).</p>
          <p>Definition 6 (Directional Flip Ratio (DFR)). Compares the number of favorable flips (from 0 to 1) with
DFR =
 favorable flips
 unfavorable flips</p>
          <p>A DFR closer to 1 indicates balanced flip directions, suggesting that the debiasing algorithm is not
disproportionately flipping predictions in one direction (e.g., systematically downgrading or upgrading
individuals).</p>
          <p>Values greater than 1 suggest more favorable than unfavorable flips, and values lower than 1 will
suggest the higher occurrence of unfavorable flips. The desired values for this metric should be close to
1 indicating balanced flips in both directions.</p>
          <p>Taking into account unfavorable flips (those in which the outcome was changed to an unfavorable
value), we can establish metrics of the impact on individuals when achieving fairness. These flips are
considered harmful because they represent a tangible loss for the afected individuals; in the previous
example, when a prediction changes from job candidate acceptance to rejection. While such changes
may be necessary to achieve overall system fairness, they represent real negative consequences for the
individuals whose predictions are flipped, making it crucial to measure and minimize their occurrence,
especially when they disproportionately afect specific demographic groups.</p>
          <p>Definition 7 (Harmful Flip Proportion(HFP)). This metric calculates the proportion of harmful flips
among all flips. The HFP is then defined as:</p>
          <p>HFP =
 unfavorable flips</p>
        </sec>
        <sec id="sec-5-1-4">
          <title>Where  unfavorable flips was defined in Equation 8 and  flips is the total number of flips defined in Equation</title>
          <p>4.</p>
          <p>A harmful flip was defined as a change in prediction from a positive to a negative outcome, which is
interpreted as a detrimental change for the individual instance. A lower HFR indicates that fewer flips
result in harmful outcomes, suggesting that the debiasing algorithm is less likely to produce negative
efects on the predictions.</p>
          <p>The metrics introduced until now allow us to characterize the flips made by a debiasing algorithm
that transforms the output in the post-processing stage. After their calculations, practitioners will have
a general overview of the flips applied to the solution. Furthermore, these metrics can be independently
applied to each distinct group, providing practitioners with an understanding of the overall incidence
of flips in each group. A resume of the metrics proposed to describe the flips in the solution is detailed
in Table 2 of the Appendix A.
(8)
(9)
(10)
(11)</p>
          <p>Where (  = 1) is an indicator function equal to 1 if the instance  belongs to the privileged group, and
0 otherwise and (  = 0) is an indicator function equal to 1 if the instance  belongs to the unprivileged
group, and 0 otherwise.</p>
          <p>The same way, the metric Harmful Flip Proportion can be calculated separately for diferent groups.
For instance, HFPunprivileged represents the HFP for the unprivileged group, and HFPprivileged represents
the HFP for the privileged group:</p>
          <p>HFPprivileged =
HFPunprivileged =


∑=1 Unfavorable Flip ⋅ (  = 1)</p>
          <p>∑=1 Flip ⋅ (  = 1)
∑=1 Unfavorable Flip ⋅ (  = 0)</p>
          <p>∑=1 Flip ⋅ (  = 0)</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>4.2. Group-Based Flip Proportionality Metrics</title>
        <p>In this section we propose flip proportionality metrics to calculate the diferences between the flips
inflicted on the groups. For group-based metrics, let us consider that instances are characterized by
their belonging to a certain binary protected feature, where 1 represents individuals with historical
advantage (privileged) and 0 represents individuals in historical disadvantage (unprivileged). The Flip
Rate for each group can be calculated separately.</p>
        <p>Let us define  ∈ {0, 1}  as the vector indicating the membership of the protected group for each
instance, where   = 1 indicates a privileged individual and   = 0 indicates an unprivileged individual.</p>
        <p>The Flip Rate for the privileged group (FRprivileged) and the unprivileged group (FRunprivileged) can be
formulated as:</p>
        <p>A value close to 0 for FRD or HFPD indicates a more proportional treatment between groups, so the
desirable value is zero.</p>
        <p>These metrics capture the similarities between the flip rates and the proportion of harmful flips
between the groups. However, if the size of one group is significantly smaller, these metrics might
overemphasize disparities due to variance in smaller sample sizes. Therefore, they should be
accompanied by the analysis of the individual measures for both groups independently defined in Equations 12
and 13, as it will capture the proportions between the flip rates and the proportion of harmful flips of
the groups separately. This distinction allows for the identification of scenarios that can have small
diferences but high values separately that suggest a high incidence of flips in the overall solution.</p>
        <p>FRprivileged =
FRunprivileged =



∑
=1 Flip ⋅ (  = 1)
∑</p>
        <p>=1 (  = 1)
∑
=1 Flip ⋅ (  = 0)
∑</p>
        <p>=1 (  = 0)</p>
        <p>Taking the definition of flips rates and harmful flip proportions for each group, we define a set of
proportionality measures that quantify the disparity in the flips between the groups.
Definition 8 (Flip Rate Diference ( FRD) &amp; Harmful Flip Proportion Diference ( HFPD)). This metric
calculates the absolute diference in the flip rates or the harmful flip rates between two groups:</p>
        <p>FRD = |FRprivileged − FRunprivileged|
HFPD = |HFPprivileged − HFPunprivileged|
(12a)
(12b)
(13a)
(13b)
(14a)
(14b)
Definition 9 (Disparity Index (DI) &amp; Harmful Disparity Index (HDI)). The disparity index highlights
the disparity between the flip rates or harmful flip proportions of two groups.</p>
        <p>DI =
HDI =
max(FRprivileged, FRunprivileged)
min(FRprivileged, FRunprivileged)
max(HFPprivileged, HFPunprivileged)
min(HFPprivileged, HFPunprivileged)</p>
        <p>A DI or HDI of 1 indicates perfect proportionality, while values greater than 1 indicate the extent of
disproportionality. These metrics use ratios, and if the denominator (flip rate or harmful flip proportion)
is very small, the disparity index can become disproportionately large, particularly for groups with
fewer flips. Therefore, an analysis in conjunction with the characterization of the flips is required to
better understand the overall flip context.</p>
        <p>Definition 10 (Flip Rate Disparity (FD) &amp; Harmful Flip Proportionality Disparity (HFD)). This metric
computes the diference in the flip rates or harmful flip proportion between diferent groups in relation to
the overall Flip Rate defined in Equation 5.</p>
        <p>FD = |
HFD = |</p>
        <sec id="sec-5-2-1">
          <title>FRprivileged</title>
          <p>Flip Rate</p>
        </sec>
        <sec id="sec-5-2-2">
          <title>HFPprivileged</title>
          <p>Flip Rate
−
−</p>
        </sec>
        <sec id="sec-5-2-3">
          <title>FRunprivileged</title>
          <p>Flip Rate
|</p>
        </sec>
        <sec id="sec-5-2-4">
          <title>HFPunprivileged</title>
          <p>Flip Rate
|
(15a)
(15b)
(16a)
(16b)
(17a)
(17b)
Where the values of FRD and HFPD are defined in equations 14a and 14b respectively.</p>
          <p>These metrics are beneficial for understanding the relative diference in flip rates (or harmful flip
rates) in a way that is proportionate and comparable across diferent scenarios. A value closer to 0
indicates that the flip rates or harmful flip proportions between the groups are proportionally similar,
implying that the debiasing process similarly afects both groups. Higher values of these measures
indicate a greater disparity between the groups, suggesting that one group is experiencing flips at a
significantly diferent rate than the other.</p>
          <p>These metrics normalize disparities based on the sum of group-specific rates. If the overall rates are
small or one group dominates, normalized metrics might amplify the disparities. For clarity, we give a
summary of the proposed metrics in Table 3 of the Appendix A.</p>
          <p>If the overall flip rate or the harmful flip proportion is close to 0, the normalized rates might become
very large, potentially magnifying the value of the measures. This should be taken into account for
better interpretability of the measure; also these measures do not explicitly account for group sizes.
If one group (e.g., the privileged group) is much smaller, its flip rate might have a higher variance,
potentially skewing the value of the measures. Values close to 0 are the desirable values for these
measures.</p>
          <p>Definition 11 (Relative Flip Disparity (RFD) &amp; Relative Harmful Flip Disparity (RHFD)). These metrics
provide a normalized measure of the disparity between the flip rates or the harmful flip proportions between
groups.</p>
          <p>RFD =
RHFD =</p>
        </sec>
        <sec id="sec-5-2-5">
          <title>FRunprivileged + FRprivileged</title>
        </sec>
        <sec id="sec-5-2-6">
          <title>HFPunprivileged + HFPprivileged</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>5. Applying the Proposed Metrics: Methodology and Example</title>
      <p>In this Section, we discuss how to apply the measures we defined, while analyzing the results of a
post-processing dibiasing method that works by flipping the output of the solution. Figure 3 illustrates
a step-by-step debiasing strategy to achieve fairness in the predictions while evaluating proportionality.
The process begins by computing the predicted labels   . The fairness of these predictions is
evaluated by comparing the true ( ) and predicted labels. If fairness criteria are not met, a debiasing
post-processing step is applied, producing corrected labels (  ). Then these corrected labels are
again evaluated for fairness. Furthermore, the method assesses whether the correction of the predicted
values is proportional, ensuring that the changes applied do not disproportionately afect specific groups.
The analysis ends when both fairness and proportionality are achieved; otherwise, practitioners should
evaluate whether to change the debiasing strategy or the solution. It is important to note that any
group fairness metric can be used appropriately to evaluate classification outcomes.
y</p>
      <p>We demonstrate the application of the proposed methodology through a specific example. The results
and algorithms in this section are not intended as contributions to this paper, but rather they serve to
demonstrate the application of the proposed proportionality metrics.</p>
      <p>
        First, we solved a toy problem using a DecisionTreeClassifier implemented in the Scikit-learn library
[
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]; the accuracy of the classification is 0.725. Then, we calculated two fairness metrics for the solution:
Statistical Parity (SP) [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and Equalized Odds (EO) [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] using the AI Fairness 360 (AIF360) toolkit [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
The results of these metrics were −0.31 and 0.28 respectively which indicate the disparity between the
privileged and unprivileged groups. After that, we applied the EqOddsPostprocessing algorithm based on
[
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. The results of SP and EO after the post-processing step are −0.071 and 0.025, respectively, which
are accepted values in the fair interval [−0.1, 0.1].
      </p>
      <p>The first interpretation is that the post-processing method has solved the fairness problem.
Nevertheless, when a closer look is taken to the flips occurring in the debias process, the appropriateness of
the results may be revised. To achieve this, we have implemented the proposed metrics in Python3, as
output we ofer a proportionality report with the results of the measures as is illustrated in Table 1 and
a visualization of the main metrics as observed in Figure 4.</p>
      <p>We used the predicted and corrected results of our toy problem and show the results of the metrics
applied to the example in Table 1. The table lists the metric name (values in bold in the column Metric),
their value (Result) and a Short Analysis of the calculation. This short analysis is also produced by the
code implemented based on the computation of the metric. The table also informs about the dataset,
groups, and flipped totals.</p>
      <p>When we analyze the values in Table 1, we observe that while the flip rate value for the overall
dataset is 13% the harmful flip proportion constitutes 78% of the flipped instances, raising concerns
about the fairness of the process generating the flips. Analyzing the flips in each group reveals that
most of them occur in Group 0, and all flips in this group are harmful. In contrast, Group 1 experiences
fewer flips, all of which are beneficial. The directional flip rate highlights the disparity between the
groups in terms of the nature of their flips.
3The sources for this article are available via GitHub
1320
799
521</p>
      <p>Collective analysis of flip proportionality metrics demonstrates a systematic disparity in flip rates
between the groups. This issue is exacerbated when examining harmful flip proportionality metrics,
which expose significant fairness disparities between the groups, highlighted by the concentration of
all harmful flips within Group 0.</p>
      <p>We have commented on the desirable values for each proportionality metric, but have not proposed
thresholds in which the results of the metrics can be considered acceptable, moderate, or disproportionate.
We consider that these thresholds will also depend on the context of the problem analyzed since factors
such as the size of the dataset may influence the metrics values.</p>
      <p>In our implementation of the proposed metrics, we provide an example of threshold values that can
be configured. We consider that a diference lower than 0.1 from each proportionality metric result with
Flip Proportionality Metrics</p>
      <p>Harmful Flip Proportionality Metrics
respect to their ideal value can be considered acceptable, a diference in a range of [0.1, 0.3] indicate
manageable disparities requiring review, and beyond that diference, we consider the values to show
imbalance that may indicate problems with the proportionality of the debiasing strategy applied. We
applied these ranges for all the metrics except by the FRD &amp; HFPD (Equation 14) in which we consider
a diference of 0.05 for acceptable values and between 0.05 and 0.15 for moderate values.</p>
      <p>We encourage practitioners to consider the adaptation of these values to their particular problem.
The proposed thresholds are empirical and are not mathematically derived or rigorously proved; instead,
they are based on practical considerations to guide the interpretation of results.</p>
      <p>For an easier interpretation of the results, we have implemented a simple visualization of the main
metrics. The visual analysis can be seen in Figure 4. The figure presents a visual analysis of flips and
lfip proportionality measures. The upper graph displays the Flip Rate and the Harmful Flip Proportion
as percentages. The middle graph presents the same metrics per group. The lower graph focuses on flip
proportionality metrics. Each metric is color-coded according to the thresholds explained previously
as Acceptable (green), Moderate (yellow), and Disproportionate (red). We also limit the ∞ values to a
maximum for better visualization.</p>
      <p>The analysis of Figure 4 reinforces the results already analyzed from Table 1. As an inference of
this toy problem, in real-world scenarios, practitioners should analyze the suitability of the debiasing
strategy, as well as the possibility of applying other methods to solve the problem; otherwise, it is
necessary to provide a justification for the disproportionate adverse treatment experienced by a specific
group. This example is instrumental in illustrating the unintended consequences of debiasing strategies,
particularly in terms of the harm experienced by a particular group that is masked within the fairness
metrics results.</p>
    </sec>
    <sec id="sec-7">
      <title>6. Discussion and Limitations</title>
      <p>Our analysis of proportionality metrics underscores crucial trade-ofs inherent in algorithmic debiasing,
highlighting the complexity of ensuring equitable impacts across demographic groups. Although
disproportional flipping might sometimes be unavoidable due to underlying data distributions or
historical biases, practitioners should transparently document and ethically justify such occurrences.
Furthermore, alternative debiasing strategies, such as pre-processing or in-processing methods, should
be actively explored to achieve fairness without disproportionate impacts.</p>
      <p>We recognize that evaluating proportionality solely through prediction flips does not fully capture
the nuanced interplay between fairness interventions, predictive accuracy, and ground truth labels.
Some flips initiated to improve fairness may incidentally align predictions with actual outcomes,
thus improving accuracy; conversely, others may inadvertently degrade predictive quality. Therefore,
integrating proportionality metrics alongside traditional accuracy indicators and fairness measures is
essential. This evaluation allows practitioners to better diferentiate between beneficial corrections and
fairness-driven errors, facilitating more informed and ethically sound decision-making.</p>
      <p>Despite their utility, proportionality metrics exhibit certain limitations. Specifically, the metrics may
be overly sensitive in scenarios with low overall flip rates or imbalanced group sizes. In such cases, even
minor disparities could appear exaggerated, potentially misrepresenting the true fairness landscape.
The application of proportionality metrics should always be contextualized within specific normative
frameworks relevant to the domain in question. For example, in healthcare, proportionality might
entail accepting a certain level of disparity to prioritize the most urgent cases, while in employment or
education, a more egalitarian proportionality might be ethically justified to correct historical inequities.</p>
    </sec>
    <sec id="sec-8">
      <title>7. Conclusions and Future Work</title>
      <p>This study explores the dynamics of bias mitigation within algorithmic decision-making systems,
particularly emphasizing the unintended consequences arising from post-processing fairness interventions.
We introduce a novel set of metrics explicitly designed to evaluate the proportionality of prediction
lfips resulting from these interventions. These metrics serve as safeguards, promoting responsible
and ethically justified deployments of algorithmic systems. Furthermore, we propose an actionable
methodology that integrates these proportionality metrics into existing machine learning workflows,
enhancing transparency and accountability in algorithmic decisions.</p>
      <p>Future research directions include expanding and strengthening empirical evaluations. Specifically,
we plan comprehensive experiments involving diverse real-world datasets, multiple classification models,
and various post-processing fairness interventions to rigorously validate and generalize our metrics.
Furthermore, exploring alternative normalization techniques (such as weighting proportionality metrics
by group sizes or employing statistical validation methods) would further enhance the reliability of
the metric. Extending our proportionality framework to multiclass classification settings and multiple
protected attributes will also be crucial. Lastly, integrating these metrics into widely adopted fairness
toolkits, such as AIF360 or Fairlearn, would significantly streamline fairness assessments, enabling
practitioners to evaluate fairness, proportionality, and predictive accuracy within a unified analytical
framework.</p>
      <p>Additionally, we intend to extend our proportionality analysis by incorporating neighborhood-based
individual metrics, enabling a detailed assessment of unintended consequences at the instance level and
improving transparency and accountability in post-processing bias correction strategies.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Writefull to improve grammar and spelling
check. After using this tool, the authors reviewed and edited the content as needed and assumed full
responsibility for the content of the publication.</p>
    </sec>
    <sec id="sec-10">
      <title>A. Summary of the proposed metrics</title>
      <p>This appendix presents a summary of the proposed metrics used to analyze prediction flips after debiasing
interventions. Table 2 outlines the metrics that characterize the nature and impact of individual flips,
capturing aspects such as frequency, directionality, and potential harm. Table 3 extends this resume to
evaluate the proportionality of these flips across diferent groups, highlighting potential disparities in
how changes afect various subpopulations. Each metric is presented with its mathematical boundaries,
reference to the equation, and a brief description, including how it behaves under edge cases. This
summary is intended to provide a quick reference for understanding and interpreting the behavior of
lfips in group-level analyzes.</p>
      <p>Metric</p>
      <p>Boundaries</p>
      <p>Equation</p>
      <p>
        Short Description &amp; Edge Cases
FR
DFR
HFP
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]
[0, ∞)
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]
5
10
11
      </p>
      <p>Measures the proportion of instances where predictions
change after debiasing. Is 0 when there is no flip in the
post-processing stage. Values close to 1 suggest more
flips.</p>
      <p>Ratio of beneficial to harmful flips. Indicates the balance
between favorable and unfavorable flips. Values close to
1 are desirable. Returns ∞ when there are no harmful
flips. Returns 0 when there are no beneficial flips and 1
in the absence of flips.</p>
      <p>Proportion of flips leading to unfavorable outcomes.
Values close to 1 indicate a higher incidence of harmful flips.</p>
      <p>Is 0 when there are no harmful flips. Take 1 if all the flips
present are harmful.</p>
      <p>Metric
FRD &amp; HFPD</p>
      <p>
        Boundaries
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]
      </p>
      <p>Equation
14a &amp; 14b
DI &amp; HDI
[1, ∞)</p>
      <p>15a &amp; 15b
FD &amp; HFD
[0, ∞)</p>
      <p>
        16a &amp; 16b
RFD &amp; RHFD
[
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ]
17a &amp; 17b
      </p>
      <p>Short Description &amp; Edge Cases
Absolute diferences in flips rates or harmful flips
proportions between the groups. Take 0 value when FR or
HFP are equal in both groups. Values close to 0 indicates
greater proportionality.</p>
      <p>Ratio of flip rates between groups (DI). Ratio of harmful
flip proportions between groups (HDI). Returns ∞ when
the minimum value of the flip rate or the harmful flip
proportion is 0. Returns 1 when both values are equal or
0.</p>
      <p>Quantify the disparities in the flip rates or the harmful
flip proportion between the groups, normalized by the
overall flip rate. It has the potential to become
significantly large when the overall flip rate or harmful flip
proportion approaches 0. Returns 1 when both values
are 0 and ∞ when one of them is 0.</p>
      <p>Relative disparity in flip rates between groups (RFD).
Relative disparity in harmful flips between groups (NHFD).</p>
      <p>It is a normalized measure of disparity, making it easier
to compare in diferent scenarios. Returns 1 when one
rate is 0 and the other is not, and returns 0 when there
are no flips in any of the groups.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Tversky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kahneman</surname>
          </string-name>
          ,
          <source>Judgment under Uncertainty: Heuristics and Biases</source>
          , Springer Netherlands, Dordrecht,
          <year>1975</year>
          , pp.
          <fpage>141</fpage>
          -
          <lpage>162</lpage>
          . doi:
          <volume>10</volume>
          .1007/
          <fpage>978</fpage>
          -94-
          <fpage>010</fpage>
          -1834-0_
          <fpage>8</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Castelluccia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Métayer</surname>
          </string-name>
          ,
          <article-title>Understanding algorithmic decision-making: Opportunities and challenges</article-title>
          ,
          <source>Technical Report, European Parliament Study</source>
          ,
          <year>2019</year>
          . URL: https://www.europarl.europa. eu/thinktank/en/document/EPRS_STU(
          <year>2019</year>
          )
          <fpage>624261</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Siddique</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Haque</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>George</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. D.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. J. H.</given-names>
            <surname>Faruk</surname>
          </string-name>
          ,
          <source>Survey on machine learning biases and mitigation techniques, Digital</source>
          <volume>4</volume>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>68</lpage>
          . doi:
          <volume>10</volume>
          .3390/digital4010001.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mehrabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Morstatter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galstyan</surname>
          </string-name>
          ,
          <article-title>A survey on bias and fairness in machine learning</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>54</volume>
          (
          <year>2021</year>
          ). doi:
          <volume>10</volume>
          .1145/3457607.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Caton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Haas</surname>
          </string-name>
          ,
          <article-title>Fairness in machine learning: A survey</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>56</volume>
          (
          <year>2024</year>
          ). doi:
          <volume>10</volume>
          .1145/3616865.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Barocas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. D.</given-names>
            <surname>Selbst</surname>
          </string-name>
          ,
          <article-title>Big data's disparate impact</article-title>
          , Calif. L. Rev..
          <source>California Law Review</source>
          <volume>104</volume>
          (
          <year>2016</year>
          )
          <article-title>671</article-title>
          . URL: http://lawcat.berkeley.edu/record/1127463. doi:
          <volume>10</volume>
          .15779/Z38BG31.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>C. O'Neil</surname>
          </string-name>
          , Weapons of Math Destruction:
          <article-title>How Big Data Increases Inequality</article-title>
          and
          <string-name>
            <given-names>Threatens</given-names>
            <surname>Democracy</surname>
          </string-name>
          , Crown Publishing Group, USA,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Suresh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guttag</surname>
          </string-name>
          ,
          <article-title>A framework for understanding sources of harm throughout the machine learning life cycle</article-title>
          ,
          <source>in: Proceedings of the 1st ACM Conference on Equity and Access in Algorithms</source>
          , Mechanisms, and Optimization, EAAMO '21,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2021</year>
          . doi:
          <volume>10</volume>
          .1145/3465416.3483305.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Olteanu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Castillo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. F.</given-names>
            <surname>Diaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Kıcıman</surname>
          </string-name>
          ,
          <article-title>A social data: Biases, methodological pitfalls, and ethical boundaries</article-title>
          ,
          <source>Frontiers on Big Data</source>
          <volume>2</volume>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .3389/fdata.
          <year>2019</year>
          .
          <volume>00013</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>G.</given-names>
            <surname>Pleiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Raghavan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kleinberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. Q.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          ,
          <article-title>On fairness and calibration</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          , NIPS'17, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2017</year>
          , p.
          <fpage>5684</fpage>
          -
          <lpage>5693</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>Kamiran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Karim</surname>
          </string-name>
          ,
          <string-name>
            <surname>X. Zhang,</surname>
          </string-name>
          <article-title>Decision theory for discrimination-aware classification</article-title>
          ,
          <source>in: 2012 IEEE 12th International Conference on Data Mining</source>
          ,
          <year>2012</year>
          , pp.
          <fpage>924</fpage>
          -
          <lpage>929</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICDM.
          <year>2012</year>
          .
          <volume>45</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nabi</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Shpitser</surname>
          </string-name>
          ,
          <article-title>Fair inference on outcomes</article-title>
          ,
          <source>in: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence</source>
          , AAAI'18/IAAI'18/EAAI'18, AAAI Press,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Calders</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kamiran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pechenizkiy</surname>
          </string-name>
          ,
          <article-title>Building classifiers with independency constraints</article-title>
          ,
          <source>in: 2009 IEEE International Conference on Data Mining Workshops</source>
          ,
          <year>2009</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>18</lpage>
          . doi:
          <volume>10</volume>
          .1109/ ICDMW.
          <year>2009</year>
          .
          <volume>83</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H.</given-names>
            <surname>Weerts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Royakkers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pechenizkiy</surname>
          </string-name>
          ,
          <article-title>Are there exceptions to goodhart's law? on the moral justification of fairness-aware machine learning</article-title>
          ,
          <source>in: Proc. ACM EAAMO</source>
          ,
          <year>2022</year>
          . ArXiv:
          <volume>2202</volume>
          .
          <fpage>08536</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>J.</given-names>
            <surname>Fishkin</surname>
          </string-name>
          ,
          <source>Bottlenecks: A New Theory of Equal Opportunity</source>
          , Oxford University Press,
          <year>2014</year>
          . doi:
          <volume>10</volume>
          .1093/acprof:oso/9780199812141.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F.</given-names>
            <surname>Arif Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Manis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Stoyanovich</surname>
          </string-name>
          ,
          <article-title>Towards substantive conceptions of algorithmic fairness: Normative guidance from equal opportunity doctrines</article-title>
          ,
          <source>in: Proceedings of the 2nd ACM Conference on Equity and Access in Algorithms</source>
          , Mechanisms, and Optimization, EAAMO '22,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2022</year>
          . doi:
          <volume>10</volume>
          .1145/3551624.3555303.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Miller</surname>
          </string-name>
          , Principles of Social Justice, Harvard University Press,
          <year>1999</year>
          . doi:
          <volume>10</volume>
          .2307/j.ctv1pdrq04.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Arneson</surname>
          </string-name>
          ,
          <article-title>Luck egalitarianism and prioritarianism</article-title>
          ,
          <source>Ethics</source>
          <volume>110</volume>
          (
          <year>2000</year>
          )
          <fpage>339</fpage>
          -
          <lpage>349</lpage>
          . doi:
          <volume>10</volume>
          .1086/ 233272.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>B.</given-names>
            <surname>Mittelstadt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wachter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Russell</surname>
          </string-name>
          ,
          <article-title>The unfairness of fair machine learning: Levelling down and strict egalitarianism by default</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2302.02404. arXiv:
          <volume>2302</volume>
          .
          <fpage>02404</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>E. D. P.</surname>
          </string-name>
          <article-title>supervisor, EDPS Guidelines on assessing the proportionality of measures that limit the fundamental rights to privacy and to the protection of personal data</article-title>
          ,
          <source>Technical Report, European Union</source>
          ,
          <year>2019</year>
          . URL: https://www.edps.europa.eu
          <article-title>/data-protection/our-work/publications/guidelines/ edps-guidelines-assessing-proportionality-measures_en.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>European</surname>
            <given-names>Parliament</given-names>
          </string-name>
          ,
          <article-title>Charter of fundamental rights of the european union</article-title>
          , https://eur-lex.europa. eu/eli/treaty/char_2016/oj,
          <year>2016</year>
          . Art.
          <volume>21</volume>
          -23, general proportionality test.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>UK</given-names>
            <surname>Parliament</surname>
          </string-name>
          ,
          <source>Equality act</source>
          <year>2010</year>
          , https://www.legislation.gov.uk/ukpga/2010/15/contents,
          <year>2010</year>
          .
          <article-title>Section 19-indirect discrimination and proportionality</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pedregosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Vanderplas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cournapeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brucher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Perrot</surname>
          </string-name>
          , E. Duchesnay,
          <article-title>Scikit-learn: Machine learning in python</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          . URL: http://jmlr.org/papers/v12/pedregosa11a.html.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Price</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Srebro</surname>
          </string-name>
          ,
          <article-title>Equality of opportunity in supervised learning</article-title>
          ,
          <source>in: Proceedings of the 30th International Conference on Neural Information Processing Systems</source>
          , NIPS'16, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2016</year>
          , p.
          <fpage>3323</fpage>
          -
          <lpage>3331</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>R. K. E. Bellamy</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Dey</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hind</surname>
            ,
            <given-names>S. C.</given-names>
          </string-name>
          <string-name>
            <surname>Hofman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Houde</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Kannan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Lohia</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martino</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Mehta</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Mojsilović</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Nagar</surname>
            ,
            <given-names>K. N.</given-names>
          </string-name>
          <string-name>
            <surname>Ramamurthy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Richards</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Saha</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Sattigeri</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>K. R.</given-names>
          </string-name>
          <string-name>
            <surname>Varshney</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , Ai fairness
          <volume>360</volume>
          :
          <article-title>An extensible toolkit for detecting and mitigating algorithmic bias</article-title>
          ,
          <year>2019</year>
          . doi:
          <volume>10</volume>
          .1147/JRD.
          <year>2019</year>
          .
          <volume>2942287</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>