<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Kraków, Poland
* Corresponding author.
" ludwig.bothmann@stat.uni-muenchen.de (L. Bothmann); susanne.dandl@stat.uni-muenchen.de (S. Dandl);
michael.schomaker@stat.uni-muenchen.de (M. Schomaker)
~ https://www.slds.stat.uni-muenchen.de/people/bothmann/ (L. Bothmann)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Causal Fair Machine Learning via Rank-Preserving Interventional Distributions</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ludwig Bothmann</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Susanne Dandl</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Schomaker</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Center for Infectious Disease Epidemiology, School of Public Health, University of Cape Town</institution>
          ,
          <country country="ZA">South Africa</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Statistics, LMU Munich</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>A decision can be defined as fair if equal individuals are treated equally and unequals unequally. Adopting this definition, the task of designing machine learning models that mitigate unfairness in automated decision-making systems must include causal thinking when introducing protected attributes. Following a recent proposal, we define individuals as being normatively equal if they are equal in a fictitious, normatively desired (FiND) world, where the protected attribute has no (direct or indirect) causal efect on the target. We propose rank-preserving interventional distributions to define an estimand of this FiND world and a warping method for estimation. Evaluation criteria for both the method and resulting model are presented and validated through simulations and empirical data. With this, we show that our warping approach efectively identifies the most discriminated individuals and mitigates unfairness.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;fairness in ML</kwd>
        <kwd>causal thinking</kwd>
        <kwd>interventional distributions</kwd>
        <kwd>stochastic interventions</kwd>
        <kwd>rank-preserving interventions</kwd>
        <kwd>quasi-individual fairness</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        ML models in ADM systems. Following Aristotle [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], they define a treatment as being fair “if
equals are treated equally and if unequals are treated unequally”. Furthermore, they distinguish
between descriptively unfair treatment (which can occur without PAs) and normatively unfair
treatment (which is a causal notion). For this, they conceive a fictitious, normatively desired
(FiND) world, where the PA has no (direct or indirect) causal efect on the target variable.
Individuals are normatively considered equal if they are equal in the FiND world.
      </p>
      <p>We build upon this work by proposing concrete estimands and estimation procedures. As a
starting point, we define a directed acyclic graph (DAG) that describes the causal relations in
the real world. The DAG in the FiND world is then created by deleting all arrows that constitute
paths from the PA to the target. We achieve this through specific stochastic interventions
leading to rank-preserving interventional distributions. This intervention is rank-preserving in
the following sense: Individuals of the disadvantaged group maintain the rank they have in the
real world (compared with other individuals of the disadvantaged group) as population-wide
rank in the FiND world (compared with all individuals), see Section 3.1.</p>
      <p>After identifying the estimand, we propose a warping method for estimation that maps
realworld data to a warped world which in turn approximates the FiND world, see Section 3.2.
We call this a quasi-individual approach because individual “merits” are pulled through to the
warped world. Finally, an ML model is trained on the warped data that can be used at prediction
time after warping the new observation. Since final prediction models are trained and evaluated
in the warped world, our approach can be considered to be a pre-processing approach [see 4,
for a categorization of diferent approaches in fairML]. We propose evaluation metrics both for
evaluating the warping method in a simulation study and for evaluating an ML model using
warped data in an applied use case in Section 4. In a simulation study, we show that our warping
method is able to approximate the FiND world, identify the most discriminated individuals, and
eliminate the efects of the PA in the warped world (Section 5). Finally, we apply the proposed
methodology to German Credit data, showing how to use our framework in practice (Appendix
A).</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        In addition to group fairness concepts (see, e.g., [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for an overview), approaches of (non-causal)
individual fairness have been proposed, starting with [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], who require that similar individuals
should be treated similarly (see also [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">7, 8, 9</xref>
        ]). An early notion of causal fairness was made by
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], who conceive a fictitious world where an individual belongs to a diferent subgroup of
the PA, defining a decision as fair if it is equal in the real and fictitious world. For a thorough
explanation of how this difers from our FiND world, see [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Including causality in the fairness
debate and conceiving a fictitious world was also proposed by, e.g., [
        <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14 ref15 ref16">11, 12, 13, 14, 15, 16</xref>
        ], where
diferent ideas underlie those fictitious worlds. With the fairness concept introduced by [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], we
distinguish the real world and the FiND world by the idea that in the FiND world, there must be
no causal efects from the PA on the target – neither indirectly, nor directly. This means that we
delete all arrows starting in the PA and eventually leading to the target variable (dashed arrows
in Figure 1). This idea difers from what the literature on path-specific efects [e.g., 17, 18, 19]
conceives. However, we believe that this more adequately captures the legal requirements of
many laws that demand that individuals must not be discriminated against based on the PA1 –
rendering it irrelevant which path in the DAG this efect follows.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>
        As derived in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], in order to derive the decision basis for fair decisions, we must conceive
a “fictitious, normatively desired (FiND) world in which the PA has no causal efect” on the
target variable, “neither directly nor indirectly”. In the following, we adopt this idea, elaborate
it further by concretely specifying causal and statistical estimands, and derive an estimation
method, thereby building concrete and actionable algorithms for approximating the FiND world
by what we call a “warped world” and for using Causal Fair ML (cfML) in applied use-cases.
      </p>
      <p>Our method consists of four fundamental steps: (i) We first define the estimand as the
joint distribution in the FiND world, described by stochastic interventions, leading to
rankpreserving interventional distributions (see Section 3.1); (ii) We then estimate the joint/conditional
distributions of interest in the FiND world, based on a specific -formula type-of factorization
that follows from specifying the respective identification assumptions and allows us to “warp”
the real-world data into the warped world (see Section 3.2.1); With this, we can (iii) train an ML
model (for predicting the target) in this warped world (see Section 3.2.2), and (iv) predict on
a new observation in the warped world using the above warping models and ML model (see
Section 3.3).</p>
      <sec id="sec-3-1">
        <title>3.1. Estimand</title>
        <sec id="sec-3-1-1">
          <title>3.1.1. DAGs in the Real and the FiND World</title>
          <p>
            Deriving a DAG falls into the realm of Causal Discovery (see, e.g., [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ] for a review of current
methods). Since this is a notoriously hard challenge in practice, the alternative is to define
the DAG with expert knowledge, as is typically done in epidemiology and medicine, where
knowledge from human decision-makers is readily available [
            <xref ref-type="bibr" rid="ref21">21</xref>
            ]. In the remainder, we are
agnostic to the question of how the DAG was constructed and will assume that all DAGs are
correct; note that this may be an optimistic (and untestable) assumption and can hamper success
in practice.
          </p>
          <p>Two DAGs must be defined: the DAG in the real world and the DAG in the FiND world –
where the PAs have no causal efect on the target. Figure 1 shows the two DAGs that we assume
for the example of the German Credit data. Note that these DAGs are chosen for illustrative
purposes and not because there is empirical evidence or expert knowledge that justifies exactly
those DAGs. We reduced the feature set for a clearer presentation: Age (a confounder )
is the numerical age of an individual; Gender (the PA ) is assumed to be binary (classes
female and male) in the remainder, but note that an extension on multi-categorical gender is
methodologically straightforward; Savings (feature ) is a binary variable, indicating if the
person has small savings (1) or not (0); Amount (feature ) is the amount of credit applied
for; and Risk (target  ) is the binary risk category with values good (1) and bad (0).
1E.g., Charter of Fundamental Rights of the European Union: https://www.citizensinformation.ie/en/
government-in-ireland/european-government/eu-law/charter-of-fundamental-rights/
Gender ()</p>
          <p>Amount ()</p>
          <p>Savings ()</p>
          <p>Risk ( )</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>3.1.2. Rank-Preserving Interventional Distributions</title>
          <p>
            There are several possible interventions that can delete the dashed arrows in Figure 1 and, hence,
lead to the FiND world DAG. We propose the following idea of “rank-preserving interventional
distributions”, which we believe to be the best way of defining those interventions when aiming
to mitigate unfairness. We assume that the given DAGs (as shown in Figure 1) correctly mirror
the causal relationships in both the real world and the FiND world. Slightly adapting the
notation and terminology of [
            <xref ref-type="bibr" rid="ref22">22</xref>
            ], a general structural causal model (SCM) is given by
 :=  (( ),  ),  ∈ {1, . . . , },
where 1, . . . ,  denote exogeneous independent random variables, and () are parent
nodes of . In our example, the SCM in the real world (i.e., pre-intervention) is given by
 :=  ()
 :=  ( )
 := (, , )
 := (, , )
          </p>
          <p>
            :=  (, , , ,  ),
which entails a joint distribution that can be factorized according to our working order:
 (, , , , ) =  ( |, , , )(|, )(|, ) ()(). (1)
For the FiND world, we must make all descendants from the PA neutral w.r.t. the PA. We
achieve this by a fictitious intervention rule  on the mediators and outcome only, i.e., no
“modification” of the potentially sensitive PA is required (Eq. 2). This intervention leads to a
joint post-intervention distribution (, ,   ,   ) in which the dashed arrows have

been removed; thus, no efect of Gender on the mediators and the outcome exists – but the
distributions of males and females are comparable and still in line with the data-generating
process on which we want to train our ML model. Additionally, our suggested intervention
is “rank-preserving” in the sense that the quantile of female customers within their strata is
transported into the FiND world (see Figure 2a). Thereby, all PA-dependent quantities are
transformed into their FiND-world counterparts. Note that we can factorize the joint
postintervention in line with the pre-intervention distribution (Eq. 1), where the mediators and
outcomes are replaced by their post-intervention counterparts. This leads to a -formula
type of factorization, which we can use for plug-in estimation of the relevant counterfactual
distributions. A similar, quantile-based approach, can be found earlier in [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ] which uses
quantile regression forests for estimation.
 =
⎧⎪ () = ˜()
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪ () = ˜()
⎪⎪⎪ 
⎪
⎪
⎪
⎨
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪⎪  () = ˜()
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎪
⎩
where ˜() is the (() × 100)% quantile of the conditional
mediator distribution among the reference PA value, i.e.,
 ( ≤ ˜()| = (),  = ) = (), and () is determined
by the pre-intervention quantile of unit , i.e.,
() =  ( ≤ () |  = (),  = ()).
where ˜() is the (() × 100)% quantile of the conditional
mediator distribution among the reference PA value, i.e.,
 ( ≤ ˜()| = (),  = ) = (), and () is determined
by the pre-intervention quantile of unit , i.e.,
() =  ( ≤ ()| = (),  = ()).
where ˜() is the (() × 100)% quantile of the counterfactual
outcome distribution for the reference PA value, i.e.,
 ( ≤ ˜() | ˜(), ˜(),  = (),  = ) = (), and () is
based on the pre-intervention quantile of unit , i.e.,
() =  ( ≤ () |  = (),  = (),  = (),  = ()).
(2)
          </p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Estimation</title>
        <p>
          We base our estimation algorithm on the factorization derived above, i.e., we use the empirical
distributions of both  and  to implement the intervention to sequentially obtain the respective
post-intervention distributions of , , and  . To determine the distributions and quantiles
needed to facilitate the intervention implementation, our proposed algorithm uses the empirical
distributions for the PA reference group (i.e., male customers) and a residual-based approach
for the non-reference group (i.e., female customers). Alternatively, we could data-adaptively
estimate the quantiles from the conditional distributions [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ], but we do not pursue this more
complicated approach further in this manuscript.
        </p>
        <p>More generally, we approximate the FiND world by “warping” the target and the features
afected by the PA (see Figure 2b). Once a preprocessed dataset containing “cleaned” features and
target variables is available, standard ML techniques can be applied, prioritizing high predictive
performance. This approach does not necessitate the incorporation of “fairness metrics” in the
training process.2 The three key steps of our proposed algorithm are:
2For a more in-depth exploration of the philosophical rationale and the potential implications regarding the
introP(XA | f)
5%</p>
        <p>x(Ai)
P(XA | m)
5%
~(i)
xA</p>
        <p>Amount
Amount
(a)
(b)
(1) Derive a warping from the real world to the warped world (see Section 3.2.1)
(2) Train and test an ML model using the warped data (see Section 3.2.2).</p>
        <p>(3) At the time of prediction, use warping and trained ML model (see Section 3.3)</p>
        <sec id="sec-3-2-1">
          <title>3.2.1. Warping for Approximating the FiND World</title>
          <p>
            We propose to implement the interventions defined above (see Eq. 2) by the following
residualbased estimation method. For determining the female intervention values, we must estimate
– for each variable to be warped – (i) the individual probability rank of female  (e.g., () for
variable ) and (ii) the corresponding quantile of the male distribution (e.g., ˜()). This means
that we must estimate full distributions (not just location parameters) of | = ,  = 
for all values of  and  (analogously for  and  ), which becomes prohibitively complex
in situations with finite data and numeric confounders  or features * . In our algorithm
proposed below, we reduce estimation complexity by only estimating models for the location
parameters of these distributions and derive individual probability ranks by using residuals of
those models. The five steps of this warping algorithm are explained for feature Amount ( ),
warping for other variables works analogously:
(1) Estimate prediction models  () for the female and  () for the male population, where
we are agnostic on the model class and can choose any ML model, since we only rely on point
duction of unfairness to the ADM system utilizing the trained ML model, please refer to [
            <xref ref-type="bibr" rid="ref2">2</xref>
            ].
predictions and model residuals on training data.
(2) Compute residuals as
() =  (())
          </p>
          <p>− () ∀ ∈  ,
() =  (())</p>
          <p>− () ∀ ∈ ,
() =

|{ ∈  : 
()
 ≤ 
()
 |}
| |
.
where  and  are the female and male index set, respectively.
(3) Compute the individual probability rank of female  as ranked within the female residuals,
i.e., telling us how “exceptionally high or low” her value is in comparison to other females, by
(4) Set () to the empirical ()-quantile of the residuals of the male model  , i.e.,
() = min{ ∈  : |{ ∈  :  ≤ }|</p>
          <p>≥ ()},
||
where  = {() :  ∈ } is the set of male residuals.
(5) Warp 
() to the sum of male prediction and warped residual, i.e.,</p>
          <p>ˆ() =  (()) + ().</p>
          <p>
            Analogously, we can warp  and  (where in the latter case, warped values of Amount and
Savings must be plugged into the male prediction in step (5)). However, note that for warping
of non-continuous variables (such as Savings and Risk), we define the models to predict the
probability scores, not the hard labels. That way, the warped values, e.g., ˆ
binary, but may be ∈ [
            <xref ref-type="bibr" rid="ref2">− 1, 2</xref>
            ]. If we need hard labels – e.g., for learning a binary prediction
model, such as for the target variable  in Section 3.2.2 – we can simply threshold these scores.
On the other hand, for use in further warping steps (such as for warping of ()), we can directly
use the raw values by plugging them into the prediction function, thereby pulling through finer

(), are no longer
information than if we would threshold earlier in the process.
          </p>
          <p>Now, we have warped all Gender-dependent quantities ((), 
()

, ()) of female
individuals to their warped world counterparts (ˆ , ˆ
()
()
, ˆ()), approximating their FiND world
counterparts (˜ , ˜
()</p>
          <p>()
︁(
xˆ(1), ˆ(1))︁ , . . . , xˆ(), ˆ())︁
︁(
, ˜()) ∀ ∈  . To have a complete warped world data set  =</p>
          <p>, we set warped male values and values of non-warped features
(e.g., Age) to their real-world value. Additionally to having warped the training data, we have
also estimated warping functions that can be applied for new test data at the time of prediction.</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>3.2.2. Training ML Models in the Warped World</title>
          <p>We can now use the warped world data  to train a prediction model for the warped target ˆ .
Assuming that the warping cleaned the data from any PA discrimination, we do not have to
account for any fairness metrics in this training step but can just focus on training a model that
has high predictive performance. Since we assume that all Gender-related discrimination was
eliminated through the warping, we do not use Gender  as a feature in this model (see Section
5 for an investigation of what happens if this assumption is wrong, e.g., due to a misspecified
DAG). As a result, we obtain a trained model  (xˆ) which can be used for prediction.</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Prediction</title>
        <p>Warp New Data. Consider a new observation x* = (* , *, * , * ). If this is a male
observation, no warping must be done; if this is a female observation, we use the estimated warping
functions of Section 3.2.1 as follows for  and analogously for  (but not for  ):
(1) Compute individual residual * w.r.t. female model  (* ) as * =  (* ) − *.
(2) Compute individual probability rank * w.r.t. female population  as above.
(3) Set * to the empirical * -quantile of training data residuals of male model   as above.
(4) Warp * to the sum of male prediction and warped residual, i.e., ˆ* =  (* ) + *.
After carrying out the same steps for warping  , we finally obtain the warped observation
xˆ* = (ˆ*, ˆ* , * ) (recall that we do not use Gender as a feature in the prediction model).
Predict New Data. For predicting the target in the warped world ˆ* , we plug the warped
observation xˆ* into the prediction model trained on the warped world data, i.e., ˆ* =  (xˆ* ).</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Evaluation</title>
      <p>We propose evaluation criteria that can be used for two purposes: Section 4.1 describes how
to evaluate our proposed warping method for rank-preserving interventional distributions in
a simulation study. Section 4.2 describes how the warped data and resulting ML models can
be evaluated in an applied use-case. We denote with ˆ(), ˆ(), and  ˆ () the predicted target of
individual  in the real, warped, and FiND world, respectively.</p>
      <sec id="sec-4-1">
        <title>4.1. Evaluation of Warping Method</title>
        <p>We can evaluate our warping method w.r.t. (i) the warped data, asking, e.g., if the FiND world is
recovered by the warping and w.r.t (ii) the final ML model – using the warped data.
(W1) Recovering FiND world. In a simulation study we can compare the warped and the
FiND world distributions for investigating if the warping procedure recovers the FiND world.
For numerical features, we compare warped world and FiND world empirical distributions by
Kolmogorov-Smirnov (KS) tests, and for binary features, we use binomial tests. Additionally,
we use a t-test to test the null hypothesis that there is no discrimination in the warped world
between male and female subgroups w.r.t. risk predictions. If the method works, p-values of
these tests should be consistently high, indicating that the null hypotheses cannot be rejected.
(W2) Identifying strongest discriminated individuals. In addition to the
populationwide perspective of (W1), we are interested in the individual perspective, i.e., if the warping
method also recovers the individual ranks of the FiND world w.r.t. the target variable prediction.
If this would be the case, we could identify individuals who are most strongly afected by
discrimination in the real world by comparing real world and warped world predictions in
an applied use case. For the warped class of the PA, we compute individual risk prediction
diferences between the real world and the warped world, (1) = ˆ() − ˆ() and between the
real world and the FiND world, (2) = ˆ() −  ˆ (), respectively. We use a t-test to test the null
hypothesis that the means of these diferences are equal. If the method works, p-values of
these tests should be consistently high, and diferences () 2
1 − () should be small. Correlation
between ranks of () and () should be high, too.</p>
        <p>1 2</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Evaluation in an Applied Use-Case</title>
        <p>How can the model be evaluated in an applied use case, i.e., how can we know if the warping
method worked and if it removed unfairness? In our opinion, this cannot be answered by
evaluating the final ML model w.r.t some “classical” fairML metrics. 3 Once we have successfully
warped the data from the real to the warped world (approximating the FiND world), we reduced
the problem to finding a model with good predictive performance. However, we can train
models in the real and the warped world and then compare their behavior:
(UC1) Comparing performance. Test performance of the ML models in the real world
ˆ(· ) and the warped world ˆ(· ) can be compared, assuming that both models tfi “their” world
equally well. However, this must not be misinterpreted as either of those models being better
than the other one, as the models are merely modeling diferent worlds.
(UC2) Comparing predictions and identifying strongest discriminated individuals.
For each individual , the predictions in the real and the warped worlds can be compared
by computing the diference ()</p>
        <p>
          1 . As for (W2), this analysis can reveal individuals that are
discriminated most in the real world (either positively or negatively). Additionally, these
diferences can be aggregated on the subgroup level, and tests can be computed to test the null
hypothesis that predictions do not change between the two worlds for the respective subgroup.
(UC3) Identifying strongest warped individuals. We can also ask which individuals are
afected most by the warping. These individuals’ feature vectors have the largest distance
between the real and the warped world, i.e., we can compare x() and x˜() by a suitable distance
metric for each individual  ∈ {1, . . . , }.
3As elaborated on thoroughly in [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] (main points were summarized above), these kinds of metrics (such as
demographic parity, equalized odds, etc.) do not reflect a clearly defined concept of fairness and, hence, are not suitable
for deciding if an ML model entails unfairness. However, as readers might still be interested in the respective values
– in the sense of a descriptive or explorative analysis –, we provide the resulting metrics in Appendix B.
(UC4) Identifying important features. For each feature, we can compare the empirical
distributions in the real and the warped world, i.e., of  and ˜ for each  ∈ {1, . . . , }.
We compute distances for each (normalized) feature and, thereby, can identify features that
vary most between the two worlds, indicating that these features carry most of the real-world
discrimination w.r.t. the PA.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Simulation Study</title>
      <p>For investigating the behavior of the proposed method, we first conduct a simulation study
where we know the true DAG in both the real and the FiND world. Subsequently, we apply
the methods to the German Credit dataset in Appendix A. We seek to answer the following
research questions:
(RQ1) Does our warping method work as expected? In other words, does this method recover
the distributions in the FiND world (W1), and is it able to correctly identify the individual
ranks of the target in the FiND world (W2)?
(RQ2) How does misspecification of the DAG afect the results?
(RQ3) What efects does the direction of warping have on performance (e.g., if subgroup A of
the PA is warped to subgroup B, versus the other way around)?</p>
      <sec id="sec-5-1">
        <title>5.1. Simulation Setup</title>
        <p>Data simulation setup. We simulate data from the DAGs depicted in Figure 1. Here, the
real-world data simulation contains all arrows, while the FiND world data simulation only
contains solid arrows by setting Gender to male for all observations, thereby eliminating the
Gender efect. The distributions utilized here are (left: real-world, right: FiND world): 4
 ∼ B( )
 ∼ Ga(  ,   )
|,  ∼ Ga( (, ),  (, ))
|,  ∼ B( (, ))
 ∼ B( )
 ∼ Ga(  ,   )
˜| ∼ Ga( (, ),  (, ))
˜| ∼ B( (, ))
 |, , ,  ∼ B(  (, , , ))
˜ |˜, ˜,  ∼ B(  (˜, ˜, , )),
where we use linear combinations of the features combined with a log- and logit-link for the
Gamma and Binomial models, respectively, and mirror the Gender distribution of the German
Credit data with   = 31% females. We perform  = 1, 000 simulations on data sets of size
 = 10, 000 for training and of size  = 1, 000 for test, for each world, using the same
seed for the two worlds to ensure comparability. Note that Gender and Age are then identical in
both worlds, and only the descendants of Gender have difering values. We refer to this setup
as (SIM1). To answer the misspecification behavior question (RQ2), we modify the simulation
slightly by sampling Age from  ∼ Ga(  (),   ()) but ignoring this efect for warping.
We refer to this setup as (SIM2).
4Concrete values can be found in simulation_study.R in https://github.com/slds-lmu/paper_2023_cfml
Warping and prediction models. For warping models, we estimate models following the
same distributional assumptions as in the simulation, i.e., by estimating the parameter vectors of
the Gamma and logistic regressions separately for male and female observations of the training
data. With these models, we apply the above warping strategy. As prediction models for the
target variable, we train logistic regression models on the training data, warped training data,
and FiND world training data, separately.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Results</title>
        <p>With these models, we can now answer the above research questions (using a significance
threshold of  = 5% for all tests):</p>
        <p>(RQ1) Figure 3a shows the distribution of Amount  in the diferent worlds for male and
female observations, aggregated over all iterations of the simulation study. The null hypothesis
of equal distributions in the warped and the FiND world is only rejected in 0%, 0.4%, 0%
of the iterations for Amount, Savings, and Risk, respectively. The mean diference between
male and female risk predictions in the real world is 0.1122 (95% CI: (0.1117, 0.1127)). In
the warped world, this is reduced to − 0.0016 (− 0.0021, − 0.0012), meaning that even if the
diference between subgroups is still significantly non-zero, it is smaller by a factor of 70, i.e.,
we efectively reduced PA discrimination (and the direction switched from positive to negative).</p>
        <p>Investigating individual predictions (W2), we see that correlations between ranks in the
warped and the FiND world are high (0.892). Figure 3b shows individual risk prediction
diferences between the real world and the warped world as well as between the real world and
the FiND world for females in one iteration (males are identical in the real world and the FiND
world). The most strongly negatively afected individuals are at the upper end of the distribution.
As shown, the most discriminated individuals (large diference between the FiND world and
the real world) are correctly identified (large diference between the warped world and the
real world). In 81% of iterations, the null hypothesis of equal diferences 1 and 2 cannot be
rejected. In cases with  &lt; 0.05, the mean diference is − 0.0023 – meaning that the deviation
between the warped world and the FiND world is also minimal in these cases. We conclude that
warping (i) recovers the marginal distributions in the FiND world, (ii) diminishes discrimination
to a very small value, and (iii) correctly identifies the most discriminated individuals.</p>
        <p>(RQ2) The null hypothesis of equal distribution in the warped world and the FiND world is
rejected in 17%, 4%, 0% of the iterations for Amount, Savings, and Risk, respectively. In the real
world, the mean diference between risk predictions is 0.1723 (0.1718, 0.1728), which is higher
than in (SIM1). In the warped world, this is reduced to 0.0355 (0.0350, 0.0360) (reducing by a
factor of 4.9) – far less than above. The correlation of ranks (0.9518) is higher than above, since
discrimination in the FiND world is higher in (SIM2). In 6.9% of iterations, the null hypothesis
of equal diferences 1 and 2 cannot be rejected. In cases with  &lt; 0.05, the mean diference
is 0.026 – far higher than above. We conclude that misspecification of the DAG is a relevant
factor for degrading the performance of our approach.</p>
        <p>(RQ3) By switching the warping direction and warping male to female values, we observe
the following: Recovering of marginal FiND world distributions is equally successful as when
warping female to male values. The null hypothesis is rejected in 0%, 0.3%, 0% of the iterations
for Amount, Savings, and Risk, respectively (see also Figure 3c). The mean diference between
risk predictions in the warped world is reduced to 0.0065 (0.0060, 0.0071) – slightly worse than
in the analysis of RQ1, which is due to the imbalance of the data. The mean correlation of ranks
compared with ranks of RQ1 is high (0.9595), meaning that individual ranks are comparable
for both warping directions. In 34% of iterations, the null hypothesis of equal diferences 1
and 2 cannot be rejected. In cases with  &lt; 0.05, the mean diference is 0.0073, meaning that
the deviation between the warped world and the FiND world is also minimal in these cases
(although a bit higher than in the analysis of RQ1, due to data imbalance). This means we can
also mitigate discrimination and preserve individual ranks with changing the warping direction.
Most interestingly, the general level of the risk predictions changes, as shown in Figure 3d. This
also makes sense, since we are now warping male to female values.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Discussion</title>
        <p>We have shown that for the simulation setup above, our proposed method works as expected,
recovering the marginal distributions in the FiND world and individual ranks; direction of
warping does not make a relevant diference. However, as this is just an initial study, these
investigations should be extended by follow-up work. As subsequent investigations, we would
propose to (at least): (i) consider other, diverse DAGs, (ii) compare diferent ML models for
warping and target prediction, and (iii) investigate behavior on other empirical data sets.</p>
        <p>A general limitation of our method is that it depends on knowing the true DAG. As shown
in RQ2, misspecifying the DAG degrades the performance of the method. Hence, special care
should be given to identifying the true DAG in an applied use case by strongly connecting
expert knowledge on the subject matter and rigorous application of causal discovery methods.</p>
        <p>Practical feasibility: The computational costs of our method are rather small for the presented
analysis. The computations for the German credit data (learning warping models, training
models in both worlds, warping, and predicting in the warped world) took 0.17 seconds on a
3,4 GHz Intel Core i5, using one core of the CPU. This increases with, e.g., (i) the data size, (ii)
the complexity of the DAG (since more models have to be trained), and (iii) the ML models used
(e.g., training a multilayered neural networks takes more time than training the logit models
used here). We do not expect the computational time to be a relevant constraint – also because
learning the diferent models for warping can be parallelized. From a practical viewpoint, the
much more important challenge is finding the DAG, since this involves carefully interweaving
expert knowledge and application of methods of causal discovery [see, e.g., 20].</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion and Outlook</title>
      <p>We have presented rank-preserving interventional distributions as a framework to identify a
FiND world where no causal efects of a PA exist. Additionally, we have proposed a warping
method for estimating FiND world distributions with real-world data. A simulation study showed
that the method works for the investigated simulation setup (see Section 5.3 for limitations), and
we demonstrated in the Appendix how the method can be applied to empirical data (Appendix
A). Analyses can be reproduced via a public GitHub repository, which also contains code for
applied use cases.5 Apart from extending the study as outlined in Section 5.3, further work
should compare our method to other methods that conceive a fictitious world for tackling
fairness issues of ML models (see references in Section 2).</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We thank Holger Löwe for helping with visualizations.
5https://github.com/slds-lmu/paper_2023_cfml</p>
    </sec>
    <sec id="sec-8">
      <title>A. German Credit Data</title>
      <p>DAG, warping, and prediction models. We assume the same DAG as in the simulation
study, depicted in Figure 1. For warping and prediction models, we use the same models as in
the simulation study.</p>
      <p>Evaluation. For the evaluation of the behavior of our method for this applied use case, we
use the evaluation strategies defined in Section 4.2. Models are trained on randomly sampled
80% of the training data (i.e., 800 from 1,000 observations) and tested on the remaining 20%.</p>
      <p>(UC1) Test accuracy in the real world is 71% for both the male and the female subgroup. In
the warped world, male accuracy is comparable, with test accuracy of 70%. However, female
accuracy increases to 75%, showing increasing performance for the discriminated subgroup.</p>
      <p>(UC2) Table 1 shows individuals whose predictions difer most in the two worlds, either
positively or negatively. Regressing this diference on features reveals that the risk prediction of
young women grows strongly through warping, indicating that this subgroup was discriminated
against most strongly in the real world (see Figure 4a). Figure 4b compares female predictions
in both worlds and shows the most strongly afected individuals. Figure 4c shows prediction
diferences for female and male subgroups. While mean diferences for females are significantly
positive ( &lt; 10− 12), the mean diferences for males do not change significantly (  = 0.26).
However, individual predictions and ranks of males do change: Figure 4d shows partial efects
of Age and Amount on the prediction diference.
(UC3) Investigating the efect of warping on the individuals reveals similar results as
investigating the prediction diferences in (UC2) and are omitted for the sake of concise presentation.</p>
      <p>(UC4) The normalized feature diferences between the real world and the warped world for
Age, Amount, and Savings are 0.00, 0.01, 0.24, respectively. This reveals that Savings is afected
most by the warping and, hence, carries the strongest discrimination efect in the real world.
n
o
iit
c
repd .06
k
s
i
R</p>
      <p>Real world</p>
      <p>Warped world
(a) Female Age efect on predictions.</p>
      <p>(b) Risk predictions for females in two worlds.</p>
      <p>Prediction difference warped−real</p>
      <p>Partial effect</p>
      <p>Partial effect
5
0
) .0
2
2
.,
5
e
g
(sA .000
e
c
feden .010
ifr
n
o
iit
c
d
reP .005
0
1
.
0
5
0
.
0
−
0
2
.
0
5
1
.
0
0
0
.
0</p>
      <p>Female</p>
      <p>Male
20 30 40 50 60 70</p>
      <p>Age</p>
    </sec>
    <sec id="sec-9">
      <title>B. Classical FairML Metrics</title>
      <p>
        As elaborated on thoroughly in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], “classical“ fairML metrics such as demographic parity,
equalized odds, etc., do not reflect a clearly defined concept of fairness and, hence, are not
suitable for deciding if an ML model entails unfairness. Hence, they can also not be used as
quality criteria for evaluating our warping method. However, since some of these metrics are
still popular, one might be interested – from a descriptive or explorative point of view – how
these metrics change after applying the above proposed warping approach. For this reason, we
provide respective results in the following, strongly emphasizing that such results are neither
suitable for proving nor for disproving that our method works.
      </p>
      <p>
        Simulation study. For the simulation study described above, Tables 2 – 4 summarize some
group fairness metrics (see, e.g., [
        <xref ref-type="bibr" rid="ref4 ref5">5, 4</xref>
        ] for an overview). We display ratios of diferent metrics,
comparing the male and female subgroup, where a value smaller than 1 indicates that the
respective metric in the male subgroup is larger than in the female subgroup (i.e., female value
divided by male value). The tables show the following mean values (averaged over simulation
iterations)6:
• ACC: Ratio of subgroup-specific accuracies, a.k.a. Overall accuracy equality
• PPV: Ratio of subgroup-specific positive predictive values (precisions), a.k.a. Predictive
parity
• FPR: Ratio of subgroup-specific false positive rates, a.k.a. Predictive equality
• TPR: Ratio of subgroup-specific true positive rates, a.k.a. Equal opportunity
• STP: Ratio of subgroup-specific positively predicted rates, a.k.a. Statistical parity or
      </p>
      <p>Demographic parity
• No. checks passed: In each iteration, we check for each of the values of ACC, PPV,
FPR, TPR, STP if it is inside the interval (, 1 ), where we use  = 0.95 as tolerance
value. This number reports the total number of checks passed, which is ∈ {0, 1, 2, 3, 4, 5}
for each iteration, i.e., the mean (as reported in the tables) is ∈ (0, 5).</p>
      <p>
        Table 2 shows that for the scenario of correctly specified DAG and using the larger subgroup
(male) as a reference group, the metrics are considerably closer to 1 for the warped and FiND
world, compared with the real world. Warped and FiND world values are very close, only for
FPR there seems to be a (small) diference between warped and FiND world values.
6We used the R Package fairmodels [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
female
      </p>
      <p>STP</p>
      <p>No. checks passed
STP
German Credit Data. Figure 5 shows the same metrics for the analysis of the German credit
data. For the real world, 1 check is passed (equal opportunity ratio), where for the warped
world, all 5 checks are passed.</p>
      <p>1
score
1
score</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Barocas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Narayanan</surname>
          </string-name>
          ,
          <source>Fairness and Machine Learning</source>
          ,
          <year>2019</year>
          . URL: http: //www.fairmlbook.org.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Bothmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bischl</surname>
          </string-name>
          , What Is Fairness? Philosophical Considerations and Implications For FairML,
          <year>2023</year>
          . doi:
          <volume>10</volume>
          .48550/arXiv.2205.09622.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Aristoteles</surname>
          </string-name>
          , Aristotelis Opera, volume
          <volume>2</volume>
          of ex rec. Immanuelis Bekkeri ed.
          <source>Acad. Regia Borussica</source>
          , de Gruyter,
          <year>1831</year>
          . Book V.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Caton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Haas</surname>
          </string-name>
          ,
          <article-title>Fairness in Machine Learning: A Survey, ACM Computing Surveys (</article-title>
          <year>2023</year>
          )
          <article-title>3616865</article-title>
          . doi:
          <volume>10</volume>
          .1145/3616865.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Verma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rubin</surname>
          </string-name>
          ,
          <article-title>Fairness definitions explained</article-title>
          ,
          <source>in: Proceedings of the International Workshop on Software Fairness</source>
          ,
          <string-name>
            <given-names>ACM</given-names>
            ,
            <surname>Gothenburg</surname>
          </string-name>
          <string-name>
            <surname>Sweden</surname>
          </string-name>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .1145/3194770. 3194776.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Dwork</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hardt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pitassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Reingold</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zemel</surname>
          </string-name>
          ,
          <article-title>Fairness through awareness</article-title>
          ,
          <source>in: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference</source>
          , Association for Computing Machinery, New York, NY, USA,
          <year>2012</year>
          , pp.
          <fpage>214</fpage>
          -
          <lpage>226</lpage>
          . doi:
          <volume>10</volume>
          . 1145/2090236.2090255.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bechavod</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Jung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Metric-Free Individual Fairness in Online Learning</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M. F.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2020</year>
          , pp.
          <fpage>11214</fpage>
          -
          <lpage>11225</lpage>
          . URL: https://proceedings.neurips.cc/paper/2020/file/ 80b618ebcac7aa97a6dac2ba65cb7e36-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Chouldechova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roth</surname>
          </string-name>
          ,
          <source>The Frontiers of Fairness in Machine Learning</source>
          , arXiv:
          <year>1810</year>
          .
          <volume>08810</volume>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1810</year>
          .
          <volume>08810</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Friedler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scheidegger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Venkatasubramanian</surname>
          </string-name>
          ,
          <article-title>On the (im)possibility of fairness</article-title>
          ,
          <source>arXiv:1609.07236</source>
          (
          <year>2016</year>
          ). doi:
          <volume>10</volume>
          .48550/arXiv.1609.07236.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>M. J. Kusner</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Loftus</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Russell</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Silva</surname>
            , Counterfactual Fairness, in: I. Guyon,
            <given-names>U. V.</given-names>
          </string-name>
          <string-name>
            <surname>Luxburg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bengio</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wallach</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Fergus</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Vishwanathan</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>30</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2017</year>
          . URL: https://proceedings.neurips.cc/paper/2017/file/ a486cd07e4ac3d270571622f4f316ec5-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. Bareinboim,
          <article-title>Equality of Opportunity in Classification: A Causal Approach</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>31</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2018</year>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2018/hash/ f1418e8cc993fe8abcfe3ce2003e5c5-Abstract.html.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. Bareinboim,
          <article-title>Fairness in Decision-Making - The Causal Explanation Formula</article-title>
          ,
          <source>in: Proceedings of the AAAI Conference on Artificial Intelligence</source>
          ,
          <year>2018</year>
          . doi:
          <volume>10</volume>
          .1609/ aaai.v32i1.
          <fpage>11564</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nabi</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Shpitser</surname>
          </string-name>
          ,
          <article-title>Fair inference on outcomes</article-title>
          ,
          <source>in: Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence</source>
          , AAAI'18/IAAI'18/EAAI'18, AAAI Press, New Orleans, Louisiana, USA,
          <year>2018</year>
          , pp.
          <fpage>1931</fpage>
          -
          <lpage>1940</lpage>
          . doi:
          <volume>10</volume>
          .5555/3504035.3504270.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Malinsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Shpitser</surname>
          </string-name>
          , Learning Optimal Fair Policies,
          <source>in: Proceedings of the 36th International Conference on Machine Learning, PMLR</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4674</fpage>
          -
          <lpage>4682</lpage>
          . URL: https://proceedings.mlr.press/v97/nabi19a.html.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Malinsky</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Shpitser</surname>
          </string-name>
          ,
          <article-title>Optimal Training of Fair Predictive Models</article-title>
          ,
          <source>in: Proceedings of the First Conference on Causal Learning and Reasoning</source>
          , PMLR,
          <year>2022</year>
          , pp.
          <fpage>594</fpage>
          -
          <lpage>617</lpage>
          . URL: https://proceedings.mlr.press/v177/nabi22a.html.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S. R.</given-names>
            <surname>Pfohl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Duan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Y.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. H.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <article-title>Counterfactual Reasoning for Fair Clinical Risk Prediction</article-title>
          ,
          <source>in: Proceedings of the 4th Machine Learning for Healthcare Conference</source>
          , PMLR,
          <year>2019</year>
          , pp.
          <fpage>325</fpage>
          -
          <lpage>358</lpage>
          . URL: https://proceedings.mlr.press/v106/pfohl19a.html.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Chiappa</surname>
          </string-name>
          ,
          <string-name>
            <surname>Path-Specific Counterfactual</surname>
            <given-names>Fairness</given-names>
          </string-name>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>33</volume>
          (
          <year>2019</year>
          )
          <fpage>7801</fpage>
          -
          <lpage>7808</lpage>
          . doi:
          <volume>10</volume>
          .1609/aaai.v33i01.
          <fpage>33017801</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chikahara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sakaue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Fujino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kashima</surname>
          </string-name>
          ,
          <article-title>Learning Individually Fair Classifier with Path-Specific Causal-Efect Constraint</article-title>
          ,
          <source>in: Proceedings of The 24th International Conference on Artificial Intelligence and Statistics</source>
          , PMLR,
          <year>2021</year>
          , pp.
          <fpage>145</fpage>
          -
          <lpage>153</lpage>
          . URL: https: //proceedings.mlr.press/v130/chikahara21a.html.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N.</given-names>
            <surname>Weinberger</surname>
          </string-name>
          ,
          <string-name>
            <surname>Path-Specific</surname>
            <given-names>Efects</given-names>
          </string-name>
          ,
          <source>The British Journal for the Philosophy of Science</source>
          <volume>70</volume>
          (
          <year>2019</year>
          )
          <fpage>53</fpage>
          -
          <lpage>76</lpage>
          . doi:
          <volume>10</volume>
          .1093/bjps/axx040.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A. R.</given-names>
            <surname>Nogueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pugnana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ruggieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pedreschi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gama</surname>
          </string-name>
          ,
          <article-title>Methods and tools for causal discovery and causal inference</article-title>
          ,
          <source>WIREs Data Mining and Knowledge Discovery</source>
          (
          <year>2022</year>
          ). doi:
          <volume>10</volume>
          .1002/widm.1449.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Hernán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Robins</surname>
          </string-name>
          , Causal Inference: What If, Boca Raton: Chapman &amp; Hall/CRC,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Pearl</surname>
          </string-name>
          ,
          <source>Causality: Models, Reasoning and Inference</source>
          , 2nd ed., Cambridge University Press,
          <year>2009</year>
          . URL: https://yzhu.io/courses/core/reading/04.causality.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>D.</given-names>
            <surname>Plečko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Meinshausen</surname>
          </string-name>
          ,
          <article-title>Fair Data Adaptation with Quantile Preservation</article-title>
          ,
          <source>Journal of Machine Learning Research</source>
          <volume>21</volume>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>44</lpage>
          . URL: http://jmlr.org/papers/v21/
          <fpage>19</fpage>
          -
          <lpage>966</lpage>
          .html.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Hejazi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Benkeser</surname>
          </string-name>
          , I. Díaz,
          <string-name>
            <surname>M. J. van der Laan</surname>
          </string-name>
          ,
          <article-title>Eficient estimation of modified treatment policy efects based on the generalized propensity score</article-title>
          ,
          <year>2022</year>
          . doi:
          <volume>10</volume>
          .48550/ arXiv.2205.05777.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wiśniewski</surname>
          </string-name>
          , P. Biecek,
          <article-title>fairmodels: A Flexible Tool For Bias Detection</article-title>
          ,
          <string-name>
            <surname>Visualization</surname>
          </string-name>
          , And
          <string-name>
            <surname>Mitigation</surname>
          </string-name>
          ,
          <year>2022</year>
          . URL: http://arxiv.org/abs/2104.00507, arXiv:
          <fpage>2104</fpage>
          .00507 [cs, stat].
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>