<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>On the Correlation between Individual Fairness and Predictive Accuracy in Probabilistic Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alessandro Antonucci</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eric Rossetto</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Duvnjak</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>IDSIA (USI-SUPSI)</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lugano</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Switzerland</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>SUPSI</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lugano</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Switzerland</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>We investigate individual fairness in generative probabilistic classiefirs by analysing the robustness of posterior inferences to perturbations in private features. Building on established results in robustness analysis, we hypothesise a correlation between robustness and predictive accuracy-specifically, instances exhibiting greater robustness are more likely to be classified accurately. We empirically assess this hypothesis using a benchmark of fourteen datasets with fairness concerns, employing Bayesian networks as the underlying generative models. To address the computational complexity associated with robustness analysis over multiple private features with Bayesian networks, we reformulate the problem as a most probable explanation task in an auxiliary Markov random field. Our experiments confirm the hypothesis about the correlation, suggesting novel directions to mitigate the traditional trade-off between fairness and accuracy.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Algorithmic Fairness</kwd>
        <kwd>Bayesian Networks</kwd>
        <kwd>Markov Random Fields</kwd>
        <kwd>Robustness Analysis</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The widespread adoption of machine learning and knowledge-based systems for decision support and
recommendation has made the development of tools and metrics for algorithmic fairness increasingly
urgent [1]. Algorithmic fairness generally refers to the identicfiation and mitigation of biases in model
predictions with respect to sensitive, or private, features. However, simply omitting these private
features from the model is often insufficient: it can significantly degrade predictive accuracy and fails
to guarantee fairness, as other features may remain highly correlated with the excluded ones [2].</p>
      <p>In this paper, we focus on generative probabilistic models that produce predictions in the form of
posterior distributions over the target variable, conditioned on a complete assignment of both private
and public features. For such models, biases can be naturally characterized through the robustness
of predictions to perturbations in private features. We demonstrate how this notion of robustness
aligns with established formulations of inference under non-ignorable missingness [3]. Prior work
has further shown that robustness can be positively correlated with predictive accuracy [4]. This
connection is particularly compelling, as it suggests a mitigation strategy for the classical trade-off
between fairness and accuracy [5]: at the individual level [6], fair classifications may also be accurate,
allowing standard predictive models to be employed without modification for the robust instances.</p>
      <p>To validate such a conjecture we consider an extensive benchmark including fourteen datasets
related to fairness. In our study, we adopt Bayesian networks as probabilistic generative models,
compute the robustness, based on a fairness deviation measure [7] for each test instance, and evaluate
the correlation with the predictive performance by using the Brier score. The empirical results we
observe are in line with those reported in the earlier literature on robustness (e.g., [8]): the most robust,
and hence, in our setup, fair, instances are also “easier” to classify. To the best of our knowledge, our
work is the first attempt to apply concepts from robust probabilistic methods to fairness analysis.</p>
      <p>We leverage this novel connection to offer a more interpretable perspective on fairness. Future work
will investigate how this link might be harnessed to actively control fairness during training, as well as
evaluate its effectiveness in comparison to existing methods (e.g., [9]).</p>
      <p>The robustness measure we adopt could, in principle, require a brute-force computation over all
possible joint assignments of the private features. However, we reformulate the inference as a most
probable explanation (MPE) task within an auxiliary Markov random field derived from the original
Bayesian network. This reformulation yields computational benefits, which we both formally establish
and validate empirically.</p>
      <p>The code used for the experiments is available for reproducibility in a dedicated Github repository.1
The repository includes references to the original datasets, preprocessing scripts, discretised versions
of the datasets, and a custom implementation of some inference algorithms.</p>
      <p>The paper is organised as follows. In Section 2, we define the necessary background on probabilistic
graphical models. Our approach to fairness analysis is discussed in Section 3. The mapping to Markov
random fields is derived in Section 4. The benchmark we adopt is detailed in Section 5, while the
results of our experiments are in Section 6. A critical discussion on the limitations of the present
work and possible strategies to bypass them is in Section 7, while our conclusions are in Section 8.
Technical results and additional experimental analyses are gathered in the appendix.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Basics</title>
      <p>Let us rfist review the necessary background information on probabilistic graphical models. We
remind the reader to [10] for deeper insights.</p>
      <p>Variables, Potentials, and Tables. We denote variables by uppercase letters and their states with
lowercase letters. A subscripted variable in Ω is used instead for the sample space. Thus v ∈ ΩV
is a generic state of the variable V . We focus on discrete variables only. A non-negative function
φ : ΩV → R is called a potential over V . A potential over V that is also normalised to one is called
probability mass function (PMF) and denoted as P (V ). PMF P gives the probability P (v) to V = v
for each v ∈ ΩV . A conditional probability table (CPT) for V given W is a collection of PMFs over V
indexed by the values of W and it is denoted as P (V |W ). We also denfie potentials over joint variables.
As an example, the CPT P (V |W ) is a special example of a multi-variate potential φ(V,W ).
Bayesian Networks (BNs). Consider a set of variables  . A BN B over  is a collection of CPTs, one
for each V ∈  , say {P (V |PaV )}V ∈ where PaV ⊆  \ {V } for each V ∈  . From the BN specification B ,
we induce a directed graph GB whose nodes are in one-to-one correspondence with the variables in 
and such that there are directed arcs from W to V for each W ∈ PaV and V ∈  . Here we only consider
acyclic BNs, i.e., BNs such that GB does not contain directed loops. An acyclic BN B compactly denfies
a joint PMF P ( ) as follows:</p>
      <p>
        P () = ∏︂ P (v |paV ) , (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
      </p>
      <p>
        V ∈
for each  ∈ Ω , where the values v and paV are those consistent with . Given a dataset of i.i.d. joint
observations of  , there are various structural learning techniques to obtain a directed acyclic graph
over the variables. Typical approaches are maximising likelihood-based scores. Once the graph is
detected, the BN CPTs can be quantiefid from the data by means of a count function n, i.e.,
s
n(v, paV ) + |ΩV | ,
P (v |paV ) = ∑︁v∈ΩV n(v, paV ) + s
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
where a Laplace smoothing with equivalent sample size s &gt; 0 is preferred to a frequentist estimate.
      </p>
      <sec id="sec-2-1">
        <title>1github.com/IDSIA-papers/fairness.</title>
        <p>Given a node W , the Markov blanket W of W is a subset of  including the variable W , its parents
PaW , the children ChW of W , and the parents of those children. We denote as GW the subgraph of G
including only the variables in W . An example is in Figure 1. A BN over W and based on GW can be
obtained by taking the CPTs of W and its children from B and arbitrarily setting the marginal PMFs for
the parents of W and the parents of its children. In the following we will use uniform PMFs for these
marginals and denote such a BN as BW .</p>
        <sec id="sec-2-1-1">
          <title>Taxes</title>
        </sec>
        <sec id="sec-2-1-2">
          <title>Capital</title>
        </sec>
        <sec id="sec-2-1-3">
          <title>Race</title>
          <p>Age</p>
        </sec>
        <sec id="sec-2-1-4">
          <title>Education</title>
        </sec>
        <sec id="sec-2-1-5">
          <title>Gender</title>
        </sec>
        <sec id="sec-2-1-6">
          <title>Income</title>
        </sec>
        <sec id="sec-2-1-7">
          <title>Occupation</title>
          <p>Markov Random Fields (MRFs). A MRF M over  is a collection of k potentials {φi (i )}ik=1,
with i ⊆  for each i = 1, . . . , k and such that for each V ∈  there is at least an index i for which
V ∈ i . Given the MRF M , we induce an undirected graph GM whose k nodes are in correspondence
with the joint variables {i }ik=1 and such that there is an edge between two nodes if and only if the
corresponding joint variables share at least one variable. An example is in Figure 2. A MRF defines a
joint PMF P ( ) as follows:</p>
          <p>n
P () ∝ ∏︂ φi (i ) ,</p>
          <p>i=1
for each possible value  of  , where the values of i are those consistent with .
φ1(Education,Race)</p>
          <p>φ2(Occupation,Education)
Education
Income , Age</p>
          <p>Occupation
φ4(Capital,Taxes,Income,Age)
φ3(Income,Occupation,Gender,Age)</p>
          <p>Inference. A typical inference tasks in MRFs and BNs is updating. This consists in the computation
of the posterior PMF P (VQ |E ) for a queried variable VQ ∈  , given an evidence about some other
variables E = E with E ⊆  \ {VQ }. Marginal MAP (MMAP) consists instead of finding the most
probable congfiuration of a joint queried variable  Q ⊆  given the evidence, i.e.,
Q∗ = arg max P (Q |E ) .</p>
          <p>
            Q
If, in particular,  = (Q , E ), i.e., all the variables are either queried either observed, the task is
called most probable explanation (MPE). Moreover, notice that the maximisation in Eq. (
            <xref ref-type="bibr" rid="ref4">4</xref>
            ) is with
respect to Q , the probability of the evidence P (E ) is a positive constant and the task in Eq. (
            <xref ref-type="bibr" rid="ref4">4</xref>
            ) can be
equivalently solved by replacing the conditional probability in the r.h.s. with a joint probability.
(
            <xref ref-type="bibr" rid="ref3">3</xref>
            )
(
            <xref ref-type="bibr" rid="ref4">4</xref>
            )
          </p>
          <p>Both updating and MPE tasks can be solved exactly by variable elimination schemes. Given an
elimination order over  , this consists in processing the potential obtained by multiplying all the
potentials, including the variable we want to eliminate. The processing consists in the marginalisation
of the variables not involved in the query, the instantiation of the observed variables and the queries
of the updating task, and the maximisation of the queried variables of the MPE. The maximum arity of
the potentials created in this way is called treewidth. Inference with this scheme is exponential with
respect to this parameter.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. (Probabilistic) Fairness Analysis</title>
      <p>We are interested in evaluating the presence of biases in a training dataset D over the discrete variables
 := (Y , , ). The dataset refers to a classification task where the target variable is Y and the features
are distinguished between private, those in , and public, those in . A test dataset D′ over the same
variables is also assumed available. We learn from D a generative probabilistic model in the form of
a joint PMF P (Y , , ). In our classicfiation task, we want to identify the most probable value of Y
given a test instance (ˆ, ˆ) of the features taken from D′, i.e.,
y∗ := arg max P (y|ˆ, ˆ) ,</p>
      <p>y∈ΩY
where a zero-one loss function is considered just for the sake of light notation. Besides achieving
a good predictive level on the test instances, we might be interested in bounding the sensitivity of
inferences to changes in the values of the private features and, ideally, keep these changes as small as
possible. This could be done at the predictive level, i.e., checking whether the same class as in Eq. (5) is
obtained for perturbed values of the private features ˆ. However, a deeper analysis can be achieved by
a dissimilarity measure δ for the PMFs over the target Y . We thus consider the following optimisation:
 =
 =
arg max P (y0|, ˆ) ,</p>
      <p>∈Ω
arg min P (y0|, ˆ) .</p>
      <p>∈Ω
This is possible because of the following result whose proof is in the appendix.
and call fairness robustness level (FRL) of the instance (ˆ, ˆ) the corresponding maximum, i.e.,
∗ := arg max δ [P (Y |, ˆ), P (Y |ˆ, ˆ)] ,</p>
      <p>∈Ω
ρ(ˆ, ˆ) := δ [︁P (Y |∗, ˆ), P (Y |ˆ, ˆ)]︁ .</p>
      <p>It is a simple remark that ρ = 0 implies a context-specific irrelevance of the private features to the
target given a particular instance of the public features.</p>
      <p>Let us consider, for the sake of simplicity, the case of a Boolean target Y and take as distance δ the
Manhattan distance, i.e.,</p>
      <p>δ [︁P (Y ), P ′(Y )]︁ := |P (y0) − P ′(y0)| +2 |P (y1) − P ′(y1)| = |P (y0) − P ′(y0)| ,
where y0 is a short form for Y = 0 and similarly for y1. Because of Eq. (8), Eq. (6) rewrites as:
∗ := arg max ⃓⃓P (y0|, ˆ) − P (y0|ˆ, ˆ)⃓⃓ .</p>
      <p>∈Ω
It is a trivial exercise to notice that considering y1 instead of y0 in Eq. (9) would not affect ∗. This
implies that the FRL corresponds to the fairness deviation measure often used in the literature to rank
unfairness [11]. Rather than directly addressing the optimisation in Eq. (9), we can focus on the two
following optimisations:
(5)
(6)
(7)
(8)
(9)
(10)
(11)
Proposition 1. The solutions of the optimisations in Eqs. (9), (10) and (11) are linked by the following
relation:
∗ :=
{︃  if P (y0|ˆ, ˆ) &lt; 21 [︁P (y0|, ˆ) + P (y0|, ˆ)]︁ ,</p>
      <p>otherwise .</p>
      <p>
        It is important to notice that the task in Eq. (10) is substantially different from an MPE task as in Eq. (
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
because the queried variables in Eq. (10) are after the conditioning bar (see discussion in Section 2).
Thus, while in MPE we can equivalently address the task in the joint model instead of the conditional
one becase the evidence is always the same, the same cannot be done with Eq. (10), and similarly for
Eq. (11). These tasks can be regarded as a conservative inference based on the law proposed in [3]. Yet,
no standard BN algorithm can solve the task, while a brute-force approach requires a number of BN
inferences exponential with respect to ||. In [12], the MMAP-like optimisation has been mapped to
a credal network inference, whose complexity might be still demanding [13]. Here we leverage on the
fact that we cope with complete observations to map the task to a proper MPE task in an auxiliary
MRF. This is detailed in the next section.
4. Robustness of BN Classifiers as a MRF MPE
Thanks to Proposition 1, we can solve the optimisation in Eq. (9) and hence compute the FRL in Eq. (7)
by solving the optimisations in Eqs. (10) and (11). In this section we show how these BN tasks can be
equivalently addressed in an auxiliary MRF M whose specification is discussed here below.
      </p>
      <p>First, it is a trivial exercise to notice that, for the classification task in Eq. (5), the BN can be
equivalently restricted to the BN over the Markov blanket of Y . In the following, we therefore assume that the
BN coincides with the Markov blanket of Y .</p>
      <p>Let ′ and ′ denote, respectively, the private and public parents of Y and by ′′ and ′′ the
private and public children of Y . We define a potential φ Y (′, ′) as:
for each ′ ∈ Ω′ and ′ ∈ Ω′ . Similarly, we denfie a potential for each X ∈ ′′ and Z ∈ ′′ as follows:
(12)
(13)
(14)
(15)
(16)
(17)</p>
      <p>P (y1|′, ′)
φY (′, ′) := P (y0|′, ′)
φX (x, pa′X ) :=
φZ (z, pa′Z ) :=</p>
      <p>P (x|pa′X , y1)
P (x|pa′X , y0)
P (z|pa′Z , y1)
P (z|pa′Z , y0)
,
Solving the task in the MRF has the same worst-case complexity, exponential with respect to ||, as
trying a brute-force in the BN according to Eqs. (10) and (11). Yet, generally speaking the treewidth of
the MRF can be smaller than ||. In particular, any X ′′ ∈ ′′ that is not a parent of one of the children
of Y can be eliminated by processing φY alone, this potentially inducing a significant reduction of
the computational complexity required to solve the task. This claim will be empirically validated in
Section 6.
for each x ∈ ΩX , pa′X ∈ ΩPa′X , z ∈ ΩZ , and pa′Z ∈ ΩPa′Z , where Pa′X := PaX \ {Y } and the same notation
is also used for Z . Overall, Eq. (13) together with Eqs. (14) and (15) define a MRF φ(, ) based on
1 + |′′| + |′′| potentials. The following result, whose proof is in the appendix, holds.
Proposition 2. Consider the MRF induced by the potentials in Eq. (13) and Eqs. (14) and (15). Let
φ(, ) denote the joint PMF induced by such a MRF. The BN optimisation in Eq. (11) is equivalent to
a MPE task in this MRF, namely:
For the task in Eq. (10) we have instead:
 = arg max φ(, ˆ) .</p>
      <p>∈Ω
 = arg min φ(, ˆ) .</p>
      <p>∈Ω
5. A Benchmark of Datasets for Fairness Analysis
The main goal of this paper is to investigate whether the FRL defined in Eq. (7) has a relation with
the predictive performance of a probabilistic classifier. For a first empirical analysis, we consider
fourteen datasets providing a suitable benchmark for fairness-aware machine learning analyses. These
datasets have been also considered in [14], where the choice of their private features is discussed and
detailed. Table 1 contains some relevant information about the datasets. Apart from Communities
&amp; Crime, all datasets have a binary target variable Y . The target of that dataset is instead an integer
corresponding to the number of violent crimes in a selected area. A binary Y is obtained by a
medianbased discretisation (sample median of 374 crimes). Continuous features are similarly processed by
quantile-based discretisation leading to quaternary variables. Rows with missing values are finally
ignored.</p>
      <p>For the experiments, we consider a ten-fold cross validation scheme. The folds are obtained by
stratification with respect to the target values. For the sake of reproducibility, the code used to process
the original datasets as well as the corresponding discretised (and anonymised) datasets are available
in the Github repository of the paper (see Footnote 1). The results of our experiments on these data
are discussed in the next section.</p>
    </sec>
    <sec id="sec-4">
      <title>6. Experiments</title>
      <p>For a deeper characterisation of the relation between the FRL in Eq. (7) and the predictive performance
of the classifier, together with the zero-one accuracy of the single prediction in Eq. (5), we also consider
the more informative Brier score, that is defined here as the square of one minus the probability
assigned by the classifier to the actual class, i.e.,
β(yˆ, ˆ, ˆ) := [1 − P (yˆ|ˆ, ˆ)]2 .
(18)
A zero Brier score indicates perfect confidence in the correct class. It is easy to notice that, with binary
classes, the prediction is correct if and only if β(yˆ, ˆ, ˆ) &lt; 0.25.</p>
      <p>We adopt BNs as probabilistic classifiers and the Pyagrum 2 Python library to learn BNs from the
data and compute the queries of interest. In particular, we employ a tabu local search with list size
equal to 10 and a maximum number of consequent iterative changes equal to 100.</p>
      <p>We focus on individual fairness by evaluating complete feature instances. In this context (see
Section 4), features that lie outside the Markov blanket of the target variable are irrelevant for prediction.
When private features are entirely irrelevant to the target, the classiefir achieves fairness by design,
without any trade-off against the accuracy, and the FRL is trivially zero on all test instances. In our
cross-validation scheme, with a same dataset, we might obtain different BN topologies for the different
folds. The last column of Table 1 reports the number of folds for which the BN is fair by design.</p>
      <p>We compute the Brier score of each test instance. The low number of private features (|| ≤ 3)
allows to obtain the corresponding FRL by a brute-force approach on the BN according to Eq. (6). Yet,
for a comparison, we obtain the same results by also solving the MPE task in the auxiliary MRF as
discussed in Section 4. In the latter case, we use a custom implementation of the variable elimination
algorithm, that also allows to solve the minimisation task in Eq. (17). To solve also the latter task, we
replace each potential with its opposite.</p>
      <p>1.0
LFR0.5
0.0
1.0
0.8
0.6
L
R
F
0.4
0.2
0.0
0.4</p>
      <p>0.5</p>
      <p>Brier Score
0.0
0.1
0.2
0.3
0.6
0.7
0.8
0.9
1.0 0.0
0.21
0.32
0.50
0.61
0.73
0.84
0.91
0.95
0.98
0.5
Accuracy
1.0</p>
      <p>A first example of our empirical analysis is provided by Figure 3, where the results for the Adult
dataset are summarised. The scatter plot displays the Brier score versus the FRL value for each test
instance. Instances satisfying β(yˆ, ˆ, ˆ) &lt; 0.25—positioned in the left region of the plot—correspond to
cases that are accurately classified (approximately 84% of the total sample). For these points, because
of Eq. (9), the FRL cannot exceed the value of P (yˆ|ˆ, ˆ) = 1 − √︁β(yˆ, ˆ, ˆ). This explain why all these
points are located under a parabolic curve. Similar considerations can be done for the incorrectly
classified instances corresponding to β(yˆ, ˆ, ˆ) &gt; 0.25. This also explains the smaller FRL ranges we
observe for less extreme values of the Brier score, as they can be observed in the boxplots obtained by
equal-frequency binning with respect to the Brier score and depicted on the upper part of the figure.</p>
      <p>The most notable aspect of Figure 3 is the set of histograms on the right. These depict predictive
accuracy across ten equally sized subgroups of instances, denfied by the quantiles of the FRL
distribution. The results reveal a clear decline in prediction accuracy as FRL increase. This trend is consistent
with the conjectured relationship between accuracy and robustness/fairness introduced in Section 1.</p>
      <p>Figure 4(a) presents the same accuracy histograms alongside the corresponding Brier score levels.
Predictive performance deteriorates with increasing FRL values, as indicated by both the decline in
accuracy and the rise in Brier losses (dashed lines). A similar pattern emerges in Figure 4(b), where
the KDD Census-Income dataset is considered.
0.5</p>
      <p>FRL
0.25
0.75</p>
      <p>1
(b) KDD Census-Income</p>
      <p>As shown in Table 1, Adult and KDD Census-Income are the only datasets of our benchmark not
achieving fairness by design on any fold (i.e., such that k = 0), this meaning that no BN learned from
these datasets makes all the private features irrelevant to the target.</p>
      <p>1
0.75</p>
      <p>0.5
0.25
0</p>
      <p>0
1
0.75
0
0
0.25
0.5</p>
      <p>FRL
(a) Adult</p>
      <p>Let us now consider the datasets for which fairness by design is achieved only on some folds, but
not on all of them (i.e., 0 &lt; k &lt; 10). On these folds we have FRL= 0 by construction. For an effective
visualisation of the relation between FRL and the predictive accuracy (or Brier score) we use the same
kind of histograms as in Figure 4, with an additional bar/line of fixed width on the left side, depicting
the predictive performance on the instances with FRL= 0. The results are in Figure 5. Also for these
models we observe a clear degradation of the predictive performance for increasing FRL levels.</p>
      <p>1
0
0</p>
      <p>1
0.75
0.5
0.25</p>
      <p>1
0.75
0.5
0.25
0</p>
      <p>Accuracy (FRL &gt; 0)
Accuracy (FRL = 0)</p>
      <p>Brier Score
0.75</p>
      <p>1
Accuracy (FRL &gt; 0)
Accuracy (FRL = 0)</p>
      <p>Brier Score
0.25
0
0
0.25
0.5</p>
      <p>FRL
(b) Bank marketing
0.5
0</p>
      <p>0.25</p>
      <p>FRL
(d) COMPAS violent recid.</p>
      <p>On the remaining eight datasets, fairness is achieved by design on each fold (i.e., k = 10, see Table 1)
and we have zero FRL on all the instances in the dataset. Yet, for a further validation of our conjecture,
we re-train the BN structures by forcing arcs between the target and all the features. This can be
trivially done by constraining the local search meta-heuristics we adopt. The results, depicted in
Figure 6, confirm the trend observed in the previous experiments. We notice that degradation of the
performance for higher FRLs is typically less evident on datasets on which the predictive performance
of the BNs is poorer (such as those in Figs. 6(e)-6(h)).</p>
      <p>0.25 FRL 0.5
(a) Communities &amp; Crime
0.5
0.25
00</p>
      <p>Accuracy
Brier Score</p>
      <p>1
0.75
Accuracy
Brier Score
Accuracy</p>
      <p>Brier Score
0.75
Accuracy
Brier Score</p>
      <p>0.75
0.25 (g) OUF0LR.5ALD 0.75 1 (0h.2)5 DiabFeRtLe0s.5 0.75
Figure 6: FRL vs. accuracy/Brier for datasets where the target is forced to be a parent of the features.</p>
      <p>TFRL−BN
TFRL−MRF 3
5
1
300
100
BN− FRM−</p>
      <p>Finally, to evaluate the computational advantages of the MRF-based approach to the computation
of the FRL with respect to the brute-force BN method, Figure 7 shows boxplots of the ratio between
the execution times for each test instance across the two methods. The low number of private features
(i.e., || ≤ 3) makes the difference in performance between the two approaches small. Nevertheless,
as expected, the datasets with three private features exhibits the largest discrepancy.
9 10 11
Figure 8: Relative execution times for FRL computation in the BN (Eq. (7)) with respect to the MRF approach
(Section 4) for in increasing number of features treated as private in the Student-Mathematics dataset.</p>
      <p>
        Adult(3a)kMarketinngi(t2i)es&amp;CCOrMiPmAeS(r1)ecid.(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
n
B mu
      </p>
      <p>Cm
o</p>
      <p>Finally, Figure 8 presents the boxplots of the same ratio while treating as private an increasing
number of public features of the Student-Mathematics dataset. These results underscore the
computational efcfiiency of the MRF-based method, particularly as the number of private features
grows.</p>
    </sec>
    <sec id="sec-5">
      <title>7. Limitations and Outlooks</title>
      <p>The analysis conducted in this study is purely correlational, and the directed graphical models
employed here are not intended to capture causal relationships. Accordingly, the fairness deviation
measure we derived via the FRL does not carry a causal interpretation, and the different instantiations
of the private features should not be interpreted as interventions. The approach of Plecˇko and
Bareinboim [15] on counterfactual fairness offers a promising foundation for integrating causal reasoning
into the models discussed here.</p>
      <p>Our work focuses on discrete features for the sake of a simple presentation. An extension to a
hybrid setup mixing discrete and continuous features could be obtained within the framework of
linear-Gaussian BNs [10], for which also extensions to the causal setup have been already discussed.
The situation is different for the target variable: extending the framework to regression tasks with a
continuous target would require a substantial reformulation of the concepts introduced in this paper.
In contrast, our focus on binary classes can be naturally extended to non-binary targets by conducting
classification and FRL evaluations through pairwise dominance tests applied to each pair of classes.</p>
      <p>Regarding the measure considered to evaluate the robustness of the posterior inferences, the current
focus on the Manhattan distance has clear computational advantages related to the linear structure of
such measure. Coping with non-linear alternatives (such as the KL) could be much more challenging,
while an extension to variance and similar measure of variation could be more manageable, and
naturally allow for the extension to continuous targets.</p>
      <p>Once our framework will be extended to hybrid models that incorporate both continuous and
discrete variables, the experimental evaluation should be revisited. A more renfied representation
of the model variables is expected to enhance the overall predictive accuracy. Assessing whether the
discriminative power of the FRL remains stable under these conditions constitutes a necessary future
investigation. A validation under controlled bias scenarios3 should be also considered.</p>
      <p>Finally, extending the framework from individual to group fairness would be conceptually
straightforward, although it would preclude the use of the MRF mapping. To maintain computational tractability
and avoid costly inference procedures, more efficient models-such as probabilistic circuits [ 16]—could
serve as practical alternatives to BNs.</p>
    </sec>
    <sec id="sec-6">
      <title>8. Conclusions</title>
      <p>We examined whether the correlation between robustness and predictive accuracy—commonly
observed in the efild of robust probabilistic models—also holds in a fairness-aware setting. The results
of an extensive empirical evaluation are promising: predictive accuracy consistently declines for less
robust instances. This observation opens new avenues for mitigating the classical fairness–accuracy
trade-off. Specifically, instances with high fairness scores can be handled by standard models, while
a fairness-constrained, potentially less accurate model could be selectively applied to non-robust
instances. Exploring this targeted approach to fairness mitigation represents an exciting and necessary
direction for future research in machine learning.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>The research of Eric Rossetto was funded by Swiss Post AG as part of a Doctoral Studies Grant on AI
and Robustness.</p>
      <sec id="sec-7-1">
        <title>3See, e.g., github.com/rcrupiISP/BiasOnDemand.</title>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>For the preparation of this work, the author(s) used ChatGPT, Grammarly in order to: Grammar and
spelling check, Paraphrase and reword. After using this tool/service, the author(s) reviewed and edited
the content as needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-9">
      <title>A. Proofs</title>
      <p>In this appendix, we gather the proofs of the two technical results of the paper.</p>
      <p>Proof of Proposition 1. The following straightforward result:
mtax ⃓⃓ f (t )⃓⃓ = max {︂max [︁ f (t )]︁ , max [︁− f (t )]︁}︂ ,</p>
      <p>t t
allows to rewrite the maximum of the optimisation in Eq. (9) as follows:
max
{︃
max [︁P (y0|, ˆ) − P (y0|ˆ, ˆ)]︁ , max [︁P (y0|ˆ, ˆ) − P (y0|, ˆ)]︁
∈Ω ∈Ω
}︃
As P (y0|ˆ, ˆ) is a constant, we can move the maximisation with respect to  and obtain the maximum
and the minimum in Eqs. (10) and (11), i.e.,</p>
      <p>max {︁[︁P (y0|, ˆ) − P (y0|ˆ, ˆ)]︁ , [︁P (y0|ˆ, ˆ) − P (y0|, ˆ)]︁}︁ .</p>
      <p>The above maximum between two quantities can be clearly rewritten as:</p>
      <sec id="sec-9-1">
        <title>And hence:</title>
        <p>{︃ P (y0|, ˆ) − P (y0|ˆ, ˆ) if P (y0|, ˆ) − P (y0|ˆ, ˆ) &gt; P (y0|ˆ, ˆ) − P (y0|, ˆ) ,</p>
        <p>P (y0|ˆ, ˆ) − P (y0|, ˆ) otherwise .</p>
        <p>{︃ P (y0|, ˆ) − P (y0|ˆ, ˆ) if P (y0|, ˆ) + P (y0|, ˆ) &gt; 2P (y0|ˆ, ˆ) , ,</p>
        <p>P (y0|ˆ, ˆ) − P (y0|, ˆ) otherwise
which can be seen as an equivalent formulation of Eq. (12), where the values of the maxima appear
instead of the points where these maxima are achieved.</p>
        <p>Proof of Proposition 2. As we cope with a Boolean target variable, i.e., |ΩY | = 2, the objective function
of the optimisations in Eqs. (10) and (11) rewrites as follows:
As f (t ) = (1 + t )−1 is a monotone decreasing function of t , we can address Eq. (11) as follows:
P (y0|, ˆ) =</p>
        <p>P (y0, , ˆ)</p>
        <p>P (, ˆ)</p>
        <p>P (y0, , ˆ)
= P (y0, , ˆ) + P (y1, , ˆ) =</p>
        <p>1</p>
        <p>P(y1,,ˆ)
1 + P(y0,,ˆ)
 := arg max
∈Ω P (y0, , ˆ)</p>
        <p>.</p>
        <p>
          P (y1, , ˆ)
The BN factorisation in Eq. (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) allows to rewrite the objective function in Eq. (25) as follows:
P (y1, , ˆ) ∏︁V ∈(Y ,,) P1(v|pav ) ∏︂
P (y0, , ˆ) = ∏︁V ∈(Y ,,) P0(v|paV ) = V ∈(Y ,,) P0(v|paV )
        </p>
        <p>P1(v|pav ) ,
where P1 denotes the elements of the BN CPTs when the states (v, paV ) are those consistent with (y1, , ˆ)
and P0 when consistent with (y0, , ˆ). The product in the rightmost hand side of Eq. (26) can be
restricted to ratios of CPT values including Y , i.e., the CPTs associated with Y and its children. This
corresponds to the joint PMF of the MRF and proves Eq. (16). The result in Eq. (17) follows analogously.
(19)
(20)
(21)
(22)
(23)
(24)
(25)
(26)</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>B. FRL-vs-Brier-Score Curves</title>
      <p>The scatter plot in Figure 9 details the output of an experiment with the Bank marketing dataset in
terms of relation between FRL levels and Brier scores. Here, as well as in similar plots we obtained for
the other datasets, the points seem to lie on different curves. In the gfiure, two groups of seven points,
each belonging to a different curve, have been highlighted in black. The points in a same group have
the same value of the private feature marital status and are obtained by changing the public features.
Namely, the points with higher Brier scores are those associated with single, and the other ones with
married. We also observe that each point in a group has a corresponding point in the other group with
the same FRL value. These pairs correspond to instances with the same public states and different
private state. The reason is that, as a straightforward consequence of the definition, two instances
with the same public feature have the same FRL level.</p>
      <p>1.0
0.8
0.6
L
R
F
0.4
0.2
0.0
0.0
0.2
0.4</p>
      <p>0.6
Brier Score
Approaches to Reasoning with Uncertainty, Springer International Publishing, Cham, 2021, pp.
284–298.
[5] A. F. Cooper, E. Abrams, N. Na, Emergent unfairness in algorithmic fairness-accuracy trade-off
research, in: Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, 2021, pp.
46–54.
[6] C. Dwork, M. Hardt, T. Pitassi, O. Reingold, R. Zemel, Fairness through awareness, in: Proceedings
of the 3rd innovations in theoretical computer science conference, 2012, pp. 214–226.
[7] B. Woodworth, S. Gunasekar, M. I. Ohannessian, N. Srebro, Learning non-discriminatory
predictors, in: S. Kale, O. Shamir (Eds.), Proceedings of the 2017 Conference on Learning Theory,
volume 65 of Proceedings of Machine Learning Research, PMLR, 2017, pp. 1920–1953.
[8] L. Mattei, A. Antonucci, D. D. Mauá, A. Facchini, J. Villanueva Llerena, Tractable inference in
credal sentential decision diagrams, International Journal of Approximate Reasoning 125 (2020)
26–48.
[9] H. Xu, X. Liu, Y. Li, A. Jain, J. Tang, To be robust or to be fair: Towards fairness in adversarial
training, in: M. Meila, T. Zhang (Eds.), Proceedings of the 38th International Conference on
Machine Learning, volume 139 of Proceedings of Machine Learning Research, PMLR, 2021, pp.
11492–11501.
[10] D. Koller, N. Friedman, Probabilistic Graphical Models: Principles and Techniques, MIT, 2009.
[11] N. Konstantinov, C. H. Lampert, Fairness through regularization for learning to rank, arXiv
preprint arXiv:2102.05996 (2021).
[12] A. Antonucci, M. Zaffalon, Equivalence between Bayesian and credal nets on an updating
problem, in: J. Lawry, E. Miranda, A. Bugarin, S. Li, M. A. Gil, P. Grzegorzewski, O. Hryniewicz
(Eds.), Proceedings of third international conference on Soft Methods in Probability and Statistics
(SMPS-2006), Springer, Bristol, United Kingdom, 2006, pp. 223–230.
[13] D. D. Mauá, C. P. de Campos, A. Benavoli, A. Antonucci, Probabilistic inference in credal networks:
new complexity results, Journal of Artificial Intelligence Research 50 (2014) 603–637.
[14] T. Le Quy, A. Roy, V. Iosifidis, W. Zhang, E. Ntoutsi, A survey on datasets for fairness-aware
machine learning, WIREs Data Mining and Knowledge Discovery 12 (2022).
[15] D. Plecˇko, E. Bareinboim, Causal fairness analysis: A causal toolkit for fair machine learning,</p>
      <p>Foundations and Trends® in Machine Learning 17 (2024) 304–589.
[16] Y. Choi, A. Vergari, G. Van den Broeck, Probabilistic circuits: A unifying framework for tractable
probabilistic models, UCLA Technical Reports (2020).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mitchell</surname>
          </string-name>
          , E. Potash,
          <string-name>
            <given-names>S.</given-names>
            <surname>Barocas</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. D'Amour</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Lum</surname>
          </string-name>
          , Algorithmic fairness: Choices, assumptions, and definitions,
          <source>Annual review of statistics and its application 8</source>
          (
          <year>2021</year>
          )
          <fpage>141</fpage>
          -
          <lpage>163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Cummings</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kimpara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Morgenstern</surname>
          </string-name>
          ,
          <article-title>On the compatibility of privacy and fairness, in: Adjunct publication of the 27th conference on user modeling, adaptation</article-title>
          and personalization,
          <year>2019</year>
          , pp.
          <fpage>309</fpage>
          -
          <lpage>315</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G.</given-names>
            <surname>De Cooman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zaffalon</surname>
          </string-name>
          ,
          <article-title>Updating beliefs with incomplete observations</article-title>
          ,
          <source>Articfiial Intelligence</source>
          <volume>159</volume>
          (
          <year>2004</year>
          )
          <fpage>75</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. V.</given-names>
            <surname>Llerena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. D.</given-names>
            <surname>Mauá</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Antonucci</surname>
          </string-name>
          ,
          <article-title>Cautious classicfiation with data missing not at random using generative random forests</article-title>
          , in: J.
          <string-name>
            <surname>Vejnarová</surname>
          </string-name>
          , N. Wilson (Eds.), Symbolic and Quantitative
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>