<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Mitigating Fairness Risks in Machine Learning under Structured Missing Data</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tim Kochs</string-name>
          <email>tkochs@techfak.uni-bielefeld.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Castellani</string-name>
          <email>andrea.castellani@honda-ri.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Felix Lanfermann</string-name>
          <email>felix.lanfermann@honda-ri.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Barbara Hammer</string-name>
          <email>bhammer@techfak.uni-bielefeld.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Machine Learning, Missing Values, Structured Missingness, Fairness, Imputation</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AG Machine Learning, Bielefeld University</institution>
          ,
          <addr-line>33615, Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Honda Research Institute Europe GmbH</institution>
          ,
          <addr-line>Carl-Legien-Str. 30, 63073, Ofenbach am Main</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Handling missing values in machine learning presents significant challenges, impacting both model predictive performance and fairness. While unstructured missingness, where values are randomly absent, has been extensively studied, structured missingness remains relatively underexplored. Structured missingness occurs when the absence of one or more features influences the absence of others. In this paper, we investigate the impact of both structured and unstructured missingness on fairness. We introduce a novel approach for simulating structured missingness and propose a training methodology designed to enhance model robustness under these conditions. We evaluate how diferent state-of-theart imputation methods for handling missing data afect fairness across multiple benchmark tabular datasets. Our empirical results demonstrate that structured missingness leads to a degradation in model fairness, particularly when the missingness mechanism is conditioned on protected attributes such as race or gender. In these cases, minority groups experience higher error rates, contributing to disparities in metrics like equalized odds. We propose a preprocessing step, Missingness Robustness Augmentation, that is shown to increase model robustness towards the presence of missing values.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        As machine learning systems increasingly influence high-stakes decisions in healthcare [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ],
ifnance [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], and criminal justice [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ], concerns about fairness and bias have become central
to their responsible deployment [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">7, 8, 9</xref>
        ]. Data, the foundational element of these systems, is
often incomplete, leading to challenges in achieving fair and unbiased outcomes [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13">10, 11, 12, 13</xref>
        ].
In particular medical application data sets often encounter missing values due to high costs,
limited availability, and a high variety of practices among medical entities, to name just a
few mechanisms [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ]. Conventionally, missing data is classified as missing completely
at random (MCAR), missing at random (MAR), or missing not at random (MNAR) [
        <xref ref-type="bibr" rid="ref11 ref12">12, 11</xref>
        ].
However, these frameworks often overlook a more nuanced form of incompleteness known as
Structured Missingness (SM), recently introduced by [
        <xref ref-type="bibr" rid="ref15 ref17">15, 17</xref>
        ]. SM refers to scenarios where the
absence of data is not merely random but exhibits complex, systematic patterns, often reflecting
structural inequalities within the data collection pipeline.
      </p>
      <p>
        This missingness mechanism is particularly prevalent in the medical domain [
        <xref ref-type="bibr" rid="ref16 ref18 ref19">18, 19, 16</xref>
        ]. To
name just one example: in oncology, decisions regarding neoadjuvant therapya pre-surgical
intervention such as chemotherapy or radiationdepend on a range of patient-specific factors. For
instance, in rectal cancer, tumors located in anatomically challenging regions may require
preoperative radiation to enable less invasive surgery. However, patients with certain comorbidities
(e.g., cardiovascular disease or connective tissue disorders) may not be eligible for radiation,
resulting in deviations from standard treatment protocols. These clinically informed decisions
can lead to systematic and non-random patterns of missing data, particularly in treatment and
outcomes records, which may introduce bias if not appropriately modeled.
      </p>
      <p>AEQUITAS 2025: Workshop on Fairness and Bias in AI | co-located with ECAI 2025, Bologna, Italy
(CC BY 4.0).</p>
      <p>CEUR
Workshop</p>
      <p>ISSN1613-0073</p>
      <p>
        Current state-of-the-art methods for handling missing data are largely impute-then-classify
approaches [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. However, imputation methods are rarely evaluated on
downstream-taskperformance, but mostly on the reconstruction error or variations thereof [
        <xref ref-type="bibr" rid="ref21 ref22">21, 22</xref>
        ]. It is still an
active area of research how imputation impacts machine learning models [
        <xref ref-type="bibr" rid="ref23 ref24 ref25 ref26">23, 24, 25, 26</xref>
        ], and
there is even evidence, that switching imputation methods negatively impacts model performance
[
        <xref ref-type="bibr" rid="ref23">23</xref>
        ].
      </p>
      <p>
        Besides model performance, fairness concerns towards a machine learning model play an
increasingly important role. It has already been extensively studied how imputation impacts
fairness [
        <xref ref-type="bibr" rid="ref27 ref8">27, 28, 8, 29</xref>
        ]. Yet these approaches do not target SM, albeit it might play an important
role in this context since minority groups might be more inclined to not provide an answer to
certain questions or afected by a bias in the data collections process [
        <xref ref-type="bibr" rid="ref15 ref8">15, 8</xref>
        ]. Therefore, in this
contribution, we address the widely open research question of how SM impacts the fairness of
ML-models. Our main contributions are threefold:
1. We propose a novel procedure to introduce SM in data, which allows extensive evaluation
of its efect across multiple datasets.
2. We empirically evaluate the impact of SM on the fairness of state-of-the-art imputation
methods and show that fairness can deteriorate even when SM is present only during
inference.
3. We present a novel preprocessing method, Missingness Robustness Augmentation, designed
to improve robustness to test-time SM. Our method is model-agnostic and demonstrates
consistent improvements in fairness metrics without degrading predictive performance.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Types of missingness. The fundamental theoretical background of missingness mechanisms has
been defined in [
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ], coining the mechanism missing completely at random (MCAR), missing
at random (MAR) and missing not at random (MNAR). In the recent work [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], the concern was
raised that these definitions do not form a complete characterization of missingness mechanisms.
This observation led to the definition of structured missingness (SM) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], which extends the
definition of missingness to also incorporate weak structures (WS) and strong structures (SS),
i.e., probabilistic or deterministic relationships of features with missing values. These additional
structures can occur with each of the missingness types in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] i.e., MAR, MNAR.
Missing data imputation. Ignoring missing data can introduce a bias into statistical analyses.
As demonstrated in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], valid inference generally requires that the missing data mechanism is
MAR and the missingness process is independent of the parameter of interest. Consequently,
imputation techniques are commonly employed to address missing data [
        <xref ref-type="bibr" rid="ref20">30, 20</xref>
        ]. There is a
wide variety of imputation methods in the literature [
        <xref ref-type="bibr" rid="ref21 ref22 ref26">31, 32, 33, 34, 35, 36, 21, 26, 22</xref>
        ], which
can be broadly classified into statistical methods, general machine learning (ML) methods,
or deep learning (DL) methods. In the statistical method, simple imputation, all missing
values of a feature are replaced by a single value such as the mean or mode [30]. This is often
used as a baseline as it might introduce bias in high dimensional datasets [37]. Methods like
Multivariate Imputation by Chained Equations (MICE) [32] ofer state-of-the-art statistical
methods. Referring to ML, popular technologies include  -nearest-neighbors (KNN) imputation
[38], MissForest [33], and SVM imputation [39]. DL methods are becoming increasingly popular
at present, yet they require large data sets in order to outperform statistical imputation methods
[
        <xref ref-type="bibr" rid="ref22">40, 34, 41, 22, 42, 43</xref>
        ].
      </p>
      <p>
        Albeit imputation ofers a versatile technique of dealing with missing values, there is some
concern about bias that might be introduced when used in conjunction with a machine learning
pipeline [
        <xref ref-type="bibr" rid="ref23">23, 44</xref>
        ]. Furthermore many imputation methods focus on imputing numerical values,
such that their application to categorical values might introduce additional bias [45]. There are
also few approaches that do not require a separate imputation step, but are capable of dealing
with missing values natively [46, 47, 48]. Those are just mentioned for completeness, since these
methods are not within the scope of our contribution.
      </p>
      <p>Fairness. There exist various (often mutually exclusive) formalizations of fairness in machine
learning which account for diferent normative intuitions and practicability [ 49]. A common
distinction refers to group fairness metrics (e.g., demographic parity, equalized odds) [50], which
refers to distinct groups of persons as characterized by a specific sensitive attribute such as
gender or ethnicity, and individual fairness metrics (e.g., counterfactual fairness, consistency)
[51]), which focus on a fair treatment of individual persons in comparisons to the whole group.
Due to their eficient evaluation, group fairness methods such as equalized odds which essentially
refer to the distribution of errors among the vulnerable group as compared to the reference group,
enjoy a wide popularity. In contrast, individual fairness measures which rely on semantically
meaningful modeling of the scenario such as a latent causal model as used within counterfactual
fairness, are sometimes praised as particularly meaningful, but might sufer from lack of causal
information, or even reduce to statistical group methods in practice [52].</p>
      <p>
        Fairness and missing values. As pointed out in [53], missing values are inherently related to
the topic of algorithmic fairness. Currently, it is common practice within the fairness literature
to simply remove missing values [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], albeit it likely introduces bias into the data [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. A growing
body of literature is concerned with missing values in fair machine learning [
        <xref ref-type="bibr" rid="ref8">54, 28, 29, 44, 8</xref>
        ].
Some fairness measures have already been adapted for their use with missing values [55], and
ifrst fairness-aware imputation methods have been proposed [ 56, 28].
      </p>
      <p>
        Yet, there is a lack of work evaluating common imputation methods on downstream task
performance, with recent work providing explorations of diferent missingness mechanism [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. In
this contribution, we extend this efort by exploring downstream task performance and fairness,
whereby we put a particular focus on SM and the impact of missing values during both training
and test-time. Moreover, as a first remedy, we ofer insights into robust training under the
assumption of structured missingness.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Methods</title>
      <sec id="sec-3-1">
        <title>3.1. Problem Statement and Notation</title>
        <p>Let X ∈ ℝ× denote a dataset with  samples and  features. The missingness in the data is
represented by a binary indicator matrix M ∈ {0, 1}× , where   = 1 if   is missing and 0
otherwise. The primary missing data mechanisms are the following:
Missing Completely at Random (MCAR): The probability of missingness is independent of both
observed and unobserved data. Formally,</p>
        <p>
          ( M ∣ X) =  ( M)
Under MCAR, the missingness does not introduce bias in parameter estimates, allowing
for unbiased analysis using complete cases [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ].
        </p>
        <p>Missing at Random (MAR): The probability of missingness may depend on observed data but
not on the missing data itself:</p>
        <p>
          ( M ∣ X) =  ( M ∣ Xobs)
This assumption permits the use of methods like multiple imputation to obtain unbiased
estimates, provided that the model for the missingness mechanism is correctly specified
[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>
          Missing Not at Random (MNAR): The probability of missingness depends on unobserved data:
 ( M ∣ X) ≠  ( M ∣ Xobs)
MNAR mechanisms require modeling the missingness process for unbiased estimates, as
the missingness is informative and cannot be ignored [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>Structured missingness (SM) refers to scenarios where the pattern of missing data exhibits
dependencies among variables or across observations. Each version of structure can be applied
to each missingness mechanism, as example we state the definition of MAR. The two notable
structure types are:
MAR with Weak Structure (MAR-WS): Missingness in one variable depends on observed data
and the missingness indicators of other variables:</p>
        <p>
          ( M ∣ Xobs, M− )
where M is the missingness indicator for variable  , and M− represents missingness
indicators for other variables [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>MAR with Strong Structure (MAR-SS): Missingness in one variable depends on observed data,
missingness indicators, and the missing values of other variables:</p>
        <p>
          ( M ∣ Xobs, Xmis, M− )
This scenario is more complex and requires advanced modeling techniques to handle the
dependencies efectively [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Generation of Benchmarks with Structured Missingness</title>
        <p>Having SM in the data can have a significant impact on the fairness of a predictive model,
especially if the protected attribute is related to the missingness structure. In order to systematically
analyze SM in a controlled environment, we introduce Structured Feature Dropout, detailed in
algorithm 1, a novel method to induce SM within datasets. This approach centers on a sensitive
attribute to simulate realistic missingness patterns pertinent to algorithmic fairness studies.
While tailored for fairness evaluations, the methodology is adaptable for broader applications
involving structured missingness.</p>
        <p>First of all, to induce a structure in the missingness mechanism, a domain knowledge of the
problem is needed. We propose to construct a Bayesian Network (BN), a probabilistic graphical
model that represents attributes as random variables and their conditional dependencies. This
approach aims to maximize the likelihood of generating the observed dataset. Specifically, we
utilize the Bayesian networks previously developed by [53]. Although BNs are used here for
their flexibility in capturing feature dependencies, other graph-based approaches may also be
suitable. We would like to stress that it is not necessary to construct a complete causal model
to implement this procedure, but any knowledge of protected features would sufice.</p>
        <p>To induce missingness, first, we generate an unstructured missingness mask by randomly
masking an  fraction of features correlated with the sensitive attribute. Next, we induce
structured missingness by masking an additional  fraction of features directly related to the
target label, reflecting the worst-case scenario in which the missingness mechanism directly
impacts predictive features.</p>
        <p>Figure 1 provides a visual illustration of the procedure described in algorithm 1. By
incorporating domain knowledge and data-driven insights, we construct a graphical model that
captures feature dependencies. Missingness is then introduced according to a specified
mechanism; for instance, in the depicted example, a Missing At Random (MAR) strategy is employed,</p>
        <p>A
3 25 ge
34 ⋮ 8</p>
        <p>Local-gov S⋮ome-colPergievate⋮ PrivaHteS-gWradorkcla1s1sth edDucaatiotna
⋯ ⋯ ⋯ ⋯ ⋯</p>
        <p>indge
ma le
o w
D o
n</p>
        <p>K
Protected 1</p>
        <p>Feature 1
Feature 2</p>
        <p>Feature 4</p>
        <p>Protected 2
Feature 3</p>
        <p>Unstructured
Missingness</p>
        <p>Feature 7</p>
        <p>Feature 5
Feature 8</p>
        <p>Feature 6</p>
        <p>Protected 2</p>
        <p>Feature 3</p>
        <p>Feature 1
Feature 2</p>
        <p>Feature 4
Label
Protected 1</p>
        <p>Feature 1
Feature 2</p>
        <p>Feature 4</p>
        <p>Protected 2
Feature 3
Feature 7</p>
        <p>Feature 5</p>
        <p>Feature 7</p>
        <p>Feature 5
Feature 8</p>
        <p>Feature 6</p>
        <p>Feature 8</p>
        <p>Feature 6
where the probability of missingness depends on an observed variable, specifically the attribute
labeled "Protected 1". Red dashed edges indicate the dependency paths used to determine
which features are selected for missingness in a given step, while red crosses denote features
in which missing values were introduced at that stage. Red crosses indicate features in which
missing values were artificially introduced according to the missingness mechanism. This process
results in a missingness mask M, with selected feature values removed. In this simulation, SM
was introduced solely based on Protected 1, ensuring that Feature 3 remained fully observed
throughout. For clarity and consistency across datasets we selected only a single protected
attribute; the approach naturally extends to settings involving multiple protected variables.
In a second iteration, missingness is induced in features that are directly associated with the
target label, thereby simulating a more structurally dependent form of missingness. Features
previously marked with black crosses remain unchanged and do not receive additional missing
values.</p>
        <sec id="sec-3-2-1">
          <title>Algorithm 1: Structured Feature Dropout</title>
          <p>Input: Dataset  ∈  × ; probability distributions  , 
Output: Modified dataset  with structured missingness (SM)
clean(X) ; // Optional preprocessing or normalization
 ← computeFeatureRelation( ) ; // Determine feature relationships
M ← missingMask( ,  ,  ) ; // Apply base-level missingness
M′ ← structuredMask(M,  , ) ; // Inject structured missingness
setMissing( , M + M′) ; // Apply total missingness mask
return D</p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Robust Training for Missing Values</title>
        <p>Whenever missing values occur during deployment of a machine learning model, it might be
susceptible to reduced performance and fairness. To counteract this efect, we propose Missingness
Robustness Augmentation, a novel method which aims to increase the robustness of an ML
model when dealing with structured missingness. The core idea of our proposed method is to
enrich the training data with artificially induced missing values. By exposing the model to
these conditions during training, we improve its ability to generalize when faced with similar
missingness during inference. Importantly, our approach enhances fairness robustness without
compromising predictive performance.</p>
        <p>The proposed method is listed in algorithm 2. Missingness Robustness Augmentation augments
the training data by randomly sampling a proportion  of the dataset and artificially removing a
predefined set of features from these samples. This induces additional structured missingness
(SM) in the training set, informed by domain knowledge. By doing so, the model is exposed
to missingness patterns that resemble those expected at deployment, thereby improving its
robustness to fairness and performance degradation under test-time SM.</p>
        <p>This algorithm is particularly efective if there is a distributional shift of the missingness
pattern from the training data to the testing data. This shift is reduced by adding randomly
drawn samples from the dataset and reinserting those into the training set, but with certain
features masked/missing.</p>
        <p>Training a model on data that includes imputed values can enhance its robustness, provided
that the same imputation strategy is consistently applied during both training and testing
phases. algorithm 2 is to be applied before the initial imputation step takes place. Exposing the
model to a suficient volume of imputed data during training, allows it to better approximate
the potentially skewed distribution it may encounter in real-world scenarios.</p>
        <p>
          The core insight underlying our method is that, in many practical applications, the patterns
of structured missingness are at least partially understood or can be inferred from domain
expertise [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Leveraging this knowledge to simulate realistic missingness during training
enables a straightforward yet efective approach to improving model robustness under structured
missingness, as supported by empirical results in this and prior work.
        </p>
        <p>In cases where this distribution is unknown, feature selection methods and uniform sampling
across the entire dataset may provide reasonable performance. However, care must be taken
when determining the number of augmented samples, as excessive additions may introduce noise
and increase the computational burden, while insuficient augmentation may fail to capture the
underlying data distribution.</p>
        <sec id="sec-3-3-1">
          <title>Algorithm 2: Missingness Robustness Augmentation</title>
          <p>
            Input: Dataset  with  rows; missingness ratio  ∈ [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ] ; feature subset ℱ ⊆ columns of 
Output: Modified dataset  ̃ with additional missingness
 ← ⌊ ⋅ ⌋ ;
 miss ← sample  rows uniformly at random from 
foreach  ∈  miss do
foreach  ∈ ℱ do
[ ] ← missing
// Number of samples to modify
Replace original rows in  with modified  miss to obtain  ̃
return  ̃
          </p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Setup</title>
      <p>To analyze how SM impacts the fairness of binary classification tasks, we set up a data processing
pipeline that was applied to a variety of imputation methods, classifiers, and datasets. The
selection of data sets is based on the survey [53], where the authors also provide the bayesian
network for each dataset. Both the imputation methods and the classifiers were selected from
state-of-the-art methods in the literature. An overview of the datasets used in our experiments
is provided in Table 1. The column with the number of entries refers to the cleaned version
Data preparation</p>
      <p>Training
data
Dataset
5 fold
crossvalidation
Validation
data</p>
      <p>Data manipulation
reVmalouveal impVuatluaetion
WWSSMMSSMNCMMAMAANRAARRARRR MiSsMKsimIFNCpoNElreest
reVmalouveal impVuatluaetion
Target Encoder
categorical
features</p>
      <p>Standard Scaler Training data</p>
      <p>MXLGmBooodsetl: Trained Model EqEuvaMalilzFuoe1addteioOlnd:ds
Validation data
of each dataset, with all samples containing missing values removed prior to experimentation.
Although several datasets include multiple protected attributes, we report only the specific
attribute used to induce missingness in our experiments, as this was the sole attribute considered
when evaluating fairness metrics. The code and additional results of this work are available at
https://github.com/fairness_SM.</p>
      <p>
        We use imputation methods from statistics and classical machine learning. Due to the small
size of the data, we do not use DL imputation [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. From the statistical realm we employed
simple imputation (using mean for numeric features and mode for categorical) and MICE [32]
while using MissForest [33] and KNN [38] from the machine learning domain.
      </p>
      <p>To employ KNN imputation, we slightly modified the approach to work with categorical
features. Every categorical feature is one-hot encoded, and during imputation, all encoded values
are set to missing and imputed individually. Then we compute their maximum, and the final
imputed value is the corresponding categorical feature.</p>
      <p>Pipeline. To set up a controlled environment we applied similar data cleaning steps as in
[53], removing all missing values which are included in the dataset. To ensure consistency
and reduce statistical noise we employed a 5-fold cross-validation-scheme. The original dataset
was first partitioned into training and testing subsets, after which missingness was artificially
introduced into both. A detailed overview of the entire pipeline can be seen in Figure 2, it
consists of 5 steps: (1) Data preparation, to split the dataset in train/validation with the 5-fold
cross validation strategy. (2) Data manipulation, to first induce the missing values in train
and validation data, with multiple percentage (0%, 10%,30%, 50%, 70%), and then impute the
missing values. (3) Data pre-processing, using target encoder [57], to encode categorical features,
and then standardize the data to zero mean and unitary varriance. (4) Training of the ML
model XGBoost, as it is the state-of-the-art in ML for tabular datasets [58]. (5) Post-processing,
to evaluate the trained model with fairness and performance metrics.</p>
      <p>Preprocessing. Following the 5-fold cross-validation split, missingness induction, and
imputation, we applied a standardized preprocessing pipeline. Given the presence of categorical features
in most datasets, we employed target encoding to transform these variables, as it is considered
a state-of-the-art approach for supervised learning tasks [59]. Finally, the resulting data were
standardized to have zero mean and unit variance to ensure consistent model performance. The
full pipeline, outlined in Figure 2, consisted of 5 steps: To ensure controlled experimentation,
we eliminate all initial missing values from the dataset before invoking algorithm 1. This step
isolates the efects of induced structured missingness (SM), allowing for clearer attribution
of observed outcomes to the experimental conditions. Specifically, we avoid mixing diferent
missingness mechanisms to focus exclusively on the impact of SM on fairness.
Missing Mechanisms. Our primary objective is to evaluate the practical implications of SM in
real-world applications. Although it is theoretically possible to construct a structured version
of MCAR, such a formulation holds limited practical significance. Consequently, we exclude
structured MCAR from our analysis. Instead, we focus on MAR and MNAR mechanisms, where
missingness is conditioned on or influenced by the protected attribute. For MNAR scenarios,
we simulate a latent confounder by removing the protected attribute from the dataset after
inducing missingness. This reflects settings in which sensitive variables afect the missingness
process but are not explicitly available at training or inference time. We evaluate both weakly
and strongly structured variants of MAR and MNAR. Missingness is introduced using a unified
procedure (algorithm 1) that generates both structured and unstructured missingness masks in
a consistent and reproducible manner.</p>
      <p>Evaluation Metrics As our primary interest lies in evaluating downstream task performance
and fairness, we do not consider reconstruction error, commonly used to assess imputation
quality, as it does not necessarily correlate with performance or fairness in predictive modeling
tasks. Instead, we focus on well-established metrics from the machine learning literature that
directly reflect model behavior: the F1 score for predictive performance and Equalized Odds
(EO) for fairness, as we are in a supervised learning setting [60]. EO evaluates fairness based on
error rate parity across groups, ofering a measure of fairness in classification tasks, particularly
in domains where label information is meaningful, and error disparities are critical. For fairness
evaluation, we focus on group-level metrics rather than individual fairness. Although individual
fairness ofers a more fine-grained notion of equitable treatment, it often relies on access to a
well-defined causal model, an assumption that is frequently impractical or unverified in real-world
applications.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Evaluation and Results Discussion</title>
      <p>In this section, we present and analyze our findings on the impact of SM. We begin by evaluating
the efect of diferent imputation strategies employed during preprocessing. Subsequently, we
investigate the model’s behavior under varying degrees of missingness in both the training and
test sets, with a particular focus on the asymmetry between the two.</p>
      <p>For simplicity, we discuss the results on the Adult dataset [61], as it is a common benchmark
in the fairness literature [53], and the trend of the results is qualitatively similar to the other
datasets investigated in this work, listed in Table 1.</p>
      <p>Shortcoming of Imputation Methods. Figure 3 presents both predictive performance
(F1score) and group-level fairness metrics (EO) for the Adult and Bank marketing dataset, under
conditions when training and test sets have matching levels of missingness. Across all imputation
strategies evaluated, a consistent trend emerges: SM leads to the most pronounced degradation
in both performance and fairness metrics. This pattern is clearly reflected in Figure 3, where
Simple imputation</p>
      <p>Mice</p>
      <p>MissForest
Simple imputation</p>
      <p>
        Simple imputation
KNN
KNN
KNN
increasing overall missingness correlates with declines in F1 score and EO. The consistency of
this trend is evident across multiple datasets, including representative results from the Adult
and Bank Marketing datasets. These results align with prior findings [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ], which suggest that
higher imputation complexity does not necessarily yield improved downstream performance.
Shifts in Missingness. The assumption that training and test sets exhibit identical levels of
missingness is rarely met in practice. To evaluate the robustness of the model under more realistic
conditions, we performed experiments across a range of missingness levels in both training and
test data. A central finding is that the presence and distribution of missing values at inference
time have a disproportionately greater impact on both predictive performance and fairness
metrics compared to missingness during training. Specifically, models trained on complete
data but evaluated on partially observed test data exhibited the most significant degradation
in fairness. Figure 4 summarizes results across three experimental conditions: missingness
introduced exclusively during training, symmetrically during both training and testing, and
exclusively at test time, all of which used simple imputation.
      </p>
      <p>Our results indicate that missingness confined to training had negligible efects on performance
and fairness, while test-time missingness introduced considerably more degradation across metrics.
Missing training data</p>
      <p>Missing test data
(b) EO on Adult dataset.</p>
      <p>Equal missing training and test data
20Missing percentage 60
40</p>
      <p>Furthermore, the magnitude of this efect scales with the degree of mismatch between training
and test missingness levels. This is particularly problematic in real-world deployments, where
ground-truth labels are unavailable at test time, limiting the ability to detect post-hoc fairness
degradation. These findings underscore the importance of anticipating structured missingness
and implementing mitigation strategies during the model development phase, rather than relying
solely on downstream audits.</p>
      <sec id="sec-5-1">
        <title>5.1. Toward Robust Fairness in Incomplete Data Settings</title>
        <p>Setup. Motivated by the observed limitations of existing techniques under structured
missingness, we developed Missingness Robustness Augmentation, a preprocessing approach aimed
at improving robustness and fairness. For clarity, recall that our proposed method is applied
in step two of the pipeline proposed in Figure 2. To evaluate its efectiveness, we conducted a
comparative analysis against three baseline strategies widely used in fair machine learning: (i) a
standard model without any fairness intervention, (ii) a reweighting scheme that adjusts the
60
40</p>
        <p>20 0 20
Missing percentage difference (test - train)
40</p>
        <p>60
influence of training samples to mitigate biasparticularly by assigning higher weights to instances
from the minority group with missing values [62], and (iii) a post-processing method that
enforces the EO criterion by modifying classifier outputs [ 63]. For our evaluation of Missingness
Robustness Augmentation we set  to about 10% of total dataset size.</p>
        <p>Evaluation. Figure 5 summarizes the overall performance of all data processing strategies
evaluated. The  -axis denotes the diference in missingness rates between the training and test
sets; higher values indicate a greater proportion of missing data in the test set relative to the
training set.</p>
        <p>The reweighting strategy utilized in our study is based on the presence of missing values in the
training data, applying more weight to data points with missing values. Similarly, when applying
Equalized Odds (EO) postprocessing, discrepancies between the distribution of missingness in
the training and test sets diminish the fairness improvements while also negatively impacting
model performance. In contrast, Missingness Robustness Augmentation remains efective under
such distribution shifts, provided the augmented features accurately reflect the missingness
patterns expected at test time. However, in scenarios where this alignment cannot be guaranteed,
our results (see Figure 4) indicate only marginal efects on both fairness and performance.
Consequently, incorporating Missingness Robustness Augmentation presents minimal risk but
ofers potential benefits, especially in deployment settings where structured missingness may
emerge.</p>
        <p>Significance. In order to assess the statistical significance of the results across multiple datasets,
we use the Friedman non-parameteric test with 0.05 confidence level, followed by Nemenyi
posthoc test [64], and the we plot the results with the Critical Distance (CD) plot in Figure 6. We
focus on where the test set exhibits a higher proportion of missing values than the training set,
we observe significant diferences among data processing methods in both predictive performance
and fairness.</p>
        <p>Although EO post-processing achieves the best fairness outcomes, it does so at the cost of
(a) Critical distance for F1 score.
2
1</p>
        <p>MRA
No Intervention
MRA
EO Postprocessing</p>
        <p>(b) Critical distance for EO.
reduced accuracy. In contrast, the proposed Missingness Robustness Augmentationofers the
second-best performance in terms of fairness, while also outperforming all other methods in
predictive accuracy. This suggests that our method provides a favorable balance between fairness
and performance under conditions of mismatched missingness.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>
        Structured missingness arises from complex real-world processes and interactions, posing unique
challenges for machine learning systems, as highlighted by [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. This study demonstrates that
SM in data can have a substantial impact on both predictive performance and fairness in machine
learning systems, particularly when it occurs during inference. Through empirical evaluation,
we show that disparities in missingness patterns between training and deployment environments
can introduce or exacerbate group-level disparities in model outcomes. To investigate this
systematically, we propose a novel methodology for inducing structured missingness, which links
the missingness mechanism explicitly to the protected attribute.
      </p>
      <p>To address the efects of SM we introduce a simple and generalizable preprocessing method
designed to improve model robustness in the presence of structured missingness. Our approach
requires no architectural changes, is easy to implement, and maintains predictive accuracy
across tasks. While improvements in fairness are modest, the method yields consistent and
stable gains across a range of missingness scenarios, and outperforms standard reweighting and
post-processing baselines in terms of reliability and sensitivity to test-time conditions. As our
approach does not rely on a complete causal model, we deliberately refrained from evaluating
counterfactual fairness, though exploring this perspective remains a promising direction for
future work.</p>
      <p>Our findings suggest that this method is most beneficial in settings where test-time missingness
is prevalent and unevenly distributed across groups, which is common in real-world deployment
contexts. Additionally, performance is determined by the alignment between the induced and
the test-time missingness. When the induced distribution deviated significantly from the true
test-time distribution, the impact on the metrics evaluated was minimal. In contrast, when the
distributions aligned closely, we observed a clear improvement in these metrics.</p>
      <p>Structured missingness remains an underexplored but pressing challenge in fair machine
learning, and our results highlight the need for proactive, pipeline-level strategies that account
for missingness during both development and deployment phases.</p>
      <p>Our proposed method relies on domain knowledge to implement realistic and efective preventive
strategies. Designing approaches that are agnostic to such prior knowledge or that leverage
learned feature representations to approximate it remains an important direction for future work.
Additionally, addressing the challenges posed by structured missingness when both the training
and test sets exhibit comparable levels of missingness remains a complex and unresolved issue.</p>
      <p>Future work should explore dynamic or adaptive methods that respond to missingness patterns
at inference time, as well as deeper theoretical frameworks to understand the interaction between
data missingness and fairness constraints. Another practical suggestion is to explore the impact
of structured missingness on deep learning systems, as such systems gain more traction when
dealing with tabular data.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <sec id="sec-7-1">
        <title>The author(s) have not employed any Generative AI tools.</title>
        <p>Imputation Strategies on Fairness in Machine Learning (2022).
[28] Y. Zhang, Q. Long, Fairness in Missing Data Imputation, 2021. URL: http://arxiv.org/abs/
2110.12002, arXiv:2110.12002 [cs].
[29] M. Fernando, F. Cèsar, N. David, H. José, Missing the missing values: The ugly
duckling of fairness in machine learning, International Journal of Intelligent Systems 36
(2021) 3217–3258. URL: https://onlinelibrary.wiley.com/doi/10.1002/int.22415. doi:10.1002/
int.22415.
[30] T. Emmanuel, T. Maupong, D. Mpoeleng, T. Semong, B. Mphago, O. Tabona, A
survey on missing data in machine learning, Journal of Big Data 8 (2021) 140. URL:
https://journalofbigdata.springeropen.com/articles/10.1186/s40537-021-00516-9. doi:10.1186/
s40537-021-00516-9.
[31] S. Alam, M. S. Ayub, S. Arora, M. A. Khan, An investigation of the imputation techniques
for missing values in ordinal data enhancing clustering and classification analysis validity,
Decision Analytics Journal 9 (2023) 100341. URL: https://linkinghub.elsevier.com/retrieve/pii/
S2772662223001819. doi:10.1016/j.dajour.2023.100341.
[32] S. V. Buuren, K. Groothuis-Oudshoorn, mice : Multivariate Imputation by Chained
Equations in R, Journal of Statistical Software 45 (2011). URL: http://www.jstatsoft.org/
v45/i03/. doi:10.18637/jss.v045.i03.
[33] D. J. Stekhoven, P. Bühlmann, MissForestnon-parametric missing value imputation
for mixed-type data, Bioinformatics 28 (2012) 112–118. URL: https://academic.oup.com/
bioinformatics/article/28/1/112/219101. doi:10.1093/bioinformatics/btr597.
[34] E. Hallaji, R. Razavi-Far, M. Saif, DLIN: Deep Ladder Imputation Network, IEEE
Transactions on Cybernetics 52 (2022) 8629–8641. URL: https://ieeexplore.ieee.org/document/
9370000/. doi:10.1109/TCYB.2021.3054878.
[35] W. Du, D. Cote, Y. Liu, SAITS: Self-Attention-based Imputation for Time Series, Expert
Systems with Applications 219 (2023) 119619. URL: http://arxiv.org/abs/2202.08516. doi:10.
1016/j.eswa.2023.119619, arXiv:2202.08516 [cs].
[36] V. Fortuin, D. Baranchuk, G. Rätsch, S. Mandt, GP-VAE: Deep Probabilistic Multivariate</p>
        <p>Time Series Imputation (2019).
[37] S. I. Khan, A. S. M. L. Hoque, SICE: an improved missing data imputation technique,
Journal of Big Data 7 (2020) 37. URL: https://journalofbigdata.springeropen.com/articles/10.
1186/s40537-020-00313-w. doi:10.1186/s40537-020-00313-w.
[38] O. Troyanskaya, M. Cantor, G. Sherlock, P. Brown, T. Hastie, R. Tibshirani, D. Botstein,
R. B. Altman, Missing value estimation methods for DNA microarrays, Bioinformatics 17
(2001) 520–525. URL: https://academic.oup.com/bioinformatics/article/17/6/520/272365. doi:10.
1093/bioinformatics/17.6.520.
[39] F. Honghai, C. Guoshun, Y. Cheng, Y. Bingru, C. Yumei, A SVM Regression Based
Approach to Filling in Missing Values, in: D. Hutchison, T. Kanade, J. Kittler, J. M.
Kleinberg, F. Mattern, J. C. Mitchell, M. Naor, O. Nierstrasz, C. Pandu Rangan, B. Stefen,
M. Sudan, D. Terzopoulos, D. Tygar, M. Y. Vardi, G. Weikum, R. Khosla, R. J. Howlett,
L. C. Jain (Eds.), Knowledge-Based Intelligent Information and Engineering Systems,
volume 3683, Springer Berlin Heidelberg, Berlin, Heidelberg, 2005, pp. 581–587. URL:
http://link.springer.com/10.1007/11553939_83. doi:10.1007/11553939_83, series Title: Lecture
Notes in Computer Science.
[40] J. Yoon, J. Jordon, GAIN: Missing Data Imputation using Generative Adversarial Nets
(2018).
[41] Z. Sun, H. Li, W. Wang, J. Liu, X. Liu, DTIN: Dual Transformer-based Imputation Nets
for multivariate time series emitter missing data, Knowledge-Based Systems 284 (2024)
111270. URL: https://linkinghub.elsevier.com/retrieve/pii/S0950705123010195. doi:10.1016/j.
knosys.2023.111270.
[42] N. B. Ipsen, P.-A. Mattei, J. Frellsen, HOW TO DEAL WITH MISSING DATA IN</p>
        <p>SUPERVISED DEEP LEARNING? (2022).
[43] S. C.-X. Li, B. Jiang, B. Marlin, MisGAN: Learning from Incomplete Data with Generative
Adversarial Networks, 2019. URL: http://arxiv.org/abs/1902.09599. doi:10.48550/arXiv.1902.
09599, arXiv:1902.09599 [cs].
[44] H. Jeong, H. Wang, F. P. Calmon, Fairness without Imputation: A Decision Tree
Approach for Fair Prediction with Missing Values, 2022. URL: http://arxiv.org/abs/2109.10431,
arXiv:2109.10431 [cs].
[45] Y. Xiao, R. Song, M. Chen, Direct and Unbiased Multiple Imputation Methods for Missing
Values of Categorical Variables, Journal of Data Science 10 (2021) 465–481. URL: https://
jds-online.org/doi/10.6339/JDS.201207_10%283%29.0007. doi:10.6339/JDS.201207_10(3).0007.
[46] T. Chen, C. Guestrin, XGBoost: A Scalable Tree Boosting System, in: Proceedings of
the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data
Mining, 2016, pp. 785–794. URL: http://arxiv.org/abs/1603.02754. doi:10.1145/2939672.2939785,
arXiv:1603.02754 [cs].
[47] C. M. Caruso, P. Soda, V. Guarrasi, Not Another Imputation Method: A Transformer-based
Model for Missing Values in Tabular Datasets, 2024. URL: http://arxiv.org/abs/2407.11540,
arXiv:2407.11540 [cs].
[48] S. M. Kia, N. M. Rad, D. v. Opstal, B. v. Schie, A. F. Marquand, J. Pluim, W. Cahn,
H. G. Schnack, PROMISSING: Pruning Missing Values in Neural Networks, 2022. URL:
http://arxiv.org/abs/2206.01640, arXiv:2206.01640 [cs].
[49] N. Mehrabi, F. Morstatter, N. Saxena, K. Lerman, A. Galstyan, A Survey on Bias and
Fairness in Machine Learning, 2022. URL: http://arxiv.org/abs/1908.09635. doi:10.48550/
arXiv.1908.09635, arXiv:1908.09635 [cs].
[50] P. Garg, J. Villasenor, V. Foggo, Fairness Metrics: A Comparative Analysis, in: 2020 IEEE
International Conference on Big Data (Big Data), IEEE, Atlanta, GA, USA, 2020, pp.
3662–3666. URL: https://ieeexplore.ieee.org/document/9378025/. doi:10.1109/BigData50022.
2020.9378025.
[51] M. J. Kusner, J. R. Loftus, C. Russell, R. Silva, Counterfactual Fairness, 2018. URL:
http://arxiv.org/abs/1703.06856. doi:10.48550/arXiv.1703.06856, arXiv:1703.06856 [stat].
[52] L. Rosenblatt, R. T. Witter, Counterfactual Fairness Is Basically Demographic Parity, 2023.</p>
        <p>URL: http://arxiv.org/abs/2208.03843. doi:10.48550/arXiv.2208.03843, arXiv:2208.03843 [cs].
[53] T. L. Quy, A. Roy, V. Iosifidis, W. Zhang, E. Ntoutsi, A survey on datasets for fairness-aware
machine learning, WIREs Data Mining and Knowledge Discovery 12 (2022) e1452. URL:
http://arxiv.org/abs/2110.00530. doi:10.1002/widm.1452, arXiv:2110.00530 [cs].
[54] Q. Feng, M. Du, N. Zou, X. Hu, Fair Machine Learning in Healthcare: A Review, 2024. URL:
http://arxiv.org/abs/2206.14397. doi:10.48550/arXiv.2206.14397, arXiv:2206.14397 [cs].
[55] R. Feng, F. P. Calmon, H. Wang, Adapting Fairness Interventions to Missing Values (2023).
[56] D. H. Lina, A. Silva, Better Fair than Sorry: Adversarial Missing Data Imputation for
Fair GNNs, 2025. URL: http://arxiv.org/abs/2311.01591. doi:10.48550/arXiv.2311.01591,
arXiv:2311.01591 [cs].
[57] D. Micci-Barreca, A preprocessing scheme for high-cardinality categorical attributes in
classification and prediction problems, ACM SIGKDD explorations newsletter 3 (2001)
27–32.
[58] R. Shwartz-Ziv, A. Armon, Tabular Data: Deep Learning is Not All You Need, 2021. URL:
http://arxiv.org/abs/2106.03253. doi:10.48550/arXiv.2106.03253, arXiv:2106.03253 [cs].
[59] F. Pargent, F. Pfisterer, J. Thomas, B. Bischl, Regularized target encoding outperforms
traditional methods in supervised machine learning with high cardinality features, Computational
Statistics 37 (2022) 2671–2692. URL: https://link.springer.com/10.1007/s00180-022-01207-6.
doi:10.1007/s00180-022-01207-6.
[60] M. Hardt, E. Price, N. Srebro, Equality of Opportunity in Supervised Learning, 2016. URL:
http://arxiv.org/abs/1610.02413. doi:10.48550/arXiv.1610.02413, arXiv:1610.02413 [cs].
[61] B. Becker, R. Kohavi, Adult, UCI Machine Learning Repository, 1996. DOI:
https://doi.org/10.24432/C5XW20.
[62] J. Chai, X. Wang, Fairness with Adaptive Weights (2022).
[63] G. Pleiss, M. Raghavan, F. Wu, J. Kleinberg, K. Q. Weinberger, On Fairness and Calibration
(2017).
[64] J. Demšar, Statistical comparisons of classifiers over multiple data sets, Journal of Machine
learning research 7 (2006) 1–30.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B.</given-names>
            <surname>Balbierer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Heinlein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Zipperling</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kühl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A Multivocal</given-names>
            <surname>Literature</surname>
          </string-name>
          <article-title>Review on Privacy and Fairness in Federated Learning (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Guerra-Manzanares</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. J. L.</given-names>
            <surname>Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maniatakos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. E.</given-names>
            <surname>Shamout</surname>
          </string-name>
          ,
          <article-title>Privacy-preserving machine learning for healthcare: open challenges and future perspectives</article-title>
          , volume
          <volume>13932</volume>
          ,
          <year>2023</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>40</lpage>
          . URL: http://arxiv.org/abs/2303.15563. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -39539-
          <issue>0</issue>
          _
          <fpage>3</fpage>
          , arXiv:
          <fpage>2303</fpage>
          .15563 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>John</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Ekpenyong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. C.</given-names>
            <surname>Nworu</surname>
          </string-name>
          ,
          <article-title>Imputation of Missing Values in Economic and Financial Time Series Data Using Five Principal Component Analysis (PCA) Approaches</article-title>
          , Central Bank of Nigeria Journal of Applied Statistics (
          <year>2019</year>
          )
          <fpage>51</fpage>
          -
          <lpage>73</lpage>
          . URL: https://www.cbn. gov.ng/Out/2019/STD/51%
          <fpage>20</fpage>
          -%
          <source>2073_John%20et%20al.pdf. doi:10.33429/Cjas.10119.3/6.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>McKeown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Voss</surname>
          </string-name>
          ,
          <article-title>A New Direction for Machine Learning in Criminal Law (</article-title>
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>F.</given-names>
            <surname>Dakalbab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Abu</given-names>
            <surname>Talib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. Abu</given-names>
            <surname>Waraga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Bou</given-names>
            <surname>Nassif</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abbas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Nasir</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence &amp; crime prediction: A systematic literature review</article-title>
          ,
          <source>Social Sciences &amp; Humanities Open</source>
          <volume>6</volume>
          (
          <year>2022</year>
          )
          <article-title>100342</article-title>
          . URL: https://linkinghub.elsevier.com/retrieve/pii/S2590291122000961. doi:
          <volume>10</volume>
          .1016/j.ssaho.
          <year>2022</year>
          .
          <volume>100342</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G. V.</given-names>
            <surname>Travaini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Pacchioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bellumore</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bosia</surname>
          </string-name>
          , F. De Micco,
          <article-title>Machine Learning and Criminal Justice: A Systematic Review of Advanced Methodology for Recidivism Risk Prediction</article-title>
          ,
          <source>International Journal of Environmental Research and Public Health</source>
          <volume>19</volume>
          (
          <year>2022</year>
          )
          <article-title>10594</article-title>
          . URL: https://www.mdpi.com/1660-4601/19/17/10594. doi:
          <volume>10</volume>
          .3390/ijerph191710594.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>F.</given-names>
            <surname>Fioretto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. V.</given-names>
            <surname>Hentenryck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <article-title>Diferential Privacy and Fairness in Decisions and Learning Tasks: A Survey</article-title>
          ,
          <source>in: Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>5470</fpage>
          -
          <lpage>5477</lpage>
          . URL: http://arxiv.org/abs/ 2202.08187. doi:
          <volume>10</volume>
          .24963/ijcai.
          <year>2022</year>
          /766, arXiv:
          <fpage>2202</fpage>
          .08187 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>F.</given-names>
            <surname>Martínez-Plumed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ferri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Nieves</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hernández-Orallo</surname>
          </string-name>
          , Fairness and
          <string-name>
            <given-names>Missing</given-names>
            <surname>Values</surname>
          </string-name>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1905</year>
          .12728. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1905</year>
          .
          <volume>12728</volume>
          , arXiv:
          <year>1905</year>
          .12728 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Zafar</surname>
          </string-name>
          , I. Valera,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. P.</given-names>
            <surname>Gummadi</surname>
          </string-name>
          ,
          <article-title>Fairness Constraints: Mechanisms for Fair Classification</article-title>
          ,
          <year>2017</year>
          . URL: http://arxiv.org/abs/1507.05259, arXiv:
          <fpage>1507</fpage>
          .05259 [stat].
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Howell</surname>
          </string-name>
          ,
          <article-title>The Treatment of Missing Data, in: The SAGE Handbook of Social Science Methodology</article-title>
          ,
          <source>SAGE Publications Ltd, 1 Oliver's Yard,ă55 City Road,ăLondonăEnglandăEC1Y 1SPăUnited Kingdom</source>
          ,
          <year>2007</year>
          , pp.
          <fpage>212</fpage>
          -
          <lpage>226</lpage>
          . URL: https://methods.sagepub.com/ book/the-sage
          <article-title>-handbook-of-social-science-methodology/n11.xml</article-title>
          . doi:
          <volume>10</volume>
          .4135/9781848607958. n11.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>D. B. Rubin</surname>
            , Inference and
            <given-names>Missing</given-names>
          </string-name>
          <string-name>
            <surname>Data</surname>
          </string-name>
          (
          <year>1976</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Little</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rubin</surname>
          </string-name>
          ,
          <article-title>Statistical analysis with missing data</article-title>
          ,
          <source>Wiley series in probability and mathematical statistics. Probability and mathematical statistics</source>
          , Wiley,
          <year>2002</year>
          . URL: http://books.google.com/books?id=aYPwAAAAMAAJ.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>I. F.</given-names>
            <surname>Ilyas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chu</surname>
          </string-name>
          , Data Cleaning, Association for Computing Machinery, New York, NY, USA,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zuo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , H. Wu, W. Hui,
          <article-title>Poor reporting quality and high proportion of missing data in economic evaluations alongside pragmatic trials: a cross-sectional survey</article-title>
          ,
          <source>BMC Medical Research Methodology</source>
          <volume>25</volume>
          (
          <year>2025</year>
          )
          <article-title>61</article-title>
          . URL: https://bmcmedresmethodol.biomedcentral.com/articles/10.1186/s12874-025-02519-z. doi:
          <volume>10</volume>
          .1186/s12874-025-02519-z.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mitra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. F.</given-names>
            <surname>McGough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chakraborti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Holmes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Copping</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hagenbuch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Biedermann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Noonan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lehmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shenvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X. V.</given-names>
            <surname>Doan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Leslie</surname>
          </string-name>
          , G. Bianconi,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sanchez-Garcia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Davies</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mackintosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.-R.</given-names>
            <surname>Andrinopoulou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Basiri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Harbron</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. D.</surname>
          </string-name>
          <article-title>MacArthur, Learning from data with structured missingness</article-title>
          ,
          <source>Nature Machine Intelligence</source>
          <volume>5</volume>
          (
          <year>2023</year>
          )
          <fpage>13</fpage>
          -
          <lpage>23</lpage>
          . URL: https://www.nature.com/articles/s42256-022-00596-z. doi:
          <volume>10</volume>
          .1038/s42256-022-00596-z.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Radosavljevi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. E.</given-names>
            <surname>Nichols</surname>
          </string-name>
          ,
          <article-title>A generative model for evaluating missing data methods in large epidemiological cohorts</article-title>
          ,
          <source>BMC Medical Research Methodology</source>
          <volume>25</volume>
          (
          <year>2025</year>
          )
          <article-title>34</article-title>
          . URL: https://bmcmedresmethodol.biomedcentral.com/articles/10.1186/s12874-025-02487-4. doi:
          <volume>10</volume>
          .1186/s12874-025-02487-4.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jackson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mitra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Hagenbuch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>McGough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Harbron</surname>
          </string-name>
          , A Complete Characterisation of Structured Missingness,
          <year>2023</year>
          . URL: http://arxiv.org/abs/2307.02650, arXiv:
          <fpage>2307</fpage>
          .02650 [stat].
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>E. J.</given-names>
            <surname>Topol</surname>
          </string-name>
          ,
          <article-title>High-performance medicine: the convergence of human and artificial intelligence</article-title>
          ,
          <source>Nature Medicine</source>
          <volume>25</volume>
          (
          <year>2019</year>
          )
          <fpage>44</fpage>
          -
          <lpage>56</lpage>
          . URL: https://www.nature.com/articles/s41591-018-0300-7. doi:
          <volume>10</volume>
          .1038/s41591-018-0300-7.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Vale Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Rohr</surname>
          </string-name>
          ,
          <article-title>Pan-Cancer Prognosis Prediction Using Multimodal Deep Learning</article-title>
          ,
          <source>in: 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI)</source>
          , IEEE,
          <string-name>
            <surname>Iowa</surname>
            <given-names>City</given-names>
          </string-name>
          ,
          <string-name>
            <surname>IA</surname>
          </string-name>
          , USA,
          <year>2020</year>
          , pp.
          <fpage>568</fpage>
          -
          <lpage>571</lpage>
          . URL: https://ieeexplore.ieee.org/document/9098665/. doi:
          <volume>10</volume>
          .1109/ISBI45749.
          <year>2020</year>
          .
          <volume>9098665</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Herasymuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Protsiv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Stoyanovich</surname>
          </string-name>
          ,
          <source>Still More Shades of Null: An Evaluation Suite for Responsible Missing Value Imputation</source>
          ,
          <year>2025</year>
          . URL: http://arxiv.org/abs/ 2409.07510. doi:
          <volume>10</volume>
          .48550/arXiv.2409.07510, arXiv:
          <fpage>2409</fpage>
          .07510 [cs].
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>W.-C. Lin</surname>
            ,
            <given-names>C.-F.</given-names>
          </string-name>
          <string-name>
            <surname>Tsai</surname>
          </string-name>
          ,
          <article-title>Missing value imputation: a review and analysis of the literature (20062017)</article-title>
          ,
          <source>Artificial Intelligence Review</source>
          <volume>53</volume>
          (
          <year>2020</year>
          )
          <fpage>1487</fpage>
          -
          <lpage>1509</lpage>
          . URL: http://link.springer.
          <source>com/10.1007/s10462-019-09709-4</source>
          . doi:
          <volume>10</volume>
          .1007/s10462-019-09709-4.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Deep learning versus conventional methods for missing data imputation: A review and comparative study</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>227</volume>
          (
          <year>2023</year>
          )
          <article-title>120201</article-title>
          . URL: https://linkinghub.elsevier.com/retrieve/pii/S0957417423007030. doi:
          <volume>10</volume>
          . 1016/j.eswa.
          <year>2023</year>
          .
          <volume>120201</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>J.</given-names>
            <surname>Josse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Prost</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          , E. Scornet,
          <article-title>On the consistency of supervised learning with missing values</article-title>
          ,
          <source>Statistical Papers</source>
          <volume>65</volume>
          (
          <year>2024</year>
          )
          <fpage>5447</fpage>
          -
          <lpage>5479</lpage>
          . URL: https://link. springer.
          <source>com/10.1007/s00362-024-01550-4</source>
          . doi:
          <volume>10</volume>
          .1007/s00362-024-01550-4.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>P. J.</given-names>
            <surname>García-Laencina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Morales-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Verdú-Monedero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Larrey-Ruiz</surname>
          </string-name>
          , J.
          <string-name>
            <surname>-L. SanchoGómez</surname>
            ,
            <given-names>A. R.</given-names>
          </string-name>
          <string-name>
            <surname>Figueiras-Vidal</surname>
          </string-name>
          ,
          <article-title>Classification with Incomplete Data:</article-title>
          , in: E. S. Olivas,
          <string-name>
            <surname>J. D. M. Guerrero</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Martinez-Sober</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          <string-name>
            <surname>Magdalena-Benedito</surname>
            ,
            <given-names>A. J.</given-names>
          </string-name>
          <string-name>
            <surname>Serrano López</surname>
          </string-name>
          (Eds.),
          <source>Handbook of Research on Machine Learning Applications and Trends, IGI Global</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>147</fpage>
          -
          <lpage>175</lpage>
          . URL: http://services.igi-global.com/resolvedoi/resolve.aspx?doi=10.4018/ 978-1-
          <fpage>60566</fpage>
          -766-9.ch007.
          <source>doi:10.4018/978-1-60566-766-9</source>
          .
          <year>ch007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>D.</given-names>
            <surname>Bertsimas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Delarue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pauphilet</surname>
          </string-name>
          ,
          <article-title>Simple Imputation Rules for Prediction with Missing Data: Contrasting Theoretical Guarantees with Empirical Performance,</article-title>
          <year>2024</year>
          . URL: http://arxiv.org/abs/2104.03158. doi:
          <volume>10</volume>
          .48550/arXiv.2104.03158, arXiv:
          <fpage>2104</fpage>
          .03158 [stat].
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>M. L. Morvan</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Josse</surname>
          </string-name>
          , E. Scornet, G. Varoquaux,
          <article-title>What's a good imputation to predict with missing values?</article-title>
          ,
          <year>2021</year>
          . URL: http://arxiv.org/abs/2106.00311. doi:
          <volume>10</volume>
          .48550/arXiv.2106.00311, arXiv:
          <fpage>2106</fpage>
          .00311 [stat].
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Caton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Caton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Ie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Malisetty</surname>
          </string-name>
          , U. Edu,
          <string-name>
            <given-names>C.</given-names>
            <surname>Haas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Haas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. A.</given-names>
            <surname>At</surname>
          </string-name>
          , Impact of
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>