<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>F. Sabbatini); roberta.calegari@unibo.it (R. Calegari)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Unmasking the Shadows: Leveraging Symbolic Knowledge Extraction to Discover Biases and Unfairness in Opaque Predictive Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Federico Sabbatini</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberta Calegari</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Alma Mater Studiorum-University of Bologna</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Urbino Carlo Bo</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This work explores the eficacy of symbolic knowledge-extraction (SKE) techniques in identifying biases and unfairness within opaque predictive models. Logic rules extracted from black-box predictors make it possible to verify if decisions are influenced by protected or sensitive features. In particular, the identification of biased or unfair decisions can be achieved through the evaluation of if-then rules, detecting the inclusion of protected and/or sensitive information in the rules' precondition. The efectiveness of SKE in this regard is demonstrated here by conducting various simulations on a well-known data set for loan grant prediction. Our findings highlight the potential of SKE as a valuable tool to reveal biases and discrimination in opaque predictions, ultimately contributing to the pursuit of fair and transparent decision-making systems.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Fairness in AI</kwd>
        <kwd>Bias in AI</kwd>
        <kwd>Explainable artificial intelligence</kwd>
        <kwd>XAI</kwd>
        <kwd>Symbolic knowledge extraction</kwd>
        <kwd>PSyKE</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        As predictive models become increasingly integrated into various domains, ensuring their
fairness and transparency is of paramount importance [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Opaque predictive models in machine
learning (ML), often referred to as black-box models, pose challenges in understanding the
underlying mechanisms by which they make predictions. Consequently, biases and
discrimination can inadvertently permeate these models, leading to unfair or prejudiced outcomes
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. To address this critical issue, the present paper investigates the application of symbolic
knowledge-extraction (SKE) techniques in uncovering biases and discrimination within opaque
predictive models.
      </p>
      <p>
        SKE ofers a promising avenue to extract interpretable logic rules from black-box models,
enabling a deeper understanding of decision-making [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]. By distilling complex model behaviours
into human-readable rules, SKE facilitates the identification of specicfi conditions under which
biases may arise. This approach proves particularly valuable when examining whether
protected features play a role in decision-making since the presence of protected information in the
preconditions of extracted rules can provide direct evidence of bias. The same considerations
may also extend to sensitive features, e.g., those that are not protected themselves but are
related to other features identified as protected (e.g., name or height, which allow ML models
to infer race and/or gender of individuals; [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]). We point out that identifying correlations
between protected/sensitive features and other input variables is not within the scope of SKE
techniques, nor is the recognition of protected/sensitive attributes in the rule preconditions1.
The classification of input features into “unfairness-enablers” and “potentially-fairness-neutral”
should be performed by human users as an independent task.
      </p>
      <p>The main objective of this paper is to demonstrate the efectiveness of SKE in identifying
unfairness and discrimination within opaque predictions. To achieve this, we employ a
wellknown classification data set aimed at predicting loan grants. We conduct various simulations to
illustrate how SKE can be exploited to extract logic rules and evaluate their fairness implications.
Through these examples, we aim to shed light on the potential of SKE as a practical tool for
highlighting biases and promoting fairness in predictive modelling.</p>
      <p>By revealing biases and discrimination present in opaque predictive models, this research
contributes to the broader discourse on fairness, accountability, and transparency in algorithmic
decision-making. Understanding and rectifying biases in these models are crucial steps towards
building equitable systems that mitigate the perpetuation of societal inequalities. The insights
gained from this study serve as a foundation for developing strategies to enhance fairness in
predictive models and promote the responsible deployment of artificial intelligence (AI) in
critical domains.</p>
      <p>In the following sections, we will discuss the methodology employed for SKE, present the
results of our experiments, and discuss the implications and potential future directions of
this research. By critically examining the power of SKE in identifying biases, we hope to
provide practical insights and actionable recommendations for researchers, practitioners, and
policymakers working towards fair and transparent predictive models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>Several studies have explored diferent approaches and methodologies to address bias in AI.</p>
      <p>
        One line of research focuses on rule-based techniques for bias detection and explanation
[
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. These studies aim to extract interpretable rules from black-box models and analyse
them for potential biases. For instance, in [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] the authors have proposed algorithms mining
association rules or decision trees to identify discriminatory patterns in the rule sets generated
by predictive models. These approaches often leverage fairness criteria or sensitive attribute
definitions to guide the rule extraction process.
      </p>
      <p>
        Another area of related work involves the use of fairness-aware machine learning techniques
[
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ]. These approaches aim to incorporate fairness considerations during the model training
phase, ensuring that the resulting predictions are less likely to be biased. Fairness-aware
algorithms often employ mathematical optimisation techniques to balance predictive accuracy
1In the following, we adopt the terms “protected” and “sensitive” as synonyms, since the considerations discussed in
this work apply to both categories
and fairness objectives, taking into account various fairness definitions such as demographic
parity [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], equalised odds [13], or individual fairness [14].
      </p>
      <p>Furthermore, researchers have explored post-hoc methods to detect and mitigate biases in
predictive models [15, 16]. These methods involve analysing the outcomes of model predictions
on diferent subgroups defined by sensitive attributes, such as race, gender, or age. By quantifying
and comparing the disparities in prediction outcomes across subgroups, these techniques can
help identify and address discriminatory behaviour in models.</p>
      <p>SKE techniques, including rule extraction and logic rule analysis, have been used in various
domains to interpret and understand black-box models [17, 18, 19, 20]. However, thus far, their
specific application for bias and discrimination identification in opaque predictions has not
gained much attention. The proposed research aims to contribute to this body of work by
demonstrating the efectiveness of SKE in uncovering biases and discrimination and providing
insights into its practical application for fairness assessment in predictive models.</p>
      <p>Through a comprehensive review of existing related work, this paper will situate SKE methods
within the broader context of bias detection and fairness assessment in predictive modelling.
It will build upon and extend the current knowledge by showcasing the unique capabilities
of SKE techniques in addressing biases and discrimination in opaque predictive models, thus
contributing to the growing literature on fair and transparent algorithmic decision-making.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Symbolic Knowledge Extraction: Methods and Methodology</title>
      <p>SKE is a methodology aiming to extract interpretable and logic rules from complex black-box
models, enabling a deeper understanding of their decision-making processes. There are two
main approaches within SKE: pedagogical and decompositional [21].</p>
      <p>In the pedagogical approach, the focus is on extracting human-readable rules providing
meaningful explanations of the model’s behaviour. These rules are often represented in if-then format,
making them easily understandable by both humans and machines. The pedagogical approach
prioritises generality, allowing stakeholders to gain insights into the decision criteria employed
by any predictive model, even though the explanations may lose some of the underlying model’s
complexity and performance.</p>
      <p>On the other hand, the decompositional approach aims to decompose the black-box model
into simpler, more interpretable sub-models or components that are typically easier to
understand and analyse individually. The inner black-box structure is carefully analysed and the
resulting explanations may be more adherent to the underlying model behaviour. However,
these techniques are strictly tailored to narrow categories of predictors, thus lacking flexibility
and generality.</p>
      <p>Since both approaches generate intuitive explanations that can be easily communicated and
understood by a broader audience, this work prioritises bias evaluations independent of the
underlying predictive model. Therefore, we exploit pedagogical approaches as the main tools
for our experiments.</p>
      <p>In the following, we provide a summary of some state-of-the-art pedagogical SKE techniques
– namely, GridEx, CART and CReEPy – ofering insights into the specific techniques employed
in the experimentation section.</p>
      <p>
        We leverage the implementations available within the PSyKE Python package2 [22, 23]. This
library encompasses all the aforementioned SKE implementations, allowing for their seamless
comparison and evaluation [24]. The PSyKE platform ofers a unified interface, enabling
the application, assessment, and comparison of various SKE techniques. Moreover, it is fully
compatible with other widely-used Python packages [
        <xref ref-type="bibr" rid="ref13">25</xref>
        ], such as Scikit-Learn [
        <xref ref-type="bibr" rid="ref14">26</xref>
        ], and provides
additional extensions for SKE [
        <xref ref-type="bibr" rid="ref15">27</xref>
        ] and functionalities for feature engineering, data manipulation
and visualisation, Semantic Web compatibility [
        <xref ref-type="bibr" rid="ref16">28</xref>
        ], and assessment of knowledge quality
[
        <xref ref-type="bibr" rid="ref17 ref18">29, 30</xref>
        ].
      </p>
      <sec id="sec-3-1">
        <title>3.1. GridEx</title>
        <p>
          GridEx [
          <xref ref-type="bibr" rid="ref19">31</xref>
          ] is a pedagogical SKE algorithm originally designed for regression tasks and based
on hypercubic partitioning of the input feature space. The partitioning is recursive, symmetric
and performed top-down to obtain human-interpretable rules describing as many disjoint,
hypercubic input space subregions. Thanks to the generalisation presented in [
          <xref ref-type="bibr" rid="ref20 ref21">32, 33</xref>
          ], it is
possible to apply GridEx to both classification and regression tasks if they are encoded via data
sets having only continuous input features.
        </p>
        <p>GridEx requires the following set of hyper-parameters to be defined by users:
recursion depth defining the maximum number of recursions to perform during the
knowledge extraction;
splitting strategy to partition the input space. It may be fixed, if each input dimension is
split into a fixed number of partitions, or adaptive if the number of splits depends on the
relevance of the features;
number of splits defining how many slices have to be performed along each input dimension;
error threshold used to decide on which regions the recursive step of the algorithm has to be
performed. In particular, only regions with a predictive error greater than the user-defined
threshold are recursively split.</p>
        <p>
          This set of parameters may be automatically tuned with the PEDRO procedure [
          <xref ref-type="bibr" rid="ref22">34</xref>
          ].
3.2. CART
The CART algorithm [
          <xref ref-type="bibr" rid="ref23">35</xref>
          ] is based on the induction of a classification or regression binary
decision tree. It may be directly applied to a data set to build a human-interpretable predictor
(if the induced tree is not deep) or it may be adopted as an SKE technique to produce
humaninterpretable rules mimicking the behaviour of an opaque ML model. Human-interpretable rules
are obtained by reading the complete paths from the tree root to each distinct leaf, given that
internal nodes represent constraints on input variables and leaves contain output predictions.
        </p>
        <p>The most important parameters to consider for CART are:
maximum depth defining the maximum allowed depth for the decision tree;</p>
        <sec id="sec-3-1-1">
          <title>2https://github.com/psykei/psyke-python</title>
          <p>maximum number of leaves defining the maximum allowed number of tree leaves.
These two parameters are intertwined and both the predictive accuracy and the
humanreadability extent of the tree critically depend on them. In particular, deep trees usually exhibit
higher predictive performance but smaller human-readability extent than shallow ones. The
same holds for trees with a large number of leaves compared to trees with fewer leaves.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.3. CReEPy</title>
        <p>
          The CReEPy algorithm [
          <xref ref-type="bibr" rid="ref24 ref25">36, 37</xref>
          ] is a pedagogical SKE technique applicable to opaque classifiers
and regressors. It relies on underlying explainable clustering procedures aimed at identifying
hypercubic human-interpretable regions within the input feature space [
          <xref ref-type="bibr" rid="ref26 ref27">38, 39</xref>
          ]. At the end of
the knowledge extraction, each hypercubic region is translated into a Prolog rule describing the
boundaries of the region and the corresponding output prediction.
        </p>
        <p>
          Suitable explainable clusterings adopted by CReEPy are CREAM [
          <xref ref-type="bibr" rid="ref28">40</xref>
          ] and ExACT [
          <xref ref-type="bibr" rid="ref29">41</xref>
          ]. They
both perform hierarchical clustering according to diferent recursive strategies and require the
following parameters, possibly tuned with the OrCHiD automated procedure [
          <xref ref-type="bibr" rid="ref28">40</xref>
          ]:
recursion depth defining the maximum number of performed recursions;
maximum number of Gaussian components defining the maximum number of
components to use in the Gaussian mixture model clustering performed within ExACT and
CREAM;
error threshold used to pre-emptively stop the recursive clustering when clusters exhibit a
predictive error smaller than the threshold.
        </p>
        <p>To execute CReEPy users have to provide the parameters required by the underlying instance
of ExACT or CREAM as well as an optional feature relevance threshold used to drop from
the output Prolog rules all the antecedents involving input features with relevance below the
threshold.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>4.1. Running Example: the Loan Data Set Case Study</title>
        <p>We selected the Loan data set3 as a case study to carry out experiments and verify if SKE
techniques are efective tools to identify discriminative predictions provided by opaque models.
The data set is composed of 11 input features representing relevant variables to decide if a
loan should be granted or not. The final decision is the binary output feature. The data set is
completed by an additional feature representing a unique identification code for each loan. The
data set counts 614 instances. Only 480 have no missing values. The names, types, and values
of the features are reported in Table 1.</p>
        <sec id="sec-4-1-1">
          <title>3https://www.kaggle.com/datasets/burak3ergun/loan-data-set</title>
          <p>In conducting the experiments presented in this study, instances in the data set that contained
missing values were excluded, and nominal attributes were converted into discrete numeric
features.</p>
          <p>
            To evaluate the fairness of the data sets and opaque predictors, we employed the disparate
impact index [
            <xref ref-type="bibr" rid="ref30">42</xref>
            ]. This metric measures the extent of diferential treatment between two distinct
groups, specifically by quantifying the proportion of individuals from each group who receive
positive outcomes. The disparate impact index serves as a quantitative measure of the disparate
treatment experienced by individuals from diferent classes.
          </p>
          <p>The calculation of the disparate impact index involves grouping the instances in a data set 
into two subgroups: a privileged (or base) group  and an unprivileged (or protected) group
 , typically afected by fairness concerns. Formally,
 = {︁ :  = (1 , 2 , . . . ,  ) ,
 }︁
where 1 , 2 , . . . ,  are the  features of instance  and
 = { :  ∈  ∧  = ⊕} ,
 = { :  ∈  ∧  = ⊖} ,
by assuming that the sensitive feature  have values in {⊕ , ⊖} , with  = ⊕ representing the
membership to the privileged group.</p>
          <p>For each group, the ratio of positive outcomes to the total number of individuals is computed.
Subsequently, the disparate impact index, denoted as , is defined as follows:
⃒⃒ {︀  :  ∈  ∧  () = ⊙ ︀} ⃒⃒</p>
          <p>| |
⃒⃒ {︀  :  ∈  ∧  () = ⊙ ︀} ⃒⃒</p>
          <p>| |
where  () represents the output of instance  and ⊙ is the positive output.</p>
          <p>In our experimental setup, we specifically focus on the scenario of gender discrimination
( = ). Consequently, we designate female individuals as the unprivileged group
(⊖ =  ) and male individuals as the privileged group (⊕ = ). This choice allows us
to investigate and analyse potential biases and disparities that may afect females within the
context of the studied predictive models. In our case study  is a function denoting the approval
or denial of a loan (therefore, () =  corresponds to a positive outcome). The  is
accordingly defined as follows:
⃒⃒⃒ {︁ :  ∈  ∧ 
}︁⃒
 =   ∧ () =  ⃒⃒
⃒⃒⃒ {︁ :  ∈  ∧</p>
          <p>}︁⃒
 =   ⃒⃒
⃒⃒⃒ {︁ :  ∈  ∧ 
}︁⃒
 =  ∧ () =  ⃒⃒
⃒⃒⃒ {︁ :  ∈  ∧  ⃒</p>
          <p>= }︁⃒⃒
.</p>
          <p>(2)</p>
          <p>As reported in the first row of Table 2, 394 out of 480 instances describe loans demanded by
male applicants. The remaining 86 instances correspond to female applicants. Even if the gender
attribute is not balanced, it is possible to observe that loans are fairly granted to female and male
applicants. Indeed, 278 out of 394 male applicants receive the loan, as well as 54 out of 86 female
applicants. This corresponds to the 71% and 63% of male and female applicants, respectively.
By applying Equation (2) it is possible to find a disparate impact score of 0.89, corresponding
to a quite fair situation. We recall here that  = 1 denotes a perfectly fair situation. Lower
score values are associated with unfair conditions. A score of 0.8 is usually considered the
threshold to divide fairness ( &gt; 0.8) from unfairness ( &lt; 0.8). As a consequence of all
these observations, we consider the Loan data set fair from the gender standpoint.</p>
          <p>The distribution of the output feature of the data set with respect to the gender attribute is
visually presented in Figure 1a. The x-axis represents the credit history input feature, which is
considered the most significant for classification purposes. Gender is reported in the y-axis. The
size of the circles corresponds to the number of instances in each subregion of the input feature
space. Orange circles indicate granted loans, whereas green circles indicate denied loans.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. SKE on the Loan Data Set: Uncovering Insights and Patterns</title>
        <p>A random forest (RF) classifier has been trained upon the Loan data set. The data set has been
randomly split into training (85%) and test (15%) sets. The RF predictor was composed of 50
base decision trees having a maximum depth of 5 and achieved a classification accuracy equal
to 0.79. The decision boundaries of the RF classifier are reported in Figure 1e as a bidimensional
projection on the credit history and gender input features.</p>
        <p>The RF can be considered a fair predictor since its disparate impact score is equal to 0.99
(cf. first row of the bottom part of Table 2). It is worth mentioning that fairness is not directly
associated with classification accuracy. In this particular case, despite the RF classifier’s
predictive performance not being excellent, it is noteworthy that it demonstrates a high level
of fairness from a gender perspective. Fairness, in this context, refers to the absence of bias
or discrimination based on gender, regardless of the classifier’s overall accuracy in making
predictions.</p>
        <p>The goal of our experiments is to demonstrate if SKE techniques can be used to detect unfair
opaque predictors. To this purpose, we use the CART, GridEx and CReEPy algorithms to
perform knowledge extraction on the trained RF classifier. Extractors have been parametrised
as summarised in Table 3. The number of extracted rules, as a proxy of the human-readability
extent of the models, and the fidelity measured for each extractor with respect to the RF
predictions, expressed as classification accuracy, have been reported in Table 4. All extractors
can achieve a fidelity of 0.99 with 2 rules.</p>
        <p>The decision boundaries obtained via CART, GridEx and CReEPy are reported in Figures 1i,
1m and 1q, respectively. The corresponding Prolog rules are shown in Listings 1 to 3, respectively.
(a) Loan data set
(original, fair  score).</p>
        <p>(b) Loan data set (28%
perturbed, unfair
 score).</p>
        <p>(c) Loan data set (56%
perturbed, unfair
 score).</p>
        <p>(d) Loan data set (83%
perturbed, unfair
 score).
(e) RF (original, fair
).</p>
        <p>(f) RF (28% perturbed,
fair  score).</p>
        <p>(g) RF (56% perturbed,
unfair  score).</p>
        <p>(h) RF (83% perturbed,
unfair  score).
(i) CART (original).</p>
        <p>(j) CART
turbed).</p>
        <p>(28%
per(k) CART (56%
turbed).</p>
        <p>per(l) CART
turbed).</p>
        <p>(83%
per(m) GridEx (original).</p>
        <p>(n) GridEx (28%
perturbed).</p>
        <p>(o) GridEx (56%
perturbed).</p>
        <p>(p) GridEx (83%
perturbed).
(q) CReEPy (original).</p>
        <p>(r) CReEPy (28%
perturbed).</p>
        <p>(s) CReEPy (56%
perturbed).</p>
        <p>(t) CReEPy (83%
perturbed).
Listing 1: Rules extracted with CART for the Loan data set (original and 28% perturbed data set).
loan(Gender, Married, ..., CreditHistory, PropertyArea, NO)
:</p>
        <p>CreditHistory &lt; 0.5.
loan(Gender, Married, ..., CreditHistory, PropertyArea, YES).
Listing 2: Rules extracted with GridEx for the Loan data set (original and 28% perturbed data
set).
loan(Gender, Married, ..., CreditHistory, PropertyArea, NO)
:</p>
        <p>CreditHistory in [0.00, 0.33].
loan(Gender, Married, ..., CreditHistory, PropertyArea, YES)
:</p>
        <p>CreditHistory in [0.67, 1.00].</p>
        <p>Listing 3: Rules extracted with CReEPy for the Loan data set (original).
loan(Gender, Married, ..., CreditHistory, PropertyArea, NO)
:</p>
        <p>CreditHistory in [0.00, 0.00].
loan(Gender, Married, ..., CreditHistory, PropertyArea, YES).</p>
        <p>The three SKE algorithms reveal that the predictions made by the RF model are solely
influenced by the credit history input feature. Irrespective of the applicants’ gender, loans are
granted to individuals with a positive credit history (credit history = 1), while they are denied
to those with a negative credit history (credit history = 0). The SKE techniques confirm the RF’s
fair behaviour concerning the applicants’ gender, as the predictions are solely driven by the
credit history attribute and are independent on gender.</p>
        <p>Listing 4: Rules extracted with CReEPy for the Loan data set (28% perturbed).</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Injecting Bias in the Loan Data Set</title>
        <p>To inject bias, we perturbed the output feature of the Loan data set, which was originally fair
with respect to gender. The perturbation involved changing the loan status from ‘Yes’ to ‘No’ for
a variable number of female applicants. Specifically, we conducted three diferent perturbations,
modifying the positive loan outcome for 15, 30, and 45 female applicants. These numbers
correspond to 28%, 56%, and 83% respectively, of the total female applicants who originally had
a positive loan outcome in the unaltered data set.</p>
        <p>The output feature distribution for the biased data sets is reported in Figures 1b to 1d and in
the top part of Table 2. The corresponding disparate impact measurements are reported in the
rightmost column of the same table. As expected, the score values decrease by increasing the
introduced bias, down to 0.15 for the most perturbed data set. Each data set has been used to
train an RF classifier with 50 base predictors having maximum depth equal to 5 and a measured
predictive accuracy on the test set varying between 0.75 and 0.79 (cf. bottom part of Table 2).
4.3.1. 28% Perturbed Data Set
The RF classifier trained upon the Loan data set with a perturbation involving 28% of the
positive female applicants has  = 0.79, even though the biased data set has a lower score
( = 0.64). This diference is due to the predictive error of the RF. There are no noticeable
diferences in the decision boundaries of this RF compared to the one trained on the original
Loan data set (cf. Figures 1e and 1f). Also CART, GridEx and CReEPy applied to the RF provide
outputs similar to those obtained for the unbiased case study (see Figures 1j, 1n and 1r). The
only diference is the Prolog theory obtained via CReEPy, having the same semantics as the
unbiased counterpart, but diferent clauses. The theory is listed in Listing 4.</p>
        <p>Also in this case the human-interpretable rules extracted via SKE techniques do not identify
discriminative predictions based on gender for the RF classifier and this is in agreement with
the corresponding disparate impact scores.
4.3.2. 56% Perturbed Data Set
A diferent situation is evident if we modify the data set in order to refuse the loan to the 56%
of female applicants that conversely should have received it. In this case, the disparate impact
score drops to 0.40 for the data set and to 0.07 for the corresponding trained RF. These values
highlight strong unfairness, especially for the RF predictions. The corresponding decision
boundaries are reported in Figure 1g. It is clearly visible that the loan is granted only to male
applicants having a positive credit history.</p>
        <p>Decision boundaries obtained via CART, GridEx and CReEPy and the corresponding extracted
rules expressed as Prolog theories are reported in Figures 1k, 1o and 1s and Listings 5 to 7,
Listing 5: Rules extracted with CART for the Loan data set (56% perturbed).
Listing 6: Rules extracted with GridEx for the Loan data set (56% perturbed).</p>
        <p>Listing 7: Rules extracted with CReEPy for the Loan data set (56% perturbed).</p>
        <p>Listing 8: Rules extracted with CART for the Loan data set (83% perturbed).
respectively.</p>
        <p>In this scenario, the credit history of the applicant remains the primary feature considered
during the prediction phase of the RF model. For instance, the initial split in the input feature
space performed by CART focuses on this attribute. However, the gender feature also plays a
role in predicting the outcomes for a subset of instances, specifically those with a good credit
history. As a result, the SKE techniques demonstrate their efectiveness in identifying unfair
predictors by revealing the influence of the gender attribute on outcomes within specific credit
history subgroups.
4.3.3. 83% Perturbed Data Set
Finally, we report here the results obtained for the Loan data set with a perturbation involving
83% of the female applicants receiving positive outcomes. The disparate impact scores for
this data set and the corresponding trained RF are equal to 0.15 and 0.01, respectively. The
scores highlight severe unfairness. Decision boundaries identified by the RF, CART, GridEx
Listing 9: Rules extracted with GridEx for the Loan data set (83% perturbed).</p>
        <p>Listing 10: Rules extracted with CReEPy for the Loan data set (83% perturbed).
and CReEPy are reported in Figures 1h, 1l, 1p and 1t, respectively. Prolog rules provided by the
SKE techniques are reported in Listings 8 to 10.</p>
        <p>The extracted rules clearly emphasise the significant reliance of the RF predictions on the
gender feature. Despite the decision boundaries being the same as in the previous case study
with a 56% perturbation, in this instance gender is employed as the primary feature for
decisionmaking, followed by credit history as the secondary feature. Essentially, loans are primarily
granted or denied based on gender, with credit history playing a secondary role. The SKE
techniques efectively identify and reveal this unfair behaviour, presenting it to human users in
the form of interpretable logic rules.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This paper provides preliminary insights into the value of leveraging SKE techniques for studying
biases in AI predictors. The findings demonstrate the potential of SKE techniques, particularly
in analysing the relationships between decision outcomes and sensitive input attributes. This
work highlights the importance of considering the correlation between decisions and sensitive
attributes, such as gender, and how SKE can efectively identify and highlight these dependencies.</p>
      <p>Looking ahead, future research will focus on further testing and refining the proposed
approach. This will involve exploring the application of SKE techniques with proxy variables,
investigating intersectional discrimination, and employing counterfactual techniques.
Additionally, the study will delve into the evaluation of diferent fairness metrics to gain a more
comprehensive understanding of bias and discrimination within predictive models.</p>
      <p>Merging the field of AI fairness with explainable AI seems to be a promising approach. By
doing so, we can develop robust methodologies to mitigate biases and promote fairness in AI
systems. The ongoing exploration of SKE techniques holds great promise in fostering a more
equitable and unbiased landscape for AI decision-making.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work has been supported by the EU ICT-48 2020 project TAILOR (No. 952215) and the
European Union’s Horizon Europe AEQUITAS research and innovation programme under grant
number 101070363.
classifiers: A comprehensive survey, CoRR abs/2207.07068 (2022). URL: https://doi.org/10.
48550/arXiv.2207.07068. doi:10.48550/ARXIV.2207.07068. arXiv:2207.07068.
[13] M. Hardt, E. Price, N. Srebro, Equality of opportunity in supervised learning, in: D. D. Lee,
M. Sugiyama, U. von Luxburg, I. Guyon, R. Garnett (Eds.), Advances in Neural Information
Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016,
December 5–10, 2016, Barcelona, Spain, 2016, pp. 3315–3323. URL: https://proceedings.
neurips.cc/paper/2016/hash/9d2682367c3935defcb1f9e247a97c0d-Abstract.html.
[14] J. Chakraborty, K. Peng, T. Menzies, Making fair ML software using trustworthy
explanation, in: 35th IEEE/ACM International Conference on Automated Software Engineering,
ASE 2020, Melbourne, Australia, September 21–25, 2020, IEEE, 2020, pp. 1229–1233. URL:
https://doi.org/10.1145/3324884.3418932. doi:10.1145/3324884.3418932.
[15] D. Madras, T. Pitassi, R. Zemel, Predict responsibly: improving fairness and accuracy by
learning to defer, Advances in Neural Information Processing Systems 31 (2018).
[16] C. Dwork, M. Hardt, T. Pitassi, O. Reingold, R. Zemel, Fairness through awareness, in:
Proceedings of the 3rd innovations in theoretical computer science conference, 2012, pp.
214–226.
[17] G. Bologna, C. Pellegrini, Three medical examples in neural network rule extraction,</p>
      <p>Physica Medica 13 (1997) 183–187. URL: https://archive-ouverte.unige.ch/unige:121360.
[18] L. Franco, J. L. Subirats, I. Molina, E. Alba, J. M. Jerez, Early breast cancer prognosis
prediction and rule extraction using a new constructive neural network algorithm, in:
Computational and Ambient Intelligence (IWANN 2007), volume 4507 of LNCS, Springer,
2007, pp. 1004–1011. doi:0.1007/978-3-540-73007-1_121.
[19] R. Setiono, B. Baesens, C. Mues, Rule extraction from minimal neural networks for credit
card screening, International Journal of Neural Systems 21 (2011) 265–276. doi:10.1142/
S0129065711002821.
[20] F. Sabbatini, C. Grimani, R. Calegari, Bridging machine learning and diagnostics of the esa
lisa space mission with equation discovery via explainable artificial intelligence, Advances
in Space Research 74 (2024) 505–517. URL: https://www.sciencedirect.com/science/article/
pii/S0273117724003880. doi:https://doi.org/10.1016/j.asr.2024.04.041.
[21] R. Guidotti, A. Monreale, S. Ruggieri, F. Turini, F. Giannotti, D. Pedreschi, A survey
of methods for explaining black box models, ACM Computing Surveys 51 (2018) 1–42.
doi:10.1145/3236009.
[22] F. Sabbatini, G. Ciatto, R. Calegari, A. Omicini, Symbolic knowledge extraction from
opaque ML predictors in PSyKE: Platform design &amp; experiments, Intelligenza Artificiale
16 (2022) 27–48. URL: https://doi.org/10.3233/IA-210120. doi:10.3233/IA-210120.
[23] R. Calegari, F. Sabbatini, The PSyKE technology for trustworthy artificial
intelligence 13796 (2023) 3–16. URL: https://doi.org/10.1007/978-3-031-27181-6_1. doi:10.1007/
978-3-031-27181-6_1, xXI International Conference of the Italian Association for
Artiifcial Intelligence, AIxIA 2022, Udine, Italy, November 28 – December 2, 2022, Proceedings.
[24] F. Sabbatini, C. Sirocchi, R. Calegari, Symbolic knowledge comparison: Metrics and
methodologies for multi-agent systems, in: M. Alderighi, M. Baldoni, C. Baroglio, R.
Micalizio, S. Tedeschi (Eds.), Proceedings of the 25th Workshop “From Objects to Agents”, Bard
(Aosta), Italy, July 8–10, 2024, volume 3735 of CEUR Workshop Proceedings, CEUR-WS.org,
2024, pp. 202–216. URL: https://ceur-ws.org/Vol-3735/paper_17.pdf.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Vollmer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. A.</given-names>
            <surname>Mateen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Bohner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. J.</given-names>
            <surname>Király</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Jonsson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Cumbers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jonas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. S.</given-names>
            <surname>McAllister</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Myles</surname>
          </string-name>
          , et al.,
          <article-title>Machine learning and artificial intelligence research for patient benefit: 20 critical questions on transparency, replicability, ethics, and efectiveness</article-title>
          , bmj
          <volume>368</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mehrabi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Morstatter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Saxena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Galstyan</surname>
          </string-name>
          ,
          <article-title>A survey on bias and fairness in machine learning</article-title>
          ,
          <source>ACM Computing Surveys (CSUR) 54</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>On the integration of symbolic and sub-symbolic techniques for XAI: A survey</article-title>
          ,
          <source>Intelligenza Artificiale</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>7</fpage>
          -
          <lpage>32</lpage>
          . doi:
          <volume>10</volume>
          .3233/IA-190036.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agiollo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Magnini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>Symbolic knowledge extraction and injection with sub-symbolic predictors: A systematic literature review</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>56</volume>
          (
          <year>2024</year>
          )
          <volume>161</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>161</lpage>
          :
          <fpage>35</fpage>
          . URL: https://doi.org/10.1145/3645103. doi:
          <volume>10</volume>
          .1145/3645103.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Quy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Iosifidis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , E. Ntoutsi,
          <article-title>A survey on datasets for fairnessaware machine learning</article-title>
          ,
          <source>WIREs Data Mining and Knowledge Discovery</source>
          <volume>12</volume>
          (
          <year>2022</year>
          ). URL: https://doi.org/10.1002/widm.1452. doi:
          <volume>10</volume>
          .1002/WIDM.1452.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>T. van Nuenen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ferrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Such</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Coté</surname>
          </string-name>
          ,
          <article-title>Transparency for whom? assessing discriminatory artificial intelligence</article-title>
          ,
          <source>Computer</source>
          <volume>53</volume>
          (
          <year>2020</year>
          )
          <fpage>36</fpage>
          -
          <lpage>44</lpage>
          . URL: https://doi.org/10. 1109/
          <string-name>
            <surname>MC</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <volume>3002181</volume>
          . doi:
          <volume>10</volume>
          .1109/
          <string-name>
            <surname>MC</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <volume>3002181</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B.</given-names>
            <surname>Wiggins</surname>
          </string-name>
          , Calculating Race:
          <article-title>Racial Discrimination in Risk Assessment</article-title>
          , Oxford University Press,
          <year>2020</year>
          . URL: https://doi.org/10.1093/oso/9780197504000.001.0001. doi:
          <volume>10</volume>
          .1093/oso/ 9780197504000.001.0001.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Calders</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Verwer</surname>
          </string-name>
          ,
          <article-title>Three naive bayes approaches for discrimination-free classification, Data mining and knowledge discovery 21 (</article-title>
          <year>2010</year>
          )
          <fpage>277</fpage>
          -
          <lpage>292</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Pedreshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ruggieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Turini</surname>
          </string-name>
          ,
          <article-title>Discrimination-aware data mining</article-title>
          ,
          <source>in: Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          , KDD '08,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA,
          <year>2008</year>
          , p.
          <fpage>560</fpage>
          -
          <lpage>568</lpage>
          . URL: https://doi.org/10.1145/1401890.1401959. doi:
          <volume>10</volume>
          .1145/1401890.1401959.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R.</given-names>
            <surname>Zemel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Swersky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Pitassi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Dwork</surname>
          </string-name>
          ,
          <article-title>Learning fair representations</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>325</fpage>
          -
          <lpage>333</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kamishima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Akaho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Asoh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sakuma</surname>
          </string-name>
          ,
          <article-title>Fairness-aware classifier with prejudice remover regularizer</article-title>
          ,
          <source>in: Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD</source>
          <year>2012</year>
          ,
          <article-title>Bristol</article-title>
          , UK,
          <source>September 24-28</source>
          ,
          <year>2012</year>
          . Proceedings,
          <source>Part II 23</source>
          , Springer,
          <year>2012</year>
          , pp.
          <fpage>35</fpage>
          -
          <lpage>50</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sarro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Harman</surname>
          </string-name>
          ,
          <article-title>Bias mitigation for machine learning</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          , G. Ciatto,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>On the design of PSyKE: A platform for symbolic knowledge extraction</article-title>
          , in: R. Calegari,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Denti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          , G. Sartor (Eds.),
          <source>WOA 2021 - 22nd Workshop “</source>
          From Objects to Agents”, volume
          <volume>2963</volume>
          <source>of CEUR Workshop Proceedings</source>
          , Sun SITE Central Europe, RWTH Aachen University,
          <year>2021</year>
          , pp.
          <fpage>29</fpage>
          -
          <lpage>48</lpage>
          . 22nd Workshop “From Objects to Agents
          <source>” (WOA</source>
          <year>2021</year>
          ), Bologna, Italy, 1
          <article-title>-3 September 2021</article-title>
          . Proceedings.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>F.</given-names>
            <surname>Pedregosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Varoquaux</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gramfort</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Michel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Thirion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Grisel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blondel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Dubourg</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. VanderPlas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Passos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cournapeau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brucher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Perrot</surname>
          </string-name>
          , E. Duchesnay,
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          ,
          <source>Journal of Machine Learning Research (JMLR) 12</source>
          (
          <year>2011</year>
          )
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          . URL: https://dl.acm.org/doi/10.5555/ 1953048.2078195.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Achieving complete coverage with hypercube-based symbolic knowledge-extraction techniques</article-title>
          , in: S. Nowaczyk,
          <string-name>
            <given-names>P.</given-names>
            <surname>Biecek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. C.</given-names>
            <surname>Chung</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Vallati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Skruch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jaworek-Korjakowska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Parkinson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nikitas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Atzmüller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kliegr</surname>
          </string-name>
          , et al. (Eds.),
          <source>Artificial Intelligence. ECAI 2023 International Workshops - XAI 3</source>
          ,
          <string-name>
            <surname>TACTIFUL,</surname>
          </string-name>
          <article-title>XI-ML, SEDAMI, RAAIT, AI4S</article-title>
          , HYDRA, AI4AI, Kraków, Poland,
          <source>September 30 - October 4</source>
          ,
          <year>2023</year>
          , Proceedings,
          <string-name>
            <surname>Part</surname>
            <given-names>I</given-names>
          </string-name>
          , volume
          <volume>1947</volume>
          <source>of Communications in Computer and Information Science</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>179</fpage>
          -
          <lpage>197</lpage>
          . URL: https://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -50396-2_
          <fpage>10</fpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -50396-2\_
          <fpage>10</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>Semantic Web-based interoperability for intelligent agents with PSyKE</article-title>
          , in: D.
          <string-name>
            <surname>Calvaresi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Najjar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Winikof</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Främling (Eds.),
          <source>Explainable and Transparent AI</source>
          and
          <string-name>
            <surname>Multi-Agent</surname>
            <given-names>Systems</given-names>
          </string-name>
          , volume
          <volume>13283</volume>
          of Lecture Notes in Computer Science, Springer,
          <year>2022</year>
          , pp.
          <fpage>124</fpage>
          -
          <lpage>142</lpage>
          . URL: http://link.springer.com/10.1007/ 978-3-
          <fpage>031</fpage>
          -15565-
          <issue>9</issue>
          _8. doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -15565-
          <issue>9</issue>
          _
          <fpage>8</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>On the evaluation of the symbolic knowledge extracted from black boxes</article-title>
          ,
          <source>AI and Ethics</source>
          <volume>4</volume>
          (
          <year>2024</year>
          )
          <fpage>65</fpage>
          -
          <lpage>74</lpage>
          . doi:https://doi.org/10.1007/ s43681-023-00406-1.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Symbolic knowledge-extraction evaluation metrics: The FiRe score</article-title>
          , in: K.
          <string-name>
            <surname>Gal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Nowé</surname>
            ,
            <given-names>G. J.</given-names>
          </string-name>
          <string-name>
            <surname>Nalepa</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Fairstein</surname>
          </string-name>
          , R. Rădulescu (Eds.),
          <source>Proceedings of the 26th European Conference on Artificial Intelligence, ECAI</source>
          <year>2023</year>
          , Kraków, Poland.
          <source>September 30 - October 4</source>
          ,
          <year>2023</year>
          ,
          <year>2023</year>
          . URL: https://ebooks.iospress.nl/doi/10.3233/FAIA230496. doi:
          <volume>10</volume>
          .3233/FAIA230496.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Omicini,
          <string-name>
            <surname>GridEx:</surname>
          </string-name>
          <article-title>An algorithm for knowledge extraction from black-box regressors</article-title>
          , in: D.
          <string-name>
            <surname>Calvaresi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Najjar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Winikof</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Främling (Eds.),
          <source>Explainable and Transparent AI</source>
          and
          <string-name>
            <surname>Multi-Agent Systems</surname>
            . Third International Workshop, EXTRAAMAS 2021,
            <given-names>Virtual</given-names>
          </string-name>
          <string-name>
            <surname>Event</surname>
          </string-name>
          , May 3-
          <issue>7</issue>
          ,
          <year>2021</year>
          , Revised Selected Papers, volume
          <volume>12688</volume>
          <source>of LNCS</source>
          , Springer Nature, Basel, Switzerland,
          <year>2021</year>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>38</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -82017-
          <issue>6</issue>
          _
          <fpage>2</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          , G. Ciatto,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>Hypercube-based methods for symbolic knowledge extraction: Towards a unified model</article-title>
          , in: A.
          <string-name>
            <surname>Ferrando</surname>
          </string-name>
          , V. Mascardi (Eds.),
          <source>WOA 2022 - 23rd Workshop “</source>
          From Objects to Agents”, volume
          <volume>3261</volume>
          <source>of CEUR Workshop Proceedings</source>
          , Sun SITE Central Europe, RWTH Aachen University,
          <year>2022</year>
          , pp.
          <fpage>48</fpage>
          -
          <lpage>60</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3261</volume>
          /paper4.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          , G. Ciatto,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>Towards a unified model for symbolic knowledge extraction with hypercube-based methods</article-title>
          ,
          <source>Intelligenza Artificiale</source>
          <volume>17</volume>
          (
          <year>2023</year>
          )
          <fpage>63</fpage>
          -
          <lpage>75</lpage>
          . URL: https://doi.org/10.3233/IA-230001. doi:
          <volume>10</volume>
          .3233/IA-230001.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Symbolic knowledge extraction from opaque machine learning predictors: GridREx &amp; PEDRO</article-title>
          , in: G.
          <string-name>
            <surname>Kern-Isberner</surname>
          </string-name>
          , G. Lakemeyer, T. Meyer (Eds.),
          <source>Proceedings of the 19th International Conference on Principles of Knowledge Representation and Reasoning</source>
          , KR 2022, Haifa,
          <source>Israel. July 31 - August 5</source>
          ,
          <year>2022</year>
          ,
          <year>2022</year>
          . URL: https://proceedings.kr.org/
          <year>2022</year>
          /57/.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Friedman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. J.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. A.</given-names>
            <surname>Olshen</surname>
          </string-name>
          ,
          <article-title>Classification and Regression Trees</article-title>
          , CRC Press,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Unveiling opaque predictors via explainable clustering: The CReEPy algorithm</article-title>
          , in: G. Boella,
          <string-name>
            <given-names>F.</given-names>
            <surname>A. D'Asaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dyoub</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gorrieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. A.</given-names>
            <surname>Lisi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Manganini</surname>
          </string-name>
          , G. Primiero (Eds.),
          <source>Proceedings of the 2nd Workshop on Bias</source>
          ,
          <string-name>
            <surname>Ethical</surname>
            <given-names>AI</given-names>
          </string-name>
          ,
          <article-title>Explainability and the role of Logic and Logic Programming co-located with the 22nd International Conference of the Italian Association for Artificial Intelligence (AI*IA</article-title>
          <year>2023</year>
          ), Rome, Italy, November 6,
          <year>2023</year>
          , volume
          <volume>3615</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3615</volume>
          /paper1.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Untying black boxes with clustering-based symbolic knowledge extraction</article-title>
          ,
          <source>Intelligenza Artificiale</source>
          <volume>18</volume>
          (
          <year>2024</year>
          )
          <fpage>21</fpage>
          -
          <lpage>34</lpage>
          . URL: https://doi.org/10.3233/IA-240026. doi:
          <volume>10</volume>
          .3233/IA-240026.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Bottom-up and top-down workflows for hypercube- and clustering-based knowledge extractors</article-title>
          , in: D.
          <string-name>
            <surname>Calvaresi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Najjar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Omicini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aydogan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Carli</surname>
            , G. Ciatto,
            <given-names>K.</given-names>
          </string-name>
          Främling (Eds.),
          <source>Explainable and Transparent AI</source>
          and
          <string-name>
            <surname>Multi-Agent Systems</surname>
          </string-name>
          . Fifth International Workshop, EXTRAAMAS 2023, London, UK, May
          <volume>29</volume>
          ,
          <year>2023</year>
          , Revised Selected Papers, volume
          <volume>14127</volume>
          <source>of LNCS</source>
          , Springer Cham, Basel, Switzerland,
          <year>2023</year>
          , pp.
          <fpage>116</fpage>
          -
          <lpage>129</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -40878-
          <issue>6</issue>
          _
          <fpage>7</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Unlocking insights and trust: The value of explainable clustering algorithms for cognitive agents</article-title>
          , in: R.
          <string-name>
            <surname>Falcone</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Castelfranchi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Sapienza</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Cantucci</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the 24th Workshop “</source>
          From Objects to Agents”, Roma, Italy, November 6-
          <issue>8</issue>
          ,
          <year>2023</year>
          , volume
          <volume>3579</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>232</fpage>
          -
          <lpage>245</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3579</volume>
          /paper18.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <article-title>Explainable clustering with CREAM</article-title>
          , in: P. Marquis,
          <string-name>
            <given-names>C. S.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          (Eds.),
          <source>20th International Conference on Principles of Knowledge Representation and Reasoning (KR</source>
          <year>2023</year>
          ), IJCAI Organization, Rhodes, Greece,
          <year>2023</year>
          , pp.
          <fpage>593</fpage>
          -
          <lpage>603</lpage>
          . doi:
          <volume>10</volume>
          .24963/kr.2023/58.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>F.</given-names>
            <surname>Sabbatini</surname>
          </string-name>
          , R. Calegari,
          <article-title>ExACT explainable clustering: Unravelling the intricacies of cluster formation</article-title>
          , in: C.
          <string-name>
            <surname>K. Baker</surname>
            ,
            <given-names>L. Gómez</given-names>
          </string-name>
          <string-name>
            <surname>Álvarez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Heyninck</surname>
          </string-name>
          , T. Meyer, R. Peñaloza, S. Vesic (Eds.),
          <source>Joint Proceedings of the 2nd Workshop on Knowledge Diversity and the 2nd Workshop on Cognitive Aspects of Knowledge Representation co-located with 20th International Conference on Principles of Knowledge Representation and Reasoning (KR</source>
          <year>2023</year>
          ), Rhodes, Greece, September 3-
          <issue>4</issue>
          ,
          <year>2023</year>
          , volume
          <volume>3548</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2023</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3548</volume>
          /paper3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>M.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Friedler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Moeller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Scheidegger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Venkatasubramanian</surname>
          </string-name>
          ,
          <article-title>Certifying and removing disparate impact</article-title>
          , in: L.
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
            , T. Joachims,
            <given-names>G. I.</given-names>
          </string-name>
          <string-name>
            <surname>Webb</surname>
            ,
            <given-names>D. D.</given-names>
          </string-name>
          <string-name>
            <surname>Margineantu</surname>
          </string-name>
          , G. Williams (Eds.),
          <source>Proceedings of the 21th ACM SIGKDD International</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>