<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Global Model-Agnostic XAI method for the Automatic Formation of an Abstract Argumentation Framework and its Objective Evaluation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giulia Vilone</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luca Longo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>The Artificial Intelligence and Cognitive Load research lab, The applied Intelligence Research Center, School of Computer Science, Technological University Dublin</institution>
          ,
          <addr-line>Dublin</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Explainable Artificial Intelligence (XAI) aims to train data-driven, machine learning (ML) models possessing both high predictive accuracy and a high degree of explainability for humans. Comprehending and explaining the inferences of a model can be seen as a defeasible reasoning process which is expected to be non-monotonic meaning that a conclusion, linked to a set of premises, can be withdrawn when new information becomes available. Computational argumentation, a paradigm within Artificial Intelligence (AI), focuses on modeling defeasible reasoning. This research study explored a new way for the automatic formation of an argument-based representation of the inference process of a data-driven ML model to enhance its explainability by employing principles and techniques from computational argumentation, including weighted attacks within its argumentation process. An experiment was conducted on vfie datasets to test, in an objective manner, if the explanations of the proposed XAI method are more comprehensible than decision trees, which are considered naturally transparent. Findings demonstrate that usually the argument-based method can represent the logic of the model with fewer rules than a decision tree, but further work is required to achieve the same performances in terms of other characteristics, such as fidelity to the model.</p>
      </abstract>
      <kwd-group>
        <kwd>Explainable artificial intelligence</kwd>
        <kwd>Argumentation</kwd>
        <kwd>Non-monotonic reasoning</kwd>
        <kwd>Method evaluation</kwd>
        <kwd>Metrics of explainability</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        XAI, a sub-field of AI, aims to develop a unified approach to learning data-driven models that are
both highly accurate in their predictions and explainable to experts and laypeople. The explosion
in the quantity of available data and the success of ML, especially Deep Learning, have led to the
development of new models with outstanding predictive performances. However, most of these
models have complex, non-linear structures that are hard to understand and explain. Researchers
have proposed numerous XAI methods generating explanations in different formats (numerical,
rules, textual, visual or mixed) [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. The XAI methods returning rule-based explanations extract
a set of rules mimicking the inferential process of a complex ML model [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. However, these
methods do not necessarily capture and describe the actual inferential process. They just report
the relationships between inputs and outputs as learned by the model without verifying if they
are consistent with the background knowledge of the application field or are instead based on
spurious correlations of the data. Understanding the inferential process of a model should be
seen as a non-monotonic reasoning process [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This requires a mechanism replicating the way
human reasons to support humans in the comprehension of the inherent inferential process learnt
by a model. Argumentation is a multidisciplinary subfield of AI that studies how arguments
can be presented, supported or discarded in a defeasible reasoning process. It also investigates
formal approaches to evaluate the validity of the conclusions reached at the end of the reasoning
process[
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. Argumentation Theory (AT) provides the basis for implementing these processes
computationally [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and it is inspired by how humans reason. This research experiment shows that
AT can be a viable solution for building novel global model-agnostic XAI methods generating
argument-based explanations. The quality of these explanations was preliminarily tested via
an objective study based on eight quantitative metrics that assess distinct aspects of rule-based
explanations, thus providing vital insights on the inferential process of a ML model [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], and
compared to another rule-extraction XAI method generating Decision Trees (DTs), which are
considered as naturally transparent [
        <xref ref-type="bibr" rid="ref3 ref8">8, 3</xref>
        ].
      </p>
      <p>The remainder of this manuscript is organised as follows. Section 2 summarises the strategies
used by scholars to generate rule-based explanations of ML models and how to assess the quality
of these explanations. Section 3 describes the design of a primary research experiment. Section
4 discusses the findings of this experiment and its limitations. Lastly, Section 5 highlights the
contribution to the existing body of knowledge and suggests future directions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        Rule-based explanations are a structured but still intuitive format for reporting information to
humans in a compact way. They represent the logic of a ML model as a ruleset that can be
easily read, interpreted and visualised. Therefore, scholars consider rulesets and DTs as naturally
transparent and intelligible [
        <xref ref-type="bibr" rid="ref3 ref8">8, 3</xref>
        ]. However, current rule-extraction XAI methods merely produce
a rulesets mimicking the inferential process of an underlying complex model. The rules can
also be in conflict with the expert domain knowledge, thus perplexing the users of such models.
It must remember that such rules aim at faithfully representing the relationships captured by
the model during its training process between the independent variables of the input data with
its target variable. Thus, this conflict can be an essential signal of an issue occurring during
training. Similarly, the XAI methods do not provide any tool to handle potential inconsistencies
among the extracted rules, should they arise. Thus, these rules are not suitable to support a richer
reasoning process [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. AT provides formal approaches to model non-monotonic logic and assess
the validity of the conclusions reached by a set of arguments to be considered as acceptable
[
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. Non-monotonic logic consists of a family of formal frameworks devised to capture and
represent defeasible inferences. In formal logic, a defeasible concept consists of a set of pieces
of information or arguments that can be rebutted by additional information or arguments [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
Generally, arguments are designed by domain experts to create a knowledge-base in single or
multi-agent environments [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. In a single-agent environment, arguments are constructed by
an autonomous reasoner, thus conflictual information tends to be minimal. In a multi-agent
environment, multiple reasoners participate in argument construction, so more conflicts among
them usually arise, enabling in practice non-monotonic reasoning [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Defeasible argumentation
supplies a sound formalisation for reasoning with uncertain and incomplete information from a
defeasible knowledge-base [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The process of defeasible argumentation frequently requires the
recursive analysis of conflicting arguments in a dialectical setting to determine which arguments
should be accepted or discarded [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Abstract AT (AAT) is the dominant paradigm, whereby
arguments are abstractly considered in a dialogical structure. Formal semantics are habitually
adopted to identify conflict-free sets of arguments that can subsequently support decision-making,
explanations and justification [
        <xref ref-type="bibr" rid="ref14 ref6">14, 6</xref>
        ]. Existing AAT-based frameworks have common features:
[
        <xref ref-type="bibr" rid="ref13 ref15 ref16">13, 15, 16</xref>
        ]:
• a defeasible knowledge-base in the form of interactive arguments, usually formalised with
a first-order logical language;
• a set of attacks that are modelled whenever two arguments are in conflict;
• a semantic which consists of mechanism for conflict resolution. It implements in practice
non-monotonicity and provides a dialectical status to the arguments.
      </p>
      <p>
        The integration between AT and ML is still a young field. Minimal work exists on automatic
argument and attack mining from data-driven ML models, how the interpretation of these models
can be augmented via argumentation to, in turn, improve their explainability. [
        <xref ref-type="bibr" rid="ref13 ref15 ref17">17, 13, 15</xref>
        ].
In relation to this, the first issue is the automatic extraction of rules and their conflicts from
these models. The second issue is their automatic integration into an argumentation framework
that can serve as a mechanism for interpreting and explaining the inferential process of such
models without any explicit human declarative knowledge. A two-step approach for AT-ML
integration was proposed in [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. In the first step, rules are extracted from a given dataset with
the Apriori algorithm for mining association rules. In the second step, the rules are fed into
structured argumentation approaches, such as ASPIC+ [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. Using their argumentative inferential
procedures, new observations are classified by constructing arguments on top of these rules and
determining their justification status. Another study exploits argumentative graphs to depict the
structure of argument-based frameworks [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Arguments are the nodes connected by directed
edges representing attacks. The status of the arguments is provided by a label (accepted or
rejected) and is determined by using argumentation semantics [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Design</title>
      <p>The informal research hypothesis of this study is that a ruleset extracted by an XAI method from
data-driven ML models supports the automatic formation of an argumentation framework. The
expectation is that this framework possesses a higher degree of explainability when compared
to other formats of explanations considered naturally interpretable and transparent in Computer
Science, like a DT. The difference in the degree of explainability of the two methods was tested
in an objective and quantitative manner with eight metrics that measure different aspects of a
ruleset, such as number and length of its rules. The research hypothesis was tested by carrying
out a set of phases described in the following paragraphs and depicted in the diagram of Fig. 1.</p>
      <sec id="sec-3-1">
        <title>3.1. Phase 1: Dataset preparation</title>
        <p>The first step was to select a few training datasets containing multi-dimensional data built by
domain experts, so they cannot contain data produced by an algorithm. The datasets must
not present issues that can impede the successful training of a model, such as the course of
dimensionality or a significant portion of missing data. The labeled target variable, represented
by block YT in Fig. 1, must be categorical, ideally with more than two target classes, whereas the
independent features should be a mix of continuous and categorical predictors. In this study, the
experiment was carried out on vfie public datasets downloaded from Kaggleor the UCI Machine
Learning Repository (see Tab. 1).The Adult database, based on the 1994 US Census, was designed
to train ML models to predict if a person earns or not over $50K on annual basis. Avila contains
data about 800 images of a Latin copy of the Bible, called the Avila Bible, manufactured during
the XII century by 12 Italian and Spanish copyists who were individuated from a paleographic
analysis of the manuscript. The model must associate each image with the copyist who drew
it. The Credit Card Default dataset was created to train ML models that predict if Taiwanese
clients will fail to repay their credit card debts. The Hotel Bookings dataset includes booking
information for a city hotel and a resort hotel such as the booking date, length of stay, and the
number of adult and child guests, among other things. The target variable represents the final
status of the reservation, whether it was cancelled, checked-out or the client did not show up.
Online Shopper Intention records thousands of sessions on e-commerce websites. The negative
target class represents customers who did not buy anything, whilst the positive class are sessions
that ended with a purchase.</p>
        <p>
          The datasets were preprocessed to avoid data-related issues in the model’s training process.
None of the selected datasets have missing data, so no action was required. However, the input
features “fnlwgt” of the Adult dataset, which is the statistical weights measuring how many
US citizens are represented by each subject, and the Client ID from the Credit Card Default
dataset had to be discarded because they did not represent discriminative attributes. All the data
in the independent features were scaled into the range [
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ]. Features with very large values
might dominate over other in the training process of the model. Then, a correlation analysis
was performed on each dataset to detect pairs of highly correlated features and discard one of
the two to reduce the risk of multicollinearity. There is no consensus on the thresholds between
strong, moderate and weak correlations. In this study, the absolute Spearman’s rank correlation
coefcfiients were grouped into three segments: values in the range (0, 0.33) were considered
weak, (0.33, 0.66) moderate, and (0.66, 1) strong correlations. The best subset selection analysis
was carried out to chose which variable from a strongly-correlated pair had to be discarded [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ].
A linear regression model was built over each combination of the independent features excluding
one from each strongly correlated pairs. These models were then sorted in descending order
according to their R2 values and the first one was selected. The best subset selection approach was
chosen for its simplicity and because it requires little computational time and resources. Some of
the chosen datasets are unbalanced, meaning that one specific class contains more instances than
the others. This disparity can lead some learning algorithms to classify all the instances into the
majority class and ignore the minority one. To avoid this, each dataset was split into a training
and a validation subsets with the stratified vfie-fold cross-validation technique to ensure that
each class was represented with the same proportion as in the original dataset. Furthermore, the
Synthetic Minority Over-Sampling Technique (SMOTE) [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] was applied to the training datasets
to up-sample the minority classes.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Phase 2: Model training</title>
        <p>A feed-forward neural network with two fully-connected hidden layers was trained on each
datasets to fit YT . The block YM in Fig. 1 represents the predictions obtained from the trained
model (represented by block f (x)) over the evaluation dataset (test data) whose original labelled
target variable is depicted by block YE . YE is compared with YM to assess the model’s prediction
accuracy. The number of hidden nodes and the value of other model’s hyperparameters, reported
in Tab. 2, were determined with a grid search to reach the highest feasible prediction accuracy.
To avoid overfitting, the training process was early stopped when the validation accuracy did not
improve for vfie epochs in a row. The networks were trained vfie times over the vfie training
subsets extracted from the datasets with the vfie-fold cross-validation technique. The models
with the highest validation accuracy were chosen. Lastly, the not relevant input features were
pruned by recursively removing one at a time, retraining the selected model and checking if its
prediction accuracy decreased. If this was not the case, the pruned variable was removed.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Phase 3: Formation of the explainable argumentation framework</title>
        <p>
          The trained models were translated into an explainable argument-based representation which can
be easily embedded into an online interactive platform where the argumentation framework is
represented as a graph (an example can be found in [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], page 10). The process of argumentation
towards the achievement of a justifiable conclusion, as emerged from theoretical works of AT, can
be broken down into vfie layers [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], as depicted in Fig. 2 and detailed in the following subsections.
        </p>
        <p>Layer 1: definition of the internal structure of arguments. In standard logic, an argument
consists of a set premises leading to a conclusion, or more formally:
Definition 3.1 (Argument). An argument Ar is a tentative inference → that links one or more
premises Pi to a conclusion C and can be written as Ar : P1, . . . , Pn → C.</p>
        <p>
          In this study, an argument corresponds to an IF-THEN rule, thus the premises and conclusion
of an argument correspond to the rule’s antecedents and conclusion. The ML models and the
evaluation datasets were fed into a bespoke rule-extraction method that generates a set of
IFTHEN rules by using a two-step algorithm. First, each dataset was divided into groups according
to the target class as predicted by the model. In other words, all the instances assigned by the
model to the same class were grouped together. Second, the Ordering Points To Identify the
Clustering Structure (OPTICS) [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ] algorithm was exploited to further split the groups into
clusters that coincide with areas of the input space having a high density of samples. Then, each
cluster was translated into a rule by finding, for each relevant feature, the minimum and maximum
values that include all the samples in the cluster. These ranges determine the rule’s antecedents,
whereas the conclusion corresponds to the predicted class of the cluster’s samples. A typical rule
is:
        </p>
        <p>IF m1 ≤</p>
        <p>X1 ≤</p>
        <p>M1 AND . . . AND mN ≤</p>
        <p>XN ≤</p>
        <p>MN T HEN ClassX
(1)
where Xi, i = 1, . . . , N are the N independent relevant features, mi and Mi, i = 1, . . . , N are the
minimum and maximum values w.r.t the i-th independent feature of the samples included in the cluster.</p>
        <p>
          Layer 2: definition of the attacks between arguments. The inconsistencies between the formed
arguments were modelled via the notion of attack. Generally, attacks are binary relations between
two conflicting arguments. They can be of different kinds [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], but only the following two types
were considered in this study.
        </p>
        <p>Definition 3.2 (Rebutting attack). Given two distinct arguments A, B ∈ AR, where AR represents
the set of all the arguments, with A : P1, . . . , Pn → C1, B : P1, . . . , Pm → C2, A is rebuttal of B and
is denoted as (A, B) if C1 logically contradicts C2. A rebuttal attack is symmetrical, so it holds
that iff (A, B), then ∃(B, A).</p>
        <p>Definition 3.3 (Undercutting attack). Given an argument A ∈ AR that challenges some or all of
the premises used to construct another argument B ∈ AR, A undercuts B and is denoted as (A, B)
when A claims there is a special case that does not allow the application of the inference rule
(→) of argument B.</p>
        <p>
          Attacks are usually specified by domain experts, but in this study they can be automatically
extracting by identifying conflicting rules. Two rules are conflictual if they are overlapping
and reach different conclusions. Two rules overlap if their covers intersect. The cover of a rule
corresponds to the set of input instances whose attribute values satisfy the rule’s antecedents [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ].
As depicted in Fig. 3, two rules can be 1) fully overlapping, with one rule including the second
one (part a), 2) partially overlapping (part b) or 3) sharing the same cover (part c). The first case
could be seen as an undercutting attack because the internal rule represents an exception of the
external one. The remaining two cases could be equivalent to a rebutting attack as two rules start
from the same premises, at least in part, but reach different conclusions.
        </p>
        <p>(a) Undercutting attack
(b) Rebutting attack
(c) Rebutting attack</p>
        <p>
          Layer 3: evaluation and definition of valid attacks. Once arguments and attacks are embodied in
a dialogical structure, the formalised knowledge-base, a fundamental characteristics of
argumentbased systems is their ability to determine the success of an attack. Different approaches can be
found in the literature to decide if an attack is successful, thus valid, including a) binary attacks,
b) strengths of arguments, and c) strengths of attacks [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. In this study, a weighted notion of
attack is considered; weights represent the strength of the attacks. There are various ways to
compute these weights [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. Here, they are computed as the percentage of instances belonging to
the intersection of the covers of two conflictual rules that are assigned by the model to the same
target class of the conclusion of the attacking rule:
w(A,B) = |{x ∈ cover(A) ∩ cover(B) : f (x) = CA}|
|{x ∈ cover(A) ∩ cover(B)}|
(2)
where x represents an input instance of the training dataset, CA is the conclusion of the attacking
rule (argument) A, and | • | is the cardinality function. For example, two conflicting rules have
respectively target classes Q and S as conclusion and there are in their cover intersection 20
instances classiefid by the model in class Q and 30 in class S. In this case, the attack from the
second rule with conclusion S is stronger than the attack from the first rule and has a weight
equal to 3500 . The weight of the reciprocal attack is 2500 . It might happen that the difference in the
number of instances per class is small, like 20 versus 21. In this case, is it fair to say that the rule
with conclusion S is actually stronger than the other rule? As a consequence, the concept of
inconsistency budget [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ] was used to set a threshold on the fraction of supporting instances
of the attacking rules. In this study, it was set equal to 0.55, meaning that an attack must be
supported by at least 55% of the samples in the cover intersection. Future work will involve a
study to fine-tune it. It is important to underline that all the arguments formed in layer one have
the same importance and the notion of weight of argument is not used in this study. Not all the
arguments are activated by each training instance since not all their premises might be satisfied.
The activated portion of the knowledge-base is considered for the next computations.
        </p>
        <p>
          Layer 4: definition of the dialectal status of arguments. Dung-style acceptability semantics
investigate the inconsistencies that might emerge from the interaction of arguments [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. Given a
set of arguments where some attack others, it must be decided which arguments can be accepted.
In Dung’s theory, the internal structure of arguments is not considered. This leads to an abstract
argumentation framework (AAF) which is a finite set of arguments and attacks. In Dung’s terms,
usually, an argument defeats another argument if and only if it represents a reason against the
second argument. It is also essential to assess whether the defeaters are defeated themselves to
determine the acceptability status of an argument. This is known as acceptability semantics:
given an AAF, it specifies zero or more conflict-free sets of acceptable arguments. However,
other semantics have been proposed in the literature, not necessarily based on the notion of
acceptability, such as the ranking-base categoriser semantic, introduced by [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] and employed in
this experiment, which consists of a recursive function that rank-orders a set of arguments from
the most to the list acceptable. The rank of an argument is inversely proportional to the number
of its attacks and the rank of the attacking arguments. This semantic deems as acceptable the
argument(s) with the lowest number of attacks and/or attacks coming from the weakest arguments.
        </p>
        <p>Layer 5: Accrual of acceptable arguments. The previous layer produces a rank of activated
arguments, and a final conclusion should be brought forward as the most rational conclusion
associable to a single input instance. The highest-ranked argument is selected as the most
representative, and its conclusion is deemed the most rationale. In the case of ties (multiple
arguments with the highest rank), these are grouped into sets according to the conclusion they
support. The set with the highest cardinality is deemed the most representative of an input record
of the dataset, and the conclusion supported by its argument(s) is deemed the most rationale. In
the case of ties with respect to cardinality, the input case is treated as undecided, as not enough
information is available to associate a possible conclusion.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Phase 4: Objective evaluation analysis</title>
        <p>
          The evaluation of the degree of explainability of the two XAI methods, the argument-based one
developed in this study and the DT created with the C4.5 learning algorithm, followed the same
process proposed in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. Eight metrics were selected to assess, objectively and quantitatively, the
degree of explainability of their rulesets (see Tab. 3). The objectivity is achieved by excluding
any human intervention in this evaluation process. Two metrics, number of rules and average rule
length, measure the syntactic simplicity of the rules and should be minimised as short rulesets are
deemed more interpretable [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]. Fraction of classes and fraction of overlap enhance the clarity
and coherence of the extracted rules. Whilst the fraction overlap should be minimised to avoid
conflicts between the rules, the fraction of classes should be maximised to guarantee that all the
target classes, even the minor ones, are considered. A ruleset must also be complete, correct,
faithful to the model’s predictions, and robust to small perturbations of the inputs. To assure that
the C4.5 algorithm returned the most compact and accurate DT, a grid search was carried out on
the following hyperparameters: 1) the criterion function to measure the quality of a split (Gini,
Entropy, Log-Loss), 2) the maximum depth of the DT (from 6 to 48), and the minimum number
of instances required to 3) split an internal node (from 2 to 16) and 4) be at a leaf node (from 1 to
8).
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and discussion</title>
      <p>The values of the metrics calculated over the two rulesets extracted with the C4.5 learning
algorithm and the proposed argument-based XAI method are summarised in Tab. 4. Both methods
generate complete rulesets, meaning that they cover the entire input space and all the output
classes. The only exception occurs in the Avila dataset where the ruleset of the argument-based
method does not consider one of the 12 output classes. This is due to the presence of several
attacks, some of which have high weights, towards the rules having this class in their conclusions.
Modifying the inconsistency budget might fix this issue. The C4.5 method scores higher in terms
of correctness, fidelity and robustness throughout the vfie datasets. It can also be considered
the most coherent method as its rulesets reach completeness without overlapping areas. On the
other hand, it generated rulesets that contains more and longer rules than the argument-based
method with only one exception represented by the Online Shopper Intention datasets where
the argument-based method extracted more rules than the C4.5. However, these rules contained
less antecedents, on average. In the other three datasets (Adult, Avila, and Hotel Bookings),
the C4.5 returns thousands of rules whereas the argument-based method never reaches the 500
rules. Such big numbers of rules would hinder the explainability of these rulesets as struggle
with reading and retaining such a big amount of information. Overall, the argument-based XAI
method generates simpler ruleset that are potentially more comprehensible than the C4.5 DT,
but there is the need to identify a way to fine tune the inconsistency budget to reach the optimal
argumentation framework.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>
        This study presented a novel XAI method to form an argumentation framework with weighted
attacks representing the inferential process of complex data-driven ML models. These models
were trained on vfie datasets with handcrafted features manually engineered by humans. Eight
quantitative and objective metrics were used to assess the degree of explainability of the rulesets
extracted by the proposed XAI method and a DT, used as baseline. The results suggested the
presence of a trade-off between completeness, number of rules and average length, measuring
the syntactic simplicity of the rulesets, and the other vfie metrics. The C4.5 algorithm usually
generate bigger rulesets, but it is more correct and faithful to the model than the argument-based
method. In conclusion, the proposed XAI method returns rulesets that are complete, simpler and
smaller in terms of rule cardinality and length, thus more comprehensible. However, they are not
as faithful to the model, correct and robust as the C4.5 DTs. Future work will extend this research
study by training deeper neural networks, employing datasets with additional types of input data,
like texts and images, fine tuning the inconsistency budget between weighted attacks to obtain
the optimal set of arguments and attacks, and using semantics designed for handling weighted
argumentation frameworks. The evaluation of the argumentation frameworks will include a
human-centred study, as done in [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], to compare the outcome of the objective metrics with users’
perception of their explainability.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Goebel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Lecue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kieseberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Holzinger</surname>
          </string-name>
          ,
          <source>Explainable artificial intelligence: Concepts</source>
          , applications, research challenges and visions, in: International Cross-Domain
          <source>Conference for Machine Learning and Knowledge Extraction</source>
          , Springer,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>Classification of explainable artificial intelligence methods through their output formats</article-title>
          ,
          <source>Machine Learning and Knowledge Extraction</source>
          <volume>3</volume>
          (
          <year>2021</year>
          )
          <fpage>615</fpage>
          -
          <lpage>661</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F. K.</given-names>
            <surname>Došilovic´</surname>
          </string-name>
          , M. Brcˇic´,
          <string-name>
            <surname>N.</surname>
          </string-name>
          <article-title>Hlupic´, Explainable artificial intelligence: A survey</article-title>
          ,
          <source>in: 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO)</source>
          , IEEE,
          <year>2018</year>
          , pp.
          <fpage>0210</fpage>
          -
          <lpage>0215</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>A novel human-centred evaluation approach and an argument-based method for explainable artificial intelligence</article-title>
          ,
          <source>in: IFIP International Conference on Artificial Intelligence Applications and Innovations</source>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>447</fpage>
          -
          <lpage>460</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Bryant</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Krause</surname>
          </string-name>
          ,
          <article-title>A review of current defeasible reasoning implementations</article-title>
          ,
          <source>The Knowledge Engineering Review</source>
          <volume>23</volume>
          (
          <year>2008</year>
          )
          <fpage>227</fpage>
          -
          <lpage>260</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>Argumentation for knowledge representation, conflict resolution, defeasible inference and its integration with machine learning</article-title>
          ,
          <source>in: Machine Learning for Health Informatics</source>
          , Springer,
          <year>2016</year>
          , pp.
          <fpage>183</fpage>
          -
          <lpage>208</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Vilone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>A quantitative evaluation of global, rule-based explanations of post-hoc, model agnostic methods</article-title>
          ,
          <source>Frontiers in artificial intelligence 4</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H. K.</given-names>
            <surname>Dam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ghose</surname>
          </string-name>
          ,
          <article-title>Explainable software analytics</article-title>
          ,
          <source>in: Proceedings of the 40th International Conference on Software Engineering: New Ideas and Emerging Results</source>
          , ACM,
          <year>2018</year>
          , pp.
          <fpage>53</fpage>
          -
          <lpage>56</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z. C.</given-names>
            <surname>Lipton</surname>
          </string-name>
          ,
          <article-title>The mythos of model interpretability</article-title>
          ,
          <source>Commun. ACM</source>
          <volume>61</volume>
          (
          <year>2018</year>
          )
          <fpage>36</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <article-title>Formalising human mental workload as a defeasible computational concept</article-title>
          , The University of Dublin, Trinity College,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <surname>L. Longo,</surname>
          </string-name>
          <article-title>An empirical evaluation of the inferential capacity of defeasible argumentation, non-monotonic fuzzy reasoning and expert systems</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>147</volume>
          (
          <year>2020</year>
          )
          <fpage>113220</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Longo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rizzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dondio</surname>
          </string-name>
          ,
          <article-title>Examining the modelling capabilities of defeasible argumentation and non-monotonic fuzzy reasoning</article-title>
          ,
          <source>Knowledge-Based Systems</source>
          <volume>211</volume>
          (
          <year>2021</year>
          )
          <fpage>106514</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Gómez</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. I. Chesnevar</surname>
          </string-name>
          ,
          <article-title>Integrating defeasible argumentation and machine learning techniques: A preliminary report</article-title>
          , in: In Procs. V Workshop of Researchers in Comp. Science,
          <year>2003</year>
          , pp.
          <fpage>320</fpage>
          -
          <lpage>324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>P. M. Dung</surname>
          </string-name>
          ,
          <article-title>On the acceptability of arguments and its fundamental role in nonmonotonic reasoning, logic programming and n-person games</article-title>
          ,
          <source>Artificial intelligence 77</source>
          (
          <year>1995</year>
          )
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Modgil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Toni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bex</surname>
          </string-name>
          , I. Bratko,
          <string-name>
            <given-names>C. I.</given-names>
            <surname>Chesnevar</surname>
          </string-name>
          , W. Dvorˇák,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Falappa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Gaggl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>García</surname>
          </string-name>
          , et al.,
          <article-title>The added value of argumentation</article-title>
          , in: Agreement technologies, Springer,
          <year>2013</year>
          , pp.
          <fpage>357</fpage>
          -
          <lpage>403</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Gómez</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. I. Chesnevar</surname>
          </string-name>
          ,
          <article-title>Integrating defeasible argumentation with fuzzy art neural networks for pattern classification</article-title>
          ,
          <source>Journal of Computer Science &amp; Technology</source>
          <volume>4</volume>
          (
          <year>2004</year>
          )
          <fpage>45</fpage>
          -
          <lpage>51</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>O.</given-names>
            <surname>Cocarascu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Toni</surname>
          </string-name>
          ,
          <article-title>Argumentation for machine learning: A survey.</article-title>
          ,
          <source>in: COMMA</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>219</fpage>
          -
          <lpage>230</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>M.</given-names>
            <surname>Thimm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kersting</surname>
          </string-name>
          ,
          <article-title>Towards argumentation-based classification, in: Logical Foundations of Uncertainty and Machine Learning</article-title>
          ,
          <source>IJCAI Workshop</source>
          , volume
          <volume>17</volume>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Modgil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Prakken</surname>
          </string-name>
          ,
          <article-title>The aspic+ framework for structured argumentation: a tutorial</article-title>
          ,
          <source>Argument &amp; Computation</source>
          <volume>5</volume>
          (
          <year>2014</year>
          )
          <fpage>31</fpage>
          -
          <lpage>62</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Riveret</surname>
          </string-name>
          , G. Governatori,
          <article-title>On learning attacks in probabilistic abstract argumentation</article-title>
          ,
          <source>in: Proceedings of the 2016 International Conference on Autonomous Agents &amp; Multiagent Systems</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>653</fpage>
          -
          <lpage>661</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>P.</given-names>
            <surname>Baroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Caminada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Giacomin</surname>
          </string-name>
          ,
          <article-title>An introduction to argumentation semantics, The knowledge engineering review 26 (</article-title>
          <year>2011</year>
          )
          <fpage>365</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Hocking</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Leslie</surname>
          </string-name>
          ,
          <article-title>Selection of the best subset in regression analysis</article-title>
          ,
          <source>Technometrics</source>
          <volume>9</volume>
          (
          <year>1967</year>
          )
          <fpage>531</fpage>
          -
          <lpage>540</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>N. V.</given-names>
            <surname>Chawla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. W.</given-names>
            <surname>Bowyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. O.</given-names>
            <surname>Hall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. P.</given-names>
            <surname>Kegelmeyer</surname>
          </string-name>
          ,
          <article-title>Smote: synthetic minority over-sampling technique</article-title>
          ,
          <source>Journal of artificial intelligence research 16</source>
          (
          <year>2002</year>
          )
          <fpage>321</fpage>
          -
          <lpage>357</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>H.-P.</given-names>
            <surname>Kriegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kröger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sander</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zimek</surname>
          </string-name>
          ,
          <article-title>Density-based clustering, Wiley interdisciplinary reviews: data mining and knowledge discovery 1 (</article-title>
          <year>2011</year>
          )
          <fpage>231</fpage>
          -
          <lpage>240</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>H.</given-names>
            <surname>Lakkaraju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Bach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <article-title>Interpretable decision sets: A joint framework for description and prediction, in: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining</article-title>
          ,
          <source>ACM</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>1675</fpage>
          -
          <lpage>1684</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>P. E.</given-names>
            <surname>Dunne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>McBurney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Parsons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wooldridge</surname>
          </string-name>
          ,
          <article-title>Weighted argument systems: Basic definitions, algorithms, and complexity results</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>175</volume>
          (
          <year>2011</year>
          )
          <fpage>457</fpage>
          -
          <lpage>486</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>P.</given-names>
            <surname>Besnard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hunter</surname>
          </string-name>
          ,
          <article-title>A logic-based theory of deductive arguments</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>128</volume>
          (
          <year>2001</year>
          )
          <fpage>203</fpage>
          -
          <lpage>235</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>