<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Workshop “From Objects to Agents” (WOA), September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Neuro-symbolic Computation for XAI: Towards a Unified Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giuseppe Pisano</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giovanni Ciatto</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Roberta Calegari</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrea Omicini</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Alma AI - Alma Mater Research Institute for Human-Centered Artificial Intelligence, Alma Mater Studiorum-Università di Bologna</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Dipartimento di Informatica - Scienza e Ingegneria (DISI), Alma Mater Studiorum-Università di Bologna</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <volume>1</volume>
      <fpage>4</fpage>
      <lpage>16</lpage>
      <abstract>
        <p>The idea of integrating symbolic and sub-symbolic approaches to make intelligent systems (IS) understandable and explainable is at the core of new fields such as neuro-symbolic computing (NSC). This work lays under the umbrella of NSC, and aims at a twofold objective. First, we present a set of guidelines aimed at building explainable IS, which leverage on logic induction and constraints to integrate symbolic and sub-symbolic approaches. Then, we reify the proposed guidelines into a case study to show their efectiveness and potential, presenting a prototype built on the top of some NSC technologies.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;XAI</kwd>
        <kwd>Hybrid Systems</kwd>
        <kwd>Neural Networks</kwd>
        <kwd>Logical Constraining</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        a natural choice for building human-intelligible IS [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. In fact, symbolic systems ofer a
humanunderstandable representation of their internal knowledge and processes: so, integrating them
into sub-symbolic models – to promote transparency of the resulting system – is the most
prominent stimulus for new research fields such as neuro-symbolic computing (NSC) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>Our work lays under the umbrella of NSC, and its contribution is twofold. On the one hand, we
present a set of guidelines aimed at building explainable IS, even when they exploit sub-symbolic
techniques. The guidelines leverage on logic induction and logic constraints as the two main
techniques integrating symbolic and sub-symbolic approaches. In particular, logic induction
makes it possible to extract the knowledge from black-box ML-based predictors – typically,
the sub-symbolic part of an IS – ofering a corresponding symbolic, logical representation.
Conversely, logic constraints are exploited to inject some logic knowledge into the black box,
thus restricting the numerical underlying model.</p>
      <p>On the other hand, we reify the proposed guidelines into a case study to show their
efectiveness and potential. In particular, we present a prototype built over some promising NSC
technologies. The resulting system is then assessed to verify its capability of being adjusted
(i.e., debugged and fixed) in case some unexpected behaviour in the sub-symbolic part of the
system is revealed. Accordingly, we show how the prototype is correctly performing w.r.t. the
proposed guidelines.</p>
      <p>The paper is structured as follows. Section 2 provides a brief overview of the field of symbolic
and sub-symbolic integration. It also includes the main related works on the use of a hybrid
system as a mean for explainability. In Section 3, we first present the guidelines for building
explainable systems; then, in Section 4 we discuss a possible instantiation of the proposed
guidelines. In Section 5, we proceed with the assessment of our prototype and the discussion of
results. Finally, Section 6 concludes the work.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>In the last years, deep and machine (ML) learning methods have become largely popular and
successful in real-world intelligent systems. However, their use raises the issue of understanding
and explaining their behaviour to humans. Neural networks in particular – which are the most
hyped and widely adopted approach within sub-symbolic AI – mostly sufer from the problem of
opaqueness: the way they obtain their results and acquire experience from data is unintelligible
to humans.</p>
      <p>
        One of the proposed approaches addressing the explainability problem [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] is the hybridisation
of symbolic and sub-symbolic techniques [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. An increasing number of authors recognises that
formal logic is capable of significantly improving humans’ understanding of data [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. In their
view, in principle, an opaque system, combined with a symbolic model, can provide a significant
result in terms of transparency. Many researches start from these assumptions.
      </p>
      <p>
        Among the others, the neuro-symbolic computing (NSC) field is a very recent and promising
research area whose ultimate goal is to make symbolic and sub-symbolic AI techniques
efortlessly work together. Due to the freshness of the topic, however, a well-established and coherent
theory of NSC is still missing. For this reason, a variety of methods have been proposed so far,
focusing on a multitude of aspects not always addressing interpretability and explainability
as their major concern. Nevertheless, some attempts to categorise XAI-related existing works
under the NSC umbrella exist [
        <xref ref-type="bibr" rid="ref3 ref4">4, 3</xref>
        ], which provide for a helpful overview of the topic.
      </p>
      <p>
        Despite NSC does not explicitly include XAI among its primary goals, in this paper we
borrow some ideas from the NSC field to show how the explainability of modern IS may benefit
from the integration of symbolic and sub-symbolic AI. In particular, our work focuses on
two main NSC sub-fields: namely, the logic as a constraint, for the constraining module, and
diferentiable programming for the induction module—since the proposed prototype exploits
logic induction via diferentiable programming [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In short, the former line of research aims
at constraining the training process of a sub-symbolic predictor in such a way that it cannot
violate the superimposed constraint at runtime. About the latter, diferentiable programming is
the combination of neural networks approaches with algorithmic modules in an end-to-end
diferentiable model, often exploiting optimisation algorithms like gradient descent [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Within the scope of logic as a constraint, most approaches exploit some sort of logic formulae
to constrain the behaviour of the sub-symbolic predictor—in most cases, a neural network.
This formula is then vectorised – i.e. translated into a continuous function over vectors of real
numbers – and exploited as a regularisation term in the loss function used for training the
subsymbolic predictor [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ]. Diferent strategies have been proposed to this end. For example,
in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] the symbolic constraints are used to modify the network structure incorporating them
into the training process. In the general case, however, logic constraining can be used to fix bias
or bugs in the behaviour of a sub-symbolic system, or, it can mitigate the situation in which
poor training data is available to correctly train a black-box system on a specific aspect.
      </p>
      <p>
        With respect to the second research area, some works laying under the umbrella of
diferentiable programming fruitfully intertwine ML and inductive logic programming (ILP) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] to
provide logic induction capabilities on top of sub-symbolic predictors. ILP is a well established
research area, laying at the intersection of ML and logic programming, which is strongly
interrelated with NSC. An ILP system is a tool able to induce (i.e., derive) – given an encoding
of some background knowledge and a set of positive and negative examples represented logic
facts –, a logic program that entails all the positive and none of the negative examples. While
traditionally these systems base their operation on the use of complex algorithms as their core
component [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] – deeply undermining their eficiency and usability – hybrid approaches exist
leveraging NSC to make the induction process more eficient [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Furthermore, as we show in
this paper, induced logic rules can be used as a means to inspect what a black-box predictor has
learned—as induction makes the predictor knowledge explicit in symbolic form.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Related Works</title>
        <p>
          As far as the intersection of logical systems and numerical models is concerned, the main
contributions come from [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] and [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. Their work can be summarised in:
• usage of a knowledge base filled automatically from training data to reason about what
has been learned and to provide explanations;
• adoption of logic rules to constrain the network and to correct its biases.
        </p>
        <p>Although these works ofer a good starting point in the search of a solution to the transparency
problem, some remarks should be pointed out. First, exploiting a knowledge base obtained
only from the training data is not suficient to acquire the knowledge required to explain the
entire network behaviour. That would lead to a system giving explanations according to the
network optimal functioning, not accounting for the training errors. Moreover, according to
these models, also the constraining part should be driven by the rules inferred from the training
data, hardly limiting the potential of those techniques. Indeed, the possibility for users to
impose their own rules would also give them the ability to mitigate the errors derived from an
incomplete or incorrect training set.</p>
        <p>The work presented here aims at building a model overcoming both these limitations. As
for the explanations’ coherence problem, the use of the black box as a data source in the logic
induction process should guarantee the correct correlation between the black box itself and the
derived logic theory. Furthermore, logic can be leveraged so as to combine the IS with the user
experience and knowledge, thus exploiting all the advantages of the constraining techniques.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. A NSC model for XAI</title>
      <p>In this paper, we present a general model for explainable data-driven IS, and a set of guidelines
supporting their construction. The novelty of our approach lays in the fruitful intertwining of
symbolic and sub-symbolic AI, which aims at providing both predictive accuracy – through the
exploitation of state-of-the-art machine learning (ML) techniques – and transparency—through
the exploitation of computational logic and logic programming (LP). The proposed model, in
particular, aims at overcoming the well-known limitations of ML-based AI w.r.t. interpretability.
Accordingly, it leverages on a number of contributions from the NSC and LP research field, as
well as two basic techniques—namely, induction and constraining.</p>
      <p>The main idea behind our work is that IS should feature both predictive precision and
interpretability features. To preserve predictive precision, IS should keep exploiting
highperformance, data-driven, black-box predictors such as (deep) neural networks. To overcome
the interpretability-related issues, IS should couple sub-symbolic approaches with logic theories
obtained by automatically extract the sub-symbolic knowledge of black boxes into symbolic
form. This would in turn enable a number of XAI-related features for IS, providing human
users with the capabilities of (i) inspecting a black box – also for debugging purposes –, and
(ii) correcting the system behaviour by providing novel symbolic specifications.</p>
      <p>Accordingly, in the reminder of this section, we provide further details about the desiderata
which led to the definition of our model. We then provide an abstract description of our model
and a set of guidelines for software engineers and developers. Finally, we provide a technological
architecture to assess both the model and the guidelines.</p>
      <sec id="sec-3-1">
        <title>3.1. Desiderata</title>
        <p>Regardless of the architectural and technological choices performed by designers and developers,
the IS adhering to our model are characterised by several common desiderata (Di) w.r.t. their
overall functioning and behaviour. Generally speaking, these are aimed at making IS both
prediction-efective and explainable. Indeed, following this purpose, IS should
D1 attain high predictive performances by leveraging ML and data-driven AI</p>
        <sec id="sec-3-1-1">
          <title>D2 provide human-intelligible outcomes / suggestions / recommendations</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>D3 acquire knowledge from both data and from high-level specifications</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>D4 make their knowledge base inspectable by human experts</title>
        </sec>
        <sec id="sec-3-1-4">
          <title>D5 let human experts override / modify their knowledge base</title>
          <p>A key enabling point in satisfying these desiderata is knowledge representation. While ML
and data-driven AI are certainly required to mine efective information from data eficiently,
they soon fall short when it comes to satisfying desiderata D2–D5. This happens because
they mostly leverage on a distributed, sub-symbolic representation of knowledge which is
hard to interpret for human beings. Therefore, to support D2 and D4, we need an alternative
human-intelligible representation of the sub-symbolic model and a procedure to perform such
a representation transformation. Furthermore, to support D3 and D5, we also need to link
symbolic and sub-symbolic representations in a bidirectional way—meaning that an inverse
procedure aimed at converting symbolic information back into sub-symbolic form is needed as
well.</p>
          <p>Accordingly, the focus of our model is both on the extraction of symbolic representation from
the black-box predictor and, vice-versa, on the injection of symbolic representation (constraints)
in the corresponding black-box predictor. The purpose of this dichotomy is twofold:
guaranteeing the comprehensibility of the black-box model for humans – as symbolic representations are
to some extent inherently intelligible –, and enabling debugging and correction of the black-box
behaviour, in case some issue is found through inspection.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Modelling</title>
        <p>Generally speaking, we model an explainable, hybrid IS as composed by a black box, a knowledge
base (KB) and an automatic logic reasoner, as depicted in Figure 1. Explainable recommendations
or suggestions are provided to the end-user via the logic reasoner – based on symbolic rules
and facts included in the KB by domain experts – and via black-box predictions—based on data.
Accordingly, the reasoner combines several sorts of inference mechanisms (e.g. deduction and
induction). Furthermore, to balance the knowledge coming from data with the domain experts
knowledge, the model exploits induction and constraining techniques to improve the black box
with the logical knowledge and vice-versa.</p>
        <p>The black-box module is the core of the IS, making it capable of mining efective information
from data. Any sub-symbolic model providing high predictive performances – e.g. neural
networks, SVM, generalised linear models, etc. – can be used for the module implementation. This,
however, may bring opaqueness-related issues. Thus, the black-box module is complemented
with the other two modules to provide explanation and debugging facilities.</p>
        <p>The reasoner module is aimed at providing explainable outcomes to the end-users. In
particular, explanations are given in terms of logic KB, capable of approximating the black box.
The construction of the logic KB relies on the induction capabilities ofered by this module.
More precisely, the outcomes generated by the black box are exploited to build a logic theory,
mimicking the work of the black-box predictor with the highest possible fidelity. An inductive
process is then fed with the resulting extended theory. This leads to a theory containing a
number of general relations, which can be exploited to provide intelligible information to the
end-users. The described workflow supports explainability in two ways: (i) it provides a global
explanation of how the black box works in terms of general relations that must always hold;
(ii) it provides local explanations, justifying each black-box conclusion by enabling the use of
deductive reasoning on the connected logical knowledge. In other words, this mechanism is
what enables IS to be interpretable and debuggable by users. Finally, the KB module aims at
storing the logical knowledge approximating the black-box behaviour. The knowledge can be
modified by domain experts. When this is the case, it becomes of paramount importance to
keep the black-box module coherent with the human-proposed edits. To this end, constraining
is performed to align the black-box behaviour with the KB. This mechanism is what enables IS
to be fixed by users.</p>
        <p>In the reminder of this section, we delve deeper into the details of these mechanisms.
3.2.1. Logic induction
While deductive reasoning moves from universal principles that are certainly true to specific
conclusions that can be mechanically derived, inductive reasoning moves from specific instances
to general conclusions. Induction is a key mechanism of our model. Assuming IS can transform
the data they leverage upon for training black boxes into theories of logic facts, inductive
reasoning can be used to extract the rules that best explain that data.</p>
        <p>
          The induction procedure is not meant to replace the black box as a learning or data-mining
tool—as it would be quite dificult to obtain the same performance in terms of accuracy and
ability to scale over huge amounts of data. Conversely, it aims at “opening the black box”,
letting humans understand how it is performing and why. In other words, following the abstract
framework presented in [
          <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
          ], logic induction is a means for explaining the functioning
of a black-box predictor via symbolic rules. More precisely, induction is fundamental for the
debuggability of our model. For example, it enables the discovery of fallacies in the learning
process of a black box even for people not used to the specific domain.
        </p>
        <p>In our model, the induction process is fed with both the raw data and the outcomes of the
black box. In the former case, induction leads to the unveiling of latent relations possibly buried
in the original data: we refer to the logic theory obtained as the reference theory. In the latter
case, induction leads to a symbolic approximation of the knowledge the black box has acquired
from data via ML: we refer to the logic theory obtained as explanation theory. Assuming the
soundness of the induction process, discrepancies between these two theories could reveal
some possible rules that have not been correctly learned by the sub-symbolic model. It is then
possible to fix it by enforcing the correct relations via logic constraining.</p>
        <p>Finally, one last point is worth to be discussed. Unless the induction process possesses the
same learning capabilities of the black box, it is impossible to detect all its learning errors. In
this case, in fact, the reference theory would be the optimal solution itself, and the sub-symbolic
model would be useless. As the induction process aims at opening the box, its use on the raw
data aims at providing insights on the accuracy of the training phase. Thus, it does not aim at
providing an optimal solution.
3.2.2. Logic constraining
While induction is the mechanism aimed at translating sub-symbolic knowledge into symbolic
form, constraining is the inverse process aimed at injecting some symbolic knowledge into a
black box. In this way, both the induced rules and those coming from domain experts are used
to constrain the black box and its outcomes. This is another key mechanism of our model.</p>
        <p>In particular, the ability to encode some prior knowledge within a model is interesting for
two main reasons. On the one side, it enables a reduction in the data needed to train the black
box. In fact, handmade rules may be exploited to model a portion of the domain not included in
the training data. So, rather than creating a more exhaustive training set, an expert may directly
encode his/her knowledge into rule and train a constrained black box. We call this procedure
domain augmentation. On the other side, one may exploit the constraining process to guide
and support a black-box learning, e.g., helping it to avoid biases. We call this procedure bias
correction.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Guidelines</title>
        <p>3.3.1. Type of logic
In the following, we discuss the main aspects to be considered when designing a system
conforming to our model. As a first step in this direction, we first handle some important
aspects concerning data representation.</p>
        <p>The first aspect concerns the type of logic used by the IS. Logic, in fact, plays a fundamental role:
it heavily afects the explainability properties of the final system. For this reason, the choice of
the most appropriate logic formalism is a primary concern.</p>
        <p>For our model to be efective, the selected logic formalism should provide high flexibility
in the description of the domain. At the same time, to keep a system as simple as possible, it
should be coherent in every part of it: from the constraining module to the induction one, every
part should share the same representation of the logical knowledge.
3.3.2. Sort of data
The second aspect concerns the sort of data used to feed the IS. For the logical induction process
to be carried out on data, the first thing to do is transforming the data itself. In other words,
the available data should be translated into a logical theory according to the chosen formalism.
For the translation to be efective, the resulting theory should preserve as much as possible the
information contained in the original data. Anyway, when dealing with some particular type of
data some problems have to be taken into account.</p>
        <p>When unstructured data – e.g., images, audio, time-series, etc – come into play, the
transformation process may become considerably harder. In that case, a two-step transformation must
be performed, involving:
1. extraction of semantically meaningful, structured features from unstructured data;
2. translation of extracted features into the desired logical formalism.</p>
        <p>Of course, this procedure adds a whole range of new problems related to the accuracy of the
features extracted. In fact, the explanation process is mainly linked to the reliability of the
data used as the basis for the induction process. The exploitation of data generated through an
automatic process makes a discrepancy between the data extrapolated by the inductive process
and the behaviour of the black box more likely.
3.3.3. Sorts of black box
In the general case, we require the architectures adhering our model to be as agnostic as possible
w.r.t. the particular sort of black box to be adopted. In fact, the choice of the most adequate
sort of black box is strongly afected by (i) the nature of the data at hand, (ii) the availability of
some symbolic extraction technique making the black box less opaque, (iii) the possibility of
constraining the black box with logic formulae.</p>
        <p>Strictly speaking, the choice of the black box should be delayed to the implementation phase.
Here we describe a number of criteria to be taken into account when performing this choice. We
recall, however, that the other guidelines described so far are agnostic w.r.t. the sub-symbolic
model to be used.</p>
        <p>
          As far as the nature of the data is concerned, traditional ML techniques [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] – e.g., decision
trees, generalised linear models, etc. – are usually exploited on structured datasets, whereas
deep learning techniques are better suited to deal with unstructured data. However, this is
not due to an inadequacy of neural networks for structured data, but rather to the greater
simplicity of the learning algorithms used by traditional ML techniques. Structured data delivers
a relatively-smaller complexity to deal with, making simpler ML algorithms a more suitable
choice for their comprehensibility and usability.
        </p>
        <p>
          As far as opaqueness is concerned, virtually any ML techniques is afected by that to some
extent: yet, neural networks remain the most critical from this point of view. Nevertheless, the
vast majority of rule induction procedures are either black-box agnostic – a.k.a. pedagogical [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]
– or neural-network-specific [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. In terms of the support for constraining, it should be possible
to guide the learning activity trough some regulariser terms attained from the constraints to be
enforced.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Technological architecture</title>
      <p>This subsection provides a technological description of an IS prototype adhering to our model.
It is based on the concrete architecture detailed in Figure 2, which specialises the abstract model
from Figure 1 through a number of technological commitments.</p>
      <p>First, we present our choices in relation to the points examined in Subsection 3.3. As for the
type of logic, within the scope of this paper, we adopt first-order logic (FOL) as our formalism
of choice, and Prolog as our reference concrete language. FOL is likely to ofer the best trade-of
between flexibility and expressiveness of the representation language. Furthermore, the choice
of FOL enables the exploitation of many existing approaches supporting both conditioning and
induction. Finally, the choice of the Prolog syntax enables the direct exploitation of the induced
theories within existing automatic reasoners—e.g. tuProlog, SWI-Prolog, etc.</p>
      <p>As far as the sort of the data is concerned, in this paper we only describe a prototype based
on structured data. In fact, the feature extraction module can be omitted without hindering
the generality of our approach. Moreover, the greater simplicity in the data helps to avoid the
problems related to the possible information loss due to their transformation. In the future,
however, we plan to extend our prototype to support arbitrary datasets including any sort of
data. This requires the creation of a module for feature extraction.</p>
      <p>In terms of the sort of the black-box used, we focus on a prototype based on neural networks,
being them the most critical from the opacity perspective. The remaining technological choices
derive consequently.</p>
      <p>
        The prototype exploits two main technologies: DL2 [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] for the conditioning part, and
NTP [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ][
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] for the induction part. DL2 is one of those models leveraging on symbolic rules
vectorisation as means to constrain the target neural network. We choose DL2 as the most
mature and user-friendly technology supporting neural-network constraining through logic
rules.
      </p>
      <p>
        The choice of NTP as the induction engine is more straightforward. Indeed, NTP is among
the few ready-to-use technologies ofering both deductive and inductive logic capabilities
in a diferentiable way. On the one hand, diferentiability is what makes induction more
computationally eficient w.r.t. similar technologies—and this is why we choose it. On the other
hand, NTP deductive capabilities are not mature enough. In fact, training a model correctly
approximating a logic theory in a reasonable amount of time is very challenging—especially
when the complexity of the theory grows. While this could be acceptable for the induction
process, it is still very limiting for deductive reasoning. Moreover, traditional logic engines
combine a very large ecosystem of supporting tools – IDEs, debuggers, libraries – that have
potential to hugely improve the efectiveness of the reasoning phase. For this reason, we adopt
tuProlog (2P) [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] – a Java-based logic engine built for use in ubiquitous contexts – as the main
tool for the manipulation of logic theories, as well for automated reasoning. This choice is
motivated by its reliability, modularity, and flexibility.
      </p>
      <p>
        While the entire system is built around the aforementioned technologies, Python is used as the
glue language to keep modules together. In particular, the sub-symbolic module is implemented
via the PyTorch learning framework [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], thus ensuring an easy integration with DL2—which
is natively built to work with this framework. The modules responsible for the extraction of the
Prolog theory from the input dataset (i.e. the Translator block in Figure 2) are created using
Python as well. The NTP integration takes place through the Prolog logic language. In fact,
NTP easily allows the use of Prolog theories as input for the induction process. The result of
the induction phase is also delivered in a logical form, allowing for an easy consultation and
analysis through the 2P technology.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Assessment</title>
      <p>In this section we assess our model from the explainability perspective leveraging on the
prototype implementation described in Section 4. Particularly, we train the sub-symbolic module
through a real-world ML problem, and we test whether and to what extent our architecture
actually provides (i) the capability of building a logical representation of sub-symbolic knowledge
acquired from data via induction, (ii) the capability of altering or fixing the system behaviour
via conditioning, and, ultimately, (iii) the inspectability and debuggability of the system as a
whole. In the following subsections we set up an ad-hoc experiment aimed at performing these
(a) Network default training
(b) Decision tree built on test data
tests. A detailed presentation of the experiments and their results follows.</p>
      <sec id="sec-5-1">
        <title>5.1. Experiment Design</title>
        <p>The proposed experiment aims at comprehensively testing the operation of the prototype
against some real classification problem. In particular, we consider a binary classification task
on structured data. Roughly speaking, the experiment works by artificially constructing a buggy
neural network for the classification task, and then showing how our architecture makes it
possible to reveal the bug. More precisely, the experiment is structured as follows:
1. a neural network is trained until reaching maximum validation-set accuracy for the
classification task;
2. by combining the training data and the predictions of the network, a coherent Prolog
theory is extracted;
3. the induction module is used to extract the latent relations from the theory, until at least
one relation which properly divides the data is found;
4. through constraining, we inject an error in the network, in such a way that it misclassify
some instances;
5. by repeating Item 2 and Item 3, we show that the approximated logic theory reveals the
injected error.</p>
        <p>This workflow lets us verify the behaviour of our prototype and of all its components. In
particular, Item 3 and Item 5 aim at demonstrating that a logic theory that optimally approximates
a neural network is actually attainable. In terms of the ability of debugging and correcting a
network, the whole procedure aims at demonstrating their feasibility. The extraction of the
correct theory in the initial part of the experiment is comparable to the one extracted from
a malfunctioning classifier. The inclusion of the fake constraint, as well as its recognition in
the theory extracted at the conclusion of the experiment, shows the feasibility of the network
correction process.</p>
        <p>Two fundamental points are worth to be taken into account for the experiment to be
meaningful. The first point is the ability to accurately evaluate the accuracy of the logic theory
recovered. In fact, in order to verify that the theory extracted from the neural network is
actually the correct one, it is either necessary (i) to have an optimal knowledge of the domain,
or (ii) to use easily analysable and verifiable datasets. As for the first case – being an expert of
the analysed domain – it turns out to be very easy, by verifying the correctness of the logical
relations. The only alternative comes from the usage of an easily verifiable dataset as the base
for the experiment. Indeed, it should be possible also for a domain novice to understand the
ratio behind the data. For instance, a simple way to verify the correctness of the rules is to use
an alternative ML tool – one that can guarantee a high level of transparency – to analyse the
data. In the case of a classifier, the best choice could be a decision tree (DT). In fact, DTs training
produces a symbolic model that can be eficiently verified by the user. Hence, DT output can be
used as a reference in the evaluation of the induction results.</p>
        <p>The second point is the exploitation of constraining as a means to inject bugs into a network
in a controlled way. In fact, in order to verify the prototype capability of revealing and correcting
bugs in the black-box module, it is first necessary to construct a buggy network. However, in
case of a poorly training, as in that case, it can not be clear where the bug is. Hence, to evaluate
the actual capabilities of the prototype, the bugs to be spotted must be a priori known in order
to have a reference.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Experiment Setup</title>
        <p>
          We base our experiment on the Congressional Voting Records data set1 from the UCI Repository
[
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. It consists of table registering the voting intentions – namely, favourable, contrary or
unknown – of 435 members of the Congress of the USA on 15 main topics, as well as their
political orientation—namely, either Democrat or Republican. The goal of the classification task
is to sort a member of the Congress as either democratic or republican depending on his/her
intention on those 15 topics.
        </p>
        <p>Given the relatively small amount of instances in the dataset, along with the small number of
attributes, an intuitive understanding of the classification problem can be assumed. However,
to further simplify the analysis of the results, we use a DT trained over the data as a reference.</p>
        <p>A neural network capable of distinguishing between Democrats and Republicans based on
voting intentions is trained and assessed as described above. The experiments code is available
at the Github repository2. We now proceed discussing the results we obtained in each step of
the experiment.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Results</title>
        <p>The training of the aforementioned neural network is performed via ordinary stochastic gradient
descent until reaching a 95% accuracy score on the validation set (containing a randomly selected
20% of whole data). The training process is depicted in Figure 3a.</p>
        <p>Through the induction process on the data generated by the network it was possible to recover
the following relationships:
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
  (1, . . . ,  ),
(1, . . . ,  ).
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
(1, . . . ,  ).
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
 (1, . . . ,  ).
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
 (1, . . . ,  ),
(1, . . . ,  ),
 (1, . . . ,  ).</p>
        <p>To verify their validity we can examine the DT generated using the original data as source
(Figure 3b)—the C4.5 algorithm has been used. As expected, the inductive process managed
to recover the relationship that most discriminates between Democrats and Republicans: the
tendency of the Democrats to be against the freezing of the cost of medical expenses.</p>
        <p>The verification of the correction process is based on the reversal of the relationship retrieved
above. Formally, the conditioning of the network occurred on this rule:
(1, . . . ,  )
:</p>
        <p>ℎ  (1, . . . ,  ).</p>
        <p>In natural language, this Prolog rule expresses that, given a set of votes (1, . . . ,  ), if these
share the opposition to the freezing of medical expenses, then the voter is likely from the
Republican Party. This constraint translates into a very simple result: all the votes should
belong to Republicans. Indeed, if initially only the Democrats were against the freezing, by
forcing the network to consider them as Republicans, we reach a situation where the whole set
of voters is Republican.</p>
        <p>More in detail, constraining rules are codified through the DSL ofered by the D2L
implementation. For example, the above rule can be expressed in the form:
dl2.Implication(
dl2.EQ(x[FeesFreeze], 0),
dl2.LT(y[Dem], y[Rep]))
This rule is then automatically converted in a numerical function. Its evaluation contributes to
the final loss function adopted by the target neural network.</p>
        <p>For the experiment, a completely-new network is trained considering the new constraint. The
results in Table 1 show the efect of the constraint on the network. The Democrats/Republican
imbalance – which is made evident by Table 1 – reflects in the result of the induction process
on the data generated by the constrained network:
(1, . . . ,  )
:  (1, . . . ,  ),
 ℎ  (1, . . . ,  ).
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
(1, . . . ,  ),
  (1, . . . ,  ).
(1, . . . ,  )
: ℎ  (1, . . . ,  ),
(1, . . . ,  ),
 (1, . . . ,  ),
  (1, . . . ,  ).</p>
        <p>The inducted knowledge base only contains rules concerning the Republican wing, thus
conifrming the footprint of the relation imposed during the conditioning process.</p>
        <p>Summarising, the results of the assessment are positive. It was first possible to obtain the
relationships implied in the operation of the classifier in a logic form. By these rules has been
possible (i) to identify the more discriminant features in the dataset; (ii) to enable the correction
of the black box using the user knowledge. The constraining part has also demonstrated efective.
Indeed, the imposition of the user-crafted rule has led to a coherent change in the black-box
behaviour—enabling its correction. The results demonstrate how the presented guidelines lead
to an IS giving explanations about its functioning, thus allowing the user intervention on its
behaviour.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>The solutions proposed so far in the literature to the opaqueness issue – one of the main problem
of today AI technologies – are disparate. In this work, we showed how symbolic logic can be a
crucial element in this panorama.</p>
      <p>On the trails of the NSC models, we presented a series of guidelines aimed at correctly
integrating a ML-based predictor – i.e., a black box – with a logic-based subsystem. In particular,
our guidelines support the creation of IS exposing clear insights about their own functioning,
thus enabling end users to intervene on the IS behaviour via a logical interface. We then tested
our guidelines against a prototype IS, in order to study if and to what extent our approach
is feasible and useful. Notably, the prototype assessment confirms our approach is feasible
exploiting technologies already available in the research scene. Nevertheless, the prototype
has been tested only on a single scenario. In order to confirm the eficacy of our approach, we
need to perform a more exhaustive range of experiments. Moreover, we plan to extend the
prototype assessment with more complex use-cases. For example, as anticipated in Section 4,
we intend to enhance the prototype with the support for unstructured data. This extension
would considerably improve the applicability of the studied approach, allowing its assessment
also on the more complex area of image classification.</p>
      <p>Through the above experimental investigation – in the case of more positive results – we aim
at introducing a rigorously formalised version of the proposed model—presented in this paper
in a more intuitive and preliminary shape. Consequently, we should also better investigate and
verify the preliminary guidelines provided. We aim at obtaining an accurate and comprehensive
guide that would allow developers eficiently integrating opaque AI systems and logic.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Bughin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Hazan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ramaswamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Allas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dahlstrom</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Henke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Trench</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence: the next digital frontier?, Discussion paper</article-title>
          ,
          <source>McKinsey Global Institute</source>
          ,
          <year>2017</year>
          . URL: https: //www.mckinsey.com/~/media/mckinsey/industries/advancedelectronics/ ourinsights/howartificialintelligencecandeliverrealvaluetocompanies/ mgi-artificial
          <article-title>-intelligence-discussion-paper</article-title>
          .ashx.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A. B.</given-names>
            <surname>Arrieta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Ser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bennetot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tabik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barbado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>García</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gil-Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Molina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Benjamins</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Chatila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Herrera</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence (XAI): concepts, taxonomies, opportunities and challenges toward responsible AI</article-title>
          ,
          <source>Information Fusion</source>
          <volume>58</volume>
          (
          <year>2020</year>
          )
          <fpage>82</fpage>
          -
          <lpage>115</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.inffus.
          <year>2019</year>
          .
          <volume>12</volume>
          .012.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Calegari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>On the integration of symbolic and sub-symbolic techniques for XAI: A survey</article-title>
          ,
          <source>Intelligenza Artificiale</source>
          <volume>14</volume>
          (
          <year>2020</year>
          )
          <fpage>7</fpage>
          -
          <lpage>32</lpage>
          . doi:
          <volume>10</volume>
          .3233/IA-190036.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>d'Avila Garcez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gori</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. C.</given-names>
            <surname>Lamb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Serafini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Spranger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N.</given-names>
            <surname>Tran</surname>
          </string-name>
          ,
          <article-title>Neuralsymbolic computing: An efective methodology for principled integration of machine learning and reasoning</article-title>
          ,
          <source>FLAP</source>
          <volume>6</volume>
          (
          <year>2019</year>
          )
          <fpage>611</fpage>
          -
          <lpage>632</lpage>
          . URL: https://arxiv.org/abs/
          <year>1905</year>
          .06088.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Muggleton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Schmid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zeller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tamaddoni-Nezhad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Besold</surname>
          </string-name>
          ,
          <article-title>Ultra-strong machine learning: comprehensibility of programs learned with ILP, Machine Learning 107 (</article-title>
          <year>2018</year>
          )
          <fpage>1119</fpage>
          -
          <lpage>1140</lpage>
          . doi:
          <volume>10</volume>
          .1007/s10994-018-5707-3.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rocktäschel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riedel</surname>
          </string-name>
          ,
          <article-title>End-to-end diferentiable proving</article-title>
          , in: I. Guyon, U. von Luxburg, S. Bengio,
          <string-name>
            <given-names>H. M.</given-names>
            <surname>Wallach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. V. N.</given-names>
            <surname>Vishwanathan</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems</source>
          <year>2017</year>
          ,
          <year>2017</year>
          , pp.
          <fpage>3788</fpage>
          -
          <lpage>3800</lpage>
          . URL: http://papers.nips.cc/paper/ 6969-end
          <article-title>-to-end-diferentiable-proving.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Hernández</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Amigó</surname>
          </string-name>
          ,
          <article-title>Diferentiable programming and its applications to dynamical systems</article-title>
          , CoRR abs/
          <year>1912</year>
          .0 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1912</year>
          .08168.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Serafini</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. S.</surname>
          </string-name>
          <article-title>d'Avila Garcez, Logic tensor networks: Deep learning and logical reasoning from data and knowledge</article-title>
          , in: T. R.
          <string-name>
            <surname>Besold</surname>
            ,
            <given-names>L. C.</given-names>
          </string-name>
          <string-name>
            <surname>Lamb</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Serafini</surname>
          </string-name>
          , W. Tabor (Eds.),
          <source>11th International Workshop on Neural-Symbolic Learning and Reasoning (NeSy'16) colocated with the Joint Multi-Conference on Human-Level Artificial Intelligence (HLAI</source>
          <year>2016</year>
          ), volume
          <volume>1768</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>34</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>1768</volume>
          /NESY16_paper3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fischer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Balunovic</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
            Drachsler-Cohen,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Gehr</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , M. T. Vechev,
          <article-title>DL2: training and querying neural networks with logic</article-title>
          , in: K. Chaudhuri, R. Salakhutdinov (Eds.),
          <source>36th International Conference on Machine Learning</source>
          ,
          <source>(ICML</source>
          <year>2019</year>
          ), volume
          <volume>97</volume>
          <source>of Proceedings of Machine Learning Research</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1931</fpage>
          -
          <lpage>1941</lpage>
          . URL: http://proceedings. mlr.press/v97/fischer19a.html.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , T. Friedman,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>V. den Broeck, A semantic loss function for deep learning with symbolic knowledge</article-title>
          , in: J. G. Dy,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Krause (Eds.),
          <source>35th International Conference on Machine Learning (ICML 2018)</source>
          , volume
          <volume>80</volume>
          <source>of Proceedings of Machine Learning Research</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>5498</fpage>
          -
          <lpage>5507</lpage>
          . URL: http://proceedings.mlr.press/v80/xu18h.html.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Srikumar</surname>
          </string-name>
          ,
          <article-title>Augmenting neural networks with first-order logic, in: 57th Annual Meeting of the Association for Computational Linguistics (ACL</article-title>
          <year>2019</year>
          ),
          <article-title>Association for Computational Linguistics (ACL</article-title>
          ),
          <year>2020</year>
          , pp.
          <fpage>292</fpage>
          -
          <lpage>302</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/p19-
          <fpage>1028</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Muggleton</surname>
          </string-name>
          ,
          <article-title>Inductive logic programming</article-title>
          ,
          <source>New Generation Computing</source>
          <volume>8</volume>
          (
          <year>1991</year>
          )
          <fpage>295</fpage>
          -
          <lpage>318</lpage>
          . doi:
          <volume>10</volume>
          .1007/BF03037089.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cropper</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dumancic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. H.</given-names>
            <surname>Muggleton</surname>
          </string-name>
          , Turning 30:
          <article-title>New ideas in inductive logic programming</article-title>
          , in: C.
          <string-name>
            <surname>Bessiere</surname>
          </string-name>
          (Ed.),
          <source>29th International Joint Conference on Artificial Intelligence (IJCAI</source>
          <year>2020</year>
          ), IJCAI,
          <year>2020</year>
          , pp.
          <fpage>4833</fpage>
          -
          <lpage>4839</lpage>
          . doi:
          <volume>10</volume>
          .24963/ijcai.
          <year>2020</year>
          /673.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>D.</given-names>
            <surname>Doran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Schulz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. R.</given-names>
            <surname>Besold</surname>
          </string-name>
          ,
          <article-title>What does explainable AI really mean? A new conceptualization of perspectives</article-title>
          , in: T. R.
          <string-name>
            <surname>Besold</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          Kutz (Eds.), 1st International Workshop on Comprehensibility and
          <article-title>Explanation in AI and ML 2017 co-located with 16th International Conference of the Italian Association for Artificial Intelligence (AI*IA</article-title>
          <year>2017</year>
          ), volume
          <volume>2071</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>15</fpage>
          -
          <lpage>22</lpage>
          . URL: http://ceur-ws.
          <source>org/</source>
          Vol-2071/CExAIIA_2017_paper_2.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bennetot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-L.</given-names>
            <surname>Laurent</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Chatila</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Díaz-Rodríguez</surname>
          </string-name>
          ,
          <article-title>Highlighting bias with explainable neural-symbolic visual reasoning</article-title>
          , CoRR abs/
          <year>1909</year>
          .0 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1909</year>
          . 09065.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Schumacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvaresi</surname>
          </string-name>
          ,
          <article-title>Agent-based explanations in AI: Towards an abstract framework</article-title>
          , in: D.
          <string-name>
            <surname>Calvaresi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Najjar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Winikof</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          Främling (Eds.), Explainable, Transparent Autonomous Agents and
          <string-name>
            <surname>Multi-Agent</surname>
            <given-names>Systems</given-names>
          </string-name>
          , volume
          <volume>12175</volume>
          of Lecture Notes in Computer Science, Springer, Cham,
          <year>2020</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>20</lpage>
          . doi:
          <volume>10</volume>
          .1007/ 978-3-
          <fpage>030</fpage>
          -51924-
          <issue>7</issue>
          _
          <fpage>1</fpage>
          , Second International Workshop, EXTRAAMAS 2020, Auckland, New Zealand, May 9-
          <issue>13</issue>
          ,
          <year>2020</year>
          , Revised Selected Papers.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>G.</given-names>
            <surname>Ciatto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Calvaresi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. I.</given-names>
            <surname>Schumacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <article-title>An abstract framework for agentbased explanations in AI</article-title>
          ,
          <source>in: 19th International Conference on Autonomous Agents and MultiAgent Systems (AAMAS</source>
          <year>2019</year>
          ),
          <source>International Foundation for Autonomous Agents and Multiagent Systems</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1816</fpage>
          -
          <lpage>1818</lpage>
          . Extended Abstract.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Quinlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Motoda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. J.</given-names>
            <surname>McLachlan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. F. M.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Steinbach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. J.</given-names>
            <surname>Hand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Steinberg</surname>
          </string-name>
          ,
          <article-title>Top 10 algorithms in data mining</article-title>
          ,
          <source>Knowledge and Information Systems</source>
          <volume>14</volume>
          (
          <year>2008</year>
          )
          <fpage>1</fpage>
          -
          <lpage>37</lpage>
          . doi:
          <volume>10</volume>
          .1007/ s10115-007-0114-2.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>R.</given-names>
            <surname>Andrews</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Diederich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. B.</given-names>
            <surname>Tickle</surname>
          </string-name>
          ,
          <article-title>Survey and critique of techniques for extracting rules from trained artificial neural networks</article-title>
          ,
          <source>Knowledge-Based Systems 8</source>
          (
          <year>1995</year>
          )
          <fpage>373</fpage>
          -
          <lpage>389</lpage>
          . doi:
          <volume>10</volume>
          .1016/
          <fpage>0950</fpage>
          -
          <lpage>7051</lpage>
          (
          <issue>96</issue>
          )
          <fpage>81920</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>M. de Jong</surname>
          </string-name>
          , F. Sha,
          <article-title>Neural theorem provers do not learn rules without exploration</article-title>
          , CoRR abs/
          <year>1906</year>
          .0 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1906</year>
          .06805.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>E.</given-names>
            <surname>Denti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Omicini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ricci</surname>
          </string-name>
          ,
          <article-title>tuProlog: A light-weight Prolog for internet applications and infrastructures</article-title>
          ,
          <source>in: Lecture Notes in Computer Science</source>
          , volume
          <volume>1990</volume>
          , Springer Verlag,
          <year>2001</year>
          , pp.
          <fpage>184</fpage>
          -
          <lpage>198</lpage>
          . doi:
          <volume>10</volume>
          .1007/3-540-45241-9_
          <fpage>13</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>A.</given-names>
            <surname>Paszke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gross</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Massa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lerer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bradbury</surname>
          </string-name>
          , G. Chanan,
          <string-name>
            <given-names>T.</given-names>
            <surname>Killeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gimelshein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antiga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Desmaison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Köpf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>DeVito</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Raison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Tejani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chilamkurthy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Steiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chintala</surname>
          </string-name>
          ,
          <string-name>
            <surname>Pytorch:</surname>
          </string-name>
          <article-title>An imperative style, high-performance deep learning library</article-title>
          , in: H.
          <string-name>
            <surname>M. Wallach</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Beygelzimer</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <article-title>d'Alché-</article-title>
          <string-name>
            <surname>Buc</surname>
            ,
            <given-names>E. B.</given-names>
          </string-name>
          <string-name>
            <surname>Fox</surname>
          </string-name>
          , R. Garnett (Eds.),
          <source>Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems</source>
          <year>2019</year>
          , NeurIPS
          <year>2019</year>
          ,
          <year>2019</year>
          , pp.
          <fpage>8024</fpage>
          -
          <lpage>8035</lpage>
          . URL: http://papers.nips.cc/paper/ 9015-pytorch
          <article-title>-an-imperative-style-high-performance-deep-learning-library.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dua</surname>
          </string-name>
          ,
          <string-name>
            <surname>C. Graf,</surname>
          </string-name>
          <article-title>UCI machine learning repository</article-title>
          ,
          <year>2017</year>
          . URL: http://archive.ics.uci.edu/ml.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>