=Paper= {{Paper |id=Vol-2742/paper4 |storemode=property |title=SeXAI: Introducing Concepts into Black Boxes for Explainable Artificial Intelligence |pdfUrl=https://ceur-ws.org/Vol-2742/paper4.pdf |volume=Vol-2742 |authors=Ivan Donadello,Mauro Dragoni |dblpUrl=https://dblp.org/rec/conf/aiia/DonadelloD20 }} ==SeXAI: Introducing Concepts into Black Boxes for Explainable Artificial Intelligence== https://ceur-ws.org/Vol-2742/paper4.pdf
          SeXAI: Introducing Concepts into Black Boxes for
                 Explainable Artificial Intelligence

                              Ivan Donadello1 and Mauro Dragoni1

                  Fondazione Bruno Kessler, Via Sommarive 18, I-38123, Trento, Italy
                               {donadello|dragoni}@fbk.eu

    The interest in Explainable Artificial Intelligence (XAI) research is dramatically
grown during the last few years. The main reason is the need of having systems that
beyond being effective are also able to describe how a certain output has been obtained
and to present such a description in a comprehensive manner with respect to the tar-
get users. A promising research direction making black boxes more transparent is the
exploitation of semantic information. Such information can be exploited from different
perspectives in order to provide a more comprehensive and interpretable representation
of AI models. In this paper, we present the first version of SeXAI, a semantic-based
explainable framework aiming to exploit semantic information for making black boxes
more transparent. After a theoretical discussion, we show how this research direction
is suitable and worthy of investigation by showing its application to a real-world use
case.


1         Introduction
1
  Explainable Artificial Intelligence (XAI) aims at explaining the algorithmic decisions
of AI solutions with non-technical terms in order to make these decisions trusted and
easily comprehensible by humans [1]. If these AI solutions are based on learning al-
gorithms and perceived as black boxes due to their complexity, XAI makes them more
transparent and interpretable too. This is of great interest for both logical reasoning in
rule engines and Machine Learning (ML) methods. The explanation of a reasoning pro-
cess can be very difficult, especially when a system is based on a set of complex logical
axioms whose logical inferences are performed with, for example, tableau algorithms
[4]. Indeed, inconsistencies in logical axioms may be not well understood by users if
the system limits to just report the violated axioms. Indeed, users are generally skilled
to understand neither formal languages nor the behavior of a whole system. This is cru-
cial for some applications, such as a power plant system where a warning message to
the user must be clear and concise to avoid catastrophic consequences. On the other
hand, ML methods are based on statistical models of the data where some explanatory
variables (i.e., the features) of the data are leveraged in order to predict a dependent
variable (i.e., a class or a numeric value). Many statistical methods (e.g., the principal
component analysis) are able to detect what are the main involved features in a ML
task. These involved features can be used to explain to user the reason of a particular
decision. These features are usually handcrafted by human experts and consequently
    1
        Copyright ©2020 for this paper by its authors. Use permitted under Creative Commons Li-
        cense Attribution 4.0 International (CC BY 4.0).
present a shared semantics. Modern Deep Neural Network (DNN) are able to learn
these features with no need of human effort. However, the semantics of these learnt
features is nor explicit or shareable with humans. Therefore, a human-comprehensible
explanation about how and why an AI system took a decision is necessary.


    A shared and agreed definition on explainability has not been reached in the AI
community so far. Here we follow the definition of Adadi and Berrada [1] that argue for
a distinction between interpretability and explainability. The former regards the study
of a mathematical mapping between the inputs and the outputs of a black-box system.
The latter regards a human comprehension of the logic and/or semantics of such a sys-
tem. Doran et al. [13] refine the notion of explainability stating that an explainable (or
comprehensible) system should provide a reason or justification about its output in-
stead of focusing solely on the mathematical mapping. Moreover, they argue that truly
explainable systems have to adopt reasoning engines that run on knowledge bases con-
taining an explicit semantics in order to generate a human comprehensible explanation.
In addition, the explainability power depends also on the background knowledge of the
users.


    To this extent, the logical reasoning associated to semantics is fundamental as it
represents a bridge between the output machine and human concepts. This differs from
other XAI works that try to analyze the activations of the hidden neurons (i.e., the learnt
features) with respect to a given output without attaching a shared semantics. However,
logical reasoning on the back-box output is not sufficient as it performs a post-hoc
explanation of the black-box guided only by the axioms of a knowledge base. Indeed,
no explicit link from the black-box learned features and the concepts in the knowledge
base is used. The contribution of the paper addresses this issue.


    We propose a novel semantic-based XAI (SeXAI) framework that generates expla-
nations for a black-box output. Differently from Doran et al., such explanations are
First-Order logic formulas whose predicates are semantic features connected to the
classes of the black box output. Logic formulas are then easy to translate in natural lan-
guage for a better human comprehension. Moreover, the semantic features are aligned
with the neurons of the black box thus creating a neural-symbolic model. This allows
reasoning between the output and the features and the improvement of both the knowl-
edge base and the black box output. In addition, the semantics in the knowledge base
is aligned with the annotation in the dataset. This is fundamental both for the neural-
symbolic alignment and for the black box performance. The latter were tested with
experiments on image classification showing that a semantic aligned with the training
set outperforms a model whose semantics is deduced from the output with only logical
reasoning. The rest of the paper follows with Section 2 that provides a state-of-the-art
of techniques for generating explanations from logical formulas. Section 3 describes
the main concepts of the SeXAI framework whereas Section 4 shows a first application
and results of the framework in an image classification task. Section 5 concludes the
paper.
2   Related Work
The research on XAI has been widely explored in the last years [17], but most of the
contributions focused only on the analysis of how learning models (a.k.a. black boxes)
work. This is a limited view of the topic since there is a school of thought arguing that
an effective explainability of learning models cannot be achieved without the use of
domain knowledge since data analysis alone is not enough for achieving a full-fledged
explainable system [8]. This statement has been further discussed recently by assert-
ing that the key for designing a completely explainable AI system is the integration of
Semantic Web technologies [19,20,29]. Semantic Web technologies enabling the the
design of strategies for providing explanations in natural language [2,26] where ex-
planations are provided through textual rule-like notation. NLG strategies have been
designed also for generating natural language text from triples [34] and for translat-
ing SPARQL queries into a natural language form understandable by non-experts [15].
Here, we focused on the integration of semantic information as enabler for improving
the comprehensiveness of XAI systems. Our aim is to generate natural language ex-
planations as result of the synergies between neural models and logic inferences for
supporting end-users in understanding the output provided by the systems.
     The explanation of the logical reasoning in an ontology is implemented with two
two orthogonal approaches: justifications and proofs. The former computes the minimal
subset of the ontology axioms that logically entails an axiom. The latter computes also
all the inference steps [25].
     One of the first user studies dealing with explanations for entailments of OWL
ontologies was performed by [24]. The study investigated the effectiveness of differ-
ent types of explanation for explaining unsatisfiable classes in OWL ontologies. The
authors found that the subjects receiving full debugging support performed best (i.e.,
fastest) on the task, and that users approved of the debugging facilities. Similarly, [28]
performed a user study to evaluate an explanation tool, but did not carry out any de-
tailed analysis of the difficulty users had with understanding these explanations. While,
[5] presents a user study evaluating a model-exploration based approach to explanation
in OWL ontologies. The study revealed that the majority of participants could solve
specific tasks with the help of the developed model-exploration tool, however, there
was no detailed analysis of which aspects of the ontology the subjects struggled with
and how they used the tool. The work [23] presents several algorithms for computing all
the justifications of an entailment in a OWL-DL knowledge base. However, nor study
or user evaluation is performed to assess the capability of the computed justifications of
the logical entailments. The work in [18] focuses on the explanation, through justifica-
tions, of the disclosure of personal data to users (patients and staff) of hospitals. This is
performed by translating SWRL rules inconsistencies into natural language utterances.
Moreover, the SWRL rules translation is performed axiom by axiom, thus generating a
quite long sentence. This could require too much time for reading and understanding.
Whereas, our method returns only a single utterance summarizing the whole justifica-
tion.
     Formal proofs are the other form of explanation for logical reasoning. In [31] the
authors present an approach to provide proof-based explanations for entailments of the
CLASSIC system. The system omits intermediate steps and provides further filtering
strategies in order to generate short and simple explanations. The work proposed in [7]
first introduced a proof-based explanation system for knowledge bases in the Descrip-
tion Logic ALC [4]. The system generates sequent calculus style proofs using an exten-
sion of a tableaux reasoning algorithm, which are then enriched to create natural lan-
guage explanations. However, there exists no user studies to explore the effectiveness of
these proofs. In [27] the authors proposed several (tree, graphical, logical and hybrid)
visualizations of defeasible logic proofs and present a user study in order to evaluate
the impact of the different approaches. These representations are hard to understand for
non-expert users. Indeed, the study is based on participants from a postgraduate course
(who have attended a Semantic Web course) and from the research staff. In general,
proof algorithms for Description Logic are based on Tableau techniques [4] whereas
proof algorithms for other logics are studied in the field of Automated Reasoning [32].
     This wide range of approaches to explanation of logical entailments is more focused
on the development of efficient algorithms than on effective algorithms for common
users. Indeed, all the computed explanations are sets of logical axioms understandable
only by expert users. The aim of our work is to provide and effective representation
to explanation for all users. This representation is based on the verbalization of the
explanation in natural language. This verbalization can be performed by using methods
that translate axioms of an OWL ontology in Attempto Controlled English [22,21] or in
standard English [3] with the use of templates. This last work also presents some users’
studies on the quality of the generated sentences. However, these works do not handle
with the reasoning results (justifications or proofs), indeed, no strategy for selecting and
rendering an explanation is studied.


3   The Framework

In the fields of Machine Learning and Pattern Recognition, a feature is a characteris-
tic or a measurable property of an object/phenomenon under observation [6]. Features
can be numeric or structured and they are crucial in tasks such as pattern detection,
classification or regression as they serve as explanatory variables. Indeed, informative
and discriminating features are combined in a simple or complex manner by the main
ML algorithms. This also holds in our everyday experience, a dish composed by pasta,
bacon, eggs, pepper and aged cheese (features) is recognized as pasta with Carbonara
sauce (the class). Diseases are recognized according to the symptoms (features), the
price of the houses is computed according to the features of, e.g., location, square
meters and years of the real estate. However, with the rise of Deep Neural Networks
(DNN), features are learnt by the system from the raw data without the necessity of
handcrafting from domain experts. This has improved the performance of such systems
with the drawback of loosing comprehensibility from users. Indeed, DNNs embed the
data in a vector space in the most discriminating way without any link to a formal se-
mantics. The aim of SeXAI is to link a DNN with a formal semantics in order to provide
a comprehensible explanation of the DNN output to everyday users.
    Following the definitions of Doran et al.[13], we ground the notion of explainable
system into the concept of a comprehensible system, that is a system that computes its
output along with symbols that allow users to understand what are the main semantic
features in the data that triggered that particular output. Here, we refine the work of
Doran et al. by introducing the concept of semantic feature. These are features that can
be expressed through predicates of a First-Order Logic (FOL) language and represent
the common and shared attributes of an object/phenomenon that allow its recognition.
Examples can be ContainsBacon(x) or ContainsEggs(x) indicating the ingredients
of a dish in a picture. Semantic features in principle can be further explained by more
fine-grained semantic features. For example, the ChoppedBacon(x) feature can be
explained by the HasCubicShape(x) and HasP inkColor(x) features. However, in a
nutritional domain, these latter features do not add further comprehension to users and
can represent an overload of information. Therefore, the knowledge engineering and/or
domain expert have to select the right granularity of the semantic features to present
to users and therefire ensuring a sort of atomic property of these features. Semantic
features are different from the learnt numeric (and not comprehensible) features of a
DNN. The aim of a comprehensible system is to find an alignment between the learnt
and the semantic features.
    The connection between a DNN output and its semantic features is formalized
through the definition of comprehension axiom.
Definition 1 (Comprehension axiom). Given a FOL language with P = {O}n1 ∪
{A}m1 the set of its predicate symbols, a comprehension axiom is a formula of the form

                                k
                                ^                l
                                                 ^
                                      Oi (x) ↔         Ai (x)
                                i=1              i=1

with {O}n1 the set of output symbols of a DNN and {A}m
                                                     1 the corresponding semantic
features (or attributes).
A comprehension axiom formalizes the main tasks of a DNN:

Multiclass Classification: the predicate Oi (x) represents a class (e.g., pasta with Car-
   bonara sauce or sushi) for x and k = 1 as a softmax is applied in the last layer of
   the DNN. The semantic features represent, for example, ingredients contained in
   the recognized dish.
Multilabel Classification: Oi (x) is part of a list of predicates being computed by the
   DNN (e.g., dinner and party) for x and k > 1 as a sigmoid is applied in the last
   layer of the DNN. The semantic features represent, for example, objects in the
   scene, such as, pizza, table, bottles, person and balloons.
Regression: Oi (x) can be part of a list of predicates being computed by the DNN
   (e.g., the asked price and the real values of house) for x. Here k ≥ 1 with a sigmoid
   applied in the last layer of the DNN. The semantic features are properties of interest
   for buying a house.

    We present the SeXAI framework for comprehensible systems in Figure 3. The
knowledge base KB contains both the predicate symbols in P for annotating the data
and the comprehension axioms. These latter are passed to the symbolic system that is in
charge of i) analyzing the output of the DNN and the associated semantic features; ii)
reasoning about them according to the comprehension axioms; iii) returning a, possibly
Fig. 1. In the SeXAI framework data are annotated with symbols of a knowledge base. A sym-
bolic system is aligned with a DNN in order to provide an output and a set of semantic features
consistent with the comprehension axioms in the knowledge base.


refined, output along with the related semantic features. This architecture extends the
ones in [13], where a reasoner computes the explanation of the output, with a semantic
module that enables several tasks that improve the comprehension and the transparency
(i.e., the interpretation) of the DNN:
Output and semantic features refinement: The DNN is trained to return both the
    output and the semantic features in a multitasking learning setting. Then, with the
    use of fuzzy reasoning or neural-symbolic systems [11,12,9,30], both outputs can
    be refined according to the comprehension axioms and to the evidence coming from
    the scores of the DNN.
Feature Alignment: Once a DNN is trained in a multitasking learning setting, it is
    possible to analyze which are the most activated neurons of the last hidden layer
    [16] for each semantic feature. In this manner, we can align the high-level features
    of the DNN with the semantic features in KB.
Knowledge base improvement: Once the features alignment is performed, the system
    can turn off the neurons corresponding to a given semantic feature and check the
    performance degradation with respect to the output. No degradation of the perfor-
    mance means that the particular semantic feature has just a correlation with the
    output and, therefore, it can be removed from the corresponding comprehension
    axiom or stated as a simple correlation. On the other hand, a degradation of the per-
    formance indicates a causality of the semantic features with respect to the output.
  The more the performance degrades the higher the causality degree for that feature
  is. Therefore, we can enrich KB with some priors in the comprehension axioms
  about the importance of the semantic features.
Model improvement: Analyzing the semantic features returned by wrong output pre-
  dictions allows the system to detect the presence of some common semantic fea-
  tures that alter some predictions. Therefore, the model can assign a lower weight to
  the neurons aligned with that semantic features.

The symbolic system in SeXAI extends the framework of Doran et al. [13] by comput-
ing the alignment of semantic and DNN features that enables the improvement of both
KB and of the model. Differently, in [13] the reasoner module is able to only generate
the output and the semantic features.


4      SeXAI in Action

Section 3 provided the general description of the SeXAI framework that we proposed
for increasing the overall comprehensiveness of AI models. In this Section, we show
how the SeXAI framework can be instantiated within a real-world scenario. In partic-
ular, we applied the SeXAI framework to image classification with the aim of demon-
strating how the integration of semantics into an AI-based classification systems trig-
gers both the generation of explanations and, at the same time, an improvement of the
overall effectiveness of the classification model.
    As described in Section 3, the SeXAI framework is composed by different modules
that, depending on the scenario in which the framework is deployed, can be instantiated
or not. Let us consider a scenario where the goal is to classify food images with respect
to the food categories contained by the represented recipe instead of the recipe itself.
Information about food categories are particularly useful in scenario where physicians
are supported by information systems concerning the diet monitoring of people affected
by nutritional diseases (e.g., diabetes, hypertension, obesity, etc.). By starting from the
SeXAI architecture shown in Figure 3, we instantiated the modules as follows.

    – The “Data” module contains our dataset of recipe images we used for training the
      classification model. A more detailed description of the dataset is provided in Sec-
      tion 4.1.
    – The “Knowledge Base” contains, beyond a taxonomy of recipes and food cate-
      gories, the composition of each recipe in terms of its food categories. Recipes
      compositions are described by object properties within the knowledge base. More
      specifically, in our scenario we adopted the HeLiS ontology [14] where we have
      the food category-based composition of more than 8,000 recipes2 .
    – As “Black-box model”, we implemented a DNN trained with recipe/food images
      annotated with the list of related food categories. Given a recipe image x, the recipe
      label represents the O(x) output neuron, while the food categories represent the
 2
     In the remaining of the paper, we will refer to some concepts defined within the HeLiS on-
     tology. We leave to the reader the task of checking the meaning of each concept within the
     reference paper.
    semantic features A(x) output neurons. In our scenario we decided to not include
    the O(x) output neurons and to classify each image by its semantic features A(x).
    Hence, each neuron of the DNN output layer indicates if one of the food categories
    contained in the dataset has been detected within the images or not.
 – Finally, in our scenario the “Symbolic System” links together the “Knowledge
   Base” and the output of the DNN for generating natural language explanations of
   the classification results.


    The evaluation of explanations quality is still an open topic within the AI research
area [19]. Moreover, in our scenario, explanations aim to provide a comprehensive de-
scription of the output rather than being a vehicle for improving the model. Hence, the
evaluation of their language content is not of interest. Instead, the SeXAI framework
evaluation provided in this work focuses on the effectiveness of exploiting semantic
features for both training and classification purposes. As baseline, we used a post-hoc
semantic-based strategy where images used for training the DNN were annotated only
with the corresponding recipe label. Here, the list of food categories has been extracted
after the classification of each images by exploiting the predicted recipe label. Figure 4
shows the building blocks of the baseline. For readability, hereafter we will refer to the
instantiation of the SeXAI framework as “multi-label classifier”, while the baseline will
be labeled as “single-label classifier”.




Fig. 2. The architecture of the baseline system we used for comparing the effectiveness of the
SeXAI framework concerning the recipe images classification task.
4.1     Quantitative Evaluation

In the considered scenario, a good performance on recognizing food categories is im-
portant as the misclassification of images could trigger wrong behaviors of the systems
in which the classifier is integrated. For example, if the framework would be integrated
into a recommendation system, a misclassification of a recipe image would lead to the
generation of wrong messages or even no message to the target user.

The Food and Food Categories (FFoCat) Dataset 3 We leverage the food and food
category concepts in HeLiS for the multi-label classification. However, current food
image datasets are not built with these concepts as labels, so it was necessary to build
a new dataset (named FFoCat) with these concepts. We start by sampling some of the
most common recipes in Recipe and use them as food labels. The food categories are
then automatically retrieved from BasicFood with a SPARQL query. Examples of food
labels are Pasta with Carbonara Sauce and Baked Sea Bream. Their associated food
categories are Pasta, AgedCheese, VegetalOils, Eggs, ColdCuts and FreshFish, Vege-
talOils, respectively. We collect 156 labels for foods (Recipe concept) and 51 for food
categories (BasicFood concept). We scrape the Web, using Google Images as search
engine, to automatically download all the images related to the food labels. Then, we
manually clean the dataset by checking if the images are compliant with the related
labels. This results in 58,962 images with 47,108 images for the training set and 11,854
images for the test set (80-20 ratio of splitting). Then, by leveraging HeLiS properties,
we enrich the image annotations with the corresponding food category labels to perform
multi-label classification. The dataset is affected by some natural imbalance, indeed the
food categories present a long-tail distribution: only few food categories labels have
the majority of the examples. On the contrary, many food categories labels have few
examples. This makes the food classification challenging.

Experimental Settings and Metrics For both multi and single-label classification we
separately train the Inception-V3 network [33] from scratch on the FFoCat training set
to find the best set of weights. The fine tuning using pre-trained ImageNet [10] weights
did not perform sufficiently. This is probably due to the fact that the learnt low-level
features of the first layers of the network belong to a general domain and do not match
properly with the specific food domain. For the multi-label classification, we use a sig-
moid as activation function of the last fully-connected layer of the Inception-V3 and
binary cross entropy as loss function. This is a standard setting for multi-label clas-
sification. Regarding the single-label classification, the activation function of the last
fully-connected layer is a softmax and the loss function is a categorical cross entropy.
We run 100 epochs of training with a batch size of 16 and a learning rate of 10−6 . At
each epoch images are resized to 299x299 pixels and are augmented by using rotations,
width and height shifts, shearing, zooming and horizontal flipping. This results in a
training set 100 times bigger than the initial one. We used early stopping to prevent
overfitting. The training has been performed with the Keras framework (TensorFlow as
backend) on a PC equipped with a NVIDIA GeForce GTX 1080.
 3
     The dataset, its comparison and the code are available at https://bit.ly/2Y7zSWZ.
    As performance metric we use the mean averagePn precision (MAP) that summarizes
the classifier precision-recall curve: M AP = i=1 (Rn − Rn−1 )Pn , i.e., the weighted
mean of precision Pn achieved at each threshold level n. The weight is the increase of
the recall in the previous threshold: Rn − Rn−1 . The macro AP is the average of the AP
over the classes, the micro instead considers each entry of the predictions as a label. We
preferred MAP instead of accuracy as the latter for sparse vectors can give misleading
results: high results for output vectors with all zeros.



Results Given an (set of) input image(s) x, the computing of the precision-recall curve
requires the predicted vector(s) y of food category labels and a score associated to each
label in y. In the multi-label method this score is directly returned by the Inception-
V3 network (the final logits). In the single-label and inference method this score needs
to be computed. We test two strategies: (i) we perform exact inference of the food
categories from HeLiS and assign the value 1 to the scores of each yi ∈ y; (ii) the
food categories labels inherit the uncertainty returned the DNN: the score of each yi
is the logit value si returned by DN N (x). Results are in Table 1. The direct multi-



                             Method                     Micro-AP (%) Macro-AP (%)
           Multi-label (SeXAI framework)                    76.24          50.12
           Single-class without uncertainty (baseline)      50.53          31.79
           Single-class with uncertainty (baseline)         60.21          42.51
Table 1. The multi-label classification of food categories outperforms in average precision (AP)
the methods based on single-label classification and logical inference.




label has very good performance (both in micro and macro AP) in comparison with the
single-label models. The micro-AP is always better than the macro-AP as it is sensible
to the mentioned imbalance of the data. This means that errors in the single recipe
classification propagate to the majority of the food categories the recipe contains. That
is, the inferred food categories will be wrong because the recipe classification is wrong.
On the other hand, errors in the direct multi-label classification will affect only few food
categories. We inspected in more detail some of the errors committed by the classifiers
in order to have a better understanding of their behaviors. In some cases, the single-label
method misclassified an image with Backed Potatoes as Backed Pumpkin thus missing
the category of FreshStarchyVegetables. Another image contains a Vegetable Pie but
the single-label method infers the wrong category of PizzaBread. In another image, this
method mistakes Pasta with Garlic, Oil and Chili Peppers with Pasta with Carbonara
Sauce, thus inferring wrong Eggs and ColdCuts. Here the multi-label method classifies
all the categories correctly. Therefore, the multi-label method allows a more fine grained
classification of the food categories w.r.t. the single-label method. The latter has better
results if the score returned by the DNN is propagated to the food categories labels w.r.t.
the exact inference.
4.2   Discussion

The experience of designing the SeXAI framework and the analysis of results obtained
from a preliminary validation within a real-world use case highlighted two important
directions towards the long-term goal of achieving a fully-explainable AI system.
    First, the integration of semantic features with black-box models enabled the gen-
eration of comprehensive explanations. SeXAI can be considered a neuro-symbolic
framework conjugating the effectiveness of black-box models (e.g., DNN) with the
transparency of semantic knowledge that, where possible, can support the generation
of explanations describing the behavior of AI systems. This aspect opens to a very
interesting and innovative research direction centered on the content of the generated
explanations. Indeed, the integration of semantic features for generating explanations
can be exploited for refining the statistical model itself (as described in Section 3). For
instance by analyzing correlations between the presence of specific semantic features
within explanations and the performance of the black-box model. Future work will fo-
cus on strengthening this liaison within the SeXAI framework in order to validate if an
inference process could improve the classification capability and, at the same time, to
observe how inference results could be exploited for refining the black-box model.
    Second, the integration of semantic features can lead to better classification per-
formance. Results presented in Table 1 show that through the integration of semantic
features, it is possible to improve the overall effectiveness of the black-box model. This
is a very interesting finding since it confirms the importance of a by-design integration
of semantic features. Future activities will further investigate this hypothesis within
other scenarios with the aim of understanding which are the boundaries and if there
exist some constraints in the application of this strategy. For instance, the granularity
of semantic features with respect to the entities that have to be classified could play
an important role. Hence, a trade-off has to be found in order to maintain the explain-
able capability of the system and, at the same time, an acceptable effectiveness of the
classification model.


5     Conclusions

The aim of Explainable Artificial Intelligence is to provide black-box algorithms with
strategies to produce a reason or justification for their outputs. This is fundamental to
make these algorithms trusted and easily comprehensible by humans. A formal seman-
tics, provided by knowledge bases, encoded in a logical language allows the connection
between the numeric features of a black box and the human concepts. Indeed, a justifica-
tion in a logical language format can be easily translated in natural language sentences
in an automatic way.
    In this paper, we presented the first version of SeXAI, a semantic-based explainable
framework aiming at exploiting semantic information for making black boxes more
comprehensible. SeXAI is a neural-symbolic system that analyses the output of a black
box and creates a connection between the learnt features and the semantic concepts of
a knowledge base in order to generate an explanation in a logical language. This allows
reasoning on the black box and its explanation, the improvement of the knowledge base
and of the black box output. The semantics in the knowledge base is aligned with the
annotations in the dataset. This improves the performance of SeXAI on a task of multi-
label image classification with respect to a system that performs solely logical reasoning
on the black box output.
    As future work, we will perform some experiments on the quality of the alignment
between the learnt and the semantic features. In particular, we will evaluate the degree
of causality of the semantic features with respect to the output and how the attention of
a black box can be moved towards the semantic features in order to improve the model
performance.


References
 1. Adadi, A., Berrada, M.: Peeking inside the black-box: A survey on explainable artificial
    intelligence (XAI). IEEE Access 6, 52138–52160 (2018)
 2. Ai, Q., Azizi, V., Chen, X., Zhang, Y.: Learning heterogeneous knowledge base embeddings
    for explainable recommendation. Algorithms 11(9), 137 (2018)
 3. Androutsopoulos, I., Lampouras, G., Galanis, D.: Generating natural language descriptions
    from OWL ontologies: the naturalowl system. J. Artif. Intell. Res. 48, 671–715 (2013)
 4. Baader, F., Calvanese, D., McGuinness, D.L., Nardi, D., Patel-Schneider, P.F. (eds.): The
    Description Logic Handbook: Theory, Implementation, and Applications. Cambridge Uni-
    versity Press (2003)
 5. Bauer, J., Sattler, U., Parsia, B.: Explaining by example: Model exploration for ontology
    comprehension. In: Description Logics. CEUR Workshop Proceedings, vol. 477. CEUR-
    WS.org (2009)
 6. Bishop, C.M.: Pattern recognition and machine learning, 5th Edition. Information science
    and statistics, Springer (2007)
 7. Borgida, A., Franconi, E., Horrocks, I.: Explaining ALC subsumption. In: Horn, W. (ed.)
    ECAI 2000, Proceedings of the 14th European Conference on Artificial Intelligence, Berlin,
    Germany, August 20-25, 2000. pp. 209–213. IOS Press (2000)
 8. Cherkassky, V., Dhar, S.: Interpretation of Black-Box Predictive Models, pp. 267–286.
    Springer International Publishing, Cham (2015)
 9. Daniele, A., Serafini, L.: Neural networks enhancement through prior logical knowledge.
    CoRR abs/2009.06087 (2020)
10. Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: A Large-Scale Hier-
    archical Image Database. In: CVPR09 (2009)
11. Diligenti, M., Gori, M., Saccà, C.: Semantic-based regularization for learning and inference.
    Artif. Intell. 244, 143–165 (2017)
12. Donadello, I., Serafini, L.: Compensating supervision incompleteness with prior knowledge
    in semantic image interpretation. In: IJCNN. pp. 1–8. IEEE (2019)
13. Doran, D., Schulz, S., Besold, T.R.: What does explainable AI really mean? A new concep-
    tualization of perspectives. In: Besold, T.R., Kutz, O. (eds.) Proceedings of the First Interna-
    tional Workshop on Comprehensibility and Explanation in AI and ML 2017 co-located with
    16th International Conference of the Italian Association for Artificial Intelligence (AI*IA
    2017), Bari, Italy, November 16th and 17th, 2017. CEUR Workshop Proceedings, vol. 2071.
    CEUR-WS.org (2017)
14. Dragoni, M., Bailoni, T., Maimone, R., Eccher, C.: Helis: An ontology for supporting healthy
    lifestyles. In: International Semantic Web Conference (2). Lecture Notes in Computer Sci-
    ence, vol. 11137, pp. 53–69. Springer (2018)
15. Ell, B., Harth, A., Simperl, E.: SPARQL query verbalization for explaining semantic search
    engine queries. In: Presutti, V., d’Amato, C., Gandon, F., d’Aquin, M., Staab, S., Tordai, A.
    (eds.) The Semantic Web: Trends and Challenges - 11th International Conference, ESWC
    2014, Anissaras, Crete, Greece, May 25-29, 2014. Proceedings. Lecture Notes in Computer
    Science, vol. 8465, pp. 426–441. Springer (2014)
16. Erhan, D., Bengio, Y., Courville, A., Vincent, P.: Visualizing higher-layer features of a deep
    network. University of Montreal 1341(3), 1 (2009)
17. Gilpin, L.H., Bau, D., Yuan, B.Z., Bajwa, A., Specter, M., Kagal, L.: Explaining expla-
    nations: An overview of interpretability of machine learning. In: Bonchi, F., Provost, F.J.,
    Eliassi-Rad, T., Wang, W., Cattuto, C., Ghani, R. (eds.) 5th IEEE International Conference
    on Data Science and Advanced Analytics, DSAA 2018, Turin, Italy, October 1-3, 2018. pp.
    80–89. IEEE (2018)
18. Hamed, R.G., Pandit, H.J., O’Sullivan, D., Conlan, O.: Explaining disclosure decisions over
    personal data. In: ISWC Satellites. CEUR Workshop Proceedings, vol. 2456, pp. 41–44.
    CEUR-WS.org (2019)
19. Holzinger, A., Biemann, C., Pattichis, C.S., Kell, D.B.: What do we need to build explainable
    AI systems for the medical domain? CoRR abs/1712.09923 (2017)
20. Holzinger, A., Kieseberg, P., Weippl, E.R., Tjoa, A.M.: Current advances, trends and chal-
    lenges of machine learning and knowledge extraction: From machine learning to explainable
    AI. In: Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E.R. (eds.) Machine Learning and
    Knowledge Extraction - Second IFIP TC 5, TC 8/WG 8.4, 8.9, TC 12/WG 12.9 International
    Cross-Domain Conference, CD-MAKE 2018, Hamburg, Germany, August 27-30, 2018, Pro-
    ceedings. Lecture Notes in Computer Science, vol. 11015, pp. 1–8. Springer (2018)
21. Kaljurand, K.: ACE view — an ontology and rule editor based on attempto controlled en-
    glish. In: OWLED. CEUR Workshop Proceedings, vol. 432. CEUR-WS.org (2008)
22. Kaljurand, K., Fuchs, N.E.: Verbalizing OWL in attempto controlled english. In: OWLED.
    CEUR Workshop Proceedings, vol. 258. CEUR-WS.org (2007)
23. Kalyanpur, A., Parsia, B., Horridge, M., Sirin, E.: Finding all justifications of OWL DL
    entailments. In: ISWC/ASWC. Lecture Notes in Computer Science, vol. 4825, pp. 267–280.
    Springer (2007)
24. Kalyanpur, A., Parsia, B., Sirin, E., Hendler, J.A.: Debugging unsatisfiable classes in OWL
    ontologies. J. Web Semant. 3(4), 268–293 (2005)
25. Kazakov, Y., Klinov, P., Stupnikov, A.: Towards reusable explanation services in protege. In:
    Description Logics. CEUR Workshop Proceedings, vol. 1879. CEUR-WS.org (2017)
26. Khan, O.Z., Poupart, P., Black, J.P.: Explaining recommendations generated by mdps. In:
    Roth-Berghofer, T., Schulz, S., Leake, D.B., Bahls, D. (eds.) Explanation-aware Comput-
    ing, Papers from the 2008 ECAI Workshop, Patras, Greece, July 21-22, 2008. University of
    Patras. pp. 13–24 (2008)
27. Kontopoulos, E., Bassiliades, N., Antoniou, G.: Visualizing semantic web proofs of defeasi-
    ble logic in the DR-DEVICE system. Knowl.-Based Syst. 24(3), 406–419 (2011)
28. Lam, J.S.C.: Methods for resolving inconsistencies in ontologies. Ph.D. thesis, University of
    Aberdeen, UK (2007)
29. Lécué, F.: On the role of knowledge graphs in explainable AI. Semantic Web 11(1), 41–51
    (2020)
30. Mao, J., Gan, C., Kohli, P., Tenenbaum, J.B., Wu, J.: The neuro-symbolic concept learner: In-
    terpreting scenes, words, and sentences from natural supervision. In: ICLR. OpenReview.net
    (2019)
31. McGuinness, D.L., Borgida, A.: Explaining subsumption in description logics. In: IJCAI (1).
    pp. 816–821. Morgan Kaufmann (1995)
32. Robinson, J.A., Voronkov, A. (eds.): Handbook of Automated Reasoning (in 2 volumes).
    Elsevier and MIT Press (2001)
33. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception archi-
    tecture for computer vision. In: CVPR. pp. 2818–2826. IEEE Computer Society (2016)
34. Vougiouklis, P., ElSahar, H., Kaffee, L., Gravier, C., Laforest, F., Hare, J.S., Simperl, E.:
    Neural wikipedian: Generating textual summaries from knowledge base triples. J. Web Se-
    mant. 52-53, 1–15 (2018)