<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>{ikram.chraibi-kaadoud, lina.fahed, philippe.lenca}@imt-atlantique.fr</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ikram Chraibi Kaadoud</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lina Fahed</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Philippe Lenca</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IMT Atlantique, Lab-STICC, UMR CNRS 6285</institution>
          ,
          <addr-line>F-29238 Brest</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <fpage>28</fpage>
      <lpage>40</lpage>
      <abstract>
        <p>EXplainable Artificial Intelligence (XAI) has recently become a very active domain, mainly due to the extensive development of black-box models such as neural networks. Recent XAI objectives have been defined in the state-of-the-art, for which specific approaches have been proposed. Implicit links can be found between XAI and other domains, especially related to knowledge and neural networks. We here aim to highlight these implicit links. We present a narrative review of research works in two domains: (i) Knowledge domain with focus on Knowledge Discovery and Representation, and (ii) Representation Learning. We discuss the similarity and joining points between these domains and XAI. We conclude that, in order to make black-boxes more transparent, XAI approaches should be more inspired and take advantage of past and recent works in Knowledge and Representation Learning domains. Through this paper, we offer an entry point to the domain of XAI for both multidisciplinary researchers and specialists in AI, as well for AI knowledgeable users.</p>
      </abstract>
      <kwd-group>
        <kwd>XAI</kwd>
        <kwd>Knowledge Discovery</kwd>
        <kwd>Knowledge representation</kwd>
        <kwd>Representation learning</kwd>
        <kwd>State representation learning</kwd>
        <kwd>Manifold representation learning</kwd>
        <kwd>Multi-view representation learning</kwd>
        <kwd>Network representation learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>During the last few years, eXplainable Artificial
Intelligence (XAI), has become a very active domain1 facing the
high development of black-box models, such as neural
networks [Guidotti et al., 2018]. A new generation of XAI
approaches have been proposed, for which several new concepts
and terms are specific to application domains, data types or
modeling. Application domains of XAI are multiple:
machine learning, robotics, multi-agent systems, computer
vision, Knowledge Representation and Reasoning, etc.</p>
      <p>[Barredo Arrieta et al., 2020] defined “Given an
audience, an explainable Artificial Intelligence is the one that
produces details or reasons to make its functioning clear
or easy to understand”. Indeed, XAI aims to make
Artificial Intelligence (AI) models more intelligible and
accessible or to directly design explainable models and
results [Buchanan and Shortliffe, 1984; Guidotti et al., 2018;
Barredo Arrieta et al., 2020]. When the first case arises, XAI
provides an explanation of the internal mechanisms and/or
the reasons behind the AI model behavior i.e. its
functioning and performance: an explanation is thus an interface
between the AI model to explain and the target audience
[Gunning, 2017]. We define an explanation as an information in
a semantically complete format, which is self-sufficient and
chosen according to the target audience regarding its
knowledge, its expectations and the context. Hence, the purpose
of an explanation is to clarify the cause, context and
consequences of described facts through a set of statements or
information [Van Fraassen, 1988].</p>
      <p>It is important to underline that an explanation by its very
nature is contextual: it is specific to a given target audience
and also to a given context [Walton, 2004]. This makes XAI
more challenging as automatic context understanding is still
a very challenging task [Bre´zillon, 1999; Lim et al., 2009;
Augusto et al., 2017; Hollister et al., 2017] and no unified
way for modelling context in intelligent environments has yet
been proposed in the literature [Brenon et al., 2018]. We
emphasize that the context (i.e users context, goal context, etc.)
is important to take into account in XAI. However, this point
is not the focus of the paper.</p>
      <p>In the state-of-the-art, an explanation can take different
formats (e.g. visual, natural language, features relevance
explanations, etc.) and combine several representations of the
same information [Barredo Arrieta et al., 2020]. Two main
XAI techniques are proposed: (i) Ante-hoc techniques which
consist in optimizing an already transparent AI model (e.g.
linear regression, decision trees, etc.) by adding constraints
or features in order to increase transparency through
metrics, data visualisation, etc. (ii) Post-hoc techniques that
aim to explain already built black-box AI models (mainly
deep neural networks). Among famous Post-hoc techniques:
LIME [Ribeiro et al., 2016], SHAP [Lundberg and Lee,
2017], visual explanations, saliency mapping, etc. XAI has
recently been covered by several reviews that reveal its
complex and intrinsically multidisciplinary aspects from a
technical, user or Human-Interaction viewpoint [Guidotti et al.,
2018; Gilpin et al., 2018; Barredo Arrieta et al., 2020; Vilone
and Longo, 2020]. As examples, we can note
technicalbased reviews, as those related to reinforcement learning
[Puiutta and Veith, 2020; Heuillet et al., 2021], data-based
reviews as those related to time series [Schlegel et al., 2019;
Rojat et al., 2021] and application-based reviews related to
healthcare [Adadi and Berrada, 2020] and banking [Burgt,
2020]. Other reviews are inspired by social science, human
psychology, sociology or cognitive sciences [Miller, 2019;
Capone and Bertolaso, 2020] in order to build ethical and fair
models [Barredo Arrieta et al., 2020].</p>
      <p>One key issue that have not been discussed in the above
cited reviews and that we would like to highlight, is the
importance of knowledge in XAI. As an interface between an AI
and a target audience, an explanation can be considered as an
interpreter between the AI knowledge and the human target
audience knowledge. Since knowledge domain is historical
in AI, this raises in turn important questions about the impact
of domains such as Knowledge Discovery and
Representation on XAI. Furthermore, regarding black-box models and
especially neural networks, it is important to mention that in
recent papers, concepts like representation learning,
knowledge/latent/hidden/abstract representation, latent space, etc.
have been studied in order to tackle issues such as
dimensionality, running time, algorithmic complexity, etc.
However, to the best of our knowledge, no explicit relation has
been defined between these concepts and XAI. We consider
that as these concepts are increasingly recurrent in the
literature, with no consensual definitions across fields, it becomes,
in turn, more difficult to apprehend the XAI domain.</p>
      <p>To address this issue, we propose a narrative review that,
contrary to the above cited literature reviews, does not
review XAI techniques. Our paper is a narrative review across
several domains: a literature-based review that synthesizes
technical research works related to domains that implicitly
inspire XAI works. Our goal is to bring original insights,
formulate new research questions and highlight promising future
directions of XAI. More precisely, in this narrative review, we
aim to address three questions. First, to centralize and
clarify concepts recurrently used in AI domains but not always
clear for XAI specialists. Second, to bring a new light to XAI
by making explicit the links between XAI and two other
domains: (i) Knowledge domain including Knowledge
Discovery Process (KDP) and Knowledge Representation (KR), and
(ii) Representation Learning (RL) more associated to deep
learning domain. Third, to offer an entry point to the XAI
domain for multidisciplinary or specialists in these domains.</p>
      <p>Actually, these domains are often perceived as
disconnected as most of the research is currently concentrated on
only one of them [Sallinger et al., 2020]. Despite this, we
believe that it is important to enhance the links and the implicit
relations that can be found between them. We thus consider
that XAI has been indirectly inspired by these domains.</p>
      <p>Figure 1 shows our vision as a schematic representation of
XAI domain and both KDP, KR and RL domains. Table 1 lists
the acronyms used. The paper is organized as follows:
definitions are presented in section 2, KDP and KR in section 3,
and RL in section 4. At the end of both last sections, we
discuss the relation between the highlighted points, related
to KDP, KR, RL and XAI. Finally, in section 5, we discuss
future directions and perspectives related to XAI.</p>
      <p>Definition 2.2 A data set is a collection of data that describes
real-word objects (such as cars, documents, animal, etc.)
through multiple properties called features [Bishop, 2006].
Definition 2.3 Once data is analyzed and correlated, it
represents information. Information can be reproduced from
data and its importance depends on the context it is generated
from/for [Grazzini and Pantisano, 2015; Malhotra and Nair,
2015].</p>
      <p>Definition 2.4 Knowledge is a set of information that is
assessed by a human, i.e. human adds a value and semantics
according to his/her own background and context [Grazzini
and Pantisano, 2015; Malhotra and Nair, 2015].</p>
      <p>Definition 2.5 In the data mining domain, a “ pattern is an
expression in some language describing a subset of the data
or a model applicable to the subset” [Fayyad et al., 1996].
Hence, Pattern extraction designates the process of finding
structures in data, fitting a model to data, or finding a
highlevel description of a data set.</p>
      <p>Many data modeling approaches have been proposed in the
state-of-the-art. We can cite reinforcement learning,
graphbased approaches, neural networks, etc. We now define some
important concepts related to these approaches.</p>
      <p>Definition 2.6 Reinforcement learning is an approach in
which an intelligent agent interacts with its environment
through trial-and-errors actions in order to reach a goal. Each
action leads to a modification of thestate of the agent and the
environment and the increase or decrease of a cumulative
reward value. Actions are chosen according to a strategy that
is called a policy [Barto and Sutton, 1995].</p>
      <p>Definition 2.7 A Manifold is a topological structure of
ndimensions. For example, a one-dimensional manifold is
a curve, a two-dimensional manifold is a surface, a
threedimensional manifold is a sphere.</p>
      <p>Definition 2.8 A network is a collection of discrete objects
called nodes, which are connected through links: it can be
viewed as a graph with vertices and edges, both with
attributes/weights or not [Fletcher et al., 1991].</p>
      <p>Definition 2.9 Neural networks are machine learning
models with several architectures, that are usually structured by
one or several layers (input, hidden and output). Each layer
is composed of one or several computational units called
artificial neurons - conceptually derived from biological
neurons [McCulloch and Pitts, 1943; Abraham, 2005].
Computational units can also be a Long Short Term Memory (well
known also as LSTM) [Hochreiter and Schmidhuber, 1997]
or Gated recurrent units [Cho et al., 2014]. A deep
neural network have many hidden layers, units, and edges with
weights. Units of layer n can be all or partially connected to
units of layer n + 1. Due to this inner complexity, deep neural
networks are a typical example of black-boxes.</p>
    </sec>
    <sec id="sec-2">
      <title>Definition 2.10 In neural networks, an activation pattern</title>
      <p>refers to units activation values of one of the layers. An
activation pattern is a numerical vector of the size of the layer it
is associated with. A hidden pattern refers to the activation
pattern of a hidden layer.</p>
      <p>In the literature of neural networks, concepts like latent
space and latent representation have been developed and
widely used. However, to the best of our knowledge, no
complete definitions have been clearly proposed for such
concepts. Due to the importance of both concepts in the rest of
this paper, we choose to formulate their definition next.
Definition 2.11 Latent space refers to the abstract
multidimensional space associated to each layer of a neural
network where the representation of the learned data is implicitly
built. Latent space contains the meaningful internal features
(definition 2.2) representations of learned data, which makes
it not directly interpretable. In a deep neural network
(definition 2.9), each hidden layer, whether it has the same number
of units or not, has its own latent space. It is thus possible
to extract several implicit representations from this network.
The latent space can be used to achieve a data
dimensionality reduction, when the hidden layer is smaller than the
input layer. This is the case for example with autoencoders
and variational autoencoders [Kingma and Welling, 2014],
models that can reduce high-dimensional inputs into efficient
and representative low-dimensional representations [Roberts
et al., 2018b].</p>
      <p>Definition 2.12 Latent or hidden representation refers to
the data representation implicitly encoded by a neural
network during the learning task and thus is
hidden-layerdependant [Bengio et al., 2013]. It is a machine-readable
data representation that contains features of the original data
that have been learned by associated hidden layer. One key
property of latent space (definition 2.11) is that real-world
objects (definition 2.2) that are semantically close (e.g. cars
of different brands), will end up grouped together in one
latent space: their respective hidden representation in the
corresponding layer, will be close to each other compared to other
objects that are not semantically close (e.g. cats) [Roberts et
al., 2018a]. Thus, a latent representation is useful for pattern
analysis (definition 2.5) and for similarity detection between
objects (definition 2.2) using clustering methods.
3</p>
      <sec id="sec-2-1">
        <title>Knowledge: discovery and representation</title>
        <p>We now present two active research domains: KDP
(section 3.1) and KR (section 3.2). Then, we discuss the relation
between them and XAI in section 3.3.
3.1</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Knowledge Discovery Process (KDP)</title>
      <p>KDP is a human-centered domain that seeks useful
knowledge (definition 2.4) through an iterative and interactive
process that involves humans [Lenca, 2002; Cios et al., 2007].
As the domains KDP, data mining, and Knowledge
Discovery in Databases (referred to as KDD) are often used in a
confused way, we consider that it is important to present a
clarification about them, as follows:
• According to [Cios et al., 2007], KDP and KDD
designate the same process. However, KDP can be
generalized to non-databases sources of data, while KDD
emphasizes databases as a primary source of data.
• KDP and data mining are related to each other as well as
to other domains like machine learning and statistics, but
are clearly distinct. Indeed, according to [Fayyad et al.,
1996] and [Cios et al., 2007], KDP is the global process
of discovering useful knowledge from data, whereas
data mining is a particular step within the KDP process
that consists in applying algorithms to extract patterns
(definition 2.5) or to build a model that fits the data.</p>
      <p>There is no consensus about the steps of a KDP: nine steps
in [Fayyad et al., 1996], eight steps in [Anand and Bu¨chner,
1998], six steps in [Wirth, 2000; Cios et al., 2007] and five
steps in [Cabena et al., 1998]. However, we emphasize that
globally KDP consists of three common main steps:
1. A pre-processing step for data collection or generation,
data preparation, cleaning, curing, etc.
2. A data processing step where several techniques from
statistics/machine learning/data mining, etc.
communities can be used.
3. A post-processing step for visualisation, evaluation and
validation.</p>
      <p>At each step, the extracted information (definition 2.3) is
usually evaluated by the human, given the context, to form
knowledge2 (definition 2.4). Thus, the target audience of
the KDP is the human: application domain experts and
decision makers. In addition, it is important to underline that
two mains goals of KDP are usually defined [Fayyad et al.,
1996]: (i) verification of a user hypothesis, and (ii)
discovery of valid and useful new knowledge that is understandable
with respect to the data (definition 2.1) from which it is
derived. These goals are thoroughly discussed in section 3.3.
3.2</p>
    </sec>
    <sec id="sec-4">
      <title>Knowledge Representation (KR)</title>
      <p>KR is a crucial question in AI [Malhotra and Nair, 2015].
Also known as “Knowledge Representation and Reasoning”,
KR aims at finding ways to efficiently structure specific
domain knowledge for automated reasoning. In this way,
intelligent machines can learn, draw inferences, make decision
and answer questions related to this knowledge [Davis et al.,
1993; Shapiro, 2006; Davis, 2015]. Thus, seen in such a way,
KR can be considered as a machine-oriented domain. The
purpose of KR is neither about storing data, nor making
actions but it is about allowing “thinking by reasoning” [Davis
et al., 1993]. Consequently, KR has been a key component
for the conception of intelligent knowledge-based systems.</p>
      <p>
        KR is also, according to [Malhotra and Nair, 2015], closely
related to the Knowledge retrieval in the shape of
ontologies
        <xref ref-type="bibr" rid="ref36 ref58 ref91">(concepts for representing, storing and accessing
knowledge [Guarino et al., 2009])</xref>
        . KR techniques have also been
widely developed and applied to semantic web [Hagedorn et
al., 2020], semantic networks [Malhotra and Nair, 2015], text
interpretation and cognitive robotics [Davis, 2015]. In
addition, from a user viewpoint, KR is important during the
development of software systems in order to perform particular
tasks, as well as for broader community of cognitive science
whose goal is to constitute and organize knowledge from
humans and machine perspectives [Das, 2003].
      </p>
    </sec>
    <sec id="sec-5">
      <title>Knowledge Representation Learning (KRL)</title>
      <p>KRL is the process of making AI algorithms model and
learn a structured representation of domain-specific
knowledge. As a consequence, concepts, relations between
them and their representations can be encoded in a
lowdimensional semantic space [Lin et al., 2018]. For
example, when knowledge is represented as a graph, the KRL
process allows graph embedding and preserves semantic
similarities [Xie et al., 2018]. Notice that the development of
deep learning algorithms and their performance on distributed
representations (i.e. representations that describe features of
the same data across layers) that reduce the computational
complexity has contributed to the emergence of several KRL
applications such as recommendation system [Zhang et al.,
2016], language modeling [Ahn et al., 2016] and question
answering [Yin et al., 2016]. We consider that KR has
recently become a more central domain in AI, and by extension
in XAI. This is mainly due to the development of
Representation Learning in neural networks (introduced in section 4).</p>
      <p>2Notice that recent approaches like AutoML tend to perform all
these steps automatically without user intervention [He et al., 2021]
3.3</p>
    </sec>
    <sec id="sec-6">
      <title>Discussion: relation between KDP, KR and XAI</title>
      <p>We now discuss and highlight several links and common
points between KDP, KR and XAI. As mentioned in
sections 3.1 and 3.2, KDP is a human-centered domain, whereas
KR is a machine-oriented domain. However, both domains
are complementary: in KDP, the main question is “How to
efficiently discover new or retrieve existing knowledge?”,
whereas in KR the tackled question is “How to represent the
knowledge efficiently to be able toreason on it?”.</p>
      <p>It is important to highlight that both KDP and KR
questions are also addressed and are crucial in XAI. Recall that the
objective of XAI is to make the reasons behind AI behavior
simple and accessible to a target audience regarding a given
task and context. We consider that this XAI objective can be
viewed and divided into two sub-objectives: (i) to discover
the reasons behind AI behavior - which is the same as in a
KDP problem -, (ii) to represent these reasons in a way that is
intelligible for the human target audience, but also sometimes
for an artificial one - which is the same as for a KR problem.</p>
      <p>Let us first detail the links between KDP and XAI. In XAI,
for black-boxes like deep neural networks [Guidotti et al.,
2018; Gilpin et al., 2018], technical approaches are used to
search the behavior of AI and make it explainable by
providing an explanation that can take several forms and be
multimodal [Barredo Arrieta et al., 2020]. Explaining an AI model
is therefore very inspired by KDP. The particular point is that
in XAI, the input data (definition 2.1) is related to the
blackbox AI model. This input data can be of several types, e.g.
activation patterns of hidden layers (definition 2.10), features or
representations, and require the same techniques as in KDP.</p>
      <p>Figure 2 represents a schematic representation of the
transformation of data into knowledge, in KDP and XAI domains.
It clarifies the similarities between both domains regarding
the human intervention, and the role of the technical part, i.e.
data mining and explainable methods.</p>
      <p>Let us now go into deep details about knowledge
representation in XAI. Two cases can be highlighted according to
target audience: (i) human who uses the knowledge
representation to reason and understand the situation, e.g. the decision
maker and the application domain expert, depending on their
expertise, role and goals, (ii) another AI system for which the
input data is provided from a complex AI architecture.</p>
      <p>Let us take an example in the domain of computer vision
and especially classification using deep neural networks.
Researchers have proposed approaches that exploit different AI
algorithms and their latent representation (definition 2.12) as
an input to the neural networks. The objective of such
approaches is to perform both classification and explainability
tasks through saliency masks applied to images and text
generation [LeCun et al., 2015]. This is one strategy among
multiple others for the representation of knowledge in order to
favor the explainability of the behavior of the initial AI model.</p>
      <p>In addition, notice that KRL has been basically associated
with deep learning algorithms, especially with techniques like
graph representation learning [Hamilton, 2020] and concept
learning [Dolgikh, 2018], which are both studied in the XAI
domain [Xu et al., 2018; Fazi, 2020].</p>
      <p>Finally, it is important to highlight the importance of the
target audience in both KDP/KR and XAI domains.
Actually, the role of the target audience is decisive: knowledge
is usually retrieved and shaped in order to answer a question
of a target audience related to a given task and context such
as verifying an hypothesis, inference and decision making.
Knowledge representation and content in both KDP/KR and
XAI domains are thus target and context dependant.</p>
      <p>As a conclusion, XAI is closely related to both KDP and
KR, and future works in XAI should take advantage of recent
works in both domains, as well as older works.</p>
      <sec id="sec-6-1">
        <title>Representation Learning (RL)</title>
        <p>We now present the RL domain, its importance in deep neural
networks, and RL sub-domains that are recurrent and popular.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>4.1 RL introduction and definition</title>
      <p>RL has been discussed as a key challenge related to different
machine learning domains [Dietterich et al., 2008] especially
to neural networks. As first demonstrated by [Rumelhart et
al., 1986], in neural networks, back-propagation algorithms
can generate useful internal representations of data in hidden
layers. Since then, different approaches have been proposed
in order to learn, analyze and visualize latent data
representations [Gilpin et al., 2018; Guidotti et al., 2018]. Thus, RL
has become an active research domain for which the objective
is to study of latent representations in order to improve deep
neural network efficiency[Bengio et al., 2013].</p>
      <p>RL - and synonyms like Data RL or Feature
Learning [Zhong et al., 2016] - focuses on “learning
representations of the data that make it easier to extract useful
information when building classifiers or other predictors” [Bengio et
al., 2013]. In other words, RL is designed to learn abstract
features that characterize data [Lesort et al., 2018].</p>
      <p>RL algorithms can be classified into two categories: global
and local RL algorithms. While the first ones tend to
preserve the data global information in the learned feature space,
the second ones focus more on preserving local similarity
between data during learning the new representations [Zhong
et al., 2016]. Representations are not task-specific but are
useful to machine learning algorithms to solve tasks, as well
as to humans to comprehend the behavior of these last
algorithms [Bengio et al., 2013]. One of the reasons that makes
RL popular is that representations express priors about the
data. The expressed priors can vary within a single learning
algorithm. Consequently, the characteristic of the priors
variations leads to different RL approaches, that we classify into
two categories: problems-oriented RL and data-oriented RL.</p>
      <p>In the following section, we first present the concept of
hierarchical representation in deep neural networks, a key
property of RL. Then we present examples of particular cases of
RL that are problems-oriented and data-oriented.
4.2</p>
    </sec>
    <sec id="sec-8">
      <title>Hierarchical representations in deep neural networks</title>
      <p>One key property of the RL domain in deep neural networks
is the ability to provide both high level features and low level
features for the same learned data. Recall that a deep
neural network will encode a latent representation at each hidden
layer (definitions 2.9, 2.12). Since the layer n units can be
all or partially connected to the layer n + 1 units, each layer
uses the previous layer as input. If the previous layer is a
hidden layer, then the input is already a latent representation,
i.e. an abstract feature representation that characterizes the
data. Thus, each layer extracts an abstract feature
representation of the previous layer. As a result, a deep neural network
learns multiple levels of abstraction and implicitly encodes a
hierarchy of latent and abstract representations that are
built progressively, layer by layer. The layers that are close
to the input layer will encode a low-level feature
representation, whereas those deeper inside the architecture will encode
a high level feature representation. In other words, the closer
the considered layer is to the output layer, the more the
representation is abstract [Bengio et al., 2013; Zhong et al., 2016;
Lesort et al., 2018], as represented in Figure 3.</p>
      <p>It has also been shown that, in deep learning algorithms,
hidden representations tend to keep dominant information
and propagate them across hidden layers, regardless the width
or depth increase of the deep neural networks [Nguyen et al.,
2021]. This characteristic of RL is also a key one for XAI:
by extracting and comparing the low-level and the high-level
representations of a deep architecture, we consider that it is
possible to explicit the inner mechanism of the architecture by
observing the differences between the representations. This
will be discussed further in section 5.
4.3</p>
    </sec>
    <sec id="sec-9">
      <title>Problems-oriented RL approaches</title>
      <p>Recall that the objective of RL algorithms is to learn
abstract features that characterize data. This objective can be
challenging according the issues that one could face such as
high dimensionnality of data or RL application to another AI
paradigm like reinforcement learning (definition 2.6). In the
following sub-sections, we describe two RL sub-domains:
Manifold RL and State RL, that have recently shown great
performances in deep learning and that deal with our core
questions. The links with XAI are also briefly discussed.</p>
    </sec>
    <sec id="sec-10">
      <title>Manifold RL</title>
      <p>Manifold RL is particularly suited for dealing with
highdimensional data sets that are very difficult to visualize and
less intuitive. However, within such data sets, data can
locally belong to a subset that can be represented by a
manifold. As stated in definition 2.7, a manifold is a
topological structure of n-dimensions. Thus, Manifold RL
corresponds to the learning of complex data representation in
several dimensions while preserving the topological
properties of the considered manifold. We consider Manifold
RL as a non-linear dimensionality reduction approach, that
can help to discover similarities in data for which
dimensions have been reduced [Cayton, 2005; Bengio, 2009; Zhang
et al., 2011]. The Manifold RL domain aims at
discovering manifold structure hidden in high dimensional data.
It seeks to discover the intrinsic structure of a given
manifold. Notice that when many manifolds are considered,
we refer to this as multi-manifold RL [Lee et al., 2016;
Torki et al., 2010]. It allows to both preserve the local
geometric structure within distinct manifolds while ensuring the
discriminability between them [Wu et al., 2020].</p>
      <p>When more neural networks transparency is required, the
visualisation of latent representations is essential: it allows
to develop an intuition about the distance between subsets
of data represented by their associated latent manifold
representations. Consequently, we consider that this
dimensionreduction characteristic is therefore of great practical interest
for XAI. Indeed, reducing the complexity due to the high
dimensions can strongly contribute in understanding the inner
mechanisms of models exploiting the data, but also the role
of the data subsets on the models behaviors.</p>
    </sec>
    <sec id="sec-11">
      <title>State RL (SRL)</title>
      <p>In addition, RL can also concern domains where data are in a
low dimensional space. SRL is “is a particular type of
representation learning that aims at building a low-dimensional
and meaningful representation of a state space, by processing
high-dimensional raw observation data (e.g., learn a position
(x, y) from raw image pixels) .” [Heuillet et al., 2021]. This
domain is thus particularly suited for learning features in
reinforcement learning, robotics and control scenarios. Thus,
learning in SRL for an artificial agent is rather related to
building a latent model of the environment and the task to
perform through interactions [Lesort et al., 2018]. In addition, it
has been shown that SRL provides three main advantages for
several research domains [Heuillet et al., 2021]:
• The learned features are of low dimensions which
improves speed and generalization of deep learning
models [Lesort et al., 2017].
• SRL helps improving performance in some
reinforcement learning steps such as policy learning [Heuillet et
al., 2021].
• Learning representations of states (definition 2.6),
actions or policies provide meaning to explain a
reinforcement learning algorithms. Indeed, SRL allows to learn
representations that capture the variation in the
environment generated by the action of the agent [Lesort et al.,
2017; Heuillet et al., 2021].</p>
      <p>It has been shown that SRL is particularly suitable to
make the behavior of an artificial agent and the reasons of
this behavior accessible for humans [Lesort et al., 2017;
Lesort et al., 2018; Heuillet et al., 2021]. Consequently, we
can consider SRL as an example of domains used for
explanation goals in reinforcement learning.
4.4</p>
    </sec>
    <sec id="sec-12">
      <title>Data-oriented RL approaches</title>
      <p>In RL, several approaches tackle the problem of increasing
data volumes, their heterogeneity and the multiplicity of their
sources. We can consider them as data-oriented approaches
and present two of them: the Multi-view RL and the Network
RL. We also highlight the link between RL applied to
realworld data-oriented problems and XAI domain.</p>
    </sec>
    <sec id="sec-13">
      <title>Multi-view RL</title>
      <p>In real-world applications, each object can be described by
multiple features (definition 2.2) [Xu et al., 2013]. It is
thus referred as Multi-view data. These features, also
referred to as views, constitute complementary and diverse
information of the same data [Xu et al., 2018]. For
example, one information can be obtained through multiple
sources, which is the case in the application where
different people are talking about the same thing. Another
example can be an image that is described via a set of visual
features such as color, shape and textures. Multi-view RL
is thus concerned with the problem of the integration of
information from multiple views and uncovers the latent
structure shared by multiple views, while preserving the
original information and the global meaning [Zhu et al., 2014;
Xu et al., 2018]. It has been shown that Multi-view RL can
facilitate extracting useful information when developing
prediction models [Li et al., 2018] and also helps encoding
concepts and semantics in deep neural network [Xu et al., 2018].
Recently, Multi-view RL has been used to design an
explainable recommendation system [Gao et al., 2019], where
authors claim that “it is difficult to model the relationships
between high-level and low-level features since they have
overlapping meaning” . To overcome this issue, a Multi-view
learning approach has been proposed by considering different
levels of features as different views. The learned
representation can then be a representation of different levels of features
of the input data. Accordingly, we consider that Multi-view
RL can be employed for explainability tasks.</p>
    </sec>
    <sec id="sec-14">
      <title>Network RL</title>
      <p>Network RL is a learning paradigm proposed to analyze
networks such as graphs, and thus allows users to deeply
understand the hidden features of graphs [Sun et al., 2020].
This domain aims at learning in a low-dimensional space of
network vertices (definition 2.8), while preserving the
structure of the network topology, the content of the vertices and
other information as vertices attributes and links attributes.
Network RL can be considered as a dimensionality
reduction technique and an intermediate step to solve a target
task [Zhang et al., 2020]. Since the information of the
original network is preserved in a new vector-based
representation, conventional vector-based machine learning algorithms
can be applied. Thus, Network analysis and mining tasks
become easier as there is no more need to use complex
algorithms directly designed for graphs.</p>
      <p>Consequently, Network RL has multiple applications such
as: vertex classification, link prediction, clustering,
visualization and recommendations [Dong et al., 2020; Zhang et al.,
2020]. Network RL approaches have been widely applied to
information networks [Sun et al., 2020; Zhang et al., 2020]
and are becoming increasingly popular for capturing complex
relationships in various real-world applications [Yang et al.,
2015; Sun et al., 2020; Zhang et al., 2020], such as social
networks, citation networks, telecommunication networks,
biological networks, recommender systems, etc.</p>
      <p>In addition, Network RL is essential in the study of
heterogeneous information networks (i.e. where vertices are of
different types), in order to capture semantic proximity
between vertices representations [Dong et al., 2020]. Given the
high scale of some networks that can range from hundred to
billions of vertices and the heterogeneity of information, we
believe that Network RL and XAI should be considered
together in order to perform efficient and explainable analytical
tasks. Also, in related applications, an in depth analysis using
XAI techniques and Network RL can help interpreting
empirical results and providing a deep understanding of the
applied black-box model. To conclude, Network RL should be
considered as a dimensionality reduction technique whenever
graph-data structure is involved in the design of XAI.
4.5</p>
    </sec>
    <sec id="sec-15">
      <title>Discussion: relation between RL and XAI</title>
      <p>We have presented several research works in RL (Manifold
RL, State RL, Multi-view RL and Network RL) and we next
highlight common points between RL and XAI.</p>
      <p>First, let us discuss the contribution of the hierarchical
RL on XAI modeling. Recall that while RL focuses on
learning a data representation in order to get a better performance
of the AI model [Bengio et al., 2013], XAI is interested in
exploring this representation to explain the performance and
behavior of the model. This representation varies according
to the techniques used in the involved AI models (e.g. an
artificial agent or a neural network). In the case of deep neural
networks models, the hierarchical level of representations is
important for XAI, as it allows to extract different types of
information that can be used in several ways:
• The study of low-level representations can help to detect
important features used by the deep network to make a
prediction. This contributes to the explanation and
understanding of the deep network by determining features
involved in a particular output (i.e. a prediction).
• The study of high-level representations can help to
detect groups of features involved in a prediction, and how
and where a deep neural architecture deals with these
groups. This is interesting to explain relevant hidden
information and their location within the architecture.</p>
      <p>For example, a hierarchical multi-scale deep recurrent
network approach has been proposed for data sequences [Chung
et al., 2016]: in order to discover temporal dependencies in
data, the latent hierarchical structure in the sequences has
been exploited without using explicit boundary information.
Accordingly, we consider that the hierarchical structure of the
latent representations is an important characteristic of deep
networks in order to propose a model-specific XAI modeling.</p>
      <p>Second, we focus now on the contribution of
problemsoriented and data-oriented RL approaches discussed above
on the explainability of AI models.</p>
      <p>• Recall that for high-dimensional data sets, Manifold
RL allows to perform dimension reduction in the latent
space while preserving the distance or similarities
between data. Consequently, one of the main advantages
is that visualisation of the data representation inside the
latent space allows to get a better intuition and
understanding of the inner mechanisms of models.
• Recall that in reinforcement learning, SRL allows to
explicit the agent state changes while performing a task in
a given environment. This is similar to the XAI objective
as it makes the behavior of an artificial agent explicit and
more intelligible for a given target audience. Also,
recent works have mentioned that State RL can be viewed
as a mean for XAI in reinforcement learning [Heuillet
et al., 2021]. Other works describe State RL as an
approach for robotics and control scenarios that provides
easier interpretation of the variation in the environment
[Lesort et al., 2017]. Consequently, we can consider that
the goals of SRL are in line with those of XAI.
• Through the presentation of Multi-view RL and
Network RL in section 4.4, we have shown that real-world
applications of RL techniques that can be more specific
to a particular data type or data organisation, are also
linked to XAI. Indeed, an AI model can learn from
multiple data sets of complex data representation such as
networks (e.g. social network modeling, biological
networks). The complexity of the learned data can also
impact the behavior of the AI model. Consequently, this
allows us to conclude that adopting RL approaches that
take into account the type of learned data, is a way to
make AI models more explicit and explainable.</p>
      <p>Figure 4 summarizes the above conclusions and questions
tackled throughout the section 4. Table 2 summarizes RL
domains and some examples of application domains.</p>
      <p>Non-exhaustive examples of application domain
Speech recognition [Liu et al., 2020]
Object recognition [Wang et al., 2020]
NLP [Mikolov et al., 2013; Be´rard et al., 2016]
Robotics [Lesort et al., 2017]
Numerical artificial agent[Madumal et al., 2020]
Data mining [Torki et al., 2010]
Concept learning [Xu et al., 2018]
Image processing [Su et al., 2011]
Recommender systems explainability [Gao et al., 2019]
Networks of concepts [Yang et al., 2015; Qi et al., 2020]
Identification of genes in biology[Ietswaart et al., 2021]</p>
      <p>Community detection in social networks [Tu et al., 2018]</p>
      <sec id="sec-15-1">
        <title>Discussion and conclusion</title>
        <p>We now summarize the highlighted points presented in
previous sections. We also present promising directions related to
the XAI domain. Since our paper is a multidisciplinary one at
the crossroad of several domains, we have first (in section 2)
centralized and clarified definitions of several concepts, that
could indeed seem basic and well-known to involved AI
experts, but are important to bridge the discussed domains. A
special focus has been made on latent space, latent
representation and hierarchical representation which are essential for
knowledge extraction in deep neural networks and thus in
XAI. To the best of our knowledge, no previous work has
established a clear definition of these concepts for XAI
community. This is necessary to allow the collaboration between
the different domains necessary to build XAI. Second, we
analysed and highlighted the existence of relations between
Knowledge domains (KDP, KR), RL and XAI.</p>
        <p>As we have shown in section 1, the goal of XAI is to
convey the most semantically complete explanation to a target
audience in order to answer a particular question within a
given context. This explanation should take into account two
important points: (i) the prior knowledge of the target
audience regarding the application context, and (ii) the
technical aspects of the AI used model that provided solutions to a
specific task, and that thus contributed, due to its
complexity/opacity, to the emergence of the question behind the need
of XAI, i.e in short, ”What are the reasons behind the results
and/or how the AI model reaches these results?”.</p>
        <p>We consider that XAI is technically at the crossroad of
at least two domains: (i) KDP and KR when viewed from
a human perspective, and (ii) RL that tackles implicitly the
same objectives as XAI, from a technical and algorithmic
perspectives. KDP, KR, RL domains, while distinct, are
overlapped. They do and should have an explicit impact on XAI
approaches:
• First, as we have previously mentioned, several XAI
approaches are indirectly inspired by the domain of
Knowledge (KDP, KR and data mining) as both tend to
express information from data. However, it is important
to recall that, in XAI the input data reflects the
internal mechanisms of the AI model, its predictions, and/or
its behavior. The evolution of the Knowledge domain is
therefore an inspiration area for XAI.
• Second, the development of AI approaches and in
particular of deep learning, has blurred the boundaries
between KR and RL, since several KR approaches involve
RL and deep learning. In addition, recall that while RL
is interested in features modeling for algorithmic issues
(performance, dimensionality, etc.), XAI is interested in
features since it contributes to explicit the inner
mechanisms behind the results. This implies that KR, RL and
XAI are indeed interested in the data representation in
order to answer different but related questions. We thus
consider that, in order to make a significant progress,
XAI future works should not forget KR and RL past and
recent works as inspirations.</p>
        <p>KDP, KR and RL have been extensively confronted with,
first, issues related to providing a data-driven explanation
to different stakeholders according to their expectations and
context, and second, issues related to biases and fairness in AI
[Nelson, 2019]. This highlights the human significant role on
data processing and bias detection in AI towards XAI. We
believe that this review is all the more topical and important
as works about the alliance between symbolic AI and
connectionist AI should be more and more important in the next
years3, e.g. injecting a priori knowledge into neural networks
to limit unethical AI [Goebel et al., 2018] and biases
[Gordon and Desjardins, 1995; Leavy, 2018; Lepri et al., 2018;
Nelson, 2019]. We are convinced that very promising
directions can be taken in XAI future works by taking advantage
of KDP, KR and RL development to help design ethical,
unbiased and human-centered XAI. To conclude, we point out
that other domains, not discussed in this paper, also impact
XAI directions such as cognitive psychology [Le Saux et al.,
2002], cognitive sciences for biases studies [Soleimani et al.,
2021], social sciences [Miller, 2019] and Human Machine
Interaction field [Le Saux et al., 1999; Mueller et al., 2021;
Ehsan et al., 2021].
6</p>
      </sec>
      <sec id="sec-15-2">
        <title>Acknowledgments</title>
        <p>Thanks to the Conseil re´gional de Bretagne and the European
Union for funding this work via the FEDER program.</p>
        <p>3The alliance of symbolic AI and connectionist approaches have
been proposed a long time ago, e.g. [Honavar, 1995].
human-centered perspectives in explainable AI. In
Extended Abstracts of the 2021 CHI Conference on Human
Factors in Computing Systems, pages 1–6, 2021.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <source>[Abraham</source>
          , 2005]
          <string-name>
            <given-names>Ajith</given-names>
            <surname>Abraham</surname>
          </string-name>
          .
          <article-title>Artificial neural networks</article-title>
          .
          <source>Handbook of measuring system design</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <source>[Adadi and Berrada</source>
          , 2020]
          <string-name>
            <given-names>Amina</given-names>
            <surname>Adadi</surname>
          </string-name>
          and
          <string-name>
            <given-names>Mohammed</given-names>
            <surname>Berrada</surname>
          </string-name>
          .
          <article-title>Explainable ai for healthcare: From black box to interpretable models</article-title>
          .
          <source>In Embedded Systems and Artificial Intelligence</source>
          , pages
          <fpage>327</fpage>
          -
          <lpage>337</lpage>
          . Springer,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [Ahn et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Sungjin</given-names>
            <surname>Ahn</surname>
          </string-name>
          , Heeyoul Choi, Tanel Pa¨rnamaa, and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <article-title>A neural knowledge language model</article-title>
          .
          <source>arXiv preprint arXiv:1608.00318</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [Anand and Bu¨chner,
          <year>1998</year>
          <article-title>] Sarabjot S Anand and Alex G Bu¨chner. Decision support using data mining</article-title>
          .
          <source>Financial Times Management</source>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [Augusto et al.,
          <year>2017</year>
          ]
          <string-name>
            <given-names>J</given-names>
            <surname>Augusto</surname>
          </string-name>
          , Asier Aztiria,
          <string-name>
            <surname>Dean Kramer</surname>
            , and
            <given-names>Unai</given-names>
          </string-name>
          <string-name>
            <surname>Alegre</surname>
          </string-name>
          .
          <article-title>A survey on the evolution of the notion of context-awareness</article-title>
          .
          <source>Applied Artificial Intelligence</source>
          ,
          <volume>31</volume>
          (
          <issue>7-8</issue>
          ):
          <fpage>613</fpage>
          -
          <lpage>642</lpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>[Barredo</surname>
          </string-name>
          Arrieta et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Alejandro</given-names>
            <surname>Barredo</surname>
          </string-name>
          <string-name>
            <surname>Arrieta</surname>
          </string-name>
          ,
          <source>Natalia Diaz-Rodriguez, Javier Del Ser</source>
          ,
          <string-name>
            <given-names>Adrien</given-names>
            <surname>Bennetot</surname>
          </string-name>
          , Siham Tabik, Alberto Barbado, Salvador Garcia, Sergio Gil-Lopez, Daniel Molina, Richard Benjamins, Raja Chatila, and Francisco Herrera.
          <article-title>Explainable artificial intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI</article-title>
          .
          <source>Information Fusion</source>
          ,
          <volume>58</volume>
          :
          <fpage>82</fpage>
          -
          <lpage>115</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <source>[Barto and Sutton</source>
          , 1995] Andrew G Barto and
          <article-title>Richard S Sutton. Reinforcement learning</article-title>
          .
          <source>Handbook of brain theory and neural networks</source>
          , pages
          <fpage>804</fpage>
          -
          <lpage>809</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [Bengio et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Vincent</surname>
          </string-name>
          .
          <article-title>Representation learning: A review and new perspectives</article-title>
          .
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          ,
          <volume>35</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1798</fpage>
          -
          <lpage>1828</lpage>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <source>[Bengio</source>
          , 2009]
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <article-title>Learning deep architectures for AI</article-title>
          . Now Publishers Inc,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [Be´rard et al.,
          <year>2016</year>
          ] Alexandre Be´rard, Christophe Servan, Olivier Pietquin, and
          <string-name>
            <given-names>Laurent</given-names>
            <surname>Besacier</surname>
          </string-name>
          .
          <article-title>Multivec: a multilingual and multilevel representation learning toolkit for NLP. In The 10th edition of the Language Resources</article-title>
          and Evaluation Conference,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[Bishop</source>
          , 2006]
          <string-name>
            <surname>Christopher</surname>
            <given-names>M</given-names>
          </string-name>
          <string-name>
            <surname>Bishop.</surname>
          </string-name>
          <article-title>Pattern recognition and machine learning</article-title>
          . springer,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [Brenon et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Alexis</given-names>
            <surname>Brenon</surname>
          </string-name>
          , Franc¸ois Portet, and
          <string-name>
            <given-names>Michel</given-names>
            <surname>Vacher</surname>
          </string-name>
          .
          <article-title>Context feature learning through deep learning for adaptive context-aware decision making in the home</article-title>
          .
          <source>In 2018 14th International Conference on Intelligent Environments (IE)</source>
          , pages
          <fpage>32</fpage>
          -
          <lpage>39</lpage>
          . IEEE,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [Bre´zillon, 1999]
          <article-title>Patrick Bre´zillon. Context in Artificial Intelligence: I. A survey of the literature</article-title>
          .
          <source>Computers and artificial intelligence</source>
          ,
          <volume>18</volume>
          :
          <fpage>321</fpage>
          -
          <lpage>340</lpage>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>[Buchanan and Shortliffe</source>
          , 1984] Bruce G Buchanan and
          <article-title>Edward H Shortliffe</article-title>
          .
          <article-title>Rule-based expert systems: the MYCIN experiments of the stanford heuristic programming project</article-title>
          .
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>[Burgt</source>
          , 2020]
          <article-title>Joost van der Burgt. Explainable AI in banking</article-title>
          .
          <source>Journal of Digital Banking</source>
          ,
          <volume>4</volume>
          (
          <issue>4</issue>
          ):
          <fpage>344</fpage>
          -
          <lpage>350</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [Cabena et al.,
          <year>1998</year>
          ]
          <string-name>
            <given-names>Peter</given-names>
            <surname>Cabena</surname>
          </string-name>
          , Pablo Hadjinian, Rolf Stadler, Jaap Verhees, and
          <string-name>
            <given-names>Alessandro</given-names>
            <surname>Zanasi</surname>
          </string-name>
          .
          <article-title>Discovering data mining: from concept to implementation</article-title>
          . PrenticeHall, Inc.,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <source>[Capone and Bertolaso</source>
          , 2020]
          <string-name>
            <given-names>Luca</given-names>
            <surname>Capone</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Bertolaso</surname>
          </string-name>
          .
          <article-title>A philosophical approach for a human-centered explainable AI</article-title>
          .
          <source>In The Italian Workshop on Explainable Artificial Intelligence, the 19th International Conference of the Italian Association for Artificial Intelligence</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <source>[Cayton</source>
          , 2005]
          <string-name>
            <given-names>Lawrence</given-names>
            <surname>Cayton</surname>
          </string-name>
          .
          <article-title>Algorithms for manifold learning</article-title>
          .
          <source>Univ. of California at San Diego Tech. Rep</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [Cho et al.,
          <year>2014</year>
          ]
          <string-name>
            <given-names>Kyunghyun</given-names>
            <surname>Cho</surname>
          </string-name>
          , Bart van Merrie¨nboer, Caglar Gulcehre, Dzmitry Bahdanau, Fethi Bougares, Holger Schwenk, and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <article-title>Learning phrase representations using RNN encoder-decoder for statistical machine translation</article-title>
          .
          <source>In The 18th Empirical Methods in Natural Language Processing</source>
          , pages
          <fpage>1724</fpage>
          -
          <lpage>1734</lpage>
          , Doha, Qatar,
          <year>October 2014</year>
          .
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [Chung et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Junyoung</given-names>
            <surname>Chung</surname>
          </string-name>
          , Sungjin Ahn, and
          <string-name>
            <given-names>Yoshua</given-names>
            <surname>Bengio</surname>
          </string-name>
          .
          <article-title>Hierarchical multiscale recurrent neural networks</article-title>
          .
          <source>arXiv preprint arXiv:1609.01704</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [Cios et al.,
          <year>2007</year>
          ] Krzysztof J Cios, Witold Pedrycz, Roman W Swiniarski,
          <article-title>and Lukasz Andrzej Kurgan. Data mining: a knowledge discovery approach</article-title>
          . Springer Science &amp; Business
          <string-name>
            <surname>Media</surname>
          </string-name>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>[Das</surname>
          </string-name>
          ,
          <year>2003</year>
          ]
          <string-name>
            <given-names>Amit</given-names>
            <surname>Das</surname>
          </string-name>
          .
          <article-title>Knowledge representation</article-title>
          . In Hossein Bidgoli, editor,
          <source>Encyclopedia of Information Systems</source>
          , pages
          <fpage>33</fpage>
          -
          <lpage>41</lpage>
          . Elsevier, New York,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [Davis et al.,
          <year>1993</year>
          ]
          <string-name>
            <given-names>Randall</given-names>
            <surname>Davis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Howard</given-names>
            <surname>Shrobe</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Peter</given-names>
            <surname>Szolovits</surname>
          </string-name>
          .
          <article-title>What is a knowledge representation? AI magazine</article-title>
          ,
          <volume>14</volume>
          (
          <issue>1</issue>
          ):
          <fpage>17</fpage>
          -
          <lpage>17</lpage>
          ,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <source>[Davis</source>
          , 2015]
          <string-name>
            <given-names>Ernest</given-names>
            <surname>Davis</surname>
          </string-name>
          .
          <article-title>Knowledge representation</article-title>
          .
          <source>In International Encyclopedia of the Social &amp; Behavioral Sciences: Second Edition</source>
          , pages
          <fpage>98</fpage>
          -
          <lpage>104</lpage>
          . Elsevier Inc.,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [Dietterich et al.,
          <year>2008</year>
          ] Thomas G Dietterich,
          <article-title>Pedro Domingos</article-title>
          , Lise Getoor, Stephen Muggleton, and
          <string-name>
            <surname>Prasad Tadepalli.</surname>
          </string-name>
          <article-title>Structured machine learning: the next ten years</article-title>
          .
          <source>Machine Learning</source>
          ,
          <volume>73</volume>
          (
          <issue>1</issue>
          ):
          <fpage>3</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <source>[Dolgikh</source>
          , 2018]
          <string-name>
            <given-names>Serge</given-names>
            <surname>Dolgikh</surname>
          </string-name>
          .
          <article-title>Spontaneous concept learning with deep autoencoder</article-title>
          .
          <source>International Journal of Computational Intelligence Systems</source>
          ,
          <volume>12</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [Dong et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Yuxiao</given-names>
            <surname>Dong</surname>
          </string-name>
          , Ziniu Hu,
          <string-name>
            <surname>Kuansan</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yizhou Sun</surname>
            , and
            <given-names>Jie</given-names>
          </string-name>
          <string-name>
            <surname>Tang</surname>
          </string-name>
          .
          <article-title>Heterogeneous network representation learning</article-title>
          .
          <source>In The 29th International Joint Conference on Artificial Intelligence</source>
          , pages
          <fpage>4861</fpage>
          -
          <lpage>4867</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [Ehsan et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Upol</given-names>
            <surname>Ehsan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Philipp</given-names>
            <surname>Wintersberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q Vera</given-names>
            <surname>Liao</surname>
          </string-name>
          , Martina Mara, Marc Streit, Sandra Wachter, Andreas Riener, and
          <string-name>
            <given-names>Mark O</given-names>
            <surname>Riedl. Operationalizing</surname>
          </string-name>
          [Fayyad et al.,
          <year>1996</year>
          ]
          <string-name>
            <given-names>Usama</given-names>
            <surname>Fayyad</surname>
          </string-name>
          , Gregory PiatetskyShapiro, and Padhraic Smyth.
          <article-title>From data mining to knowledge discovery in databases</article-title>
          .
          <source>AI magazine</source>
          ,
          <volume>17</volume>
          (
          <issue>3</issue>
          ):
          <fpage>37</fpage>
          -
          <lpage>37</lpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <source>[Fazi</source>
          , 2020]
          <string-name>
            <given-names>M</given-names>
            <surname>Beatrice</surname>
          </string-name>
          <article-title>Fazi</article-title>
          .
          <article-title>Beyond human: Deep learning, explainability and representation</article-title>
          .
          <source>Theory, Culture &amp; Society</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [Fletcher et al.,
          <year>1991</year>
          ]
          <string-name>
            <given-names>Peter</given-names>
            <surname>Fletcher</surname>
          </string-name>
          , Hughes Hoyle, and
          <string-name>
            <given-names>C</given-names>
            <surname>Wayne</surname>
          </string-name>
          <article-title>Patty</article-title>
          .
          <source>Foundations of discrete mathematics. Brooks/Cole</source>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>[Gao</surname>
          </string-name>
          et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Jingyue</given-names>
            <surname>Gao</surname>
          </string-name>
          , Xiting Wang,
          <string-name>
            <given-names>Yasha</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Xing</given-names>
            <surname>Xie</surname>
          </string-name>
          .
          <article-title>Explainable recommendation through attentive multi-view learning</article-title>
          .
          <source>The 33th AAAI Conference on Artificial Intelligence</source>
          ,
          <volume>33</volume>
          (
          <issue>01</issue>
          ):
          <fpage>3622</fpage>
          -
          <lpage>3629</lpage>
          , Jul.
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [Gilpin et al.,
          <year>2018</year>
          ]
          <string-name>
            <surname>Leilani H Gilpin</surname>
            , David Bau, Ben Z Yuan, Ayesha Bajwa,
            <given-names>Michael</given-names>
          </string-name>
          <string-name>
            <surname>Specter</surname>
            , and
            <given-names>Lalana</given-names>
          </string-name>
          <string-name>
            <surname>Kagal</surname>
          </string-name>
          .
          <article-title>Explaining explanations: An overview of interpretability of machine learning</article-title>
          .
          <source>In The 5th IEEE International Conference on data science and advanced analytics</source>
          , pages
          <fpage>80</fpage>
          -
          <lpage>89</lpage>
          . IEEE,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [Goebel et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Randy</given-names>
            <surname>Goebel</surname>
          </string-name>
          , Ajay Chander, Katharina Holzinger, Freddy Lecue, Zeynep Akata, Simone Stumpf,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Kieseberg</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Andreas</given-names>
            <surname>Holzinger</surname>
          </string-name>
          .
          <article-title>Explainable ai: the new 42? In International cross-domain conference for machine learning</article-title>
          and
          <source>knowledge extraction</source>
          , pages
          <fpage>295</fpage>
          -
          <lpage>303</lpage>
          . Springer,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <source>[Gordon and Desjardins</source>
          , 1995]
          <article-title>Diana F Gordon and Marie Desjardins. Evaluation and selection of biases in machine learning</article-title>
          .
          <source>Machine learning</source>
          ,
          <volume>20</volume>
          (
          <issue>1-2</issue>
          ):
          <fpage>5</fpage>
          -
          <lpage>22</lpage>
          ,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <source>[Grazzini and Pantisano</source>
          , 2015]
          <string-name>
            <given-names>Jacopo</given-names>
            <surname>Grazzini</surname>
          </string-name>
          and
          <string-name>
            <given-names>Francesco</given-names>
            <surname>Pantisano</surname>
          </string-name>
          .
          <article-title>Collaborative research-grade software for crowd-sourced data exploration: from context to practice - part I: Guidelines for scientific evidence provision for policy support based on big data and open technologies</article-title>
          .
          <source>EUR 27094. Luxembourg: Publications Office of the European Union, (JRC94504)</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [Guarino et al.,
          <year>2009</year>
          ]
          <string-name>
            <given-names>Nicola</given-names>
            <surname>Guarino</surname>
          </string-name>
          , Daniel Oberle, and
          <string-name>
            <given-names>Steffen</given-names>
            <surname>Staab</surname>
          </string-name>
          .
          <article-title>What is an ontology? In Handbook on ontologies</article-title>
          , pages
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          . Springer,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [Guidotti et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Riccardo</given-names>
            <surname>Guidotti</surname>
          </string-name>
          , Anna Monreale, Salvatore Ruggieri, Franco Turini, Fosca Giannotti, and
          <string-name>
            <given-names>Dino</given-names>
            <surname>Pedreschi</surname>
          </string-name>
          .
          <article-title>A survey of methods for explaining black box models</article-title>
          .
          <source>ACM computing surveys (CSUR)</source>
          ,
          <volume>51</volume>
          (
          <issue>5</issue>
          ):
          <fpage>93</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <source>[Gunning</source>
          , 2017]
          <string-name>
            <given-names>David</given-names>
            <surname>Gunning</surname>
          </string-name>
          .
          <article-title>Explainable artificial intelligence (XAI)</article-title>
          .
          <source>Defense Advanced Research Projects Agency (DARPA)</source>
          ,
          <volume>2</volume>
          (
          <issue>2</issue>
          ),
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [Hagedorn et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Hagedorn</surname>
          </string-name>
          , Mary Bone, Benjamin Kruse, Ian Grosse, and
          <string-name>
            <given-names>Mark</given-names>
            <surname>Blackburn</surname>
          </string-name>
          .
          <article-title>Knowledge representation with ontologies and semantic web technologies to promote augmented and artificial intelligence in systems engineering</article-title>
          . Insight,
          <volume>23</volume>
          (
          <issue>1</issue>
          ):
          <fpage>15</fpage>
          -
          <lpage>20</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <source>[Hamilton</source>
          , 2020] William L Hamilton.
          <article-title>Graph representation learning</article-title>
          .
          <source>Synthesis Lectures on Artifical Intelligence and Machine Learning</source>
          ,
          <volume>14</volume>
          (
          <issue>3</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>159</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [He et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Xin</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kaiyong Zhao</surname>
            ,
            <given-names>and Xiaowen</given-names>
          </string-name>
          <string-name>
            <surname>Chu</surname>
          </string-name>
          .
          <article-title>Automl: A survey of the state-of-the-art</article-title>
          .
          <source>KnowledgeBased Systems</source>
          ,
          <volume>212</volume>
          :
          <fpage>106622</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [Heuillet et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Alexandre</given-names>
            <surname>Heuillet</surname>
          </string-name>
          , Fabien Couthouis, and
          <string-name>
            <surname>Natalia</surname>
          </string-name>
          Diaz-Rodriguez.
          <article-title>Explainability in deep reinforcement learning</article-title>
          .
          <source>Knowledge-Based Systems</source>
          ,
          <volume>214</volume>
          :
          <fpage>106685</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          <source>[Hochreiter and Schmidhuber</source>
          , 1997]
          <article-title>Sepp Hochreiter and Ju¨rgen Schmidhuber. Long short-term memory</article-title>
          .
          <source>Neural computation</source>
          ,
          <volume>9</volume>
          (
          <issue>8</issue>
          ):
          <fpage>1735</fpage>
          -
          <lpage>1780</lpage>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [Hollister et al.,
          <year>2017</year>
          ] Debra L Hollister,
          <string-name>
            <surname>Avelino Gonzalez</surname>
            , and
            <given-names>James</given-names>
          </string-name>
          <string-name>
            <surname>Hollister</surname>
          </string-name>
          .
          <article-title>Contextual reasoning in human cognition and the implications for artificial intelligence systems</article-title>
          .
          <source>In The International and Interdisciplinary Conference on Modeling and Using Context</source>
          , pages
          <fpage>599</fpage>
          -
          <lpage>608</lpage>
          . Springer,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          <source>[Honavar</source>
          , 1995]
          <string-name>
            <given-names>Vasant</given-names>
            <surname>Honavar</surname>
          </string-name>
          .
          <article-title>Symbolic artificial intelligence and numeric artificial neural networks: towards a resolution of the dichotomy</article-title>
          .
          <source>In Computational architectures integrating neural and symbolic processes</source>
          , pages
          <fpage>351</fpage>
          -
          <lpage>388</lpage>
          . Springer,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [Ietswaart et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Robert</given-names>
            <surname>Ietswaart</surname>
          </string-name>
          ,
          <string-name>
            <surname>Benjamin</surname>
            <given-names>M Gyori</given-names>
          </string-name>
          , John A Bachman,
          <article-title>Peter K Sorger,</article-title>
          and
          <string-name>
            <given-names>L Stirling</given-names>
            <surname>Churchman</surname>
          </string-name>
          .
          <article-title>Genewalk identifies relevant gene functions for a biological context using network representation learning</article-title>
          .
          <source>Genome biology</source>
          ,
          <volume>22</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>35</lpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          <source>[Kingma and Welling</source>
          , 2014]
          <article-title>Diederik P Kingma and Max Welling</article-title>
          .
          <article-title>Auto-encoding variational bayes</article-title>
          .
          <source>In The 2nd International Conference on Learning Representations</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          <string-name>
            <surname>[Le Saux</surname>
          </string-name>
          et al.,
          <year>1999</year>
          ]
          <string-name>
            <given-names>Elisabeth</given-names>
            <surname>Le</surname>
          </string-name>
          <string-name>
            <surname>Saux</surname>
          </string-name>
          , Philippe Lenca, Philippe Picouet, and
          <article-title>Jean-Pierre Barthe´lemy. An anthropocentric tool for decision making support</article-title>
          .
          <source>In The International Joint Conference on Artificial Intelligence</source>
          , volume
          <volume>16</volume>
          , pages
          <fpage>338</fpage>
          -
          <lpage>343</lpage>
          . Lawrence Erlbaum Associates LTD,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          <string-name>
            <surname>[Le Saux</surname>
          </string-name>
          et al.,
          <year>2002</year>
          ]
          <string-name>
            <given-names>Elisabeth</given-names>
            <surname>Le</surname>
          </string-name>
          <string-name>
            <surname>Saux</surname>
          </string-name>
          , Philippe Lenca, and
          <string-name>
            <given-names>Philippe</given-names>
            <surname>Picouet</surname>
          </string-name>
          .
          <article-title>Dynamic adaptation of rules bases under cognitive constraints</article-title>
          .
          <source>European Journal of Operational Research</source>
          ,
          <volume>136</volume>
          (
          <issue>2</issue>
          ):
          <fpage>299</fpage>
          -
          <lpage>309</lpage>
          ,
          <year>January 2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          <source>[Leavy</source>
          , 2018]
          <string-name>
            <given-names>Susan</given-names>
            <surname>Leavy</surname>
          </string-name>
          .
          <article-title>Gender bias in artificial intelligence: The need for diversity and gender theory in machine learning</article-title>
          .
          <source>In Proceedings of the 1st international workshop on gender equality in software engineering</source>
          , pages
          <fpage>14</fpage>
          -
          <lpage>16</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          [LeCun et al.,
          <year>2015</year>
          ] Yann LeCun, Yoshua Bengio, and
          <string-name>
            <given-names>Geoffrey</given-names>
            <surname>Hinton</surname>
          </string-name>
          .
          <article-title>Deep learning</article-title>
          .
          <source>Nature</source>
          ,
          <volume>521</volume>
          (
          <issue>7553</issue>
          ):
          <fpage>436</fpage>
          -
          <lpage>444</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          <string-name>
            <surname>[Lee</surname>
          </string-name>
          et al.,
          <year>2016</year>
          ]
          <string-name>
            <surname>Chan-Su</surname>
            <given-names>Lee</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Ahmed</given-names>
            <surname>Elgammal</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Marwan</given-names>
            <surname>Torki</surname>
          </string-name>
          .
          <article-title>Learning representations from multiple manifolds</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>50</volume>
          :
          <fpage>74</fpage>
          -
          <lpage>87</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          <source>[Lenca</source>
          , 2002]
          <string-name>
            <given-names>Philippe</given-names>
            <surname>Lenca</surname>
          </string-name>
          .
          <article-title>Human centered processes</article-title>
          .
          <source>European Journal of Operational Research</source>
          ,
          <volume>136</volume>
          (
          <issue>2</issue>
          ):
          <fpage>231</fpage>
          -
          <lpage>232</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          [Lepri et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Bruno</given-names>
            <surname>Lepri</surname>
          </string-name>
          , Nuria Oliver, Emmanuel Letouze´,
          <string-name>
            <given-names>Alex</given-names>
            <surname>Pentland</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Patrick</given-names>
            <surname>Vinck</surname>
          </string-name>
          . Fair, transparent, and
          <article-title>accountable algorithmic decision-making processes</article-title>
          .
          <source>Philosophy &amp; Technology</source>
          ,
          <volume>31</volume>
          (
          <issue>4</issue>
          ):
          <fpage>611</fpage>
          -
          <lpage>627</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          [Lesort et al.,
          <year>2017</year>
          ] Timothe´e Lesort, Mathieu Seurin,
          <string-name>
            <given-names>Xinrui</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Natalia D´ıaz-Rodr ´ıguez, and David Filliat</article-title>
          .
          <article-title>Unsupervised state representation learning with robotic priors: a robustness benchmark</article-title>
          .
          <source>arXiv preprint arXiv:1709.05185</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          [Lesort et al.,
          <year>2018</year>
          ] Timothe´e Lesort,
          <string-name>
            <surname>Natalia</surname>
            <given-names>D</given-names>
          </string-name>
          ´ıazRodr´ıguez, Jean-Franc¸
          <article-title>ois Goudou, and</article-title>
          <string-name>
            <given-names>David</given-names>
            <surname>Filliat</surname>
          </string-name>
          .
          <article-title>State Representation Learning for Control: An Overview</article-title>
          .
          <source>Neural Networks</source>
          ,
          <volume>108</volume>
          :
          <fpage>379</fpage>
          -
          <lpage>392</lpage>
          ,
          <year>December 2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref57">
        <mixed-citation>
          <string-name>
            <surname>[Li</surname>
          </string-name>
          et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Yingming</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Ming</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Zhongfei</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <article-title>A survey of multi-view representation learning</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          ,
          <volume>31</volume>
          (
          <issue>10</issue>
          ):
          <fpage>1863</fpage>
          -
          <lpage>1883</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref58">
        <mixed-citation>
          [Lim et al.,
          <year>2009</year>
          ] Brian Y Lim,
          <article-title>Anind K Dey,</article-title>
          and Daniel Avrahami.
          <article-title>Why and why not explanations improve the intelligibility of context-aware intelligent systems</article-title>
          .
          <source>In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems</source>
          , pages
          <fpage>2119</fpage>
          -
          <lpage>2128</lpage>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref59">
        <mixed-citation>
          <string-name>
            <surname>[Lin</surname>
          </string-name>
          et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Yankai</given-names>
            <surname>Lin</surname>
          </string-name>
          , Xu Han,
          <string-name>
            <given-names>Ruobing</given-names>
            <surname>Xie</surname>
          </string-name>
          , Zhiyuan Liu, and
          <string-name>
            <given-names>Maosong</given-names>
            <surname>Sun</surname>
          </string-name>
          .
          <article-title>Knowledge representation learning: A quantitative review</article-title>
          .
          <source>CoRR</source>
          , abs/
          <year>1812</year>
          .10901,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref60">
        <mixed-citation>
          [Liu et al.,
          <year>2020</year>
          ] Alexander H Liu,
          <article-title>Tao Tu, Hung-yi Lee, and Lin-shan Lee</article-title>
          .
          <article-title>Towards unsupervised speech recognition and synthesis with quantized speech representation learning</article-title>
          .
          <source>In The 45th IEEE International Conference on Acoustics, Speech and Signal Processing</source>
          , pages
          <fpage>7259</fpage>
          -
          <lpage>7263</lpage>
          . IEEE,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref61">
        <mixed-citation>
          <source>[Lundberg and Lee</source>
          , 2017]
          <article-title>Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions</article-title>
          .
          <source>In Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          , pages
          <fpage>4765</fpage>
          -
          <lpage>4774</lpage>
          . Curran Associates, Inc.,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref62">
        <mixed-citation>
          [Madumal et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Prashan</given-names>
            <surname>Madumal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Tim</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Liz</given-names>
            <surname>Sonenberg</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Frank</given-names>
            <surname>Vetere</surname>
          </string-name>
          .
          <article-title>Explainable reinforcement learning through a causal lens</article-title>
          .
          <source>In The 34th AAAI Conference on Artificial Intelligence</source>
          , pages
          <fpage>2493</fpage>
          -
          <lpage>2500</lpage>
          , Apr.
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref63">
        <mixed-citation>
          <source>[Malhotra and Nair</source>
          , 2015]
          <article-title>Meenakshi Malhotra and TR Gopalakrishnan Nair</article-title>
          .
          <article-title>Evolution of knowledge representation and retrieval techniques</article-title>
          .
          <source>International Journal of Intelligent Systems and Applications</source>
          ,
          <volume>7</volume>
          (
          <issue>7</issue>
          ):
          <fpage>18</fpage>
          -
          <lpage>28</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref64">
        <mixed-citation>
          <source>[McCulloch and Pitts</source>
          , 1943]
          <article-title>Warren S McCulloch</article-title>
          and
          <string-name>
            <given-names>Walter</given-names>
            <surname>Pitts</surname>
          </string-name>
          .
          <article-title>A logical calculus of the ideas immanent in nervous activity</article-title>
          .
          <source>The bulletin of mathematical biophysics</source>
          ,
          <volume>5</volume>
          (
          <issue>4</issue>
          ):
          <fpage>115</fpage>
          -
          <lpage>133</lpage>
          ,
          <year>1943</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref65">
        <mixed-citation>
          [Mikolov et al.,
          <year>2013</year>
          ]
          <article-title>Toma´s Mikolov, Kai Chen</article-title>
          , Greg Corrado, and
          <string-name>
            <given-names>Jeffrey</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <article-title>Efficient estimation of word representations in vector space</article-title>
          .
          <source>In Yoshua Bengio and Yann LeCun</source>
          , editors,
          <source>1st International Conference on Learning Representations, Scottsdale</source>
          , Arizona, USA, May 2-
          <issue>4</issue>
          ,
          <year>2013</year>
          , Workshop Track Proceedings,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref66">
        <mixed-citation>
          <source>[Miller</source>
          ,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Tim</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Explanation in artificial intelligence: Insights from the social sciences</article-title>
          .
          <source>Artificial intelligence</source>
          ,
          <volume>267</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref67">
        <mixed-citation>
          [Mueller et al.,
          <year>2021</year>
          ] Shane T Mueller, Elizabeth S Veinott, Robert R Hoffman, Gary Klein, Lamia Alam, Tauseef Mamun, and
          <string-name>
            <surname>William</surname>
          </string-name>
          J Clancey.
          <article-title>Principles of explanation in human-ai systems</article-title>
          .
          <source>arXiv preprint arXiv:2102.04972</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref68">
        <mixed-citation>
          <source>[Nelson, 2019] Gregory S Nelson. Bias in artificial intelligence</source>
          .
          <source>North Carolina medical journal</source>
          ,
          <volume>80</volume>
          (
          <issue>4</issue>
          ):
          <fpage>220</fpage>
          -
          <lpage>222</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref69">
        <mixed-citation>
          [Nguyen et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Thao</given-names>
            <surname>Nguyen</surname>
          </string-name>
          , Maithra Raghu, and
          <string-name>
            <given-names>Simon</given-names>
            <surname>Kornblith</surname>
          </string-name>
          .
          <article-title>Do wide and deep networks learn the same things? Uncovering how neural network representations vary with width and depth</article-title>
          .
          <source>In 9th International Conference on Learning Representations, Virtual Event, Austria, May 3-7</source>
          ,
          <year>2021</year>
          . OpenReview.net,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref70">
        <mixed-citation>
          <source>[Puiutta and Veith</source>
          , 2020]
          <string-name>
            <given-names>Erika</given-names>
            <surname>Puiutta and Eric</surname>
          </string-name>
          <string-name>
            <given-names>M. S. P.</given-names>
            <surname>Veith</surname>
          </string-name>
          .
          <article-title>Explainable reinforcement learning: A survey</article-title>
          . In Andreas Holzinger,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Kieseberg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A Min</given-names>
            <surname>Tjoa</surname>
          </string-name>
          , and Edgar R. Weippl, editors,
          <source>Machine Learning and Knowledge Extraction - 4th IFIP TC 5, TC 12, WG 8.4, WG 8</source>
          .9,
          <string-name>
            <surname>WG</surname>
          </string-name>
          <year>12</year>
          .9 International
          <string-name>
            <surname>Cross-Domain</surname>
            <given-names>Conference</given-names>
          </string-name>
          ,
          <string-name>
            <surname>CDMAKE</surname>
          </string-name>
          <year>2020</year>
          , Dublin, Ireland,
          <source>August 25-28</source>
          ,
          <year>2020</year>
          , Proceedings, volume
          <volume>12279</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>77</fpage>
          -
          <lpage>95</lpage>
          . Springer,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref71">
        <mixed-citation>
          [Qi et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Zhaobo</given-names>
            <surname>Qi</surname>
          </string-name>
          , Shuhui Wang,
          <string-name>
            <surname>Chi Su</surname>
            , Li Su,
            <given-names>Qingming</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            , and
            <given-names>Qi</given-names>
          </string-name>
          <string-name>
            <surname>Tian</surname>
          </string-name>
          .
          <article-title>Towards more explainability: Concept knowledge mining network for event recognition</article-title>
          .
          <source>In The 28th ACM International Conference on Multimedia, page 3857-3865</source>
          , New York, NY, USA,
          <year>2020</year>
          .
          <article-title>Association for Computing Machinery</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref72">
        <mixed-citation>
          [Ribeiro et al.,
          <year>2016</year>
          ]
          <article-title>Marco Tu´lio Ribeiro, Sameer Singh, and Carlos Guestrin. ”Why Should I Trust you?”: Explaining the predictions of any classifier</article-title>
          .
          <source>InThe 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          , pages
          <fpage>1135</fpage>
          -
          <lpage>1144</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref73">
        <mixed-citation>
          [Roberts et al., 2018a]
          <string-name>
            <given-names>Adam</given-names>
            <surname>Roberts</surname>
          </string-name>
          , Jesse Engel, Colin Raffel, Curtis Hawthorne, and
          <string-name>
            <given-names>Douglas</given-names>
            <surname>Eck</surname>
          </string-name>
          .
          <article-title>A hierarchical latent vector model for learning long-term structure in music</article-title>
          .
          <source>In The 35th International Conference on Machine Learning</source>
          , pages
          <fpage>4364</fpage>
          -
          <lpage>4373</lpage>
          . PMLR,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref74">
        <mixed-citation>
          [Roberts et al., 2018b]
          <string-name>
            <given-names>Adam</given-names>
            <surname>Roberts</surname>
          </string-name>
          , Jesse H Engel, Sageev Oore, and
          <string-name>
            <given-names>Douglas</given-names>
            <surname>Eck</surname>
          </string-name>
          .
          <article-title>Learning latent representations of music to generate interactive musical palettes</article-title>
          .
          <source>In IUI Workshops</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref75">
        <mixed-citation>
          [Rojat et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Rojat</surname>
          </string-name>
          , Raphae¨l Puget, David Filliat,
          <source>Javier Del Ser</source>
          ,
          <string-name>
            <given-names>Rodolphe</given-names>
            <surname>Gelin</surname>
          </string-name>
          , and
          <article-title>Natalia D´ıaz-Rodr ´ıguez. Explainable artificial intelligence (XAI) on Time Series Data: A Survey</article-title>
          .
          <source>arXiv preprint arXiv:2104.00950</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref76">
        <mixed-citation>
          [Rumelhart et al.,
          <year>1986</year>
          ]
          <string-name>
            <given-names>David E Rumelhart</given-names>
            , Geoffrey E Hinton, and
            <surname>Ronald J Williams</surname>
          </string-name>
          .
          <article-title>Learning representations by back-propagating errors</article-title>
          .
          <source>Nature</source>
          ,
          <volume>323</volume>
          (
          <issue>6088</issue>
          ):
          <fpage>533</fpage>
          -
          <lpage>536</lpage>
          ,
          <year>1986</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref77">
        <mixed-citation>
          [Sallinger et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Emanuel</given-names>
            <surname>Sallinger</surname>
          </string-name>
          , Sahar Vahdati, and Mojtaba Nayyeri.
          <source>Workshop knowledge representation &amp; representation learning</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref78">
        <mixed-citation>
          [Schlegel et al.,
          <year>2019</year>
          ]
          <string-name>
            <given-names>Udo</given-names>
            <surname>Schlegel</surname>
          </string-name>
          , Hiba Arnout, Mennatallah
          <string-name>
            <surname>El-Assady</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Oelke</surname>
            , and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Keim</surname>
          </string-name>
          .
          <article-title>Towards a rigorous evaluation of XAI methods on time series</article-title>
          .
          <source>The 2019 IEEE/CVF International Conference on Computer Vision Workshop</source>
          , pages
          <fpage>4197</fpage>
          -
          <lpage>4201</lpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref79">
        <mixed-citation>
          <source>[Shapiro</source>
          ,
          <year>2006</year>
          ] Stuart C Shapiro.
          <article-title>Knowledge representation</article-title>
          .
          <source>Encyclopedia of cognitive science</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref80">
        <mixed-citation>
          <source>[Shortliffe</source>
          , 1974]
          <string-name>
            <surname>Edward H Shortliffe</surname>
          </string-name>
          .
          <article-title>A rule-based computer program for advising physicians regarding antimicrobial therapy selection</article-title>
          .
          <source>In Proceedings of the 1974 annual ACM conference-Volume</source>
          <volume>2</volume>
          , pages
          <fpage>739</fpage>
          -
          <lpage>739</lpage>
          ,
          <year>1974</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref81">
        <mixed-citation>
          [Soleimani et al.,
          <year>2021</year>
          ]
          <string-name>
            <given-names>Melika</given-names>
            <surname>Soleimani</surname>
          </string-name>
          , Ali Intezari, Nazim Taskin, and David Pauleen.
          <article-title>Cognitive biases in developing biased Artificial Intelligence recruitment system</article-title>
          .
          <source>In Proceedings of the 54th Hawaii International Conference on System Sciences, page 5091</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref82">
        <mixed-citation>
          [Su et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>Ya</given-names>
            <surname>Su</surname>
          </string-name>
          , Yun Fu,
          <string-name>
            <given-names>Xinbo</given-names>
            <surname>Gao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Qi</given-names>
            <surname>Tian</surname>
          </string-name>
          .
          <article-title>Discriminant learning through multiple principal angles for visual recognition</article-title>
          .
          <source>IEEE transactions on image processing</source>
          ,
          <volume>21</volume>
          (
          <issue>3</issue>
          ):
          <fpage>1381</fpage>
          -
          <lpage>1390</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref83">
        <mixed-citation>
          [Sun et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Ke</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <surname>Lei</surname>
            <given-names>Wang</given-names>
          </string-name>
          , Bo Xu,
          <string-name>
            <given-names>Wenhong</given-names>
            <surname>Zhao</surname>
          </string-name>
          , Shyh Wei Teng, and
          <string-name>
            <given-names>Feng</given-names>
            <surname>Xia</surname>
          </string-name>
          .
          <article-title>Network representation learning: From traditional feature learning to deep learning</article-title>
          .
          <source>IEEE Access</source>
          ,
          <volume>8</volume>
          :
          <fpage>205600</fpage>
          -
          <lpage>205617</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref84">
        <mixed-citation>
          [Torki et al.,
          <year>2010</year>
          ]
          <string-name>
            <given-names>Marwan</given-names>
            <surname>Torki</surname>
          </string-name>
          , Ahmed Elgammal, and
          <string-name>
            <given-names>Chan Su</given-names>
            <surname>Lee</surname>
          </string-name>
          .
          <article-title>Learning a joint manifold representation from multiple data sets</article-title>
          .
          <source>In The 20th International Conference on Pattern Recognition</source>
          , pages
          <fpage>1068</fpage>
          -
          <lpage>1071</lpage>
          . IEEE,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref85">
        <mixed-citation>
          [Tu et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Cunchao</given-names>
            <surname>Tu</surname>
          </string-name>
          , Xiangkai Zeng, Hao Wang, Zhengyan Zhang, Zhiyuan Liu, Maosong Sun, Bo Zhang, and
          <string-name>
            <given-names>Leyu</given-names>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>A unified framework for community detection and network representation learning</article-title>
          .
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          ,
          <volume>31</volume>
          (
          <issue>6</issue>
          ):
          <fpage>1051</fpage>
          -
          <lpage>1065</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref86">
        <mixed-citation>
          <string-name>
            <surname>[Van Fraassen</surname>
          </string-name>
          ,
          <year>1988</year>
          ] Bas Van Fraassen.
          <article-title>The pragmatic theory of explanation</article-title>
          .
          <source>Theories of Explanation</source>
          ,
          <volume>8</volume>
          :
          <fpage>135</fpage>
          -
          <lpage>155</lpage>
          ,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref87">
        <mixed-citation>
          <source>[Vilone and Longo</source>
          , 2020]
          <string-name>
            <given-names>Giulia</given-names>
            <surname>Vilone</surname>
          </string-name>
          and
          <string-name>
            <given-names>Luca</given-names>
            <surname>Longo</surname>
          </string-name>
          .
          <article-title>Explainable artificial intelligence: a systematic review</article-title>
          .
          <source>arXiv preprint arXiv:2006.00093</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref88">
        <mixed-citation>
          <source>[Walton</source>
          , 2004]
          <string-name>
            <given-names>Douglas</given-names>
            <surname>Walton</surname>
          </string-name>
          .
          <article-title>A new dialectical theory of explanation</article-title>
          .
          <source>Philosophical Explorations</source>
          ,
          <volume>7</volume>
          (
          <issue>1</issue>
          ):
          <fpage>71</fpage>
          -
          <lpage>89</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref89">
        <mixed-citation>
          <string-name>
            <surname>[Wang</surname>
          </string-name>
          et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Jingdong</given-names>
            <surname>Wang</surname>
          </string-name>
          , Ke Sun, Tianheng Cheng, Borui Jiang, Chaorui Deng,
          <string-name>
            <surname>Yang</surname>
            <given-names>Zhao</given-names>
          </string-name>
          , Dong Liu, Yadong Mu, Mingkui Tan, Xinggang
          <string-name>
            <surname>Wang</surname>
          </string-name>
          , et al.
          <article-title>Deep high-resolution representation learning for visual recognition</article-title>
          .
          <source>IEEE transactions on pattern analysis and machine intelligence (accepted)</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref90">
        <mixed-citation>
          <source>[Wirth</source>
          , 2000]
          <article-title>Ru¨diger Wirth. Crisp-dm: Towards a standard process model for data mining</article-title>
          .
          <source>In The 4th International Conference on the Practical Application of Knowledge Discovery and Data Mining</source>
          , pages
          <fpage>29</fpage>
          -
          <lpage>39</lpage>
          ,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref91">
        <mixed-citation>
          <string-name>
            <surname>[Wu</surname>
          </string-name>
          et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>Lirong</given-names>
            <surname>Wu</surname>
          </string-name>
          , Zicheng Liu, Zelin Zang, Jun Xia,
          <string-name>
            <given-names>Siyuan</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>and Stan Z Li</article-title>
          .
          <article-title>Deep clustering and representation learning that preserves geometric structures</article-title>
          . arXiv e-prints, pages
          <fpage>arXiv</fpage>
          -
          <lpage>2009</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref92">
        <mixed-citation>
          [Xie et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Ruobing</given-names>
            <surname>Xie</surname>
          </string-name>
          , Zhiyuan Liu,
          <string-name>
            <given-names>Fen</given-names>
            <surname>Lin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Leyu</given-names>
            <surname>Lin</surname>
          </string-name>
          .
          <article-title>Does William Shakespeare REALLY Write Hamlet? Knowledge Representation Learning With Confidence</article-title>
          .
          <source>In Sheila A. McIlraith</source>
          and
          <string-name>
            <surname>Kilian Q</surname>
          </string-name>
          . Weinberger, editors,
          <source>Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence</source>
          , pages
          <fpage>4954</fpage>
          -
          <lpage>4961</lpage>
          . AAAI Press,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref93">
        <mixed-citation>
          [Xu et al.,
          <year>2013</year>
          ]
          <string-name>
            <given-names>Chang</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Dacheng</given-names>
            <surname>Tao</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Chao</given-names>
            <surname>Xu</surname>
          </string-name>
          .
          <article-title>A survey on multi-view learning</article-title>
          .
          <source>arXiv preprint arXiv:1304.5634</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref94">
        <mixed-citation>
          [Xu et al.,
          <year>2018</year>
          ]
          <string-name>
            <given-names>Cai</given-names>
            <surname>Xu</surname>
          </string-name>
          , Ziyu Guan,
          <string-name>
            <surname>Wei</surname>
            <given-names>Zhao</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Yunfei</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Quan</given-names>
            <surname>Wang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Zhiheng</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <article-title>Deep multi-view concept learning</article-title>
          .
          <source>In The 27th International Joint Conference on Artificial Intelligence</source>
          , pages
          <fpage>2898</fpage>
          -
          <lpage>2904</lpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref95">
        <mixed-citation>
          [Yang et al.,
          <year>2015</year>
          ] Cheng Yang, Zhiyuan Liu,
          <string-name>
            <given-names>Deli</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Maosong</given-names>
            <surname>Sun</surname>
          </string-name>
          , and Edward Y Chang.
          <article-title>Network representation learning with rich text information</article-title>
          .
          <source>In The 24th International Joint Conference on Artificial Intelligence</source>
          , volume
          <volume>2015</volume>
          , pages
          <fpage>2111</fpage>
          -
          <lpage>2117</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref96">
        <mixed-citation>
          [Yin et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Jun</given-names>
            <surname>Yin</surname>
          </string-name>
          , Xin Jiang, Zhengdong Lu, Lifeng Shang,
          <string-name>
            <given-names>Hang</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Xiaoming</given-names>
            <surname>Li</surname>
          </string-name>
          .
          <article-title>Neural generative question answering</article-title>
          .
          <source>In The 25th International Joint Conference on Artificial Intelligence</source>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref97">
        <mixed-citation>
          [Zhang et al.,
          <year>2011</year>
          ]
          <string-name>
            <given-names>Zhenyue</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Jing Wang, and
          <string-name>
            <given-names>Hongyuan</given-names>
            <surname>Zha</surname>
          </string-name>
          .
          <article-title>Adaptive manifold learning</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          ,
          <volume>34</volume>
          (
          <issue>2</issue>
          ):
          <fpage>253</fpage>
          -
          <lpage>265</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref98">
        <mixed-citation>
          [Zhang et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Fuzheng</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Nicholas Jing Yuan, Defu Lian, Xing Xie, and
          <string-name>
            <surname>Wei-Ying Ma</surname>
          </string-name>
          .
          <article-title>Collaborative knowledge base embedding for recommender systems</article-title>
          .
          <source>In The 22nd ACM SIGKDD international conference on knowledge discovery and data mining</source>
          , pages
          <fpage>353</fpage>
          -
          <lpage>362</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref99">
        <mixed-citation>
          [Zhang et al.,
          <year>2020</year>
          ]
          <string-name>
            <given-names>D.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          , and
          <string-name>
            <surname>C. Zhang.</surname>
          </string-name>
          <article-title>Network representation learning: A survey</article-title>
          .
          <source>IEEE Transactions on Big Data</source>
          ,
          <volume>6</volume>
          (
          <issue>01</issue>
          ):
          <fpage>3</fpage>
          -
          <lpage>28</lpage>
          , jan
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref100">
        <mixed-citation>
          [Zhong et al.,
          <year>2016</year>
          ]
          <string-name>
            <given-names>Guoqiang</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <surname>Li-Na</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiao Ling</surname>
            , and
            <given-names>Junyu</given-names>
          </string-name>
          <string-name>
            <surname>Dong</surname>
          </string-name>
          .
          <article-title>An overview on data representation learning: From traditional feature learning to recent deep learning</article-title>
          .
          <source>The Journal of Finance and Data Science</source>
          ,
          <volume>2</volume>
          (
          <issue>4</issue>
          ):
          <fpage>265</fpage>
          -
          <lpage>278</lpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref101">
        <mixed-citation>
          [Zhu et al.,
          <year>2014</year>
          ]
          <string-name>
            <given-names>Zhenfeng</given-names>
            <surname>Zhu</surname>
          </string-name>
          , Linlin Du, Lei Zhang, and
          <string-name>
            <given-names>Yao</given-names>
            <surname>Zhao</surname>
          </string-name>
          .
          <article-title>Shared subspace learning for latent representation of multi-view data</article-title>
          .
          <source>Journal of Information Hiding and Multimedia Signal Processing</source>
          ,
          <volume>5</volume>
          (
          <issue>3</issue>
          ):
          <fpage>546</fpage>
          -
          <lpage>554</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>