<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Explaining Trained Neural Networks with Semantic Web Technologies: First Steps</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Md Kamruzzaman Sarker</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ning Xie</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Derek Doran</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Raymer</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pascal Hitzler</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Data Science and Security Cluster, Wright State University</institution>
          ,
          <addr-line>Dayton, OH</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The ever increasing prevalence of publicly available structured data on the World Wide Web enables new applications in a variety of domains. In this paper, we provide a conceptual approach that leverages such data in order to explain the input-output behavior of trained arti cial neural networks. We apply existing Semantic Web technologies in order to provide an experimental proof of concept.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Trained neural networks are usually imagined as black boxes, in that they do
not give any direct indications why an output (e.g., a prediction) was made by
the network. The reason for this lies in the distributed nature of the information
encoded in the weighted connections of the network. Of course, for applications,
e.g., safety-critical ones, this is an unsatisfactory situation. Methods are therefore
sought to explain how the output of trained neural networks are reached.</p>
      <p>
        This topic of explaining trained neural networks is not a new one, in fact there
is already quite a bit of tradition and literature on the topic of rule extraction
from such networks (see, e.g., [
        <xref ref-type="bibr" rid="ref16 ref2 ref9">2,9,16</xref>
        ]), which pursued very similar goals. Rule
extraction, however, utilized propositional rules as target logic for generating
explanations, and as such remained very limited in terms of explanations which
are human-understandable. Novel deep learning architectures attempt to retrieve
explanations as well, but often the use-case is only for computer vision tasks
like object or scene recognition. Moreover, explanations in this context actually
encode greater details about the images provided as input, rather than explaining
why or how the neural network was able to recognize a particular object or scene.
      </p>
      <p>
        Semantic Web [
        <xref ref-type="bibr" rid="ref12 ref4">4,12</xref>
        ] is concerned with data sharing, discovery, integration,
and reuse. As eld, it does not only target data on the World Wide Web, but its
methods are also applicable to knowledge management and other tasks o the
Web. Central to the eld is the use of knowledge graphs (usually expressed using
the W3C standard Resource Description Framework RDF [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]) and type logics
attached to these graphs, which are called ontologies and are usually expressed
using the W3C standard Web Ontology Language OWL [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        This paper introduces a new paradigm for explaining neural network
behavior. It goes beyond the limited propositional paradigm, and directly targets
the problem of explaining neural network activity rather than the qualities of
Copyright © 2017 for this paper by its authors. Copying permitted for private and academic purposes.
the input. The paradigm leverages advances in knowledge representation on the
World Wide Web, more precisely from the eld of Semantic Web technologies. It
in particular utilizes the fact that methods, tool, and structured data in the
mentioned formats are now widely available, and that the amount of such structured
data on the Web is in fact constantly growing [
        <xref ref-type="bibr" rid="ref18 ref5">5,18</xref>
        ]. Prominent examples of
large-scale datasets include Wikidata [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] and data coming from the schema.org
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] e ort which is driven by major Web search engine providers. We will utilize
this available data as background knowledge, on the hypothesis that background
knowledge will make it possible to obtain more concise explanations. This
addresses the issue in propositional rule extraction that extracted rulesets are often
large and complex, and due to their sizes di cult to understand for humans.
      </p>
      <p>While the paper only attempts to explain input-output behavior, the authors
are actively exploring ways to also explain internal node activations.</p>
      <p>
        An illustrative example
Let us consider the following very simple example which is taken from [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>Assume that the input-output mapping P of the neural network without
background knowledge could be extracted as
p1 ^ q ! r</p>
      <p>p2 ^ q ! r:
Now assume furthermore that we also have background knowledge K in form of
the rules
p1 ! p</p>
      <p>p2 ! p:
The background knowledge then makes it possible to obtain the simpli ed
inputoutput mapping PK , as</p>
      <p>p ^ q ! r:</p>
      <p>The simpli cation through the background knowledge is caused by p acting
as a \generalization" of both p1 and p2. For the rest of the paper it may be
bene cial to think of p, p1 and p2 as classes or concepts, which are hierarchically
related, e.g., p1 being \oak," p2 being \maple," and p being \tree."</p>
      <p>Yet this example is con ned to propositional logic.1 In the following, we show
how we can bring structured (non-propositional) Semantic Web background
knowledge to bear on the problem of explanation generation for trained
neural networks, and how we can utilize Semantic Web technologies in order to
generate non-propositional explanations. This work is at a very early stage, i.e.,
we will only present the conceptual architecture of the approach and minimal
experimental results which are encouraging for continuing the e ort.</p>
      <p>
        The rest of the paper is structured as follows. In Section 2 we introduce
notation as needed, in particular regarding description logics which underly the OWL
standard, and brie y introduce the DL-Learner tool which features prominently
in our approach. In Section 3 we present the conceptual and experimental setup
1 How to go beyond the propositional paradigm in neural-symbolic integration is one
of the major challenges in the eld [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
for our approach, and report on some rst experiments. In Section 4 we conclude
and discuss avenues for future work.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Preliminaries</title>
      <p>
        We describe a minimum of preliminary notions and information needed in order
to keep this paper relatively self-contained. Description logics [
        <xref ref-type="bibr" rid="ref1 ref12">1,12</xref>
        ] are a major
paradigm in knowledge representation as a sub eld of arti cial intelligence. At
the same time, they play a very prominent role in the Semantic Web eld since
they are the foundation for one of the central Semantic Web standards, namely
the W3C Web Ontology Language OWL [
        <xref ref-type="bibr" rid="ref11 ref12">11,12</xref>
        ].
      </p>
      <p>Technically speaking, a description logic is a decidable fragment of rst-order
predicate logic (sometimes with equality or other extensions) using only unary
and binary predicates. The unary predicates are called atomic classes,2 while the
binary ones are refered to as roles,3 and constants are refered to as individuals.
In the following, we formally de ne the fundamental description logic known as
ALC, which will su ce for this paper. OWL is a proper superset of ALC.</p>
      <p>Desciption logics allow for a simpli ed syntax (compared to rst-order
predicate logic), and we will introduce ALC in this simpli ed syntax. A translation
into rst-order predicate logic will be provided further below.</p>
      <p>Let C be a nite set of atomic classes, R be a nite set of roles, and N be a
nite set of individuals. Then class expressions (or simply, classes ) are de ned
recursively using the following grammar, where A denotes atomic classes from
A and R denotes roles from R. The symbols u and t denote conjunction and
disjunction, respectively.</p>
      <p>C; D ::= A j :C j C u D j C t D j 8R:C j 9R:C</p>
      <p>A TBox is a set of statements, called (general class inclusion) axioms, of
the form C v D, where C and D are class expressions { the symbol v can be
understood as a type of subset inclusion, or alternatively, as a logical implication.
An ABox is a set of statements of the forms A(a) or R(a; b), where A is an
atomic class, R is a role, and a; b are individuals. A description logic knowledge
base consists of a TBox and an ABox. The notion of ontology is used in di erent
ways in the literature; sometimes it is used as equivalent to TBox, sometimes as
equivalent to knowledge base. We will adopt the latter usage.</p>
      <p>We characterize the semantics of ALC knowledge bases by giving a
translation into rst-order predicate logic. If is a TBox axiom of the form C v D,
then ( ) is de ned inductively as in Figure 1, where A is a class name. ABox
axioms remain unchanged.</p>
      <p>
        DL-Learner [
        <xref ref-type="bibr" rid="ref17 ref6">6,17</xref>
        ] is a machine learning system inspired by inductive logic
programming [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Given a knowledge base and two sets of individuals from the
knowledge base { called positive respectively negative examples { DL-Learner
2 or atomic concepts
3 or properties
(C v D) = (8x0)( x0 (C) ! x0 (D))
      </p>
      <p>
        xi (A) = A(xi)
xi (:C) = : xi (C)
xi (C u D) = xi (C) ^ xi (D)
xi (C t D) = xi (C) _ xi (D)
xi (8R:C) = (8xi+1)(R(xi; xi+1) ! xi+1 (C))
xi (9R:C) = (9xi+1)(R(xi; xi+1) ^ xi+1 (C))
attempts to construct class expressions such that all the positive examples are
contained in each of the class expressions, while none of the negative examples
is. DL-Learner gives preference to shorter solutions, and in the standard setting
returns approximate solutions if no fully correct solution is found. The inner
workings of DL-Learner will not matter for this paper, and we refer to [
        <xref ref-type="bibr" rid="ref17 ref6">6,17</xref>
        ] for
details. However, we exemplify its functionality by looking at Michalski's trains
as an example, which is a symbolic machine learning task from [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], and which
was presented also in [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>
        For purposes of illustrating DL-Learner, Figure 2 shows two sets of trains, the
positive examples are on the left, the negative ones are on the right. Following
[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], we use a simple encoding of the trains as a knowledge base: Each train is an
individual, and has cars attached to it using the hasCar property, and each car
then falls into di erent categories, e.g., the top leftmost car would fall into the
classes Open, Rectangular and Short, and would also have information attached
to it regarding symbol carried (in this case, square), and how many of them (in
this case, one). Given these examples and knowledge base, DL-Learner comes
up with the class
      </p>
      <sec id="sec-2-1">
        <title>9hasCar:(Closed u Short)</title>
        <p>which indeed is a simple class expression such that all positive examples fall
under it, while no negative example does.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Approach and Experiments</title>
      <p>In this paper, we follow the lead of the propositional rule extraction work
mentioned in the introduction, with the intent of improving on it in several ways.
1. We generalize the approach by going signi cantly beyond the propositional
rule paradigm, by utilizing description logics.
2. We include signi cantly sized and publicly available background knowledge
in our approach in order to arrive at explanations which are more concise.</p>
      <p>More concretely, we use DL-Learner as the key tool to arrive at the
explanations. Figure 3 depicts our conceptual architecture: The trained arti cial
neural network (connectionist system) acts as a classi er. Its inputs are mapped
to a background knowledge base and according to the networks' classi cation,
positive and negative examples are distinguished. DL-Learner is then run on
the example sets and provides explanations for the classi cations based on the
background knowledge.</p>
      <p>In the following, we report on preliminary experiments we have conducted
using our approach. Their sole purpose is to provide rst and very preliminary
insights into the feasibility of the proposed method. All experimental data is
available from http://daselab.org/projects/human-centered-big-data.</p>
      <p>
        We utilize the ADE20K dataset [
        <xref ref-type="bibr" rid="ref23 ref24">23,24</xref>
        ]. It contains 20,000 images of scenes
which have been pre-classi ed regarding scenes depicted, i.e., we assume that the
classi cation is done by a trained neural network.4 For our initial test, we used
six images, three of which have been classi ed as \outdoor warehouse" scenes
(our positive examples), and three of which have not been classi ed as such (our
negative examples). In fact, for simplicity, we took the negative examples from
among the images which had been classi ed as \indoor warehouse" scenes. The
images are shown in Figure 4.
      </p>
      <p>The ADE20K dataset furthermore provides annotations for each image which
identify information about objects which have been identi ed in the image. The
annotations are in fact richer than that and also talk about the number of
objects, whether they are occluded, and some more, but for our initial experiment
we only used presence or absence of an object. To keep the initial experiment
simple, we furthermore only used those detected objects which could easily be
mapped to our chosen background knowledge, the Suggested Upper Merged
Ontology (SUMO).5 Table 1 shows, for each image, the objects we kept. The
Suggested Upper Merged Ontology was chosen because it contains many, namely
about 25,000 common terms which cover a wide range of domains. At the same
4 Strictly speaking, this is not true for the training subset of the ADE20K dataset,
but that doesn't really matter for our demonstration.
5 http://www.adampease.org/OP/
image p1: road, window, door, wheel, sidewalk, truck, box, building
image p2: tree, road, window, timber, building, lumber
image p3: hand, sidewalk, clock, steps, door, face, building, window, road
image n1: shelf, ceiling, oor
image n2: box, oor, wall, ceiling, product
image n3: ceiling, wall, shelf, oor, product
time, the ontology arguably structures the terms in a relatively straightforward
manner which seemed to simplify matters for our initial experiment.</p>
      <p>In order to connect the annotations to SUMO, we used a single role called
\contains." Each image was made an individual in the knowledge base.
Furthermore, for each of the object identifying terms in Table 1, we either identi ed
a corresponding matching SUMO class, or created one and added it to SUMO
by inserting it at an appropriate place within SUMO's class hierarchy. We
furthermore created individuals for each of the object identifying terms, including
duplicates, in Table 1, and added them to the knowledge base by typing them
with the corresponding class. Finally, we related each image individual to each
corresponding object individual via the \contains" role.</p>
      <p>To exemplify { for the image p1 we added individuals road1, window1,
door1, wheel1, sidewalk1, truck1, box1, building1, declared Road(road1),
Window(window1), etc., and nally added the ABox statements contains(p1; road1),
contains(p1; window1), etc., to the knowledge base. For the image p2, we added
contains(p2; tree2), contains(p2; road2), etc. as well as the corresponding type
declarations Tree(tree2), Road(road2), etc.</p>
      <p>The mapping of the image annotations to SUMO is of course very simple,
and this was done deliberately in order to show that a straightforward approach
already yields interesting results. As our work progresses, we do of course
anticipate that we will utilize more complex knowledge bases and will need to
generate more complex mappings from picture annotations (or features) to the
background knowledge.</p>
      <p>Finally, we ran DL-Learner on the knowledge base, with the positive and
negative examples as indicated. DL-Learner returns 10 solutions, which are listed
in Figure 5. Of these, some are straightforward from the image annotations,
such as (1), (5), (8, (9) and (10). Others, such as (2), (4), (6), (7) are much
more interesting as they provide solutions in terms of the background knowledge
without using any of the terms from the original annotation. Solution (3) looks
odd at rst sight, but is meaningful in the context of the SUMO ontology:
SelfConnectedObject is an abstract class which is a direct child of the class
Object in SUMO's class hierarchy. Its natural language de nition is given as
\A SelfConnectedObject is any Object that does not consist of two or more
disconnected parts." As such, the class is a superclass of the class Road, which
explains why (3) is indeed a solution in terms of the SUMO ontology.
(6)
(7)
(8)
(9)
(10)</p>
      <p>We have conducted four additional experiments along the same lines as
described above. We brie y describe them below { the full raw data and results
are available from http://daselab.org/projects/human-centered-big-data.</p>
      <p>In the second experiment, we chose four workroom pictures as positive
examples, and eight warehouse pictures (indoors and outdoors) as negative examples.
An example explanation DL-Learner came up with is</p>
      <sec id="sec-3-1">
        <title>9contains:(DurableGood u :ForestProduct):</title>
        <p>On of the outdoor warehouse pictures indeed shows timber. DurableGoods in
SUMO include furniture, machinery, and appliances.</p>
        <p>In the third experiment, we chose the same four workroom pictures as
negative examples, and the same eight warehouse pictures (indoors and outdoors)
as positve examples. An example explanation DL-Learner came up with is</p>
      </sec>
      <sec id="sec-3-2">
        <title>8contains:(:Furniture u :IndustrialSupply);</title>
        <p>i.e., \contains neither furniture nor industrial supply". IndustrialSupply in SUMO
includes machinery. Indeed it turns out that furniture alone is insu cient for
distingushing between the positive and negative exaples, because \shelf" is not
classi ed as funiture in SUMO. This shows the dependency of the explanations
on the conceptualizations encoded in the background knowledge.</p>
        <p>In the fourth experiment, we chose eight market pictures (indoors and
outdoors) as positive examples, and eight warehouse pictures (indoors and outdoors)
as well as four workroom pictures as negative examples. An example explanation
DL-Learner came up with is
And indeed it turns out that people are shown on all the market pictures. There
is actually also a man shown on one of the warehouse pictures, driving a forklift,
however \man" or \person" was not among the annotations used for the picture.
This example indicates how our approach could be utilized: A human monitor
inquiring with an interactive system about the reasons for a certain classi cation
may notice that the man was missed by the software on that particular picture,
and can opt to interfere with the decision and attempt to correct it.</p>
        <p>In the fth experiment, we chose four mountain pictures as positive examples,
and eight warehouse pictures (indoors and outdoors) as well as four workroom
pictures as negative examples. An example explanation DL-Learner came up
with is</p>
      </sec>
      <sec id="sec-3-3">
        <title>9contains:BodyOfWater:</title>
        <p>Indeed, it turns out that all mountain pictures in the example set show either a
river or a lake. Similar to the previous example, a human monitor may be able
to catch that some misclassi cations may occur because presence of a body of
water is not always indicative of presence of a mountain.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions and Further Work</title>
      <p>We have laid out a conceptual sketch how to approach the issue of explaining
arti cial neural networks' classi cation behaviour using Semantic Web
background knowledge and technologies, in a non-propositional setting. We have
also reported on some very preliminary experiments to support our concepts.</p>
      <p>
        The sketch already indicates where to go from here: We will need to
incorporate more complex and more comprehensive background knowledge, and
if readily available structured knowledge turns out to be insu cient, then we
foresee using state of the art knowledge graph generation and ontology
learning methods [
        <xref ref-type="bibr" rid="ref13 ref19">13,19</xref>
        ] to obtain suitable background knowledge. We will need to
use automatic methods for mapping network input features to the background
knowledge [
        <xref ref-type="bibr" rid="ref21 ref7">7,21</xref>
        ], while the features to be mapped may have to be generated
from the input in the rst place, e.g. using object recognition software in the
case of images. And nally, we also intend to apply the approach to sets of hidden
neurons in order to understand what their activations indicate.
      </p>
      <p>Acknowledgements. This work was supported by the Ohio Federal Research
Network project Human-Centered Big Data.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patel-Schneider</surname>
            ,
            <given-names>P.F</given-names>
          </string-name>
          . (eds.):
          <article-title>The Description Logic Handbook: Theory, Implementation, and Applications</article-title>
          . Cambridge University Press, 2nd edn. (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bader</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Dimensions of neural-symbolic integration { A structured survey</article-title>
          . In: Artemov,
          <string-name>
            <given-names>S.N.</given-names>
            ,
            <surname>Barringer</surname>
          </string-name>
          , H.,
          <string-name>
            <surname>d'Avila Garcez</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lamb</surname>
            ,
            <given-names>L.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Woods</surname>
            ,
            <given-names>J</given-names>
          </string-name>
          . (eds.)
          <article-title>We Will Show Them! Essays in Honour of Dov Gabbay</article-title>
          , Volume One. pp.
          <volume>167</volume>
          {
          <fpage>194</fpage>
          .
          <string-name>
            <surname>College Publications</surname>
          </string-name>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Beckett</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berners-Lee</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prud</surname>
          </string-name>
          'hommeaux, E.,
          <string-name>
            <surname>Carothers</surname>
          </string-name>
          , G.:
          <article-title>RDF 1.1. Turtle { Terse RDF Triple Language</article-title>
          .
          <source>W3C Recommendation (25 February</source>
          <year>2014</year>
          ), available at http://www.w3.org/TR/turtle/
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Berners-Lee</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lassila</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>The Semantic Web</article-title>
          .
          <source>Scienti c American</source>
          <volume>284</volume>
          (
          <issue>5</issue>
          ),
          <volume>34</volume>
          {43 (May
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bizer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Heath</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berners-Lee</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Linked Data { The Story So Far</article-title>
          .
          <source>International Journal on Semantic Web and Information Systems</source>
          <volume>5</volume>
          (
          <issue>3</issue>
          ),
          <volume>1</volume>
          {
          <fpage>22</fpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. Buhmann, L.,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Westphal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <string-name>
            <surname>DL-Learner</surname>
          </string-name>
          {
          <article-title>A framework for inductive learning on the semantic web</article-title>
          .
          <source>Journal of Web Semantics</source>
          <volume>39</volume>
          ,
          <issue>15</issue>
          {
          <fpage>24</fpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Euzenat</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shvaiko</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          : Ontology Matching,
          <source>Second Edition</source>
          . Springer (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Garcez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Besold</surname>
          </string-name>
          , T.,
          <string-name>
            <surname>de Raedt</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , Foldiak,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Hitzler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Icard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            , Kuhnberger, K.U.,
            <surname>Lamb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            ,
            <surname>Miikkulainen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Silver</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          :
          <article-title>Neural-symbolic learning and reasoning: Contributions and challenges</article-title>
          . In: Gabrilovich,
          <string-name>
            <given-names>E.</given-names>
            ,
            <surname>Guha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>McCallum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Murphy</surname>
          </string-name>
          ,
          <string-name>
            <surname>K</surname>
          </string-name>
          . (eds.)
          <source>Proceedings of the AAAI 2015 Spring Symposium on Knowledge Representation and Reasoning: Integrating Symbolic and Neural Approaches. Technical Report</source>
          , vol. SS-
          <volume>15</volume>
          -
          <fpage>03</fpage>
          . AAAI Press, Palo Alto, CA (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>d'Avila Garcez</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaverucha</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>The connectionist inductive lerarning and logic programming system</article-title>
          .
          <source>Applied Intelligence</source>
          <volume>11</volume>
          (
          <issue>1</issue>
          ),
          <volume>59</volume>
          {
          <fpage>77</fpage>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Guha</surname>
            ,
            <given-names>R.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brickley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macbeth</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Schema.
          <article-title>org: evolution of structured data on the web</article-title>
          .
          <source>Commun. ACM</source>
          <volume>59</volume>
          (
          <issue>2</issue>
          ),
          <volume>44</volume>
          {
          <fpage>51</fpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Krotzsch,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Parsia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            ,
            <surname>Patel-Schneider</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.F.</given-names>
            ,
            <surname>Rudolph</surname>
          </string-name>
          , S. (eds.):
          <article-title>OWL 2 Web Ontology Language Primer (Second Edition)</article-title>
          .
          <source>W3C Recommendation (11 December</source>
          <year>2012</year>
          ), http://www.w3.org/TR/owl2-primer/
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , Krotzsch,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Rudolph</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.</surname>
          </string-name>
          :
          <article-title>Foundations of Semantic Web Technologies</article-title>
          . CRC Press/Chapman &amp; Hall (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Ji</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grishman</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>Knowledge base population: Successful approaches and challenges</article-title>
          . In: Lin,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Matsumoto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . (eds.)
          <article-title>The 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies</article-title>
          ,
          <source>Proceedings of the Conference</source>
          ,
          <volume>19</volume>
          -
          <fpage>24</fpage>
          June,
          <year>2011</year>
          , Portland, Oregon, USA. pp.
          <volume>1148</volume>
          {
          <fpage>1158</fpage>
          .
          <article-title>The Association for Computer Linguistics (</article-title>
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Labaf</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Evans</surname>
            ,
            <given-names>A.B.</given-names>
          </string-name>
          :
          <article-title>Propositional rule extraction from neural networks under background knowledge</article-title>
          .
          <source>In: Proceedings of the Twelfth International Workshop on Neural-Symbolic Learning and Reasoning</source>
          , NeSy'
          <fpage>17</fpage>
          , London, UK,
          <year>July 2017</year>
          (
          <year>2017</year>
          ), to appear
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Larson</surname>
          </string-name>
          , J., Michalski, R.S.:
          <article-title>Inductive inference of VL decision rules</article-title>
          .
          <source>SIGART Newsletter</source>
          <volume>63</volume>
          ,
          <issue>38</issue>
          {
          <fpage>44</fpage>
          (
          <year>1977</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bader</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Extracting reduced logic programs from articial neural networks</article-title>
          .
          <source>Applied Intelligence</source>
          <volume>32</volume>
          (
          <issue>3</issue>
          ),
          <volume>249</volume>
          {
          <fpage>266</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hitzler</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Concept learning in description logics using re nement operators</article-title>
          .
          <source>Machine Learning</source>
          <volume>78</volume>
          (
          <issue>1-2</issue>
          ),
          <volume>203</volume>
          {
          <fpage>250</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Isele</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jakob</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jentzsch</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kontokostas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mendes</surname>
            ,
            <given-names>P.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hellmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Morsey</surname>
            , M., van Kleef,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Auer</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bizer</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>DBpedia { A largescale, multilingual knowledge base extracted from Wikipedia</article-title>
          .
          <source>Semantic Web</source>
          <volume>6</volume>
          (
          <issue>2</issue>
          ),
          <volume>167</volume>
          {
          <fpage>195</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , Volker, J.:
          <source>Perspectives on Ontology Learning, Studies on the Semantic Web</source>
          , vol.
          <volume>18</volume>
          . IOS Press (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Muggleton</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Raedt</surname>
            ,
            <given-names>L.D.:</given-names>
          </string-name>
          <article-title>Inductive logic programming: Theory and methods</article-title>
          .
          <source>Journal of Logic Programming</source>
          <volume>19</volume>
          /20, 629{
          <fpage>679</fpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Uren</surname>
            ,
            <given-names>V.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cimiano</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iria</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Handschuh</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vargas-Vera</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Motta</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ciravegna</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Semantic annotation for knowledge management: Requirements and a survey of the state of the art</article-title>
          .
          <source>J. Web Sem</source>
          .
          <volume>4</volume>
          (
          <issue>1</issue>
          ),
          <volume>14</volume>
          {
          <fpage>28</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Vrandecic</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , Krotzsch, M.:
          <article-title>Wikidata: a free collaborative knowledgebase</article-title>
          .
          <source>Commun. ACM</source>
          <volume>57</volume>
          (
          <issue>10</issue>
          ),
          <volume>78</volume>
          {
          <fpage>85</fpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Puig</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fidler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barriuso</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Torralba</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Semantic understanding of scenes through the ADE20K dataset</article-title>
          .
          <source>arXiv preprint arXiv:1608.05442</source>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Puig</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fidler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barriuso</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Torralba</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Scene parsing through ADE20K dataset</article-title>
          .
          <source>In: Proceedings of the IEEE Conference on Computer Vision</source>
          and Pattern
          <string-name>
            <surname>Recognition</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>