<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Interweaving Knowledge Representation and Adaptive Neural Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ilianna Kollia</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nikolaos Simou</string-name>
          <email>nsimou@image.ntua.gr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Giorgos Stamou</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas Stafylopatis</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Electrical and Computer Engineering, National Technical University of Athens</institution>
          ,
          <addr-line>Zographou 15780</addr-line>
          ,
          <country country="GR">Greece</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Both symbolic knowledge representation systems and machine learning techniques, including arti¯cial neural networks, play a signi¯cant role in Arti¯cial Intelligence. Interweaving these techniques, in order to achieve adaptation and robustness in classi¯cation problems is of great importance. In this paper we present a novel architecture that can provide e®ective connectionist adaptation of ontological knowledge. The proposed architecture can be used to improve performance of multimedia analysis and man machine interaction systems.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Intelligent systems based on symbolic knowledge processing, on the one hand,
and arti¯cial neural networks, on the other, di®er substantially. Nevertheless,
they are both standard approaches to arti¯cial intelligence and it is very
desirable to combine the robustness provided by neural networks, especially when
data are noisy, with the expressivity of symbolic knowledge representation. This
has contributed decisively to the growing interest in developing neural-symbolic
systems [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 6, 5</xref>
        ]. This integration can be realised by an incremental work°ow for
knowledge adaptation. Symbolic knowledge bases can be embedded into a
connectionist representation, where the knowledge can be adapted and enhanced
from raw data. This knowledge may in turn be extracted into symbolic form,
where it can be further used.
      </p>
      <p>In this paper we focus on developing connectionist adaptation of ontological
knowledge. Section 2 presents the proposed architecture that mainly consists
of the formal knowledge and the knowledge adaptation components, which are
described in sections 3 and 4 respectively. Conclusions and ongoing research
involving semantic multimedia analysis applications are reported in section 6.
Knowledge stores the knowledge base components that describe the problem
under analysis. More speci¯cally, the Ontologies module formally represents the
general knowledge about the problem (TBox) generated during the Development
Phase by knowledge engineers and experts.</p>
      <p>Moreover, the Formal Knowledge contains the World Description that is
actually a representation of all objects and individuals of the world, as well as
their properties and relationships in terms of the Ontology (ABox). It is evident
that most of the above data involve di®erent types of uncertain information and,
thus, they can be represented as formal (fuzzy) assertions connecting the objects
and individuals of the world with the concepts and relationships of the Ontology.
This operation is performed by the Semantic Interpretation module.</p>
      <p>In real environments however, this is a rather optimistic claim. Unfortunately,
there may be lot of reasons that cause inconsistencies in the Formal Knowledge.
For example, it is impossible to model all speci¯c environments and thus, in some
cases, con°icting assertions can arise. In such cases, the Knowledge Adaptation
component of the system tries to resolve the inconsistency through a recursive
learning process. Adaptation improves the knowledge of system by changing the
world description and to some degree the axioms of the terminology. The new
information as represented in a connectionist model and, with the aid of learning
algorithms, is adapted and then re-inserted in the knowledge base through the
Knowledge Extraction and Semantic Interpretation modules for system
adaptation.</p>
    </sec>
    <sec id="sec-2">
      <title>The Formal Knowledge Component</title>
      <p>
        The focus of the proposed system architecture in Figure 1 is on adaptation of the
knowledge base, so as to deal with contextual information and raw data
peculiarities obtained from multimodal inputs. In the paper we deal with interweaving
of formal knowledge representation and neural-symbolic integration. In
particular, we use recent results that extract parameter kernel functions for individuals
within ontologies [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">3, 2, 1</xref>
        ]. Exploitation of these kernels permits inducing
classi¯ers for individuals in Semantic Web (OWL) ontologies. In this paper, extraction
of kernel functions is the main outcome of the Formal Knowledge component
assisted by a reasoning engine - for feeding the connectionist-based Knowledge
Adaptation task.
      </p>
      <p>The family of kernel functions kpF : Ind(A)£Ind(A) ! [0; 1], for a knowledge
base K = hT; Ai consisting of the TBox T (set of terminological axioms of
concept descriptions) and the ABox A (assertions on the world state); Ind(A)
indicates the set of individuals appeared in A, and F = fF1; F2; : : : ; Fmg is a
set of concept descriptions. These functions are de¯ned as the Lp mean of the,
say m, simple concept kernel functions ·i , i = 1; : : : ; m, where, for every two
individuals a,b, and p &gt; 0,
·i(a; b) =
8 1 (Fi(a) 2 A ^ Fi(b) 2 A) _ (:Fi(a) 2 A ^ :Fi(b) 2 A)
&lt;</p>
      <p>0 (Fi(a) 2 A ^ :Fi(b) 2 A) _ (:Fi(a) 2 A ^ Fi(b) 2 A)
: 21 otherwise
8a; b 2 Ind(A)
kpF (a; b) :=</p>
      <p>i=1
· Xm ¯¯ ·i(a; b) p¯¸1=p</p>
      <p>¯
¯ m ¯
(1)
(2)</p>
      <p>The rationale of these kernels is that similarity between individuals is
determined by their similarity with respect to each concept Fi , i.e, if they both are
instances of the concept or of its negation. It has been shown that kpF is a valid
kernel, based on its composition through simpler matching kernels and on the
closure property with respect to sum, multiplication by a constant and kernel
multiplication.</p>
      <p>
        It should be stressed that the reasoning engine, included in Figure 1, is of
major importance for the whole procedure, because it assists the operation of all
knowledge related components. First, during the knowledge development phase,
it is responsible for enriching manual generation of concepts and relations. In
the operation phase, it interacts with the semantic interpretation layer and the
connectionist system for knowledge adaptation to local environments. Both crisp
and fuzzy reasoners can form this engine, we use the FIRE engine [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] that is
based on the fuzzy extension of the DL SHIN [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        We use fuzzy reasoning because a fuzzy assertional component permits more
detailed descriptions of a domain. In order to compute (1), (2) the greatest lower
bound (GLB) reasoning service of FiRE de¯ned in [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] is used, but the resulting
greatest lower bound is treated crisply. In other words, if GLB for Fi(a) &gt; 0,
then Fi(a) 2 A, while if GLB for Fi(a) = 0, then :Fi(a) 2 A. We intend to
incorporate the fuzzy element in the estimation of kernel functions using fuzzy
operations like fuzzy aggregation and fuzzy weighted norms for the evaluation
of the individuals.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>The Adaptation Mechanism</title>
      <p>
        In the proposed architecture of Figure 1, let us assume that the set of
individuals (and corresponding features), that have been used to generate the formal
knowledge, is provided, by the Semantic Interpretation Layer, to the
Knowledge Adaptation component. Support Vector Machines constitute a well known
method which can be based on kernel functions[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] to e±ciently induce classi¯ers
that work by mapping the instances into an embedding feature space, where
they can be discriminated by means of a linear classi¯er. Kernel Perceptrons
can be also applied to this linearly separable classi¯cation problem.
      </p>
      <p>
        Let the system be in its -real life- operation phase. Due to local or user
oriented characteristics, real life data can be quite di®erent from those of the
individuals used in the training phase; thus they may be not well represented by
the existing formal knowledge. Whenever a new individual is presented to the
system, it should be related, through the kernel function to each individual of
the knowledge base w.r.t to a speci¯c concept - category; the input data domain
is, thus, transformed to another domain - taking into account the semantics that
have been inserted to the kernel function. However, the kernel function in (1), (2)
is not continuous w.r.t individuals. Consequently, the values of the kernel
functions when relating a new individual to any existing one should be computed.
To cope with this problem, it is assumed that the semantic relations, that are
expressed through the above kernel functions, also hold for the syntactic
relations of the individuals, as expressed by their corresponding low level features,
estimated and presented at the system input. Under this assumption, a feature
based matching criterion using a k-means algorithm, is used to relate the new
individual to each one of the cluster centers w.r.t the low level feature vector.
The SVM or Kernel Perceptron can be retrained - including the new individuals
in the training data set, while getting the corresponding desired responses by
the User or by the Semantic Interpretation Layer - adapting its knowledge to
the speci¯c context and use. An hierarchical, multilayer kernel perceptron, the
input layer of which is identical to the trained kernel perceptron can also be used
[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        Knowledge extraction from trained neural networks has been a topic of
extensive research [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. One way is to transfer the - most characteristic - new
individuals obtained in the local environment, together with the corresponding
desired outputs - concepts, to the knowledge development module of the main
system (Figure 1), so that with the assistance of the reasoning engine, the system
formal knowledge, i.e., both the T-Box and the A-Box, be updated, w.r.t the
speci¯c context or user. A methodology that can be used to adapt a knowledge
base K = hT; Ai for a concept Fi, is the following. Check all related concepts,
denoted as RFi F1 : : : RFi Fi under the speci¯c context, count the number jRFi Fij
of occurrences of RFi Fi 2 A, as well as the axioms de¯ned for the concept Fi
in the knowledge base (i.e. Axiom(Fi) 2 T ). Let RFi Fi 2 Axiom(Fi) when the
concept RFi Fi is used in Axiom(Fi) and RFi Fi 62 Axiom(Fi) when it is not used.
The related concepts with the highest occurrence are selected for adaptation of
the terminology, while the insigni¯cant ones are removed.
5
      </p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>
        In this paper we presented a novel architecture for connectionist adaptation
of ontological knowledge. We are currently performing experimentation of the
system performance for solving an image/video segmentation and classi¯cation
problem [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ]. Future work, includes the incorporation of fuzzy set theory in
the kernel evaluation. Additionally, we intend to further examine the adaptation
of the knowledge base using the connectionist architecture, mainly focusing on
the selection of the appropriate DL constructors and on inconsistency handling.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>S.</given-names>
            <surname>Bloehdorn</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sure</surname>
          </string-name>
          .
          <article-title>Kernel methods for mining instance data in ontologies</article-title>
          .
          <source>In Proceedings of the 6th International Semantic Web Conference (ISWC)</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d
          <string-name>
            <surname>Amato</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Randomised metric induction and evolutionary conceptual clustering for semantic knowledge bases</article-title>
          .
          <source>In Proceedings of CIKM 07</source>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>N.</given-names>
            <surname>Fanizzi</surname>
          </string-name>
          , C. d
          <string-name>
            <surname>Amato</surname>
            , and
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Esposito</surname>
          </string-name>
          .
          <article-title>Statistical learning for inductive query answering on owl ontologies</article-title>
          .
          <source>In Proceedings of the 7th International Semantic Web Conference (ISWC)</source>
          , pages
          <fpage>195</fpage>
          {
          <fpage>212</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Artur</surname>
            <given-names>S. Avila</given-names>
          </string-name>
          <string-name>
            <surname>Garcez</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Broda</surname>
            , and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Gabbay</surname>
          </string-name>
          .
          <article-title>Symbolic knowledge extraction from trained neural networks: A sound approach</article-title>
          .
          <source>Arti¯cial Intelligence</source>
          ,
          <volume>125</volume>
          :
          <fpage>155</fpage>
          {
          <fpage>207</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Barbara</given-names>
            <surname>Hammer</surname>
          </string-name>
          and Pascal Hitzler, editors.
          <source>Perspectives of Neural-Symbolic Integration</source>
          , volume
          <volume>77</volume>
          <source>of Studies in Computational Intelligence</source>
          . Springer,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>P.</given-names>
            <surname>Hitzler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Holldobler</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Seda</surname>
          </string-name>
          .
          <article-title>Logic programs and connectionist networks</article-title>
          .
          <source>Journal of Applied Logic, page 245272</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>I.</given-names>
            <surname>Horrocks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Sattler</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Tobies</surname>
          </string-name>
          .
          <article-title>Reasoning with Individuals for the Description Logic SHIQ</article-title>
          . In David MacAllester, editor,
          <source>CADE-2000</source>
          ,
          <article-title>number 1831 in LNAI</article-title>
          , pages
          <volume>482</volume>
          {
          <fpage>496</fpage>
          . Springer-Verlag,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>E.</given-names>
            <surname>Kolman</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Margaliot</surname>
          </string-name>
          .
          <article-title>Are arti¯cial neural networks white boxes</article-title>
          ?
          <source>IEEE Trans. on Neural Networks</source>
          ,
          <volume>16</volume>
          (
          <issue>4</issue>
          ):
          <volume>844</volume>
          {
          <fpage>852</fpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>N.</given-names>
            <surname>Simou</surname>
          </string-name>
          , Th. Athanasiadis,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kollias</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Stamou, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Stafylopatis</surname>
          </string-name>
          .
          <article-title>Semantic adaptation of neural network classi¯ers in image segmentation</article-title>
          .
          <source>18th ICANN</source>
          <year>2008</year>
          , pp.
          <fpage>907</fpage>
          -
          <lpage>916</lpage>
          ,
          <year>September 2008</year>
          , Prague, Czech Republic.,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10. G. Stamou and
          <string-name>
            <given-names>S.</given-names>
            <surname>Kollias</surname>
          </string-name>
          .
          <article-title>Multimedia Content and the Semantic Web: Methods, Standards and Tools</article-title>
          . John Wiley &amp; Sons Ltd,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Giorgos</surname>
            <given-names>Stoilos</given-names>
          </string-name>
          , Nikos Simou, Giorgos Stamou, and
          <string-name>
            <given-names>Stefanos</given-names>
            <surname>Kollias</surname>
          </string-name>
          .
          <article-title>Uncertainty and the semantic web</article-title>
          .
          <source>IEEE Intelligent Systems</source>
          ,
          <volume>21</volume>
          (
          <issue>5</issue>
          ):
          <volume>84</volume>
          {
          <fpage>87</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <given-names>U.</given-names>
            <surname>Straccia</surname>
          </string-name>
          .
          <article-title>Reasoning within fuzzy description logics</article-title>
          .
          <source>Journal of Arti¯cial Intelligence Research</source>
          ,
          <volume>14</volume>
          :
          <fpage>137</fpage>
          {
          <fpage>166</fpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>