<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marco Schorlemmer</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IIIA - Artificial Intelligence Research Institute CSIC - Spanish National Research Council Bellaterra (Barcelona)</institution>
          ,
          <addr-line>Catalonia</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Ontology matching is currently a key technology to achieve the semantic alignment of ontological entities used by knowledge-based applications, and therefore to enable their interoperability in distributed environments such as multi-agent systems. Most ontology matching mechanisms, however, assume matching prior integration and rely on semantics that has been coded a priori in concept hierarchies or external sources. We present a formal model for a semantic alignment procedure that incrementally aligns differing conceptualisations of two or more agents relative to their respective perception of the environment or domain they are acting in. It hence makes the situation in which the alignment occurs explicit in the model. We resort to Channel Theory to carry out the formalisation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        An ontology is commonly defined as a specification of the conceptualisation of a particular domain. It
fixes the vocabulary used by knowledge engineers to denote concepts and their relations, and it constrains
the interpretation of this vocabulary to the meaning originally intended by knowledge engineers. As such,
ontologies have been widely adopted also in multi-agent systems. Ontology matching is commonly defined
as the process that computes the semantic relationships between entities of separate ontologies. With the
proliferation of ontologies used in multi-agent systems caused by different conceptualisations of even the
same domain —and their subsequent specification using varying terminology— ontology matching is now
seen as a necessary technology to achieve semantic alignment and to enable interoperability of separately
engineered knowledge-based agents [
        <xref ref-type="bibr" rid="ref11 ref5">5, 11</xref>
        ].
      </p>
      <p>
        Until recently, most ontology matching mechanisms developed so far have taken a classical functional
approach to the semantic heterogeneity problem, in which ontology matching is seen as a process taking
two or more ontologies as input and producing a semantic alignment of ontological entities as output [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Furthermore, matching often has been carried out at design-time, before integrating knowledge-based agents
or making them interoperate. This might have been successful for clearly delimited and stable domains but
is untenable and even undesirable for the kind of applications that are currently deployed on the Web.
Webservice composition, multi-agent communication and peer-to-peer information sharing are all of a highly
distributed, dynamic and open-ended nature, and they require ontology matching to be performed during
run-time. In addition, one can imagine many situations in which ontologies are not even open for inspection
(e.g., when they are based on commercially confidential information).
      </p>
      <p>
        Certainly, there exist efforts to efficiently match ontological entities at run-time, taking only those
ontology fragment that are necessary for the task at hand [
        <xref ref-type="bibr" rid="ref10 ref13 ref8 ref9">10, 13, 9, 8</xref>
        ]. However, the techniques used by these
systems to establish the semantic relationships between ontological entities, even though applied at run-time,
they still exploit a priori defined concept taxonomies as they are represented in the graph-based structures of
the ontologies to be matched, use previously existing external sources such as thesauri (e.g. WordNet) and
upper-level ontologies [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], or resort to additional background knowledge repositories or shared instances.
      </p>
      <p>We claim that semantic alignment is ultimately relative to the particular situation in which it is carried
out and that this situation should be made explicit and brought into the ontology matching mechanism.
Even two systems with identical conceptualisation capabilities and using exactly the same vocabulary to
specify their respective conceptualisations may fail to interoperate because of their differing perception of
the domain. Imagine a situation in which two agents are facing each other on a checker board. Agent A1
may conceptualise a figure on the board as situated on the “left”, while agent A2 may conceptualise the
same figure as situated on the “right”. Although the conceptualisation of “left” and “right” is done in exactly
in the same manner by both agents, they still will need to align their respective vocabularies if they want
to successfully communicate to each other actions that move figures on the checker board. Their semantic
alignment, however, will only be valid in the scope of their communication within this particular situation
or environment. The same agents situated differently may lead to a different alignment.</p>
      <p>
        This scenario is reminiscent to those in which a group of distributed agents adapt to form an ontology
and a shared lexicon in an emergent, bottom-up manner, with only local interactions and no central control
authority [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. This sort of self-organised emergence of shared meaning is namely ultimately grounded on
the physical interaction of agents with the environment. In this paper, however, we address the case in which
agents are already endowed with a top-down engineered ontology (it can even be the same one), which they
do not adapt or refine, but for which they want to find the semantic relationships with separate ontologies of
other agents on the grounds of their communication within a specific situation. In particular, we provide a
formal model that formalises situated semantic alignment as a sequence of information-channel refinements
in the sense of Barwise and Seligman’s theory of information flow [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. This theory is particularly useful
for our endeavour because it models the flow of information occurring in distributed systems due to the
particular situations —or tokens— that carry information.1 Analogously, the semantic alignment that will
allow information to flow ultimately will be carried by the particular situation agents are acting in. We
shall therefore consider a scenario with two or more agents situated in an environment. Each agent will
have its own viewpoint of the environment so that, if the environment is in a concrete state, both agents
may have different perceptions of this state. Because of these differences there may be a mismatch in
the meaning of the syntactic entities by which agents describe their perceptions (and which constitute the
agents’ respective ontologies). We state that these syntactic entities can be related according to the intrinsic
semantics provided by the existing relationship between the agents’ viewpoint of the environment. The
existence of this relationship is precisely justified by the fact that the agents are situated and observe the
same environment.
2
2.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Semantic Alignment in an Environment</title>
      <sec id="sec-2-1">
        <title>The Information Channel linking Agent Perspectives</title>
        <p>
          Let us consider two agents A1 and A2 situated in an environment E.2 We associate a numerable set S of
states to E and, at any given instant, we suppose E to be in one of these states. We further assume that each
agent is able to observe the environment and has its own perception of it. This ability is faithfully captured
by a surjective function seei : S → Pi, where i ∈ {1, 2}. According to Channel Theory, information is only
viable where there is a systematic way of classifying some range of things as being this way or that. This
idea is captured in the notion of classification:
Definition 1 ([
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]) A classification is a tuple A = htok(A), typ(A), |=Ai where tok(A) is a set of tokens,
typ(A) is a set of types and |=A is a binary relation between tok(A) and typ(A). If a |=A α then a is said
to be of type α.
        </p>
        <p>For each agent Ai, i ∈ {1, 2}, we consider a classification Ai that models Ai’s viewpoint of E. First,
tok(Ai) is composed of Ai’s perceptions of E states, that is, tok(Ai) = Pi. Second, typ(Ai) comprises the
syntactic entities by which Ai describes its perceptions, the ones constituting the ontology of Ai. Finally,
|=Ai synthesizes how Ai relates its perceptions with these syntactic entities.</p>
        <p>Now, in order to associate the environment E with a classification E we choose the power classification
of S as E. which is the classification whose set of types is 2S , whose tokens are the elements of S, and
for which a is of type α if a ∈ α. The reason for taking the power classification is because, in general,
there is no global conceptualisation of the environment and hence there are no sytactic entities that may play
the role of types for E. However, the set of types of the power classification includes all possible token
configurations potentially described by types. Thus tok(E) = S, typ(E) = 2S and e |=E ε iff e ∈ ε.</p>
        <p>1We do not assume any knowledge of Channel Theory; we restate basic definitions and theorems in this paper, but any detailed
exposition of the theory is outside the scope of this paper.</p>
        <p>2The generalization to any numerable set of agents is straightforward.</p>
        <p>
          In Channel Theory, the information flow between the components of a distributed system is modelled
in terms of a channel. The relationships between the components are expressed via infomorphisms which
provide a way of moving information between them:
Definition 2 ([
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]) An infomorphism f : A → B from A to B is a contravariant pair of functions f =
hfˆ, fˇi, where fˆ : typ(A) → typ(B) and fˇ : tok(B) → tok(A), satisfying the following fundamental
property:
        </p>
        <p>fˇ(b) |=A α iff b |=B fˆ(α)
for each token b ∈ tok(B) and each type α ∈ typ(A).3
We omit the superscripts of informorphisms if no confusion arises. Types are usually denoted by greek
letters and tokens by latin letters so f (α) ≡ fˆ(α) and f (a) ≡ fˇ(a).</p>
        <p>
          Definition 3 ([
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]) A channel C consists of two infomorphisms {fi : Ai → C}i∈{1,2} with a common
codomain C, called the core of C. The tokens of C are called connections and a connection c is said to
connect the tokens fˇ1(c) and fˇ2(c).4
        </p>
        <p>typ(C)
sssfˆs1sssss9 eKKKKKfKˆ2KKK
typ(A1) |=C typ(A2)
|=A1</p>
        <p>yss
tok(A1)</p>
        <p>tok(C) |=A2
sssfsˇ1sss KKKfˇK2KKKKK%
tok(A2)
The information flow of our scenario is accurately described by the channel E = {fi : Ai → E}i∈{1,2}
defined as follows:
• fˆi(α) = {e ∈ tok(E) | seei(e) |=Ai α} for each α ∈ typ(E),
• fˇi(e) = seei(e) for each e ∈ tok(E),
where i ∈ {1, 2}. The fundamental property of the infomorphisms is fulfilled indeed:
fˇi(e) |=Ai α
seei(e) |=Ai α
e ∈ fˆi(α)
e |=E fˆi(α)
{by definition of fˇi}
{by definition of fˆi}
{by definition of |=E}
In this way, E is the core of the channel E and a state e ∈ tok(E) connects the agents’ perceptions fˇ1(e) and
fˇ2(e).</p>
        <p>A1</p>
        <p>E
}f}1}}}}}}&gt; `AAAAAAfA2A</p>
        <p>
          A2
Example. This example is taken from [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] and we will resort to it throughout the paper. There
are two observers, M r.1 and M r.2, each having a partial viewpoint of a box. This box consists
of six sectors and each sector can enclose a ball. The box is “magic” because observers are
not able to distinguish the depth inside it. Figure 1 shows the scenario we are describing and
illustrates schematically what M r.1 and M r.2 can observe.
        </p>
        <p>3Strictly speaking, an infomorphism f comprehends on the one hand a pair of classifications hA, Bi and on the other hand a
contravariant pair of functions hfˆ, fˇi defined as above.</p>
        <p>4Actually, this is the definition of a binary channel. A channel can be defined with an arbitrary index set as well as a distributed
system can have an arbitrary number of components.</p>
        <p>The magic box
Mr.2's viewpoint
With respect to our approach, the magic box will play the role of E and M r.1 and M r.2 will
be the agents. The states of E are the possible configurations of the box so a state e can be
represented as a 3 × 2 binary matrix (eij ). In this way:</p>
        <p>S =
(
1 0 !
Intuitively, the state 1 0 , for instance, means that there are only two balls in the box,
0 0
one in the left up sector and another in the left centered sector. Notice that M r.1 sees a ball on
the left if there is a ball in at least one of the sectors of the left column of the box and sees a ball
on the right provided that there is ball in at least one of the sectors of the right column. Then
M r.1’s perceptions of the states of E can be represented as two-dimensional binary vectors,
and the function see1 can be defined formally as follows:
see1(e) =

 (0 0)


</p>
        <p>(1 0)
 (0 1)


 (1 1)
if
if
if
if
e =
where e ∈ S. The function see2 can be defined analogously on three-dimensional binary vectors
as representations of M r.2’s perceptions. Note that M r.1 has the notions of a ball being on the
left or on the right and so has M r.2, but this last one has also the notion of a ball being in the
center. After these considerations, A1 is defined by:
•
•
• tok(A1) = (1 1), (1 0), (0 1), (0 0)
• typ(A1) = {left, right}
(v1 v2) |=A1 left iff
(v1 v2) |=A1 right iff
• typ(A2) = {left, center, right}
(v1 v2 v3) |=A2 left
(v1 v2 v3) |=A2 center
(v1 v2 v3) |=A2 right
• tok(A2) = (1 1 1), (1 1 0), (1 0 1), (0 1 1), (1 0 0), (0 1 0), (0 0 1), (0 0 0)
Finally, a channel E = {fi : Ai → E}i∈{1,2} as defined above explains the information flow of
this scenario.
In the previous subsection, we have shown how channel E explains the information flow of our scenario.
Now, we want to relate the agents’ syntactic entities, i.e., the agents’ types, and obtain meaningful
relationships justified by E .</p>
        <p>In Channel Theory, the sum operation gives us a way of putting the two components of a channel
C = {fi : Ai → C}i∈{1,2} together into a single classification A1 + A2 (see Definition 4), and the
infomorphisms together into a single infomorphism f = f1 + f2 (see Proposition 1).</p>
        <p>
          Definition 4 ([
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]) Let A1 and A2 be two classifications. The sum of A1 and A2 denoted by A1 + A2 is
the classification with tok(A1 + A2) = tok(A1) × tok(A2) = {ha, bi | a ∈ tok(A1) and b ∈ tok(A2)},
typ(A1 + A2) = typ(A1) ] typ(A2) = {hi, γi | i = 1 and γ ∈ typ(A1) or i = 2 and γ ∈ typ(A2)} and
relation |=A1+A2 defined by:
ha, bi |=A1+A2 h1, αi
ha, bi |=A1+A2 h2, βi
if
if
a |=A1 α
b |=A2 β
There exist two infomorphisms σ1 : A1 → A1 + A2 and σ2 : A2 → A1 + A2 defined in a natural
way: on types, σ1(α) = h1, αi and σ2(β) = h2, βi; on tokens, σ1(ha, bi) = a and σ2(ha, bi) = b. It is
straightforward that σ1 and σ2 are infomorphisms.
        </p>
        <p>
          Proposition 1 ([
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]) Given a channel C = {fi : Ai →
f = f1 + f2 such that the following diagram commutes5:
        </p>
        <sec id="sec-2-1-1">
          <title>C}i∈{1,2}, there exists a unique informorphism</title>
          <p>
            uuufu1uuuu
uu: CO dIII
f
PROOF: We just give the definition of f (see [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ] for the rest of the proof): on types, f (h1, αi) = f1(α) and
f (h2, βi) = f2(β); on tokens, f (c) = hf1(c), f2(c)i.
          </p>
          <p>Note that typ(A1 + A2) is the disjoint union of typ(A1) and typ(A2). We attach importance to take the
disjoint union because A1 and A2 could use some identical types in order to describe their perceptions of E
(for instance, in the Magic Box example, left and right are both M r.1’s types and M r.2’s types). Therefore,
A1 + A2 is the place where we should look for relationships between agents’ types. Moreover, as we will
see f = f1 + f2 is of use to find meaningful relationships justified by E .</p>
          <p>
            Channel Theory provides a way to make these relationships explicit in a logical fashion by means of
“theories” and “local logics”:
Definition 5 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) Given a set Σ, a sequent of Σ is a pair hΓ, Δi of subsets of Σ. A binary relation ` between
subsets of Σ is called a consequence relation on Σ. A theory T is a pair hΣ, `i where ` is a consequence
relation on Σ. A sequent hΓ, Δi of Σ for which Γ ` Δ is called a constraint of the theory T .
          </p>
          <p>When it comes to writing constraints of a theory T = hΣ, `i, we will use the usual notational
conventions. For example, we write α ` β, γ instead of {α} ` {β, γ} and Γ, Γ0 ` Δ, α instead of Γ∪Γ0 ` Δ∪{α}.</p>
          <p>
            A theory is regular if it satisfies the properties Identity, Weakening and Global Cut (see [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]).
Regularity refers to the purely structural properties that any theory must satisfy. From here on, all the theories
considered are regular. Every classification generates a theory in a natural way:
Definition 6 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) Let A be a classification. A token a ∈ tok(A) satisfies a sequent hΓ, Δi of typ(A)
provided that if a is of every type in Γ then it is of some type in Δ. The theory generated by A, written
T h(A), is the theory htyp(A), `Ai where Γ `A Δ if every token in A satisfies hΓ, Δi.
          </p>
          <p>The notion of local logic puts a classification and a theory together but with an important added twist.
In order to model reasonable but unsound inferences, the notion of‘ “normal token” is introduced.</p>
          <p>5That is, fi = f ◦ σi, i.e., fˆi = fˆ◦ σˆi and fˇi = σˇi ◦ fˇ, for i ∈ {1, 2}.</p>
          <p>
            Definition 7 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) A local logic L is a tuple htok(L), typ(L), |=L, `L, NLi where:
1. htok(L), typ(L), |=Li is a classification denoted by Cla(L),
2. htyp(L), `Li is a regular theory denoted by T h(L),
3. NL is a subset of tok(L), called the normal tokens of L, which satisfy all the constraints of T h(L).
Definition 8 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) A local logic L is sound if every token in Cla(L) is normal, that is, NL = tok(L). L is
complete if every sequent of typ(L) satisfied by every normal token is a constraint of T h(L).
Just like with theories, a classification generates a local logic in a natural way:
Definition 9 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) Given a classification A, the local logic generated by A, written Log(A), is the local
logic on A, i.e., Cla(Log(A)) = A, such that T h(Log(A)) = T h(A) and all its tokens are normal, i.e.,
NLog(A) = tok(A).
          </p>
          <p>
            The local logic generated by a classification is sound and complete. Furthermore, given a classification A
and a local logic L on A, L is sound and complete if and only if L = Log(A) (see [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]). The main notion of
Channel Theory is the notion of “distributed logic”:
Definition 10 ([
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]) Let f : A → B be an infomorphism. Given a local logic L on B, the inverse image of
L under f , denoted f −1[L], is the local logic on A with Γ `f−1[L] Δ if f (Γ) `L f (Δ) and Nf−1[L] = {a ∈
tok(A) | a = fˇ(b) for some b ∈ NL}. Given a channel C = {fi : Ai → C}i∈{1,2} and a local logic L on
C, the distributed logic of C generated by L, written DLogC(L), is the logic f −1[L] where f = f1 + f2.
DLogC(L) represents the reasoning about relations among the components of the channel C justified by the
logic L on the core. Consider L = Log(C). We denote DLogC(Log(C)) by Log(C). Log(C) then captures
in a logical fashion the information flow inherent in the channel.
          </p>
          <p>Now, let us return to our scenario modeled by E . Log(E ) explains the relationships between the agents’
viewpoints of the environment. On the one hand, T h(Log(E )) contains the most meaningful relationships
between agents’ types according to the channel E . On the other hand, NLog(E) contains those pairs of agents’
perceptions connected together by the environment state of which they come from.</p>
          <p>uuufu1uuuu
uu: EO dIII</p>
          <p>
            f
A1
Log(E ) is complete since Log(E) is complete but it is not necessarily sound because although Log(E) is
sound, f = f1 + f2 is not token surjective in general (see [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]):
          </p>
          <p>NLog(E) = {hfˇ1(e), fˇ2(e)i | e ∈ tok(E)} ⊆ tok(A1 + A2)
If Log(E ) is also sound then Log(E ) = Log(A1 + A2), as we comment above. It means that there is
no meaningful compatibility between agents’ viewpoints of the environment according to E . It is just the
fact that Log(E ) is unsound what allows a meaningful compatibility between the agents’ viewpoints. This
compatibility relation is expressed at the type level in terms of constraints by T h(Log(E )) and at the token
level in terms of pairs of tokens by NLog(E).</p>
          <p>Example. Following the previous example, Log(E ) includes among its constraints:
h1, lefti `Log(E) h2, lefti, h2, centeri, h2, righti
h1, righti `Log(E) h2, lefti, h2, centeri, h2, righti
Let us check the first constraint (the second one can be verified in a similar way). We have to
prove:</p>
          <p>fˆ(h1, lefti) `Log(E) fˆ(h2, lefti), fˆ(h2, centeri), fˆ(h2, righti)
Let e = (eij ) ∈ tok(E) and assume that e |=E fˆ(h1, lefti). Then e ∈ fˆ(h1, lefti), that
is, see1(e) |=A1 left. Therefore e11 + e21 + e31 ≥ 1 and we can assure that there exists
i ∈ {1, 2, 3} such that ei1 = 1. Hence ei1 + ei2 ≥ 1. If i = 3 then e |=E fˆ(h2, lefti), if i = 2
then e |=E fˆ(h2, centeri) and if i = 1 then e |=E fˆ(h2, righti).</p>
          <p>The necessary and sufficient conditions for a token ha1, a2i of A1 + A2 being normal are the
following:
if a1 6= (0 0) then a2 6= (0 0 0)
if a2 6= (0 0 0) then a1 6= (0 0)
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Aligning Agent Perspectives through Communication</title>
      <p>We have stated that T h(Log(E )) contains the most meaningful relations between agents’ types. The problem
is that neither agent can make use of this theory because they do not know the channel E completely. We
present a method by which agents obtain approximations to T h(Log(E )). These approximations gradually
become more reliable as the method is applied.</p>
      <p>A1 and A2 communicate by exchanging information about their perceptions of the environment states.
This information is expressed in terms of their own classification relations. Specifically, if E is in a
concrete state e, we assume that agents can convey to each other which types are satisfied by their respective
perceptions of e and which are not. This exchange generates a channel C = {fi : Ai → C}i∈{1,2}.
Furthermore, T h(C) provides the logical relationships between the agents’ types justified by the fact that
agents have observed e. If E turns to another state e0 and agents proceed as before, another channel
C0 = {fi0 : Ai → C0}i∈{1,2} gives account of the new situation. The significant point is that C0 refines
C (see Definition 11). Theorem 1 ensures that the refined channel involves more reliable information.</p>
      <p>The communication ends when agents have observed all the environment states. Again this situation
can be modeled by a channel, call it C∗ = {fi∗ : Ai → C∗}i∈{1,2}. Theorem 2 states that T h(C∗) =
T h(Log(E )).</p>
      <p>
        Definition 11 ([
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]) Let C = {fi : Ai → C}i∈{1,2} and C0 = {fi0 : Ai → C0}i∈{1,2} be two information
channels with the same component classifications A1 and A2. A refinement infomorphism r from C0 to C is
an infomorphism r : C0 → C such that for each i ∈ {1, 2}, fi = r ◦ fi0 (i.e., fˆi = rˆ ◦ fˆi0 and fˇi = fˇi0 ◦ rˇ).
The channel C0 is a refinement of C if there exists a refinement infomorphism r from C0 to C.
Theorem 1 asserts that the more refined channel gives more reliable information. Although its theory has
less constraints, it has more normal tokens to which they apply.
      </p>
      <p>Theorem 1 Let C = {fi : Ai → C}i∈{1,2} and C0 = {fi0 : Ai → C0}i∈{1,2} be two channels. If C0 is a
refinement of C then:
1. T h(Log(C0)) ⊆ T h(Log(C))
PROOF: Since C0 is a refinement of C then there exists a refinement infomorphism r from C0 to C. Let us
write A = A1 + A2, f = f1 + f2 and f 0 = f10 + f20 .</p>
      <p>1. Let Γ and Δ be subsets of typ(A) and assume that Γ `Log(C0) Δ, which means fˆ0[Γ] `C0 fˆ0[Δ]. We
have to prove Γ `Log(C) Δ, or equivalently, fˆ[Γ] `C fˆ[Δ]. We proceed by reductio ad absurdum.
Suppose c ∈ tok(C) does not satisfy the sequent hfˆ[Γ], fˆ[Δ]i. Then c |=C fˆ(γ) for all γ ∈ Γ and
c 6|=C fˆ(δ) for all δ ∈ Δ. Let us choose an arbitrary γ ∈ Γ. We have that γ = hi, αi for some
α ∈ typ(Ai) and i ∈ {1, 2}. Thus fˆ(γ) = fˆ(hi, αi) = fˆi(α) = rˆ ◦ fˆi0(α) = rˆ(fˆi0(α)). Therefore:
c |=C fˆ(γ)
c |=C rˆ(fˆi0(α))
rˇ(c) |=D fˆi0(α)
rˇ(c) |=D fˆ0(hi, αi)
rˇ(c) |=D fˆ0(γ)
Consequently, rˇ(c) |=C fˆ0(γ) for all γ ∈ Γ. Since fˆ0[Γ] `C0 fˆ0[Δ] then there exists δ∗ ∈ Δ such
that rˇ(c) |=C0 fˆ0(δ∗). A sequence of equivalences similar to the above one justifies c |=C fˆ(δ∗),
contradicting that c is a counterexample to hfˆ[Γ], fˆ[Δ]i. Hence Γ `Log(C) Δ as we wanted to prove.
2. Let ha1, a2i ∈ tok(A) and assume ha1, a2i ∈ NLog(C). Therefore, there exists c token in C such
that ha1, a2i = fˇ(c). Then we have ai = fˇi(c) = fi0 ◦ rˇ(c) = fˇi0(rˇ(c)), for i ∈ {1, 2}. Hence
ˇ
ha1, a2i = fˇ0(rˇ(c)) and ha1, a2i ∈ NLog(C0). Consequently, NLog(C0) ⊇ NLog(C) which concludes
the proof.</p>
      <sec id="sec-3-1">
        <title>Refining the Semantic Alignment</title>
        <p>We assume that typ(Ai) is finite for i ∈ {1, 2} and S is infinite numerable, though the finite case can be
treated in a similar form. We also choose an infinite numerable set of symbols {cn | n ∈ N}6.</p>
        <p>Agent communication starts from the observation of E. Let us suppose that E is in state e1 ∈ S =
tok(E). A1’s perception of e1 is f1(e1) and A2’s perception of e1 is f2(e1). We take for granted that A1
can communicate A2 those properties that are and are not satisfied by f1(e1) according to its classification
A1. So can A2 do. Since both typ(A1) and typ(A2) are finite, this process eventually finishes. After this
communication a channel C1 = {fi1 : Ai → C1}i=1,2 arises.</p>
        <p>A1
{f{1{1{{{{{= C1 aCCCCCfC2C1C</p>
        <p>A2
On the one hand, C1 is defined by: tok(C1) = {c1}, typ(C1) = typ(A1 + A2) and c1 |=C1 hi, αi if
fi(e1) |=Ai α, for α ∈ typ(Ai) and i ∈ {1, 2}. On the other hand, fi1(α) = hi, αi and fi1(c1) = fi(e1),
for α ∈ typ(Ai) and i ∈ {1, 2}. Clearly, f11 and f21 are infomorphims.</p>
        <p>Note in this case both agents know C1. Hence they can compute separately T h(C1) = htyp(C1), `C1 i.
This theory represents the logical relationships between the agents’ types justified by the fact that agents
have observed e1.</p>
        <p>Now, let us assume that E turns to a new state e2. Agents can proceed as above exchanging this time
information about their perceptions of e2. Therefore, another channel C2 = {fi2 : Ai → C2}i∈{1,2} comes
up. First of all, C2 is defined by: tok(C2) = {c1, c2}, typ(C2) = typ(A1 + A2) and for 1 ≤ k ≤ 2,
ck |=C2 hi, αi if fi(ek) |=Ai α, for all α ∈ typ(Ai) and i ∈ {1, 2}. Second, fi2(α) = hi, αi and
fi2(ck) = fi(ek) again for 1 ≤ k ≤ 2, α ∈ typ(Ai) and i ∈ {1, 2}. f12 and f22 are infomorphims by
definition. The theory T h(C2) = htyp(C2), `C2 i represents the logical relationships between the agents’
types justified by the fact that agents have observed e1 and e2. The significant point is that C2 is a refinement
of C1:</p>
        <p>A1</p>
        <p>= C2
|f1|2 ||||
| |
f11 / C1 o</p>
        <p>aBB
f
1 B f2</p>
        <p>BB 2B</p>
        <p>B B
f21</p>
        <p>A2
It is easy to prove that f 1 defined as the identity function on types and the inclusion function on tokens is a
refinement infomorphism. By Theorem 1, C2 constraints are more reliable than C1 constraints.
6We write these symbols with superindexes because we limit the use of subindexes for what concerns to agents.</p>
        <p>Example. Let us assume that the environment changes to the state e2 =
perception of e2 is (1 1) and A2’s perception of e2 is (1 1 0). In this case:
h1, lefti `C2 h2, centeri
but
h1, lefti 0C2 h2, righti</p>
        <p>In the general situation, once the states e1, e2, . . . , en have been observed and a new state en+1 appears,
the channel Cn+1 = {fin+1 : Ai → Cn+1}i∈{1,2} informs about agents communication up to this moment.
First, tok(Cn+1) = {c1, c2, . . . , cn+1}, typ(Cn+1) = typ(A1 + A2) and for each 1 ≤ k ≤ n + 1,
ck |=Cn+1 hi, αi if fi(ek) |=Ai α, for all α ∈ typ(Ai) and i ∈ {1, 2}. Second, fin+1(α) = hi, αi and
fin+1(ck) = fˇi(ek) for 1 ≤ k ≤ n + 1, α ∈ typ(Ai) and i ∈ {1, 2}. Clearly, f1n+1 and f2n+1 are
infomorphims. The theory T h(Cn+1) = htyp(Cn+1), `Cn+1 i represents the logical relationships between
the agents’ types justified by the fact that agents have observed e1, . . . , en, en+1. Also Cn+1 is a refinement
of Cn. f n defined as the identity function on types and the inclusion function on tokens is a refinement
infomorphism. By Theorem 1, Cn+1 constraints are more reliable than Cn constraints.</p>
        <p>Remember we have assumed that S is infinite numerable. It is therefore unpractical to let communication
finish when all environment states have been observed by A1 and A2. At that point, the family of channels
{Cn}n∈N would inform of all the communication stages. It is therefore up to the agents to decide when
to stop communicating should a good enough approximation have been reached for the purposes of their
respective tasks. But the study of possible termination criteria is outside the scope of this paper and left for
future work. From a theoretical point of view, however, we can consider the channel C∗ = {fi : Ai →
C∗}i∈{1,2} which informs of the end of the communication after observing all environment states. C∗ is
defined as follows: tok(C∗) = {cn | n ∈ N}, typ(C∗) = typ(A1 + A2) and for n ∈ N, cn |=C∗ hi, αi if
fi(en) |=Ai α, for all α ∈ typ(Ai) and i ∈ {1, 2}. In addition, fi∗(α) = hi, αi and fi∗(cn) = fi(en) for
n ∈ N, α ∈ typ(Ai) and i ∈ {1, 2}. By definition, f1∗ and f2∗ are infomorphims.</p>
        <p>Theorem 2 below is the cornerstone of the theory exposed in this paper. In short, it ensures that at each
communication stage agents obtain a theory that approximates the most meaningful relationships between
their types:
Theorem 2 The following statements hold:</p>
        <sec id="sec-3-1-1">
          <title>1. For all n ∈ N, C∗ is a refinement of Cn.</title>
          <p>2. T h(Log(E )) = T h(C∗).</p>
          <p>PROOF:
1. It is easy to prove that for each n ∈ N, gn defined as the identity function on types and the inclusion
function on tokens is a refinement infomorphism from C∗ to Cn.
2. It follows directly from:
cn |=C∗ hi, αi
fˇi(en) |=Ai α
en |=E fˆi(α)
en |=E fˆ(hi, αi)
{by definition of |=C∗ }
{because fi is infomorphim}
{by definition of fˆ}
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Conclusions</title>
      <p>
        In this paper we have exposed a formal model of semantic alignment as a sequence of information-channel
refinements that are relative to the particular states of the environment in which two agents communicate and
align their respective conceptualisations of these states. Before us, Kent [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and Kalfoglou and Schorlemmer
[
        <xref ref-type="bibr" rid="ref10 ref4">4, 10</xref>
        ] have applied Channel Theory to formalise semantic alignment using also Barwise and Seligman’s
insight to focus on tokens as the enablers of information flow. Their approach to semantic alignment, however,
like most ontology matching mechanisms developed to date (regardless of whether they follow a functional,
design-time-based approach, or an interaction-based, run-time-based approach), still defines semantic
alignment in terms of a priori design decisions such as the concept taxonomy of the ontologies or the external
sources brought into the alignment process. Instead the model we have presented in this paper makes explicit
the particular states of the environment in which agents are situated and are attempting to gradually align
their ontological entities. In the future we plan to further refine the model presented here (e.g., to include
pragmatic issues such as termination criteria for the alignment process) and to devise concrete ontology
negotiation protocols based on this model that agents may be able to enact.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This work is supported under the UPIC project, sponsored by Spain’s Ministry of Education and Science
under grant number TIN2004-07461-C02- 02 and under the OpenKnowledge Specific Targeted Research
Project (STREP), sponsored by the European Commission under contract number FP6-027253. M.
Schorlemmer is also supported by a Ramo´n y Cajal Research Fellowship from Spain’s Ministry of Education and
Science, partially funded by the European Social Fund.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Barwise</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Seligman</surname>
          </string-name>
          . Information Flow:
          <article-title>The Logic of Distributed Systems</article-title>
          . Cambridge University Press,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Ghidini</surname>
          </string-name>
          and
          <string-name>
            <given-names>F.</given-names>
            <surname>Giunchiglia</surname>
          </string-name>
          .
          <article-title>Local models semantics, or contextual reasoning = locality + compatibility</article-title>
          .
          <source>Artificial Intelligence</source>
          ,
          <volume>127</volume>
          (
          <issue>2</issue>
          ):
          <fpage>221</fpage>
          -
          <lpage>259</lpage>
          ,
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Giunchiglia</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          .
          <article-title>Semantic matching</article-title>
          .
          <source>The Knowledge Engineering Review</source>
          ,
          <volume>18</volume>
          (
          <issue>3</issue>
          ):
          <fpage>265</fpage>
          -
          <lpage>280</lpage>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kalfoglou</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer.</surname>
          </string-name>
          IF-Map:
          <article-title>An ontology-mapping method based on information-flow theory</article-title>
          .
          <source>In Journal on Data Semantics I, LNCS 2800</source>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kalfoglou</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          .
          <article-title>Ontology mapping: The sate of the art</article-title>
          .
          <source>The Knowledge Engineering Review</source>
          ,
          <volume>18</volume>
          (
          <issue>1</issue>
          ):
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Kent</surname>
          </string-name>
          .
          <article-title>Semantic integration in the Information Flow Framework</article-title>
          .
          <source>In Semantic Interoperability and Integration, Dagstuhl Seminar Proceedings 04391</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Lenat</surname>
          </string-name>
          .
          <article-title>CyC: A large-scale investment in knowledge infrastructure</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>38</volume>
          (
          <issue>11</issue>
          ),
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Lo</surname>
          </string-name>
          ´pez, M. Sabou, and
          <string-name>
            <surname>E. Motta.</surname>
          </string-name>
          <article-title>PowerMap: Mapping the real semantic web on the fly</article-title>
          . Accepted to be presented at ISWC'
          <volume>06</volume>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F.</given-names>
            <surname>McNeill. Dynamic Ontology</surname>
          </string-name>
          <article-title>Refinement</article-title>
          .
          <source>PhD thesis</source>
          , School of Informatics, The University of Edinburgh,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schorlemmer</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kalfoglou</surname>
          </string-name>
          .
          <article-title>Progressive ontology alignment for meaning coordination: An information-theoretic foundation</article-title>
          .
          <source>In 4th Int. Joint Conf. on Autonomous Agents and Multiagent Systems</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Shvaiko</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Euzenat</surname>
          </string-name>
          .
          <article-title>A survey of schema-based matching approaches</article-title>
          .
          <source>In Journal on Data Semantics IV, LNCS 3730</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Steels</surname>
          </string-name>
          .
          <article-title>The Origins of Ontologies and Communication Conventions in Multi-Agent Systems</article-title>
          .
          <source>In Journal of Autonomous Agents and Multi-Agent Systems</source>
          ,
          <volume>1</volume>
          (
          <issue>2</issue>
          ),
          <fpage>169</fpage>
          -
          <lpage>194</lpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>J. van Diggelen</surname>
          </string-name>
          et al.
          <source>ANEMONE: An Effective Minimal Ontology Negotiation Environment In 5th Int. Joint Conf. on Autonomous Agents and Multiagent Systems</source>
          ,
          <year>2006</year>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>