=Paper= {{Paper |id=None |storemode=property |title=Decentralized Semantic Coordination Via Belief Propagation |pdfUrl=https://ceur-ws.org/Vol-918/111110266.pdf |volume=Vol-918 |dblpUrl=https://dblp.org/rec/conf/at/Vouros12 }} ==Decentralized Semantic Coordination Via Belief Propagation== https://ceur-ws.org/Vol-918/111110266.pdf
    Decentralized Semantic Coordination Through
                 Belief Propagation.?

                                  George A. Vouros

             Department of Digital Systems, University of Piraeus, Greece



        Abstract. This paper proposes a decentralized method for communities
        of agents to reach semantic agreements on subjective ontology elements’
        correspondences / mappings, via belief propagation. Agents detect dis-
        agreements on mapping decisions via feedback they receive from others,
        and they revise their decisions on correspondences with respect to their
        mapping preferences and the semantics of ontological specifications. We
        address this problem as a distributed extension of the max-plus algo-
        rithm via belief propagation. We show the potential of such an approach
        in a number of networks with varying complexity and for different cases
        of semantic disagreement between agents.


1     Introduction
Interacting entities (peers, software or human agents, things on the web, sen-
sors and actuators) aiming to effectively share and search for information in an
open and inherently distributed setting, need to agree on the meaning of the
terms they use for structuring information, annotating data and conceptualiz-
ing the world. In any open setting, these entities can not be assumed to share
the same ontology. However, each entity can only compute subjective correspon-
dences of own ontology elements to the ontology elements of its acquaintances.
Nevertheless, to propagate queries and enable effective sharing of information,
the entities in the network need to reach an agreement on a globally coherent
set of correspondences, wrt the semantics of specifications and their preferences
on correspondences. Let us for instance consider two peers with ids 1 and 2,
forming a cycle in a network. In this case, for each subjective equivalence cor-
respondence from 1 to 2 for a pair of ontology elements (Ei1 , Ej2 ), peer 2 must
agree that Ej2 is equivalent to Ei1 so that queries and results are propagated
among them effectively.
    This paper proposes a generic decentralized method for peers in a data man-
agement system to reach semantic agreements regarding ontology elements’ sub-
jective correspondences, via belief propagation. By ”generic” we mean a method
that is independent of the specific mapping methods used by the peers, and also
applicable to any possible setting of information sharing entities. It must be
noticed that although the proposed approach can be applied to any setting of
interacting entities, for the clarity of the presentation we refer to peers in data
?
    AT2012, 15-16 October 2012, Dubrovnik, Croatia. Copyright held by the author(s).
management systems and to interacting agents that implement decision-making
processes on behalf of these peers.
    According to the proposed approach, each peer computes correspondences
of its ontology elements to the elements of acquaintances’ ontologies, using any
set of mapping methods. To preserve the coherency of local mappings with re-
spect to the semantics of specifications, peers build internal dependency graphs
whose nodes (agents) are inter-constrained. These agents are responsible to com-
pute specific mapping decisions on behalf of the peer: Each for a specific ontol-
ogy element. Having computed subjective correspondences, peers aim to reach
agreements to their mapping decisions: By exploiting the transitive closure of
mappings in existing cycles in their network, peers can detect disagreements and
revise their decisions.
    The proposed method enables peers to revise their decisions with respect to
their preferences, the semantics of specifications and the feedback they receive
from others. The problem is addressed as a coordination problem where agents
target the establishment of agreements wrt to the semantics of specifications,
increasing the social welfare, i.e. the sum of the utilities of the individual agents.
Toward this target we use the max-plus [8] algorithm to generate near-to-optimal
solutions to reaching semantic agreements.


2    Problem Statement and Overall Approach

The problem of computing mappings (mapping relations or correspondences)
between ontologies can be stated as follows: Given two ontologies O1 = (S1 , A1 ),
O2 = (S2 , A2 ) (where Si denotes the signature and Ai the set of axioms that
specify the intended meaning of terms in Si ) and an element (class or property)
Ei1 in the signature S1 of O1 , locate a corresponding element Ej2 in S2 , such that
a mapping relation (Ei1 , Ej2 , r) holds between them. r can be any relation such
as the equivalence (≡) or the subsumption (v) relation. For any such correspon-
dence a mapping method may relate a value γ that represents the preference to
relating Ei1 with Ej2 via r, i.e. the confidence on this correspondence. In this case
the correspondence is denoted by (Ei1 , Ej2 , r, γ). In this paper we deal only with
equivalences between classes, although other mapping relations whose transitive
closure can be exploited, can be considered as well. The set of mapping relations
produces the mapping function f : S1 → S2 that must preserve the semantics
of representation: i.e. all models of axioms A2 must be models of the translated
A1 : i.e. A2 |= f (A1 ).
    Let us consider a network of peers represented by a directed graph G =
(V, E). Each node in V is a peer Pnode equipped with a specific ontology Onode .
Each directed edge (Pnode , Pneigh ) in E specifies the fact that Pnode has a specific
view of Pneigh ’s ontology (which can be either a complete, partial, vague or an
abstract view, at the conceptual or at the assertional level, depending on peers’
actual relation). Pnode can propagate queries/ answers to Pneigh by computing
correspondences between the elements of Onode and Oneigh . Generally, these are
correspondences from the subjective point of view of Pnode . In case there is also
an edge (Pneigh ,Pnode ), then the subjective correspondences from Pneigh to Onode
might not be symmetric to those computed by Pnode . This may happen even in
the special case that these peers use the same mapping method (because the
information that this method exploits may differ in each direction).
    In such a setting, the problem of computing a coherent set of correspondences
between peers in G is rather complicated if we consider that each peer is con-
nected to numerous other peers with distinct ontologies and mapping methods,
with the latter being arbitrarily connected with others and/or among themselves.
This problem includes: (a) The computation of locally coherent correspondences
to the ontologies of neighbor peers, and (b) the exploitation of the transitive clo-
sure of correspondences between ontologies to reach agreements between (even
distant) peers 1 . To describe our approach we distinguish between these, actually
highly intertwined, computation phases.
    Concerning the computation of locally coherent subjective correspondences
of each peer Pnode , given Onode (the ontology of Pnode ) we address the com-
putation of correspondences from Onode to any ontology Oneigh of a neighbor
peer Pneigh , as a coordination problem between self-interested agents. Each such
tuple (Onode ,Oneigh ) is considered as a specific, distinct mapping case for Pnode .
Considering that Pnode has at its disposal a mapping method Anode , for each
mapping case there is a distinct set of agents, each of which corresponds to the
mapping method and an ontology element in Onode , as it is also done in [19]:
Each agent is responsible to decide on the mapping of that specific ontology
element. We call these agents Pnode − internal (or simply internal) agents to
emphasize their role to compute correspondences on behalf of Pnode . Actually,
they may be distant and are organized in a certain way. Section 5.1 describes
how the internal agents are related in dependency graphs (one distinct graph per
mapping case) via validity constraints that must be satisfied so as the calculated
mappings to conform to the semantics of ontological specifications.
    Therefore, viewing the computation of ontology correspondences from Onode
to Oneigh as a coordination problem between M 2 Pnode -internal agents that are
related via validity constraints, we aim to find the joined state of agents (i.e.
the overall set of correspondences), X∗ , such that the social welfare of the whole
system (i.e. the sum of the individual agents’ utilities Um ) is maximized:
                                            X
                             X∗ = arg max       Um (xm )                         (1)
                                        X
                                            m

   Given the subjective local decisions concerning correspondences, each of
the peers propagate local mappings to the rest of the peers in the network
in an iterative and cooperative manner, aiming to detect cycles of correspon-
dences and get feedback on its decisions. Therefore, computed correspondences
1
  Notice that, even if correspondences are subjective, peers consider the correspon-
  dences computed by the others in conjunction to own correspondences, i order to
  exploit their transitive closure, trying to identify disagreements, get feedback and
  reach consensus.
2
  M in the general case depends on the number of elements in Onode and the number
  of alignment methods used by Pnode .
(Einode ,Ejneigh , r, m) for a mapping case (Onode ,Oneigh ) are sent to any of the
immediate neighbors Pneigh that use Oneigh . In this case the parameter m does
not denote the preference γ computed by the mapping method, but is an approx-
imation of the payoff contribution related to that correspondence: This will be
thoroughly explained in Section 5.2. Pneigh in its own turn dispatches received
correspondences to the appropriate Pneigh -internal agents (i.e. to the agents that
are responsible for computing a correspondence for Ejneigh 3 ). Computed cor-
respondences for Pneigh (Ejneigh ,Ekxyz ,r,m) are appended to (Einode ,Ejneigh ,r,m)
and are forwarded to Pneigh neighbors, and so on and so forth. The ordered list of
correspondences forwarded to a peer shows the correspondences computed along
a path and constitutes a mapping history. Although there are several strategies
for propagating correspondences, in this work we assume that mapping histories
are propagated in an unrestricted way, aiming to detect any cycle of correspon-
dences in the network. Although we may further "tune" the efficiency of the
propagation method, it must be pointed out that even with this naive propaga-
tion approach, the communication cost is not very high per ontology element,
given that each peer propagates each message only to neighbor peers that cor-
respond to a specific mapping case. As already said, each peer dispatches these
messages to its internal agents. Having said that, we have to point out that
the size of messages grow linearly with the length of the paths traversed in the
network of peers.
    Some correspondences might map an element from one ontology to a se-
mantically irrelevant element in another ontology. Our goal is not to provide
probabilistic guarantees on the correctness of a mapping operation as it is done
in [5]: We aim towards driving peers to re-consider their local mappings with
respect to their preferences, the semantics of specifications, also in accordance to
the feedback they receive from other peers via cycles of mappings, so as to reach
agreements. Towards this goal, also in accordance to [5], we take advantage of
the cycles existing in the graph of peers. Given a cycle (Pnode → Pneigh → · · · →
  0
Pneigh  → Pnode ) we consider that for each correspondence (Einode ,Ejneigh , ≡, m)
forwarded from Pnode to Pneigh (together with the mapping history) the origina-
tor Pnode must get a correspondence (Ekneigh ’,Einode , ≡, m’) from the last peer in
the cycle, Pneigh ’. In other words, when Pnode computes a local correspondence
of Einode to Ejneigh , then Pneigh , via the path from Pneigh to Pnode must propa-
gate a correspondence to the element Einode , rather than to any other element of
Onode . In such a case, for each such correspondence, Pnode counts a positive feed-
back. In case this does not happen, then there are one or more mappings through
this path that according to the subjective view of Pnode result to a disagreement.
Then, Pnode counts a negative feedback. It must be noticed that disagreements
may still exist when Pnode gets the expected correspondence but several corre-
spondences along the path compensate their errors: These are detected by the
corresponding peers as the mapping history propagates in the network.

3
    There is one such agent in each dependency graph corresponding to a mapping case.
    Summing up the above, each peer (a) gets a list of correspondences prop-
agated to it (the mapping history), and (b) inspects the mapping history to
detect a cycle. This is done by detecting in the mapping history the most recent
correspondence originated from it. Such a correspondence concerns one of its
own ontology classes and a specific mapping case. In case there is a cycle, the
peer propagates the history to the internal agents that (i) correspond to that
mapping case and, (ii) are responsible to compute correspondences for that class.
These agents incorporate the feedbacks in their decision-making. In case there
is not any cycle, histories are propagated to all internal agents corresponding to
that mapping case. (c) After internal agents’ computations, the peer propagates
mapping decisions (together with the mapping histories) to neighbor peers corre-
sponding to the mapping case. The propagation of beliefs in internal graphs and
between peers, together with the incorporation of feedback to the algorithm are
presented in detail in Section 5. It must be pointed out that belief propagation
between peers happens seamlessly to the local computations of correspondences
by internal agents. So, we may consider that a large, distributed graph compris-
ing the internal agents of peers, spans the whole network of peers.


3   Related work

There are several works aiming to preserve the consistency of mapping deci-
sions by locating (minimal) subsets of mappings that introduce inconsistencies
[13],[14],[15],[17]. However, we consider distributed methods for the computation
of agreed correspondences between ontologies of peers in a network.
    Argumentation methods support the creation and exchange of arguments,
followed by reasoning on their acceptability. There are numerous approaches
that explore the application of argumentation methods for reaching semantic
agreements between agents (e.g. [20],[21],[12]). Arguments are being used as po-
sitions for or against correspondences. These methods are proposed mostly for
cases where peers need to produce mappings between their shared ontology and
the ontology of a third party: I.e. to a rather restricted set of settings. Alterna-
tively to arguments, we propose agents to exploit estimations of correspondences’
payoffs, which are propagated to other agents and peers.
    Probabilistic message passing techniques proposed in [5] aim to detect er-
roneous mappings between peers in a peer-to-peer data management system.
This method targets cases where peers connected in a network have different
ontologies and produce mappings to each other. The proposed method takes ad-
vantage of the transitive closure of mapping operations in a cycle of peers, to
compare a query q to the returned q 0 via that cycle. To a certain extent than that
work, we aim to compute correspondences that maximize the social welfare of
the overall system (leading agents to agreement), with respect to the semantics
of specifications and to the preferences of agents.
    For the establishment or development of a common vocabulary among en-
tities, researchers have proposed cognitive models [16], bilateral communication
protocols and strategies [7][18] for agents, also studying the properties of shared
vocabularies [6], or the dynamics of their establishment [1]. In this paper we
start from the point where peers have own ontologies (assumed to be devel-
oped/learned independently from the ontologies of others). We do not deal with
how ontologies have been learned, or how effective mappings have been estab-
lished between pairs of peers: This may be achieved either using an instance
based method [7][22], a method for "tuning" the descriptions of concepts [22],
or other ontology mapping methods. The proposed method aims at reconciling
differences between computed correspondences, based on the notion of feedback,
whose effect has also been studied in [16],[21],[5].

4     The max-plus algorithm
Before proceeding to the details of the computations, this section briefly explains
how belief propagation happens via the max-plus algorithm.
    One can deal with a complicated problem by viewing it as a factorization of
several local functions, each depending on a subset of the variables that appear
in the global problem. In particular, aiming to optimize over a function that is
decomposed in local terms, we may compute the maximum a posteriori config-
uration in a graphical model using the max-product algorithm [11]. In our case,
we aim at selecting the appropriate correspondences among ontology elements so
as to maximize the social welfare of a set of agents. Message passing algorithms,
and particularly the max-plus algorithm [8,10] are directly applicable to such a
problem. In the most general case, considering a graph of interacting agents (a
dependency graph - aka coordination graph [9]- specifying how the decisions of
one agent affect the decisions of others), in order to compute the optimal joint
state X∗ according to equation (1), each agent (node in the graph) sends a mes-
sage µij to each of its neighbors Aj in N (Ai ). This message can be regarded as
a local payoff function [10] of agent Aj and it is defined as
                                                 X
           µij (xj ) = maxxi {Uij (xi , xj ) +            µki (xi )} + cij      (2)
                                                 Ak ∈N (Ai )−{Aj }

where cij is a normalization factor. This message is computed by maximizing
over the preferences of Ai , the sum of the utility Uij and all incoming messages
to Ai except from Aj . Messages are exchanged until they converge to a fixed
point 4 , or until an external signal is received (this is the case in this paper: The
algorithm iterates I times). Since payoff functions are defined over at most two
agents the utility Uij can be of the form
                           Uij (xi , xj ) = fi (xi ) + fij (xi , xj )              (3)

where fi specifies the payoff contribution of agent Ai , and fij the payoff contri-
bution of the pair of neighbor agents Ai and Aj . These functions are specified
in Section 5.1.
4
    Unfortunately, there are no guarantees that the max-plus converges to optimal so-
    lutions in graphs with cycles and no assurances can be given concerning the quality
    of the joint state X∗ .
      The agent Ai may at any time step compute
                                             X
                       gi (xi ) = fi (xi ) +    µki (xi )                                   (4)
                                                Ak ∈N (Ai )

which equals the contribution to the global payoff function achieved via Ai ’s
neighborhood. Thus, each agent can form a decision by approximating its optimal
choice by computing

                                    x∗i = arg max gi (xi )                                  (5)
                                                xi

   Messages may be updated either asynchronously or synchronously. Any node
in the graph may initially sent an identity (i.e. µij (xj ) = 1). When a node
receives messages from its neighbors it computes outgoing messages.


5      Decentralized Coordination Through Belief
       Propagation

5.1     Internal Dependency Graphs

Considering one of the mapping cases of Pnode , from Onode to Oneigh , Pnode
constructs a dependency graph for this mapping case on-the-fly: The nodes of this
graph are Pnode -internal agents. As already said, each agent Anode         i     corresponds
to an ontology class Einode of Onode , given a specific mapping method A of Pnode .
    The state of an internal agent Anode i      is represented by a tuple (xnode
                                                                               i   ,γ(xnode
                                                                                        i    )).
The range of xnode
                 i   is the  set of   classes    in  O neigh that  the method    A assesses  to
be related to Einode via a relation r (in this paper equivalence only). Let that
set be Dinode . A value selection for xnode   i      (i.e. a class in Dinode ) is denoted by
        node                                      node
value(xi ). The ordering imposed to Di                  by γ represents the preference of the
agent to each possible value of xnode
                                    i      . In case   elements  in Dinode are not associated
with confidence values, then the agent considers all possible values of xnode       i     to be
equally preferred. Such an agent Anode   i      is  depicted  in Figure  1(a). Therefore,   the
internal dependency graph corresponding to the mapping case (Onode , Oneigh )
with a single mapping method, comprises at most Mnode = |Snode | agents (Snode
is the signature of Onode ).
    Each agent Anode
                   i     interacts with a limited number of other agents: In the
most general case, given a specific mapping case, the Pnode -internal neighbors
of Anode
     i     are those agents that correspond to ontology classes that are direct
subsumees of Einode in Onode . Formally, the set of Pnode -internal neighbors of an
agent Anode
         i    is denoted by N (Anodei      ) and it is defined to be the set

        N (Anode
            i    ) = {Anode
                       n    |(n 6= i) and (Einode v Ennode in a direct way) }

   Agents in Pnode -internal graphs are interrelated with constraints that affect
their states (i.e. mapping decisions). Let us first show this with an example (the
corresponding dependency graph is shown in Figure 1): Given a mapping case
with two ontologies Onode and Oneigh and a mapping method A, the dependency
graph comprises agents (some of them are depicted in Figure 1 in round-corner
rectangular), related with the mapping method, a specific class from Onode , and
a state variable x). The method A computes for each class Einode the set Dinode




          Fig. 1. Ontologies (Mapping Case) and Dependency Graph (part)



of candidate corresponding classes from the ontology Oneigh . Each of the classes
in this set is assessed to be related to Einode via the equivalence relation, with a
specific degree of confidence. Given two classes from Onode , s.t. E2node v E1node ,
The agent for the class E1node considers correspondences to classes of Oneigh such
that value(x1 ) ≡ E1node . Similarly, the agent for the class E2node considers values
- i.e. classes of Oneigh - for the variable x2 such that value(x2 ) ≡ E2node . Clearly,
since E2node v E1node , to maintain consistency between the mappings computed
by these agents it must hold that value(x2 ) v value(x1 ). This is a validity
constraint (shown with a bold line between agents) that must be preserved by
the agents Anode
               1    and Anode
                            2   and it is specified by the constraintconstr(x2 , x1 , v).
     Generally, the constraints between two agents Anode    i   , Anode
                                                                   u    , are as follows
(given that the mapping methods considered compute equivalences, only):
     If ((Ejnode r Ewnode
                          ) and (value(xw ) ≡ Ew  node
                                                       and value(xj ) ≡ Ejnode ),
     then constr(xj , xw , r), where r ∈ {≡, v}
   Validity constraints specify how the utility of each agent is affected by its
own state and the state of its neighbors. Recalling that the utility has the generic
form specified in (3), payoff functions fi and fij are as follows:


                           fi (xi ) =γ(xi )                                         (6)
                                                                           
                                       1 if constr(xi , xk−l , r) satisfied
                fik−l (xi , xk−l ) =                                                (7)
                                              0 in any other case
    where γ(xi ) is the confidence by which the mapping method maps the class
Einode to the class value(xi ) of the other ontology. N (Ai ) is the set of Ai neigh-
bors, and constr(xi , xk , r) denotes the constraint that value(xi ) must be related
to value(xk ) via r.

5.2   Belief Propagation Among Peers
Each edge (Pnode ,Pneigh ) in the network of peers, corresponds to a distinct map-
ping case (Onode ,Oneigh ) for Pnode . Each Pnode -internal agent Anode    i    for this
mapping case ("responsible" to compute correspondences for the class Einode to
Oneigh ), computes the messages to its neighbor agents according to the formula
(2). Doing so, the agent Anodei   communicates its payoff function to the agents
with whom it shares a constraint. Nevertheless, to reach agreements, the state of
each Pnode -internal agent Anodei  must affect also the computations of (distant)
peers in cycles with Pnode , and visa-versa. Therefore, the states of Pnode -internal
agents must be propagated to neighbor peers. It must be pointed out that no
peer knows any part of the internal graph or any of its neighbor peers. There-
fore messages are propagated among peers, rather than among internal agents
of different peers.
    As already said, messages concerning correspondences for a mapping case
are propagated to all neighbor peers corresponding to that mapping case. Doing
so, peers may detect cycles, take the full advantage of the transitive nature of
cyclic mappings, and detect all positive/negative feedbacks provided by agents.
Specifically, given the edge (Pnode ,Pneigh ), Pnode collects the states of each Pnode -
internal agent corresponding to the mapping case from Pnode to Pneigh in a syn-
chronous way (i.e. after I iterations of the max-plus algorithm) and propagates
these decisions to Pneigh , and to all other peers corresponding to that mapping
case. Actually, Pnode propagates messages, each including a mapping history and
a new correspondence of the form (Einode ,Ejneigh ,≡, m). Ejneigh is computed ac-
cording to equation (5). The actual form of m is mneigh  node (Ej
                                                                 neigh
                                                                       ) and it is equal
to gi (Ejneigh ), given by formula (8) below, which refines the formula (4). The
computations are performed by the internal agents of Pnode that are responsible
to compute correspondences for any element Einode .
    Pneigh , by inspecting any mapping history that has been propagated to it,
can detect the existence of a cycle, and can spot one of its internal agents that
is interested to the propagated mappings: As already said, this is done by de-
tecting in the history the most recent correspondence originated from it. This
correspondence (it is exists) concerns one of Oneigh classes. The agent to which
the feedback and mapping history is propagated is the one that decides on cor-
respondences for that class and for that mapping case. In case there is not a
cycle, the messages are propagated to all internal agents corresponding to that
mapping case.
   Pneigh -internal agents exploit incoming messages to detect any feedback and
incorporate these into their decision. Specifically, after I iterations of the max-
plus algorithm, Pneigh -internal agents compute own decisions by refining the
computation of payoffs specified by formula (4), considering also feedback mes-
sages, as follows:

                                          X                             X
          gi (xi ) = fi (xi ) + (α ×                 µki (xi )) + (β×          Fnode (xi ))   (8)
                                       Ak ∈N (Ai )                      node

   where node denotes any Pnode that has propagated correspondences to Pneigh .
In case there is a closing cycle with a negative feedback, or in case there is
not a closing cycle, Fnode (xi ) is 0, else Fnode (xi ) is mneigh
                                                            node (xi ). In other words,
correspondences are decided by taking into account any feedback received from
neighbor peers by exploiting mapping histories corresponding to the mapping
case considered. Also, α and β are parameters (real numbers ranging in [0,1])
that have been used to experiment with the effect and necessity of feedback and
constraints on agents’ decision making.


6    Experimental Evaluation
We have contacted experiments with different networks, and with different peers’
abilities to compute correspondences.
    For each experimental setting (presented below) we measure (a) the recall
R and the precision P of the correspondences computed by each peer for each
mapping case (with respect to the reference mappings), (b) the number of posi-
tive and negative feedbacks F + and F − received by each peer, (c) the number of
messages P M received by each peer in each run, and (d) the number of messages
AM exchanged between the internal agents of each peer per run.
    To measure the degree / level of agreement achieved between peers and the
effectiveness of computations we measure the f − score (specified by f − score =
2 ∗ (P ∗ R)/(P + R) ) for each mapping case, the level of agreement achieved for
each peer, (specified by Level = (F + − F − )/P M ), the agreement accuracy for
pairs of peers, (specified by Accuracy = Level ∗ f − score ), and the M essage
Gain for achieving this accuracy (and level) given the number of inter-agent
messages, specified by Gain = (Accuracy ∗ sf )/AM , where sf is a scaling factor
set equal to 1000. The motivation for these proposed scores is as follows: While
f-score shows in a combined way the precision/recall of the correspondences
computed per mapping case, the level of agreement shows (independently of the
precision/recall achieved) the agreement achieved between peers as a function of
positive and negative feedbacks after a specific number of inter-peer messages:
Peers may agree to false positive correspondences, as well. To provide a measure
of these correspondences, the agreement accuracy measure combines the level
of accuracy and the accuracy of computed mappings. Finally, aiming to show
the efficiency of the method, we provide the gain as a function of the accuracy
achieved and the number of inter-agent messages.
    Given that the complexity of the method increases as the length and number
of cycles in the network of peers increase (rather than with the number of peers),
and in order to study the various patterns of method’s behavior, we demonstrate
the performance of the method with networks of N=6 peers, each with an in-
creasing complexity. Specifically, experiments concern four types of networks in
which peers’ average degree is 2,3,4 or 5. Each type of network comprises at
most 10 different networks. For each type of network we run experiments with
various values of α and β, distinguishing the following cases:
 – (C1) α ∈ (0, 1] and β ∈ (0, 1]: In these cases we need agents’ decision to be
   affected by both, the feedback received, and by the inter-agents constraints,
   in a balanced way. In this case we expect agents to achieve high f-scores and
   high levels of agreements. Since it is not clear which case is a "balanced"
   case, we have run several experiments with α and β ranging from 0.1 to 0.9.
   Below we present experiments from the following combinations of α and β:
     • (case1) (α = 0.1, β = 0.9),
     • (case2) (α = 0.3, β = 0.7),
     • (case3) (α = 0.5, β = 0.5),
     • (case4) (α = 0.7, β = 0.3),
     • (case5) (α = 0.9, β = 0.1).
 – (C2) α = 1 and β = 0: In this case agents’ decision is affected by the
   inter-constraints, rather than by the feedback received. In this case we may
   expect agents to achieve high f-scores but low levels of agreement. However,
   experiments show that actually, constraints play a crucial role in achieving
   high levels of agreements.
 – (C3) α = 0 and β = 1: In this case agents’ decisions are affected mostly by
   the feedback received, rather than by the inter-agents constraints. In this
   case we may expect high levels of agreement but low f-scores. Nevertheless,
   given that agents’ decisions are not inter-constrained, feedback alone does
   not drive agents to reach high levels of agreement.
 – (C4) α = 0 and β = 0: In this case agents compute their decisions in a purely
   random way. Neither inter-agents’ constraints nor feedback messages affect
   their decision. This is a baseline case where f-core and agreement measures
   are expected to be very low.
    For each of the above cases we ran 10 independent experiments for each type
of network (thus, 50 experiments for each case and sub-case, i.e. 400 experiments
in total). Below we present aggregated results: The average of the scores achieved
per case, and the average performance of one peer (over the different networks
considered) in two cases.
    In each of these cases each peer is been randomly assigned with a specific
distinct ontology with 13 classes.
    As far as the ability of peers to compute correspondences is concerned, in
order to have a control on the experimental settings, each peer is set to compute
mapping relations with a constant precision in {1,0.9,0.8,0.7,0.6,0.5}, in any of
the networks considered. The precision specifies the proportion of the decided
correspondences that are correct (true positives). False positive correspondences
map a class to a randomly chosen class with high preference, while the correct
correspondence for that class is assigned a random, low preference. As the dif-
ference between the mapping ability of neighbor peers grows, as the problem
for reconciling different views and thus, computing agreements, is expected to
become harder: As the average degree of peers increases, each peer has neighbors
with quite different mapping abilities.
    The max-plus running in each of the internal agent graphs iterates for a fixed
number of iterations (I) per round. After experimentation I has been set to 5.
    First, we have run experiments to determine what constitutes a "good" bal-
ance between constraints and feedback (i.e. to determine a good option for α
and β for the case C1). Then, given the best option, we compared this case with
the rest of the cases C2, C3, C4, where α and β range in {0,1}, excluding the
case where α and β are both equal to 1.
    Average results from the 5 sub-cases of case C1 are shown in Figure ??(a).
    As Figure 2(a) shows, the performance scores, and especially the f-score and
the level of agreement achieved in all cases, are nearly the same: In all cases
agents compute their decisions wrt their constraints with other internal agents
and to the feedback received from other peers. However, case 2 (α = 0.3, β = 0.7)
achieves a slightly higher f-score (0.86) and level of agreement (0.10) than the
other cases. Therefore, we consider this to be the "balanced" case. Subsequently,
when we refer to C1 we refer to this particular case. Having said this, we must
point out that the method is robust to the choice of the external parameters α,
β, although a "good" choice can make a slight difference to the results. Figure
2(b) shows the performance measures for a specific (representative) peer for the
C1 for each of the rounds of the overall method (measurements are the averages
for 10 independent experiments). As we may notice, the proposed method (in all
the cases and for each of the peers) manages to gradually increase the f-score and
the number of positive feedback received, thus the level of agreement between
peers. Also, the method manages to reach a plateau after round 5.
    Having determined the "balanced" C1 case, we can compare the results
achieved in that case, with the average f-score, agreement levels, accuracy and
message gain achieved in all cases where α and β range in {0,1}. Figure 2(c)
shows the results for all the cases (C1 to C4). As expected, the best case is C1
(i.e. the balanced case), followed by C2 with f-score 0.85 and level of agreement
0.09. It may be a surprise that C2 achieves a high level of agreement, but this is
due to the fact that all peers have the same ontology. In this case, peers’ internal
agents form mapping decisions that respect the semantic constraints, which are
shared by all groups of internal agents in each peer. This results to a high level
of agreement even if feedback is not considered by the agents. Although this
is the case for structurally similar ontologies, future work aims to investigate
                                                          (a)




                                                                (b)




                                                          (c)




                                                                (d)

                          Fig. 2. Experimental Results




this case more intensively, considering ontologies that can be aligned but that
are structurally dissimilar. For the rest of the cases, i.e. C3 and C4, both the
f-score and the level of agreement achieved are very low. However, in C3 the
level of agreement is slightly higher than C4. In both of these cases, the fact
that constraints are not being considered, causes the unconstrained fluctuation
of agents’ decisions among all possible correspondences, thus affecting the level
of agreement achieved. Figure 2(d) shows the (representative) performance of a
specific peer in case C3. Again, the reported measurements are the averages for
10 independent experiments.
    Concluding the above, experimental results show that semantic coordination
must consider both, the constraints and the feedback received, although semantic
constraints play a vital role in the process.


7    Concluding Remarks and Future Work
We address the semantic coordination problem in an agent-based data manage-
ment system through payoff propagation, by a distributed extension of the max-
plus algorithm. The method is shown to be effective enough, making the best of
the mappings produced by peers, helping peers increase their levels of agreement
and the agreement accuracy. Future work concerns including other mapping re-
lations whose transitive closure can be exploited by peers (e.g. subsumption and
disjointness) in cycles. Also, we aim to address subtle issues concerning commu-
nication; although - via experiments performed- various heuristics for reducing
the communication messages result to sacrificing considerably methods’ effec-
tiveness. Finally, the study of sub-optimality, as well as the integration of the
proposed method to a peer-to-peer setting seamlessly to other distributed pro-
cesses (e.g. information retrieval) are important issues.


References
1. Baronchelli A., Felici M., Caglioti E., Loreto V., L.Steels 2006 Sharp Transition
   towards Shared Vocabularies in Multi-Agent Systems. J. Stat. Mech. P06014, 2006.
2. Beun R-J, van Eijk M. R., H. Prust, 2004 Ontological Feedback in Multi-Agent
   Systems. In Proc. of AAMAS 2004.
3. Cudre-Mauroux P., et al. 2006 Viewpoints on emergent semantics. S. Spaccapietra
   et al. (Eds.): J. on Data Semantics VI, LNCS 4090, pp. 1-27, 2006.
4. Cudre-Mauroux P., et al. 2009 idMesh: Graph-Based Disambiguation of Linked
   Data. In Proc of WWW 2009.
5. Cudre-Mauroux P., Aberer K., and A. Feher 2006 Probabilistic Message Passing in
   Peer Data Management Systems, In Proc. of ICDE 2006.
6. Diggelen J., et al., 2005 Optimal Communication Vocabularies and Heterogeneous
   Ontologies. Agent Communication (Rogier M. van Eijk et al eds.), LNCS 3396, 2005.
7. Diggelen J., et al. 2006 ANEMONE: An effective minimal ontology negotiation
   environment, In Proc. of AAMAS 2006.
8. Farinelli, A., Rogers A., Petcu A. and N.R. Jennings 2008 Decentralised coordination
   of low-power embedded devices using the max-plus algorithm. In Proc. of AAMAS
   2008.
9. Guestrin, C., Koller D., Parr R. 2002 Multiagent planning with factored MDPs.
   NIPS 14, MIT Press 2002.
10. Kok J.R. and N.Vlassis 2006 Collaborative Reinforcement Learning by Payoff Prop-
   agation, JMLR, 7:1789-1828, 2006.
11. Kschischang, F.R., Frey B.J., Loeliger H-A., 2001 Factor graphs and the sum-
   product algorithm. IEEE Transactions on Information Theory, Vol 47, Issue 2, 2001.
12. Laera L., et al. 2007. Argumentation over ontology correspondences in mas. In
   Proc. of AAMAS 2007.
13. Meilicke C. and H. Stuckenschmidt 2009 An efficient method for computing align-
   ment diagnoses. In Proc. of RR 2009.
14. Meilicke C., Stuckenschmidt H. and A. Tamilin 2009 Reasoning support for map-
   ping revision. Journal of logic and Computation, 19(5):807-829, 2009.
15. Qi G., Q. Ji, and P. Haase 2009 A conflict-based operator for mapping revision, in
   Proc. of ISWC 2009.
16. Reitter D. and C.Lebiere, 2011 How groups develop a specialized domain vocabu-
   lary: A cognitive multi-agent model. Cognitive Sys. Res. 12, , pp 175-185, 2011.
17. Ruiz E.J., Grau B.C., Horrocks I. and R.Berlanga 2009 Ontology integration using
   mappings: Towards getting the right logical consequences. In Proc. of ESWC 2009.
18. Sensoy M., P. Yolum, 2009. Evolving service semantics cooperatively: a consumer-
   driven approach. J. of Aut. Agents and Multi-Agent Sys, 18, 2009.
19. Spiliopoulos, V, Vouros G. In Press. Synthesizing Ontology Alignment Methods
   Using the Max-Sum Algorithm. IEEE TKDE, Vol 24, Issue 5, pp 940-951, 2012.
20. Trojahn C., Euzenat J. 2010 Consistency-driven Argumentation for Alignment
   Agreement. Proc. of OM, 2010.
21. Trojahn C., Moraes M., Quaresma P., and R. Vieir 2008 A Cooperative Approach
   for Composite Ontology Mapping, J. on Data Semantics X, LNCS 4900, pp. 237-263,
   2008.
22. Williams B. A. 2004 Learning to Share Meaning in a Multi-Agent Systems. J. of
   Aut. Agents and Multi-Agent Sys, 8, 2004.
23. M. Yokoo and K. Hirayama. Distributed breakout algorithm for solving distributed
   constraint satisfaction problems. In Proceedings of the 2nd Int’l Conf. on Multiagent
   Systems, pages 401-408, 1996.