<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>From Signed Information to Belief in Multi-Agent Systems</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Laurent Perrussel</string-name>
          <email>laurent.perrussel@irit.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emiliano Lorini</string-name>
          <email>emiliano.lorini@irit.fr</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jean-Marc The´venin</string-name>
          <email>jean-marc.thevenin@irit.fr</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>IRIT - Universite ́ de Toulouse</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>IRIT - Universite ́ de Toulouse</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>IRIT - Universite ́ de Toulouse</institution>
        </aff>
      </contrib-group>
      <fpage>27</fpage>
      <lpage>33</lpage>
      <abstract>
        <p>-The aim of this paper is to propose a logical framework for reasoning about signed information. That is, as long as agents receive information in a multi-agent system, they keep track of the information source. The main advantage is that by considering a reliability relation over the sources of information, agents can justify their own current belief state. Agents believe at first information issued from the most reliable sources. Keeping track of belief's origin also enables agents to improve communication by asking and gaining details about exchanged information. This is a key issue in trust handling and improvement: an agent believes some statement because it may justify the statement's origin and its reliability.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>I. INTRODUCTION</title>
      <p>
        An agent embedded in a multi-agent system gets
information from multiple origins; it captures information from its
own sensors or, through some communication channels it may
receive messages issued by other agents. Based on this set
of basic information the agent then defines its beliefs and
performs actions [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. As long as it gets information, the agent
has to decide what it should believe and also which beliefs are
dropped [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. In order to decide which beliefs should hold,
the agent needs some criteria. A common criterion consists of
handling a reliability relation on its beliefs w.r.t. their origins
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. According to its opinion about the reliability of the
information source, the agent decides to adopt or not the
received piece of information. By keeping track of information
and its origin, agents can justify their beliefs: agent a believes
ϕ because agent b has provided ϕ and b is reliable [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This
explicit representation helps agents to enrich their dialogs: they
cannot only provide information but they may also mention
the third party at the origin of information. Let us consider
again agent a and information provided by b: a may then
ask b the underlying source of ϕ and a may then ask to this
source. Hence, this issue is a key one for trust characterization:
keeping track of agents involved in information broadcasting
enables agents to evaluate, from their own point of view,
whether they are all reliable, i.e. believable [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        The aim of this paper is to propose a modal framework
for representing agent’s belief state and its dynamics by
considering signed information, that is information associated
to its source. If many work has been made in order to show
how an agent can merge information issued from multiple
origins [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], very few work has focused on the explicit
representation of the origins of information [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] in the
context of BDI-based systems with communication actions.
But we advocate that this explicit representation is necessary
since it represents the underlying rationale of agents’ beliefs.
      </p>
      <p>
        The dynamics is usually described in terms of performative
actions based on KQML performatives [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] or speech acts
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Hereafter, we propose to consider tell actions as
private announcements from an agent (the sender of the
message) to another agent (the receiver of the message). Private
announcements enable to stress up how agents “restrict” their
belief state as they receive information. More precisely, they
shrink the space of information with their origins and then
according to that space, they build up their beliefs.
      </p>
      <p>The paper is structured as follows: In section II, we present
the intuitive meaning of signed information and belief state.
Next in section III, we present the technical details of the
modal logic framework. In section IV, we then represent an
intuitive and common policy for relating signed information
and belief which consists in the adoption as belief of all
consistent information. Next, in section V, we extend the
logical system with actions of the form “agent a tells to agent
b that a certain fact p is true”. We conclude the paper in
section VI by summing up the contribution and considering
some open issues.</p>
    </sec>
    <sec id="sec-2">
      <title>II. SETTING THE FRAMEWORK</title>
      <p>Handling the source of information leads to the notion of
signed statement, that is some statement is true according to
some source. From a semantics perspective, we want to be
able to represent, w.r.t. some initial state of affairs, for each
agent, what are the possible states that can be signed by each
source. Agents build their own belief state using information
signed by each source and the reliability of the source.</p>
      <sec id="sec-2-1">
        <title>Example 1 Suppose a car accident involving three cars which</title>
        <p>are blue (bc), red (rc) and yellow (yc). Now suppose a police
detective who is interviewing the witnesses of the accident.
Let po be the police detective. The first witness w1 tells to the
police detective that the blue car is responsible of the accident
while the second one (w2) states that the red car has caused
the collision. Both of them tell to the police detective that yc is
not responsible of the accident. In that context of information
gathering, the police detective does not need to assume that the
witnesses tell the truth or believe in information they provide.
The police detective just needs to assume that w1 provides
or signs information bc ∧ ¬rc ∧ ¬yc and w2 provides or
signs information ¬bc ∧ rc ∧ ¬yc. Next, based on these pieces
of information, the police detective will build his opinion,
i.e. his belief about the accident. The police detective faces
contradicting information about the blue and red cars, but
because the witnesses both agree about the yellow one, the
police detective should believe that the yellow car is not the
responsible of the collision. That is, the detective is willing to
root his belief upon the set of signed statements he handles.</p>
      </sec>
      <sec id="sec-2-2">
        <title>A. Representing signed statements</title>
        <p>Signed statements can be represented through Kripke
models using one accessibility relation per source of information.
Let Sign(b, p) be a modal operator stating that statement p is
true according to source b. Sign(b, p) is true in state w if p
holds in all states reachable from w through a relation denoted
Sb describing the possible information states issued from b.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Example 2 Let us consider the initial example. Informa</title>
        <p>tion which might be signed by the two witnesses are bc
and rc which leads to the signed statements Sign(w1, bc),</p>
        <sec id="sec-2-3-1">
          <title>Sign(w1, rc), Sign(w1, bc ∧ rc),... With respect to our exam</title>
          <p>ple, hereafter we will focus on the two signed statements
Sign(w1, bc ∧ ¬rc ∧ ¬yc) and Sign(w2, ¬bc ∧ rc ∧ ¬yc).</p>
        </sec>
      </sec>
      <sec id="sec-2-4">
        <title>B. Interpreting signed statements</title>
        <p>The aim is to represent formulas such as Bel(a, Sign(b, ϕ0))
or, in a more general way Bel(a, ϕ0), which respectively stands
for agent a believes that agent b signs ϕ0 and agent a believes
ϕ0. As for signatures, we use an accessibility relation denoted
Ba to represent the possible belief states of agent a.</p>
        <p>We assume that signed statements represent the rationales
for beliefs. That is, if agent a believes ϕ0 it is because some
signed statement Sign(b, ϕ0) holds in every possible belief
state of agent a and agent a is willing to commit to this
signed statement. Let a, b and c be three agents and p be
a propositional symbol; Figure 1 illustrates the possible belief
states of agent a w.r.t. some initial state w0 using accessibility
relation Ba (if p holds in a state, p is mentioned between
brackets). Agent a considers two possible belief states, w1 and
w2. In state w1, the two possible states given by Sb contain p
which entails that p is signed by b. On the other hand, the two
possible states given by Sc contain p and ¬p: no information
can be signed by agent c. In all states related to w2 with Sb and
Sc, p is true. From this figure we can conclude that in state w0
Sc
Sb
agent a believes that p is signed by b that is Bel(a, Sign(b, p))
while it does not believe that p or ¬p is signed by c. Since
p is signed by b and agent c says nothing about p, agent a
should believe p: Bel(a, p). Hence, it follows that in order to
prevent adoption of inconsistent statements, hereafter we will
assume that signed statements are always consistent (and thus
relation Sa is serial).</p>
        <p>
          Notice that the way we consider the link between beliefs
and signed statements differs from the way this link is defined
in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. That is, signed states are considered from each belief
state while C. Liau [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] considers informational states and
belief states in an independent way. This is due to the fact
that informational states in [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] reflect communication actions
while our notion of signed statement is more considered as an
epistemic notion.
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>Example 3 Let us pursue our motivating example. As men</title>
        <p>tioned, we assume that the detective is willing to adopt as
belief statements signed by the witnesses: Bel(po, Sign(w1, bc ∧
¬rc ∧ ¬yc)) and Bel(po, Sign(w2, ¬bc ∧ rc ∧ ¬yc)). Since both
witnesses agree on ¬yc, agent po also adopts as belief ¬yc.</p>
      </sec>
      <sec id="sec-2-6">
        <title>Meanwhile, he cannot set his belief about the two other cars since po faces contradicting signed statements.</title>
      </sec>
      <sec id="sec-2-7">
        <title>C. Preferences over information sources</title>
        <p>
          In order to know how to handle mutually inconsistent
signed statements, agents consider extra information stating
which signed statement they prefer. Agents may determine
themselves their preferences by considering the sources of
information [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], temporal aspects or the topics of the
statements [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>Sc
Sb
w0
w3
Ba
Ba</p>
        <p>Sb
b &lt; c (p) Sc
w1</p>
        <p>Sb
(p)
b &lt; cw2</p>
        <p>Sc Sc
w24</p>
        <p>
          In this paper, for the sake of conciseness and following
numerous contributions such as [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], we propose to consider extra
information about the reliability of sources of information as
illustrated by Figure 2. That is, we assume that the agents
consider information about only one topic. Consequently,
handling competencies or different kinds of reliability (such
as suggested in [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]) is out of the scope of the paper.
        </p>
        <p>That is, if agent a believes that b is more reliable than c,
then agent a adopts statement p as a belief even if agent c
has signed ¬p. Suppose that reliability is represented with the
help of a pre-order relation  (or &lt;): a  b stands for a is at
least as reliable as b. In semi formal terms, we get that:</p>
        <p>
          Bel(a, (Sign(b, p) ∧ Sign(c, ¬p) ∧ b &lt; c)) ⇒ Bel(a, p)
It follows that in each state, we do not only consider the value
of propositional symbols but also a pre-order relation which
characterizes a reliability order over information sources.
Using extra-information on reliability and by considering signed
statements rather than statements, the problem of belief change
[
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] is almost rephrased in terms close to the ones used in
belief merging [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ], [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]. Reliability order over sources of
information enables us to stratify signed information and then
by merging this stratified information in a consistent way the
agents get “justified” beliefs [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-8">
        <title>Example 4 Let us go on with our motivating example. Sup</title>
        <p>pose agent po considers that the first witness is at least as
reliable as the second one and he is himself willing to adopt
as belief the signed statements issued by the two witnesses,
i.e. we have the following belief:</p>
        <p>Bel(po, w1  w2  po)</p>
      </sec>
      <sec id="sec-2-9">
        <title>Hence, according to the previous semi formal axiom schema</title>
        <p>previously given, the police detective should believe that
the blue car (bc) has caused the accident. Notice that the
willingness attitude is translated in terms of preferences: po
has no opinion and considers as more important information
provided by w1 and w2.</p>
      </sec>
      <sec id="sec-2-10">
        <title>D. Representing tell statements</title>
        <p>
          Dynamics is viewed as restriction on agents’ belief states.
We interpret the tell performative as a private announcement
[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] rather than with help of actions and transitions between
states. A private announcement consists of an information
flow from one agent to a second one with a propositional
statement as content. Figure 3 illustrates how agent a’s belief
state changes after agent c tells p. According to this example,
after the performative T ell(c, a, p), agent a has restricted its
possible belief states to the states in which c signs p. In the
initial situation (the left part of the figure), at w0, agent a
believes Sign(b, p), does not believe Sign(c, p) and does not
believe p (since p does not hold in w2). After receiving agent
c’s message (right part of the figure), states where p is not
signed by c are no longer possible states for agent a and
thus, at w0, agent a believes Sign(b, p), Sign(c, p) and finally
also believes p. That is, the performative T ell(c, a, p) (agent
c tells to agent a that p is true) is responsible for updating
a’s beliefs in such a way that a believes that c signs p. In
other words, private announcements stress up the information
gathering aspect: possible worlds accessible through relation
Ba represent the ignorance of agent Ba and by shrinking
this set of possible believable worlds, we represent how agent
a gains information. Let us stress that this way of handling
the dynamics entails as a drawback that agent’s belief cannot
always be consistent: updating a model might lead to a model
where seriality cannot be guaranteed.
        </p>
        <p>Sc
Sb
w0
Ba
Ba
(p)
w1</p>
        <p>Sb
Sb Sc</p>
        <p>Sb Sc
w2 Sc
(p)
w11
(p)
w12
w21
w22
(p)
tell(c,a,p)</p>
        <p>Sc
Sb
w0
Ba
(p)
w1</p>
        <p>Sb
Sb Sc</p>
        <p>Sb Sc
w2 Sc
(p)
w11
(p)
w12
w21</p>
        <p>(p)
w22
Example 5 In the context of our motivating example, the
dynamics is represented by the sequence of interviews. For
instance, agent po interviews at first w1, action represented
by T ell(w1, po, bc ∧ ¬rc ∧ ¬yc) and then interviews the
second witness (T ell(w2, po, ¬bc ∧ rc ∧ ¬yc)). After these two
announcements, the detective believes: Bel(po, Sign(w1, bc ∧
¬rc ∧ ¬yc)) and Bel(po, Sign(¬bc ∧ rc ∧ ¬yc)).</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>III. FORMAL FRAMEWORK</title>
      <p>The proposed language for reasoning about signatures,
beliefs and preferences is a restricted first order language
which enables quantification over agent ids. In this section,
we focus on these three notions, tell actions will be introduced
later. Quantification allows agents to reason about anonymous
signatures. For the sake of conciseness, we restrict signed
statements to propositional statements. Let L0 be the
propositional language built over a set of propositional symbols P and
L be the logical language. Language L is based on doxastic
logic. Modal operator Bel represents beliefs: Bel(a, ϕ) means
agent a believes L-formula ϕ. Modal operator Sign represents
signed statements: Sign(t, ϕ0) means t (an agent id or a
variable of the agent sort) signs propositional statement ϕ0.
In order to represent agent’s opinion about reliability, we
introduce the notation a  b which stands for: agent a is
said to be at least as reliable as b.</p>
      <p>Definition 1 (Syntax of L) Let P be a finite set of
propositional symbols. Let A be a finite set of agent ids. Let V be a
set of variables s.t. A ∩ V = ∅. Let T = A ∪ V be the set of
agent terms. The set of formulas of the language L is defined
by the following BNF:
ϕ ::= p | ¬ϕ | ϕ ∧ ϕ | Sign(t, ϕ0) | Bel(a, ϕ) | ∀xϕ | t  t′
Writing a &lt; b stands for a is strictly more reliable than b:
a  b ∧ ¬(b  a). Writing a ∼ b means that a and b are
equally reliable. Operators → and ∃ are used according to
their usual meaning.</p>
      <sec id="sec-3-1">
        <title>A. Semantics</title>
        <p>
          The semantics of L-formulas is defined in terms of
possible states and relations between states [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. Those relations
respectively represent the notion of signatures and beliefs. In
each state, propositional symbols are interpreted and total
preorders representing agents’ reliability are set.
        </p>
        <p>Definition 2 (Model) Let M be a model defined as a tuple:
)W,  Si,  Bi, I, *+
i∈A</p>
        <p>i∈A
where W is a set of possible states. Si ∈ W × W is an
accessibility relation representing signatures, Bi ∈ W × W is
an accessibility relation representing beliefs. I is an
interpretation function of the propositional symbols w.r.t. each possible
state, I : W × P -→ {0, 1}. * is a function which represents
total pre-orders; these pre-orders are specific to each state,
that is *: W -→ 2A×A.</p>
        <p>A variable assignment is a function v which maps every
variable x to an agent id. A t-alternative v′ of v is a variable
assignment similar to v for every variable except t. For t ∈ T ,
[[t]]v belongs to A and refers to the assignment of agent terms
w.r.t. variable assignment v, such that:
if t ∈ A then [[t]]v = t
if t ∈ V then [[t]]v = v(t)
We define the satisfaction relation |= with respect to some
model M , state w and variable assignment v as follows.
Definition 3 (|=) Let M be a model and v be a variable
assignment: v : V → A. M satisfies an L-formula ϕ w.r.t.
a variable assignment v and a state w, according to the
following rules:
• M, v, w |= t  t′ iff ([[t]]v, [[t′]]v) ∈* (w).
• M, v, w |= p iff p ∈ P and I(w, p) = 1.
• M, v, w |= Sign(t, ϕ0) iff M, v, w′ |= ϕ0 for all w′ s.t.</p>
        <p>(w, w′) ∈ S[ t] v
• M, v, w |= Bel(a, ϕ) iff M, v, w′ |= ϕ for all w′ s.t.</p>
        <p>(w, w′) ∈ Ba
• M, v, w |= ∀tϕ iff for every t-alternative v′, M, v′, w |=
ϕ.</p>
        <p>We write |= ϕ iff for all M , w and v, we have M, v, w |= ϕ.
The semantics for operators ¬, →, ∨, ∧ and ∃ is defined
in the standard way. Let us now detail the constraints that
should operate on the model. We only require that signature
has to be consistent which entails that all relations Si have
to be serial. Belief operator is a K45 operator and thus all
Bi are transitive and euclidian. Interwoven relations between
signatures and beliefs are detailed in the next section.</p>
      </sec>
      <sec id="sec-3-2">
        <title>1) Constraining the Reliability Relations: We assume that</title>
        <p>
          every agent holds belief about reliability without any
uncertainty. That is, agent’s beliefs about reliability can be
represented as a total pre-order. However, it does not mean
that we consider a fixed notion of reliability: we propose to
handle multiple pre-orders by indexing reliability with worlds.
That is, in each possible world or believable world, an agent
considers how it ranks the agents. In that context, each rank is
considered as a possible rank and thus it is natural that each of
them should be total. However, we enforce a stronger notion
(KS) Sign(a, ϕ0 → ψ0) → (Sign(a, ϕ0) → Sign(a, ψ0))
(DS) Sign(a, ϕ0) → ¬Sign(a, ¬ϕ0)
(KB) Bel(a, ϕ → ψ) → (Bel(a, ϕ) → Bel(a, ψ))
(4B) Bel(a, ϕ) → Bel(a, Bel(a, ϕ))
(5B) ¬Bel(a, ϕ) → Bel(a, ¬Bel(a, ϕ))
(R) t  t
(T r) t  t′ ∧ t′  t′′ → t  t′′
(T) t  t′ ∨ t′  t
(T o) Bel(a, t  t′) ∨ Bel(a, t′  t)
(M P ) From ϕ and ϕ → ψ infer ψ
(G) From ϕ infer ∀tϕ
(NS) From ϕ0 infer Sign(t, ϕ0)
(NB) From ϕ infer Bel(a, ϕ)
of totality which states that the aggregation of all believable
ranks over agents (which are total) leads to a total preorder.
This will then help the agent to integrate all signed statements.
In other words, we require that the integration (or merging)
of signed statements should be based on an underlying total
preorder over statements (as it is commonly assumed in the
belief revision and merging areas—see [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]). In
terms of constraints on states and relations between them, it
means that:
1) for all states w, t * (w) t′ or t′ * (w) t and,
2) suppose wBiw′ and t * (w′) t′, then for all states w′′
s.t. wBiw′′, t * (w′′) t′.
        </p>
        <p>The first constraint enforces that pre-orders are total in all
states; the second constraint expresses that totality should hold
in all belief states. Moreover, preorder definition entails that
reflexivity and transitivity hold.</p>
      </sec>
      <sec id="sec-3-3">
        <title>B. Axiomatics</title>
        <p>Let us now translate these constraints in terms of proof
theory. Axiomatization of logic L includes all tautologies of
propositional calculus. Table I details the axioms and inference
rules describing the behavior of belief, signed statement and
reliability. Notice axiom schema (T o) which reflects that
reliability relations have to be believed as total. Let ⊢ denotes the
proof relation. We conclude by giving results about soundness
and completeness.</p>
        <sec id="sec-3-3-1">
          <title>Theorem 1 Logical system L is sound and complete1.</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>IV. LINKING SIGNATURES AND BELIEFS</title>
      <p>
        There are multiple ways to switch from information to
beliefs. These different ways may follow principles issued
from the belief merging principle [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] or epistemic
attitudes such as trust [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. As previously mentioned,
we do not require that an agent has to believe that others
1In this paper all proofs have been skipped; however a longer version of the
paper with all proofs is downloadable at the URL http://www.irit.fr/∼Laurent.
      </p>
      <p>Perrussel/lads2010-long.pdf.
believe in information they provide. This is a key issue when
information is propagated from one agent to another. At some
stage, an agent may just broadcast some information without
committing to that information in terms of belief.</p>
      <p>A common and rational way to proceed is to consider as
belief all non mutually inconsistent signed statements. All
signed statements are considered in an incremental way, that
is “ from the most reliable to the less reliable statements”.
To describe the signed statements adoption stage, we first
characterize agents which are equally reliable. Agents can be
ranked since we always consider a total preorder; agents which
are equally can be gathered in a same group. Each group can
then be ranked. Let us at first characterize the most reliable
set of agents; this set is denoted as C1:</p>
      <p>a ∈ C1 =def ∀t(a  t)
The formula characterizing members of C1 can then be used
for characterizing membership to a set Ci such that i &gt; 1.</p>
      <p>a ∈ Ci =def (¬(a ∈ Ci−1) ∧ ∀t¬(t ∈ Ci−1)) → (a  t)
Hence, all agents belonging to a set Ci are equally reliable and
for all a ∈ Ci, b ∈ Cj if i &lt;N j then a  b. Next, the following
definition stands for each agent tk belonging to some specific
set Ci believes statement ϕk:</p>
      <p>0
Sign(tk, φ0k) =def</p>
      <p>(tk ∈ Ci) → Sign(tk, φ0k)
tk∈Ci</p>
      <p>tk∈A
Using these shortcuts, we can now describe the merging
process. The following axiom states that if a propositional
statement ϕ0 is believed by agent a if the conjunction of the
statements signed by the agents belonging to the same set Ci
entails ϕ0 is believed by agent a (line 1), if statement ¬ϕ0 is
not already believed by a(line 2) and ¬ϕ0 cannot be entailed
with the help of statements signed by agents which are at least
as reliable as agents belonging to Ci (line 3).</p>
      <p>(Bel(a,</p>
      <p>Sign(tk, ϕ0k)) ∧ Bel(a,
ϕ0k → ϕ0)∧
tk∈Ci
(
0&lt;j&lt;i
¬Bel(a,
tl∈Cj
¬Bel(a, ¬ϕ0)∧</p>
      <p>Sign(tl, ϕl0) ∧</p>
      <p>ϕl0 → ¬ϕ0)))
→ Bel(a, ϕ0) (IB)
In terms of semantics, it means that, w.r.t. some initial state
w0, all belief states are related to some signed states. Hence, it
requires to consider the state’s interpretation, that is to express
the relation by using worlds, i.e. a state and its associated
interpretation. We represent a world as a set of propositional
symbols, symbols that hold in the associated state. Let w be
a state and [w] the associated world:</p>
      <p>[w] = {p|I(w, p) = 1}
In a more general way, if W is a set of states, then [W ] denotes
the set of associated worlds. At first, from the belief states,
we rank agent ids based on reliability relations believed by the
agent. Suppose an agent a and a world w0; using relation Ba,
we extract the total preorder representing reliability relation
believed by agent a at w0. Notice that the constraints shown
section III-A1 ensures that this preorder is total and thus agents
ids could be ranked for building a partition of set of agents.
Let C be a partition of A such that in every set Ci of C, all
agents are equally reliable and for all a ∈ Ci, b ∈ Cj if i &lt;N j
then a ≺ b. Second, from each set Ci, we consider common
information, that is statements that are signed by every agents
belonging to Ci. Let [Ci]w0 be the set of worlds commonly
signed by all agents belonging to Ci and related to w0:
[Ci]w0 =
! {[w] | (w0, w) ∈ Sa}
a∈Ci
Next all sets of worlds [Ci]w0 are merged in a consistent
way, the resulting set of worlds is denoted as reliable worlds.
By consistent way, we mean an incremental process which
considers as reliable worlds at first the whole set of possible
worlds [W ]. Next for each part Ci, the set of reliable worlds
is intersected with [Ci]w0 only if it does not lead to an empty
set, i.e. an inconsistent result.</p>
      <p>Definition 4 (Reliable worlds) Let M be a model and w0 a
state such that w0 ∈ W . The set of reliable worlds Ωw0 is
defined in an incremental way such that:
• Ω0 = [W ]
• Ωi = Ωi−1 ∩ [Ci]w0 if Ωi−1 ∩ [Ci]w0 1= ∅ and i &gt; 0
• Ωi = Ωi−1 if Ωi−1 ∩ [Ci]w0 = ∅ and i &gt; 0
The resulting set Ωw0 is equal to Ωk such that k = |C|.
Since the sets of worlds and agents are finite, we do not
have to consider the infinite case. Reliable worlds represent
information that should be actually believed. Let us consider
agent a and an initial world w0; from w0, we extract the belief
states, and from these belief states, the set of reliable worlds.
Beliefs of agent a are rational if all its believed worlds are
included in its set of reliable worlds:</p>
      <p>
(w0,w)∈Ba</p>
      <p>In the previous section, we have detailed a policy for
building belief based on signed information. This policy considers
belief and signed information from a static point of view. Let
us now consider a more dynamic view by introducing actions
of the form “agent a tells to agent b that a certain fact ϕ0
is true” (alias tell actions). This kind of action ensures that
agent b will believes that agent a signs p, that is, a tell action is
responsible for updating an agent’s beliefs about other agents’
signatures and, consequently, for the agent’s acquisition of
new information and for updating the agent’s beliefs about
objective facts. We note tell actions by Tell(a, b, ϕ0). Let
LT be the extended language which embeds tell statements.</p>
      <sec id="sec-4-1">
        <title>Definition 5 (Syntax of LT ) The set of formulas of the language LT is defined by the following BNF:</title>
        <p>ϕ ::=p | ¬ϕ | ϕ ∧ ϕ′ | Sign(t, ϕ0) | Bel(a, ϕ) |</p>
        <p>∀xϕ | t  t′ | [Tell(a, b, ϕ0)]ϕ
In other terms, LT just extends L with dynamic
operators [Tell(a, b, ϕ0)]. The intuitive meaning of statement
[Tell(a, b, ϕ0)]ϕ is after a tells ϕ0 to b, ϕ holds.</p>
        <p>
          The truth conditions are those given above for the formulas
p, ¬ϕ, ϕ ∧ ϕ′, Sign(t, ϕ0), Bel(a, ϕ), ∀xϕ, t  t′ and
[Tell(a, b, ϕ0)]ϕ. The truth condition for [Tell(a, b, ϕ0)]ϕ is
defined in a way which is closed to the semantics of dynamic
epistemic logic [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ]. More precisely, after agent a tells to agent
b information ϕ0, agent b removes from its belief state all
states in which agent a does not sign ϕ0. Therefore, after
agent a tells to agent b information ϕ0, agent b believes that
agent a signs ϕ0. In our framework, a tell action of agent
a (the sender) towards agent b (the receiver) that ϕ0 is true
is considered as a private announcement in the sense of [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ],
that is, after agent a tells to agent b information ϕ0, only agent
b’s belief state should change whereas the belief states of the
other agents are not changed. In other words, a tell action
Tell(a, b, ϕ0) characterizes a private communication from a
sender to a specific receiver of the sender’s message, where
the content of the speaker’s message is nothing else than the
content of the speaker’s signature (i.e. Sign(t, ϕ0)).
Definition 6 (Announcement Semantics) Let M =
)W, "i∈A Si, "i∈A Bi, I, *+ be a model and let w be
a state in W . We have:
        </p>
        <p>• M, v, w |= [Tell(a, b, ϕ0)]ϕ iff M |#a,b,ϕ0$, v, w1 |= ϕ.
M |#a,b,ϕ0$ = )W ∗, "i∈A Si∗, "i∈A Bi∗, I∗, *∗+ is defined as
follows:
• W ∗ = {w1|w ∈ W } ∪ {w2|w ∈ W };
• Bb∗ =
{(w1, w1′)|(w, w′) ∈ Bb and M, v, w′ |= Sign(a, ϕ0)} ∪
{(w2, w2′)|(w, w′) ∈ Bb};
• Bi∗ = {(w1, w2′)|(w, w′) ∈ Bi}∪</p>
        <p>{(w2, w2′)|(w, w′) ∈ Bi} for all i ∈ A such that i 1= b;
• Si∗ = {(w1, w2′)|(w, w′) ∈ Si}∪</p>
        <p>{(w2, w2′)|(w, w′) ∈ Si} for all i ∈ A;
• *∗ (w1) =*∗ (w2) =* (w) for all w ∈ W ;
• I∗(w1, p) = I∗(w2, p) = I(w, p) for all w ∈ W .
Basically, the effect of a’s action of telling to b that ϕ0 is
to shrink the set of belief accessible states for b to the states
in which a signs ϕ0, while keeping constant the set of belief
accessible states for all other agents. Note that a’s action of
(TAP ) [Tell(a, b, ϕ0)]p ↔ p
(TN ) [Tell(a, b, ϕ0)]¬ϕ ↔ ¬[Tell(a, b, ϕ0)]ϕ
(TC ) [Tell(a, b, ϕ0)](ϕ ∧ ϕ′) ↔</p>
        <p>([Tell(a, b, ϕ0)]ϕ ∧ [Tell(a, b, ϕ0)]ϕ′)
(TB) [Tell(a, b, ϕ0)]Bel(b, ϕ) ↔</p>
        <p>Bel(b, (Sign(a, ϕ0) → [Tell(a, b, ϕ0)]ϕ))
(TB= ) [Tell(a, b, ϕ0)]Bel(i, ϕ) ↔ Bel(i, ϕ) if i %= b
(TS) [Tell(a, b, ϕ0)]Sign(t, ϕ′0) ↔ Sign(t, ϕ′0)
(T≤) [Tell(a, b, ϕ0)](t ≤ t′) ↔ (t ≤ t′)
(T∀) [Tell(a, b, ϕ0)]∀xϕ ↔ ∀x[Tell(a, b, ϕ0)]ϕ
telling to b that ϕ0 also keeps constant agents’ signatures and
the reliability order over agents.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Theorem 3 If M is a L-model then M |#a,b,ϕ0$ is also a L</title>
        <p>model.</p>
        <p>Theorem 4 If M is a L-model in which constraint IB holds
then M |#a,b,ϕ0$ is also a L-model in which constraint IB holds.</p>
        <p>Let us now focus on the axiomatics of the logic LT . Table II
details the reduction axioms describing the behavior of the
operator [Tell(a, b, ϕ0)]. (TAP ) denotes the atomic
permanence, (TN ) denotes negation handling, and (TC ) denotes
conjunction handling. (TB ) describes the interplay between
a tell action and the beliefs of the message receiver. (TB= )
describes the interplay between a tell action and the beliefs of
all agents different from the message receiver. In particular,
(TB= ) highlights the permanence of the beliefs of all agents
different from the message receiver. (TS ) describes signature
permanence, (T≤) describes preferences permanence, and
(T∀) describes the interplay between tell action and
quantification over variable assignments.</p>
        <sec id="sec-4-2-1">
          <title>Theorem 5 The schemata in table II are valid.</title>
          <p>We then state the theorem about completeness of the logic LT .</p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>Theorem 6 The logic LT is completely axiomatized by prin</title>
        <p>ciples of the logic L together with the schemata in Table II
and the rule of replacement of proved equivalence.</p>
        <p>We write ⊢T to denote the proof relation for the logic LT
determined by the principles of the logic L, the schemata in
table II and the rule of replacement of proved equivalence.</p>
        <p>For instance, the following theorem of the logic LT captures
the essential aspect of the tell action. It says that, after agent
a tells to agent b information ϕ0, agent b believes that agent
a signs ϕ0:</p>
        <p>⊢T [Tell(a, b, ϕ0)]Bel(b, Sign(a, ϕ0))
Once agent b starts to believe that agent a signs ϕ0 (as an effect
of a’s act of telling to b that ϕ0), agent b might also start to
believe that ϕ0. As we have shown above, this depends on the
reliability of agent a according to agent b and on principles
linking signatures with beliefs such as principle (IB).</p>
        <sec id="sec-4-3-1">
          <title>Example 6 Let us go back to our initial example and let us</title>
          <p>represent in the system ⊢T , how agent po concludes that the
blue car has caused the collision. At first, assume that (IB).</p>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>Second, assume the following preferences: Bel(po, w1  w2).</title>
        <sec id="sec-4-4-1">
          <title>Then it follows that after the two announcements (we focus on the blue car), preferences are unchanged:</title>
          <p>⊢T [Tell(w1, po, bc)][Tell(w2, po, ¬bc)]Bel(po, w1  w2)</p>
        </sec>
        <sec id="sec-4-4-2">
          <title>And the detective believes the received information</title>
          <p>⊢T [Tell(w1, po, bc)][Tell(w2, po, bc)]
Bel(po, Sign(w1, bc) ∧ Sign(w2, ¬bc))</p>
        </sec>
        <sec id="sec-4-4-3">
          <title>Finally, axiom (IB) entails that</title>
          <p>⊢T [Tell(w1, po, ¬bc∧rc)][Tell(w1, po, bc∧¬rc)]Bel(po, bc)</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>VI. CONCLUSION</title>
      <p>
        In this paper we have shown how information and its
source can be processed by an agent so that at first, it just
acquires information from sensors or other agents and second,
it builds its belief state by considering signed information. By
splitting information and belief, an agent is able to handle
clear rationales to construct its belief state both from a
static and dynamic perspectives. From a static perspective we
have applied our formal framework to characterize a possible
attitude for agents in the process of building their belief
state from the basic signed information they hold. From this
perspective this work is close to what has been done in belief
merging [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The key difference with existing work
in the belief merging area is the introduction of merging in
a modal based framework at first (this is also a common
characteristic with [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]); second a clear distinction between
belief and signed statement and third a dynamic view on
belief construction. These last two characteristics differ in
two ways from existing work [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]: (i) it is usually
assumed that belief and information are almost similar; we
have shown that we do not have to assume this hypothesis;
(ii) beliefs are almost not viewed as a primitive concepts but
rather as the result of some information processing which gives
a flexible framework (e.g. axiom IB). Our work is also related
to the work of [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] in which agents’ mental attitudes and
agent’s ostensible (expressed) attitudes are distinguished and a
formalism capturing this distinction is proposed. In particular,
our notion of signed information is close to the notion of
ostensible belief of Nickles et al. However, Nickles et al.
do not consider reliability of information sources. Moreover,
their approach does not deal with dynamics of information by
means of communicative actions. The latter is a central aspect
of our proposal (see Section V).
      </p>
      <p>
        Concerning the dynamic perspective we have shown how
the basic signed information held by an agent may change as
it receives tell statements from another agent processed in a
similar way to private announcements in the sense of dynamic
epistemic logic (DEL) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
      <p>
        Our short term goal is to consider more sophisticated ways
to set the reliability relations. That is, our aim is to consider
agent skills [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] so that agent can consider multiple reliability
relations at the same time. At this time, even if agent can
consider multiple alternative reliability relations, they cannot
mixed them. Our goal is to avoid this limit.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Rao</surname>
          </string-name>
          and M. Georgeff, “
          <article-title>Modeling rational agents within a bdiarchitecture,”</article-title>
          <source>in Proc. of KR'91</source>
          ,
          <year>1991</year>
          , pp.
          <fpage>473</fpage>
          -
          <lpage>484</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ga</surname>
          </string-name>
          <article-title>¨rdenfors, Knowledge in flux: Modeling the Dynamics of Epistemic States</article-title>
          . MIT Press,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Herzig</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Longin</surname>
          </string-name>
          , “
          <article-title>Belief dynamics in cooperative dialogues</article-title>
          ,
          <source>” J. of Semantics</source>
          , vol.
          <volume>17</volume>
          , no.
          <issue>2</issue>
          ,
          <year>2000</year>
          , vol. published in
          <year>2001</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dragoni</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Giorgini</surname>
          </string-name>
          , “
          <article-title>Revising beliefs received from multiple sources,” in Frontiers of Belief Revision, Applied Logic</article-title>
          . Kluwer,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Cholvy</surname>
          </string-name>
          , “
          <article-title>A modal logic for reasoning with contradictory beliefs which takes into account the number and the reliability of the sources,” in Proc of ECSQARU'05, ser</article-title>
          .
          <source>LNCS</source>
          , vol.
          <volume>3571</volume>
          . Springer,
          <year>2005</year>
          , pp.
          <fpage>390</fpage>
          -
          <lpage>401</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Liau</surname>
          </string-name>
          , “
          <article-title>Belief, information acquisition, and trust in multi-agent systems-a modal logic formulation</article-title>
          ,
          <source>” Artificial Intelligence</source>
          , vol.
          <volume>149</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>31</fpage>
          -
          <lpage>60</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>E.</given-names>
            <surname>Lorini</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Demolombe</surname>
          </string-name>
          , “
          <article-title>From Binary Trust to Graded Trust in Information Sources: A Logical Perspective,” in Trust in Agent Societies, ser</article-title>
          .
          <source>LNAI</source>
          . Springer-Verlag,
          <year>2008</year>
          , vol.
          <volume>5396</volume>
          , pp.
          <fpage>205</fpage>
          -
          <lpage>225</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W. van der Hoek and M.</given-names>
            <surname>Wooldridge</surname>
          </string-name>
          , “
          <article-title>Towards a logic of rational agency</article-title>
          ,
          <source>” Journal of the IGPL</source>
          , vol.
          <volume>11</volume>
          , no.
          <issue>2</issue>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>157</lpage>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Perrussel</surname>
          </string-name>
          and
          <string-name>
            <surname>J.</surname>
          </string-name>
          <article-title>The´venin, “(Dis)Belief Change based on Messages Processing,”</article-title>
          <source>in Proc. of CLIMA'IV</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10] --, “
          <article-title>A logical approach for describing (dis)belief change</article-title>
          and message processing,”
          <source>in Proc. of AAMAS'04</source>
          . IEEE C.S.,
          <year>2004</year>
          , pp.
          <fpage>614</fpage>
          -
          <lpage>621</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Finin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Labrou</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J.</given-names>
            <surname>Mayfield</surname>
          </string-name>
          , “
          <article-title>KQML as an agent communication language,” in Software Agents</article-title>
          , J. Bradshaw, Ed. MIT Press,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Cohen</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Levesque</surname>
          </string-name>
          , “
          <article-title>Rational Interaction as the Basis for Communication,</article-title>
          ” in Intentions in Communication, P. Cohen,
          <string-name>
            <given-names>J.</given-names>
            <surname>Morgan</surname>
          </string-name>
          , and M. Pollack, Eds. MIT Press,
          <year>1990</year>
          , pp.
          <fpage>221</fpage>
          -
          <lpage>256</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13] --,
          <source>“Communicative actions for artificial agents,” in Proceedings of the First International Conference on Multi-Agent Systems (ICMAS'95)</source>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Lesser</surname>
          </string-name>
          and L. Gasser, Eds. San Francisco, CA, USA: The MIT Press,
          <year>1995</year>
          , pp.
          <fpage>65</fpage>
          -
          <lpage>72</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L. F.</given-names>
            <surname>del Cerro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herzig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Longin</surname>
          </string-name>
          , and
          <string-name>
            <given-names>O.</given-names>
            <surname>Rifi</surname>
          </string-name>
          , “
          <article-title>Belief reconstruction in cooperative dialogues,” in Proc. of AIMSA'98, ser</article-title>
          . LNCS,
          <string-name>
            <given-names>F.</given-names>
            <surname>Giunchiglia</surname>
          </string-name>
          , Ed., vol.
          <volume>1480</volume>
          . Springer,
          <year>1998</year>
          , pp.
          <fpage>254</fpage>
          -
          <lpage>266</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>L.</given-names>
            <surname>Cholvy</surname>
          </string-name>
          , “
          <article-title>Automated reasoning with merged contradictory information whose reliability depends on topics,” in Proc. of ECSQARU'95, ser</article-title>
          . LNCS,
          <string-name>
            <given-names>C.</given-names>
            <surname>Froidevaux</surname>
          </string-name>
          and J. Kohlas, Eds., vol.
          <volume>946</volume>
          ,
          <year>1995</year>
          , pp.
          <fpage>125</fpage>
          -
          <lpage>132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Liberatore</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Schaerf</surname>
          </string-name>
          , “
          <article-title>Arbitration (or how to merge knowledge bases</article-title>
          ),
          <source>” IEEE Transactions on Knowledge and Data Engineering</source>
          , vol.
          <volume>10</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>76</fpage>
          -
          <lpage>90</lpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Konieczny</surname>
          </string-name>
          and R. Pe´rez, “
          <article-title>Propositional belief base merging or how to merge beliefs/goals coming from several sources and some links with social choice theory,”</article-title>
          <source>European Journal of Operational Research</source>
          , vol.
          <volume>160</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>785</fpage>
          -
          <lpage>802</lpage>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Benferhat</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Garcia</surname>
          </string-name>
          , “
          <article-title>Handling locally stratified inconsistent knowledge bases,” Studia Logica</article-title>
          , vol.
          <volume>70</volume>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>104</lpage>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>H. van Ditmarsch</surname>
            , W. van der Hoek, and
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Kooi</surname>
          </string-name>
          , Dynamic Epistemic Logic, ser.
          <source>Synthese Library</source>
          . Springer,
          <year>2007</year>
          , vol.
          <volume>337</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Fagin</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y. M. J.</given-names>
            <surname>Halpern</surname>
          </string-name>
          ,
          <article-title>Reasoning About Knowledge</article-title>
          . MIT Press,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>H.</given-names>
            <surname>Katsuno</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Mendelzon</surname>
          </string-name>
          , “
          <article-title>Propositional knowledge base revision and minimal change</article-title>
          ,
          <source>” Artificial Intelligence</source>
          , vol.
          <volume>52</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>263</fpage>
          -
          <lpage>294</lpage>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Gerbrandy</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Groeneveld</surname>
          </string-name>
          , “
          <article-title>Reasoning about information change</article-title>
          ,
          <source>” J. of Logic, Language and Information</source>
          , vol.
          <volume>6</volume>
          , no.
          <issue>2</issue>
          ,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>M.</given-names>
            <surname>Nickles</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fischer</surname>
          </string-name>
          , and G. Weiss, “
          <article-title>Communication attitudes: A formal approach to ostensible intentions, and individual</article-title>
          and group opinions,”
          <source>in Proc. of LCMAS'05, ser. Electronic Notes in Computer Science</source>
          , vol.
          <volume>157</volume>
          (
          <issue>4</issue>
          ). Elsevier,
          <year>2005</year>
          , pp.
          <fpage>95</fpage>
          -
          <lpage>115</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>