<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Stochastic Belief Change Framework with an Observation Stream and Defaults as Expired Observations</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gavin Rens</string-name>
          <email>gavinrens@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Centre for Artificial Intelligence Research, University of KwaZulu-Natal, School of Mathematics</institution>
          ,
          <addr-line>Statistics and Computer Science and CSIR Meraka</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>A framework for an agent to change its probabilistic beliefs after a stream of noisy observations is received is proposed. Observations which are no longer relevant, become default assumptions until overridden by newer, more prevalent observations. A distinction is made between background and foreground beliefs. Agent actions and environment events are distinguishable and form part of the agent model. It is left up to the agent designer to provide an environment model; a submodel of the agent model. An example of an environment model is provided in the paper, and an example scenario is based on it. Given the particular form of the agent model, several 'patterns of cognition' can be identified. An argument is made for four particular patterns.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>MOTIVATION</title>
      <p>My intention with this research is to design a framework with which
an agent can deal with uncertainty about its observations and actions,
and refresh its beliefs in a relatively sophisticated way.</p>
      <p>
        Partially observable Markov decision processes (POMDPs) [
        <xref ref-type="bibr" rid="ref1 ref23 ref25">1, 25,
23</xref>
        ] are adequate for many stochastic domains, and they have the
supporting theory to update agents’ belief states due to a
changing world. But POMDPs are lacking in two aspects with respect
to intelligent agents, namely, (i) the ability to maintain and
reason with background knowledge (besides the models inherent in
POMDP structures) and (ii) the theory to revise beliefs due to
information acquisition. Traditionally, belief update consists of
bringing an agent’s knowledge base up to date when the world described
by the knowledge base changes, that is, it is a ‘change-recording’
operation, whereas belief revision is used when an agent obtains new
information about a static world, that is, it is a ‘knowledge-changing’
operation [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. I shall use the generic term belief change to including
belief update and belief revision.
      </p>
      <p>
        A different perspective on the proposed framework is from the area
of classical belief change [
        <xref ref-type="bibr" rid="ref12 ref13 ref17">12, 13, 17</xref>
        ]. The belief change
community has not given much attention to dealing with uncertainty,
especially not stochastic uncertainty. Hence, integrating POMDP theory
into belief change methods could be beneficial. And besides
bringing probability theory, POMDPs also bring decision theory, that is,
the theory for reasoning about actions and their utility.
      </p>
      <p>However, I go one or two steps farther than a straightforward
integration of POMDPs and belief change theory. The framework also
includes the notion of a stream of observations, the modeling of the
decay of the truthfulness of individual observations, and how to
integrate ‘expired’ and ‘prevalent’ observations into the agent’s beliefs.
More precisely, observations which are no longer immediately
relevant become default assumptions until overridden by newer, more
prevalent observations. Consider the following three scenarios.
Scenario 1 An acquaintance (Bianca) shows you the new,
expensive iSung-8 smartphone she bought, with a very particular cover
design. A year later, you visit Bianca at her home, where only her
teenage son lives in addition. You see two phones on the kitchen
counter, an iSung-8 with the particular cover design you remember
from a year ago and an iSung-9. Is the likelihood greater that the
iSung-8 or the iSung-9 is Bianca’s?
Scenario 2 You are at the airport in a city away from home and
you expect to land in your home city (Cape Town) in three hours’
time. You hear someone waiting for the same flight say ‘It is raining
in Cape Town’. Is the likelihood less that it will still be raining if
your flight is delayed by 3 hours than if the flight was not delayed?
Scenario 3 Your neighbour tells you he needs to visit the dentist
urgently. You know that he uses the dentist at the Wonder-mall. A
month later, you see your neighbour at the Wonder-mall. Is he there
to see the dentist?</p>
      <p>What these three scenarios have in common is that the answers to
the questions make use of the persistence of truth of certain pieces of
information. After a period has elapsed, the veracity of some kinds of
information dissipates. For instance, in Scenario 1, one might attach
an ‘expiry date’ to the information that the particular iSung-8 phone
is Bianca’s. So, by the time you visit her, the truth of that information
is much weaker, in fact, it has become defeasible by then. Hence,
we may easily argue that Bianca gave her old phone to her son and
she bought the iSung-9 for herself. However, if you had visited her
one month after she showed you her new iSung-8 and you saw it
together with the newer iSung-9 on the counter, you would probably
rather assume that the iSung-9 was Bianca’s son’s. In Scenario 2,
you could expect it to be raining in Cape Town when you get there in
three hours (because spells of rain usually last for four hours in Cape
Town), but if your flight is delayed, there will be no rain or only
drizzle when you land. The information ‘It is raining in Cape Town’
has a lifespan of four hours. With respect to Scenario 3, one would
expect a person who says they must visit the dentist urgently to visit
the dentist within approximately seven days. So your neighbour is
probably not at Wonder-mall to see the dentist. Hence ‘Neighbour
must visit dentist’ should be true for no longer than seven days, after
which, the statement becomes defeasibly true.</p>
      <p>In this paper, I attempt to formalise some of these ideas. Several
simplification are made; two main simplifications are (i) all
information packets (evidence/observations) have a meaningful period for
which they can be thought of as certainly true and (ii) the transition
of a piece of information from certainly true to defeasibly true is
immediate. Consider the following pieces of information.</p>
      <sec id="sec-1-1">
        <title>1. Bianca is an acquaintance of mine.</title>
        <p>2. It will rain in Cape Town this week.
3. My dentist is Dr. Oosthuizen.</p>
        <p>The first statement is problematic because, for instance, Bianca
might gradually become a friend. With respect to the second
statement, it is easy to set the ‘expiry period’ to coincide with the end of
the week. One might feel that it is difficult to assign a truth period to
‘My dentist is Dr. Oosthuizen’ due to lack of information. A person
typically does not have the same dentist life-long, but one can usually
not predict accurately when one will get a new dentist. On the other
hand, if, for instance, one knows exactly when one is moving to a
new city, then one can give a meaningful truth period, and the
transition of the piece of information from certainly true to defeasibly true
is immediate.</p>
        <p>
          Many of these issues are studied in temporal logics [
          <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
          ]. The
focus of the present work, however, is more on belief change with a
simple temporal aspect (and the integration of POMDP theory). One
may do well in future research to attempt combining the results of
the present work with established work on temporal logics. One
paper particularly relevant to the present work presents a probabilistic
temporal logic capable of modeling reasoning about evidence [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
        </p>
        <p>Expired observations are continually aggregated into the agent’s
set of default assumptions. Prevalent (unexpired) observations
remain in the agent’s ‘current memory stream’. Whenever the agent
wants to perform some reasoning task, it combines the prevalent
observations with its fixed beliefs, then modifies its changed beliefs
with respect to its default assumptions, and reasons with respect to
this final set of beliefs. However, the agent always reverts back to the
original fixed beliefs (hence, “fixed”). The default assumptions keep
changing as memories fade, that is, as observations expire.</p>
        <p>The rest of this paper unfolds as follows. In the next section, I
review the three formalisms on which the framework is mainly based,
namely, partial probability theory, POMDPs and the hybrid
stochastic belief change framework. Section 3 presents the formal definition
of the framework and Section 4 explains how the framework
components interact and change when used for reasoning. A discussion
about the possible patterns of cognition within the framework is
presented in Section 5. Then Section 6 provides an extensive example,
showing some of the computations which would be required in
practice. The paper ends with some final remarks and pointers to related
work.
2</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>FORMAL FOUNDATIONS</title>
      <p>Let L be a finite classical propositional language. A world is a
logical model which evaluates every propositional variable to true or
false, and by extension, evaluates every propositional sentence in L
to true or false. Given n propositional atoms, there are 2n
conceivable worlds. Let W be a set of possible worlds – a subset of the
conceivable worlds. The fact that w 2 W satisfied 2 L (or is a
model for ) is denoted by w .</p>
      <p>Let a belief state b be a probability distribution over all the worlds
in W . That is, b : W ! [0; 1], such that Pw2W b(w) = 1. For all
2 L, b( ) := Pw2W;w b(w).</p>
      <p>Let Lpc be some probability constraint language which has atoms
constraining the probability of some propositional sentence being
true, and contains all formulae which can be formed with the atoms
in combination with logical connectives. If C Lpc is a set of
formulae, then b satisfies C (denoted b C) iff 8 2 C, b
satisfies the constraints posed by . I denote as the set of all
belief states over the set of possible worlds W . Let C be the set of
belief states which satisfy the probability constraints in C. That is,</p>
      <p>C := fc 2 j c Cg. Later, the notion of the theory of a set of
belief states will be useful:</p>
      <p>Th( C ) := f
2 Lpc j b 2</p>
      <p>C ; b
g:</p>
      <p>
        I build on Voorbraak’s [
        <xref ref-type="bibr" rid="ref36">36</xref>
        ] partial probability theory (PPT), which
allows probability assignments to be partially determined, and where
there is a distinction between probabilistic information based on (i)
hard background evidence and (ii) some assumptions. An epistemic
state in PPT is defined as the quadruple h ; B; A; Ci, where is a
sample space, B Lpc is a set of probability constraints, A Lpc
is a sets of assumptions and C W “represents specific information
concerning the case at hand” (an observation or evidence).
      </p>
      <p>Voorbraak mentions that he will only consider conditioning where
the evidence does not contradict the current beliefs. He defines the set
of belief states corresponding to the conditionalized PPT epistemic
state as fb( j C) 2 j b 2 B[A; b(C) &gt; 0g.</p>
      <p>Voorbraak proposes constraining as an alternative to conditioning:
Let 2 Lpc be a probability constraint. Then, constraining B on
produces B[f g. Note that expanding a belief set reduces the
number of models (worlds) and expanding a PPT epistemic state with
extra constraints also reduces the number of models (belief states /
probability functions).</p>
      <p>In the context of belief sets, it is possible to obtain any [...
epistemic] state from the ignorant [... epistemic] state by a series of
expansions. In PPT, constraining, but not conditioning, has the
analogous property. This is one of the main reasons we prefer
constraining and not conditioning to be the probabilistic
version of expansion. [36, p. 4]</p>
      <sec id="sec-2-1">
        <title>Voorbraak provides the following example [36].</title>
        <p>Example 1. Consider a robot which has to recharge his battery. This
can be done in two rooms, let us call them room 1 and 2. The rooms
are equally far away from the robot. An example of generic
information might be: “the door of room 1 is at least 40% of the time open”.
Suppose there is no other information available, and let pi denote
the probability of door i being open. Then B = fp1 0:4g. Since
doors are typically sometimes open and sometimes closed, it might
be reasonable to include 0 &lt; pi &lt; 1 in A. However, such
additional assumptions should be invoked only when they are necessary,
for example, in case no reasonable decision can be made without
assumptions. A good example of specific evidence is information about
the state of the doors obtained by the sensors of the robot.</p>
        <p>
          The word assumption is not very informative: An assumption may
be highly entrenched (indefeasible), for instance, about the laws of
physics, or an assumption may be very tentative (defeasible) like
hearing gossip about some character trait of a new staff member.
However, implicit in the word default, is the notion of
defeasibility; default information is information which holds until stronger
evidence defeats it. I shall refer to the set B as background knowledge
and assume it to be indefeasible, and the set F as foreground
knowledge and assume it to be defeasible. Hence, referring to Voorbraak’s
Example 1, p1 0:4 would be in F and 0 &lt; pi &lt; 1 would be in
B. Indeed, in PPT, “it is intended to be understood that the
conclusions warranted by [A [ B] depend on the assumptions represented
in A,” [
          <xref ref-type="bibr" rid="ref37">37</xref>
          ]. I interpret this to mean that background knowledge (A)
dominates foreground knowledge (B).
        </p>
        <p>It is, however, conceivable that background knowledge should be
defeasible and that new (foreground) evidence should weigh stronger
due its recency and applicability. In the proposed framework, a
compromise is attempted: background knowledge is ‘dominated’ by new
evidence at the time of reasoning, but after reasoning, the new
evidence is ‘forgotten’. However, evidence is not completely forgotten:
as it becomes less applicable after some time, it gets assimilated into
the foreground knowledge as default information.</p>
        <p>
          I also use elements of partially observable Markov decision
process (POMDP) theory [
          <xref ref-type="bibr" rid="ref1 ref23 ref25">1, 25, 23</xref>
          ]. In a POMDP, the agent can only
predict with a likelihood in which state it will end up after
performing an action. And due to imperfect sensors, an agent must maintain
a probability distribution over the set of possible states.
        </p>
        <p>Formally, a POMDP is a tuple hS; A; T ; R; ; Oi with a finite
set of states S = fs1; s2; : : : ; sng, a finite set of actions A =
fa1; a2; : : : ; akg, the state-transition function, where T (s; a; s0) is
the probability of being in s0 after performing action a in state s, the
reward function, where R(a; s) is the reward gained for executing a
while in state s, a finite set of observations = fz1; z2; : : : ; zmg;
and the observation function, where O(a; z; s0) is the probability of
observing z in state s0 resulting from performing action a in some
other state. An initial belief state b0 over all states in S is assumed
given.</p>
        <p>To update the agent’s beliefs about the world, a state estimation
function SE (b; a; z) = bSaE;z is defined as
bSaE;z(s0) =</p>
        <p>O(a; z; s0) Ps2S T (s; a; s0)b(s)</p>
        <p>P r(z j a; b)
;
where a is an action performed in ‘current’ belief-state b, z is the
resultant observation and bSaE;z(s0) denotes the probability of the agent
being in state s0 in ‘new’ belief-state bSaE;z. Note that P r(z j a; b) is a
normalizing constant.</p>
        <p>Let the planning horizon h (also called the look-ahead depth) be
the number of future steps the agent plans ahead each time it selects
its next action. V (b; h) is the optimal value of future courses of
actions the agent can take with respect to a finite horizon h starting
in belief-state b. This function assumes that at each step, the action
which will maximize the state’s value will be selected. V (b; h) is
defined as
max h (a; b) +
a2A</p>
        <p>X P r(z j a; b)V (SE (b; a; z); h
z2
1)i;
where (a; b) is defined as Ps2S R(a; s)b(s), 0 &lt; 1 is a
factor to discount the value of future rewards and P r(z j a; b)
denotes the probability of reaching belief-state bSaE;z = SE (b; a; z).
While V denotes the optimal state value, function Q denotes the
optimal action value: Q (a; b; h) = (a; b) + Pz2Z P r(z j
a; b)V (SE (b; a; z); h 1) is the value of executing a in the
current belief-state, plus the total expected value of belief-states reached
thereafter.</p>
        <p>
          I also build on my stochastic belief change model [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ], which is
a structure hW; Evt ; T ; E; O; Str i, with a set of possible worlds W ,
a set of events Evt , an event-transition function where T (w; e; w0)
models the probability of a transition to world w0, given the
occurrence of event e in world w, an event likelihood function where
E(e; w) = P (e j w) is the probability of the occurrence of event
e in w, an observation function where O(z; w) models the
probability of observing z in w, and where Str (z; w) is the agent’s ontic
strength for z perceived in w.
        </p>
        <p>I proposed a way of trading off the probabilistic update and
probabilistic revision, using the notion of ontic strength. The argument is
that an agent could reason with a range of degrees for information
being ontic (the effect of a physical action or occurrence) or epistemic
(purely informative). It is assumed that the higher the informations
degree of being ontic, the lower the epistemic status of that
information. “An agent has a certain sense of the degree to which a piece of
received information is due to a physical action or event in the world.
This sense may come about due to a combination of sensor readings
and reasoning. If the agent performs an action and a change in the
local environment matches the expected effect of the action, it can be
quite certain that the effect is ontic information,” [29, p. 129].</p>
        <p>
          The hybrid stochastic change of belief state b due to new
information z with ontic strength (denoted b z) is defined as [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]
b
z := n(w; p) j w 2 W; p =
1
(1
        </p>
        <p>Str (z; w))bz (w) + Str (z; w)bz (w) o;
where is some probabilistic belief revision operator, is some
probabilistic belief update operator and is a normalizing factor so that
Pw2W bz (w) = 1.</p>
        <p>
          The principle of maximum entropy [
          <xref ref-type="bibr" rid="ref18 ref20 ref26 ref27 ref34">18, 27, 34, 26, 20</xref>
          ] says that it
is reasonable to represent a set of belief states C by the member of
        </p>
        <p>C which is most entropic or least biased (w.r.t. information theory)
of all the members of C :</p>
        <p>ME ( C ) := arg max H(c);
c2 C
where H(c) := Pw2W c(w) ln c(w).</p>
        <p>
          The principle of minimum cross-entropy [
          <xref ref-type="bibr" rid="ref22 ref5">22, 5</xref>
          ] is used to select
the belief state c 2 Y ‘most similar’ to a given belief state b 2 X ;
the principle minimizes the directed divergence between b and c with
respect to entropy. Directed divergence is defined as
        </p>
        <p>R(c; b) := X c(w) ln
w2W
c(w)
b(w)
:
R(c; b) is undefined when b(w) = 0 while c(w) &gt; 0; when c(w) =
0, R(c; b) = 0, because limx!0 ln(x) = 0.</p>
        <p>
          The knowledge and reasoning structure proposed in this paper is
presented next. It builds on the work of Voorbraak [
          <xref ref-type="bibr" rid="ref36">36</xref>
          ] and Rens [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ]
and includes elements of POMDP theory.
3
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>FRAMEWORK DEFINITION</title>
      <p>Let Lprob be a probabilistic language over L defined as Lprob :=
f [`; u] j 2 L; `; u 2 [0; 1]; ` ug. A sentence of the form
[`; u] means the likelihood of proposition is greater than or equal
to ` and less than or equal to u. Let N = f0; 1; 2; : : :g.</p>
      <p>b satisfies formula [`; u] (denoted b [`; u]) iff ` b( ) u.
Definition 1. An agent maintains
hW; B; F; A; Evt ; Z; Prs; Eng ; Mi, where</p>
      <p>W is a set of possible worlds;
a
structure
B Lprob is a background belief base of fixed assumptions;
F Lprob is a foreground belief base of default assumptions;
A is a set of (agent) actions, including a special action null ;
Evt is a set of (environment) events;
Z is the observation stream, a set of observation triples: Z :=
f(a1; t1; z1); (a2; t2; z2); : : : ; (ak; tk; zk)g, where ai 2 A, ti 2
N, zi 2 L, and such that 8ti; tj 2 N, i = j iff ti = tj (i.e., no
more than one action and observation occur at a time-point);
Prs : L W ! N is a persistence function, where Prs (z; w)
indicates how long z is expected to be true from the time it is
received, given the ‘context’ of w; it is a total function over L
W ;
Eng : L W A ! [0; 1], where Eng (z; w; a) is the agent’s
confidence that z perceived in w was caused by action a (i.e., that
z has an endogenous source);
M is a model of the environment, and any auxiliary information
required by the definition of the particular belief change operation
( ).</p>
      <p>Definition 2. The expected persistence of z perceived in belief state
b is</p>
      <p>ExpPrs (z; b) := X Prs (z; w)
b(w):
w2W
But the agent will reason with respect to a set of belief states
C , hence, ExpPrs (z; C ) must be defined. One such definition
employs the principle of maximum entropy:</p>
      <sec id="sec-3-1">
        <title>Definition 3.</title>
        <p>where bME = ME ( C ).</p>
        <p>ExpPrs ME (z;</p>
        <p>C ) :=</p>
        <p>X Prs (z; w)</p>
        <p>bME (w);
w2W
Definition 4. An observation triple (a; i; z) 2 Z, has expired at
point s if ExpPrs (z; b) &lt; s i.</p>
        <p>Let ba;z be the change of belief state b by a and z. My intention is
that “change” is a neutral term, not necessarily indicating revision or
update. Next, I propose one instantiation of ba;z. Let the environment
model M = hE; T ; Oi, where</p>
        <p>E : Evt W ! [0; 1] is the event function. E(e; w) = P (e j w),
the probability of the occurrence of event e in w;
T : W (A[Evt ) W ! [0; 1] is a transition function such that
for every 2 A [ Evt and w 2 W , Pw02W T (w; ; w0) = 1,
where T (w; ; w0) models the probability of a transition to world
w0, given the execution of action / occurrence of event in world
w;
O : W W A ! [0; 1] is an observation function such
t1h,awtfhoerreevOer(ywzw; w2; aW)maonddelas t2he Apr,oPbabwizli2tyWoOf(owbzse;rwvi;nag) =z
(a complete theory for wz) in w and where O(z; w; a) :=
Pwwzz2Wz O(wz; w; a), for all z 2 L.</p>
        <p>Observation z may be due to an exogenous event (originating and
produced outside the agent) or an endogenous action (originating and
produced within the agent). It is up to the agent designer to decide,
for each observation, whether it is exogenous or endogenous, given
the action and world.</p>
        <p>
          The hybrid stochastic belief change (HSBC) formalism of Rens
[
          <xref ref-type="bibr" rid="ref29">29</xref>
          ] defines the (exogenous) update of b with z as
bz := (w; p) j w 2 W; p =
1
        </p>
        <p>O(z; w) X
where is a normalizing factor and O(z; w) can be interpreted as
O(z; w; null ). But HSBC does not involve agent actions; HSBC
assumes that agents passively receive information. Hence, when an
agent is assumed to act and the actions are known, the POMDP state
estimation function can be employed for belief state update.</p>
        <p>Only for the propose of illustrating how the framework can be
used, the following belief change procedure is defined.
ba;z :=f(w; p) j w 2 W; p =</p>
        <p>Exg (z; w; a)bz (w) + Eng (z; w; a)bSaE;z(w)g;
(1)
where Exg (z; w; a) := 1 Eng (z; w; a) is the confidence that z is
exogenous in w, given a was executed.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>OPERATIONAL SEMANTICS</title>
      <p>The agent is always living at time-point N . Time units remain the
same and must be chosen to suit the domain of interest, for instance,
milliseconds, minutes, hours, etc. The first point in time is N = 0. N
indicated the end of the N -th time unit, in other words, at point N ,
exactly N u time has passed, where u is the time unit employed.</p>
      <p>It will be assumed that no observation can be made at time-point
0. If the agent designer feels that it is unreasonable for the agent to
have to wait until N = 1 before the first observation may be made,
then the time unit chosen for the particular domain is too large.</p>
      <p>Expired observations are continually aggregated into the agent’s
set of default assumptions F (foreground beliefs). Prevalent
(unexpired) observations remain in the agent’s ‘current memory stream’ Z.
Whenever the agent wants to perform some reasoning task, (i) it
combines the prevalent observations with its fixed beliefs B, then
modifies its changed beliefs with respect to its default assumptions, or (ii)
modifies its fixed beliefs B with respect to its default assumptions,
then combines the prevalent observations with its changed beliefs –
and then reasons with respect to this final set of beliefs. However,
the agent always reverts back to the original fixed beliefs (hence,
“fixed”). And the default assumptions keep changing as memories
fade, that is, as observations expire.</p>
      <p>Because any initial conditions specified are, generally, not
expected to hold after the initial time-point, it does not make sense
to place them in B. The only other option is to place them in F ,
which is allowed to change. The following guiding principle is thus
provided to agent designers.</p>
      <p>The agent’s initial beliefs (i.e., the system conditions at point
N = 0) must be specified in F .</p>
      <p>Let C Lprob be a belief base. At this early stage of research,
I shall suggest only three definitions of on a set of belief states.
The first is the most na¨ıve approach (denoted NV ). It is suitable for
theoretical investigations.</p>
      <p>C</p>
      <p>NV a; z := fba;z j b 2</p>
      <p>C g:</p>
      <p>A more practical approach is to reduce C to a representative
belief state, employing the principle of maximum entropy (denoted</p>
      <p>C ME a; z := fba;z j b = ME ( C )g:</p>
      <p>A third and final approach which I shall mention here is, in a
sense, a compromise between the na¨ıve and maximum entropy
approaches. It is actually a family of methods which will thus not be
defined precisely. It is the approach (denoted FS ) which finds a
finite (preferably, relatively small) proper subset FS of C which is
somehow representative of C , and then applies to the individual
belief states in FS :</p>
      <p>C</p>
      <p>FS a; z := fba;z j b 2</p>
      <p>FS</p>
      <p>Only ME and FS will be considered in the sequel, when
it comes to belief change over a set. Let set denote one
of these two operators. Then C set can be defined,
where is any stream of observation triples, where =
f(a1; t1; z1); (a2; t2; z2); : : : ; (ak; tk; zk)g and t1 &lt; t2 &lt; &lt;
tk:
and then
and then</p>
      <p>C</p>
      <p>Let Expired be derived from the triples in Z which have just
expired (at point N ). That is,
or
or
where X; Y
belief state in
respect to b.
where N is defined below. At each time-point, F is refreshed with
all the expired observation triples in the order they appear in Z. In
other words, at each point in time, F Th( F set Expired ).
As soon as the foreground has been refreshed, the expired triples are
removed from the observation stream: Z Z n Expired . I shall use
the notation F F set to clarify which operation was used to arrive at
the current .</p>
      <p>A function which selects a belief state in one set which is in some
sense closest to another set of belief states will shortly be required.
The following definition suggests three such functions.</p>
      <sec id="sec-4-1">
        <title>Definition 5.</title>
        <p>Closest absolute ( X ; Y ) :=</p>
        <p>arg min X (b(w)
b: b2 X ;c2 Y w2W
c(w))2
Closest ME ( X ; Y ) := arg min
b2 X</p>
        <p>X (b(w)
w2W</p>
        <p>ME ( Y )(w))2;
Closest MCE ( X ; Y ) :=</p>
        <p>arg min
b: b2 X ;c2 Y</p>
        <p>R(c; b);
2 Lprob , ME ( Y ) is the most entropic/least biased</p>
        <p>Y and R(c; b) is the directed divergence of c with</p>
        <p>Closest absolute simply chooses the belief state b 2 X which
minimizes the sum of the differences between probabilities of worlds
of b and some belief state in c 2 Y (considering all c 2 Y ).</p>
        <p>Closest ME picks the belief state cME 2 Y with maximum
entropy (i.e., least biased w.r.t. information in Y ), and then chooses
the belief state b 2 X which minimizes the sum of the differences
between probabilities of worlds of b and cME .</p>
        <p>Expired := f(a; i; z) 2 Z j ExpPrs (z;
N 1) &lt; N
ig;
5</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>PATTERNS OF COGNITION</title>
      <p>Closest MCE chooses the belief state b 2 X for which the
directed divergence from belief state c 2 Y with respect to entropy
is least (considering all c 2 Y ). This is an instance of the minimum
cross-entropy inference.</p>
      <p>From here onwards, I shall not analyse their definitions;
Closest ( ) will be used to refer to the abstract function.</p>
      <p>Reasoning is done with respect to the set of belief states N . I
look at two patterns of cognition to determine N .</p>
      <sec id="sec-5-1">
        <title>Definition 6.</title>
        <p>N )1 :=
(
or
(</p>
        <p>N )2 :=
( B set Z) \
fClosest ( B</p>
        <p>F set if ( B set Z) \
set Z; F set )g otherwise,</p>
        <p>F
set 6= ;
( B \ F set )B set Z if
ffClosest ( ; F set )g</p>
        <p>B</p>
        <p>F
\ set 6= ;
set Zg otherwise.</p>
        <p>The idea behind the definition of ( N )1 is that (pertinent)
observations in the stream (Z) dominate background beliefs (B), but
Z-modified beliefs ( B set Z) dominate foreground beliefs (F ).
The idea behind the definition of ( N )2 is that foreground beliefs
(F ) have slightly higher status than in ( N )1 because they modify
background beliefs (when consistent with B) before (pertinent)
observations in the stream (Z), but the stream finally dominates.
In this section, I shall explore the properties of ‘patterns of
cognition’ based on ( N )1 and ( N )2. I shall argue that there are four
reasonable candidates. First, a few axioms:
1. fbg \ C = fbg or ;.
32.. fbCg \ MfcEg Z6=a;lwa(y)sresbu=lts cin. a singleton set.
4. Closest (fbg; C ) = b.</p>
        <sec id="sec-5-1-1">
          <title>And one guiding principle:</title>
          <p>In a given pattern of cognition, set is instantiated to the
same operator when it appears in the same relative position.</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>For instance, in (</title>
          <p>N )1,
( B ME Z) \
fClosest ( B</p>
          <p>F FS if ( B ME Z) \</p>
          <p>ME Z; F FS )g otherwise,
( B ME Z) \
fClosest ( B</p>
          <p>F FS if ( B ME Z) \
FS Z; F FS )g otherwise,</p>
          <p>F
F
Note that in the definitions of ( N )12, ( N )21 and ( N )22, set
may be instantiated as either of the two operators, and in ( N )22,
is actually the ‘plain’ belief change operator.</p>
          <p>I now justify each of the four patterns of cognition.
11 If ( B FS Z) were ( B ME Z), by axiom 3, the result is a
singleton set, and by axioms 1 and 4, the information in F is
iwgenroeredF,iMnEcl,ubdyinagxiinoimtia1l,cothnediitniofonrsmoartitohneirinla(terBeffescetts.ZIf)
isFiFgSnored, or worse, the intersection is empty.
12 With this pattern, the issues of intersection (axioms 1 and 2) need
not be dealt with. If B FS Z were B ME Z, by axiom 3, the
result is a singleton set, and by axiom 4, the information in F is
isgtinllohreads.aWn hineflthueernceFosent isBinstaFnStiZateddueasto tFheMElatoterr tyFpFiSc,altlhye nsoett
being a singleton.
21 If ( B \ F FS ) were ( B \ F ME ), by axioms 1, the
information in B is ignored. Whether set Z is instantiated as ME Z or</p>
          <p>FS Z, the result will typically not be trivial – all information is
taken into account, although in different ways.
22 Either of the two instantiations of F set accommodate the
information in B and in F . Moreover, the issues of intersection
(axioms 1 and 2) need not be dealt with, and due to the simpler
pattern, the second belief change operation can also be simplified,
that is, in this pattern, ME Z and FS Z both reduce to Z.
6</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>AN EXAMPLE</title>
      <p>To keep things as simple as possible for this introductory report, I
shall illustrate ( N )22 instantiated as</p>
      <p>Closest absolute ( B ; F</p>
      <p>ME )</p>
      <p>Z:</p>
      <p>Consider the following scenario. The elves live in the forest.
Sometimes, orcs (monsters) wander into the forest to hunt or seek
shelter from a storm. Elves don’t like orcs. The elves can either do
nothing (action null) or they can evict the orcs (action evict). Orcs
tend to stay in the forest for more hours the darker it is. The sun rises
(event rise) and sets (event set) once a day (duh). Now and then,
there is a storm (event storm), but most of the time, nothing happens
(event null). The vocabulary will be fe; o; `g meaning, respectively,
the elves are in the forest, the orcs are in the forest, and it is light (it
is daytime and there isn’t a storm).</p>
      <p>Henceforth, I might write instead of ^ , and instead of : .
For ease of reading, the possible worlds will be ordered, and a belief
state f(eo`; p1), (eo`; p2), (eo`; p3), (eo`; p4), (eo`; p5), (eo`; p6),
(eo`; p7), (eo`; p8)g will be abbreviated as hp1, p2, p3, p4, p5, p6,
p7, p8i:</p>
      <p>Only observations eo; eo; `; ` are considered.</p>
      <p>Let
Eng (z; w; null) = 0, 8z 2 , 8w 2 W (Observations after null
actions are never endogenous).</p>
      <p>Eng (eo; w; evict) = 0, 8w 2 W (Orcs in the forest should
never be observed due to eviction).</p>
      <p>Eng (eo; w; evict) = 0:75 if w `, = 0:5 if w ` (One can
be more confident that the orcs are not in the forest due to the
eviction, when it is dark).</p>
      <p>Eng (`; w; evict) = 0, 8w 2 W (Lightness is independent of
eviction).</p>
      <p>I am not entirely comfortable with this model of endogeny; see the
concluding section for a short discussion.</p>
      <p>Let
Prs (eo) = f(`; 2); (`; 4)g (Once the orcs are observed in the
forest, one can rely on them remaining there for at least two hours
when light and four hours when dark).</p>
      <p>Prs (eo) = f(&gt;; 1)g (One can rely on the orcs remaining outside
the forest for one hour once one finds out that they are outside).
Prs (`) = f(&gt;; 5)g (One can rely on light for five hours once light
is perceived; a storm may darken things within five ours).
Prs (`) = f(&gt;; 3)g (There might be a daytime storm, which
may clear up within three hours from when it stared and was
perceived).</p>
      <sec id="sec-6-1">
        <title>To define the belief change operator</title>
        <p>must be defined as follows.
as in (1), M = hE; T ; Oi
If w</p>
        <p>`,
– E(rise; w) = 0,
– E(set; w) = 1=12,
– E(storm; w) = 3=12.
– E(null; w) = 8=12.</p>
        <p>If w</p>
        <p>:`,
– E(rise; w) = 1=12,
– E(set; w) = 0,
– E(storm; w) = 3=12.
– E(null; w) = 8=12.</p>
        <p>T (w; null; w) = 1 8w 2 W (for the action and event).
T (w; evict; w0) = 0 if w :e _ :o, else if w `,
– = 0:1 if w0
– = 0:9 if w0
else if w</p>
        <p>:`,
– = 0:1 if w0
e ^ o ^ `,
e ^ :o ^ `,
e ^ o ^ :`,
– = 0:9 if w0 e ^ :o ^ :`.</p>
        <p>T (w; rise; w0) = 1 if w
and o are invariant.</p>
        <p>T (w; set; w0) = 1 if w
and o are invariant.</p>
        <p>T (w; storm; w0) = 1 if w0
invariant.</p>
      </sec>
      <sec id="sec-6-2">
        <title>The probabilities</title>
        <p>XXXwX eo`
z
:` and w0
` and w0</p>
        <p>` and truth values of e
:` and truth values of e
:` and truth values of e and o are
for</p>
        <p>O(z; w; null)
are
as
follows.
that is, it is completely light and there is a 70% 90% belief that
the elves and orcs are in the forest. And the background belief base
demands that the probability that elves are outside the forest while
orcs are inside is never more than 10%,
at different time-points.</p>
        <p>At N = 1, ( 1)22 = Closest absolute ( B; F ME )
f(null; 1; eo)g (given Prs(eo) = f(`; 2); (`; 4)g, (null; 1; eo)
has not yet expired). F ME =F ME ( F ) = h:7; 0; :1; 0; :1; 0; :1; 0i.
DCuloesesttaobsoluteB( Ba;ndF ME ) iMsE simbpelyingh:7; 0m;u:1tu;a0l;ly:1; 0c;o:1n;s0isiten2t,
B. Denote h:7; 0; :1; 0; :1; 0; :1; 0i as b1.</p>
        <p>Eng(eo; w; null) = 0 for all w 2 W . Therefore, b1
f(null; 1; eo)g = b1 null; eo = (b1)eo, which was calculated
to be h:386; :579; :007; :01; :007; :01; 0; 0i (denoted b2 henceforth).
I shall only show how (b1)eo(eo`) = 0:579 is calculated. Note that,
due to the transition functions for the four events being invariant with
respect to e and o, the only departure worlds we need to consider are
(abusing notion) eo` and eo`: (b1)eo(eo`) =</p>
        <p>O(eo; eo`; null) X X b1(w0)E(e; w0)T (w0; e; eo`)
It turns out that = 0:968, resulting in 0:56=0:968 = 0:579.</p>
        <p>The agent believes to a high degree what it perceived (eo). The
reason why it believes to a higher degree that it is dark than light, is
due to the relatively high chance of a storm (which darkens things)
and a small chance of the sun setting. This scenario is a bit synthetic:
the agent has not yet perceived that it is light. In a realistic
situation, the agent will always sense the brightness of the environment,
disallowing a high degree of belief in darkness.</p>
        <p>At N = 5, Expired = f(null; 1; eo)g because
ExpPrs(eo; b1) = Pw2W Prs(eo; w) b1(w) = 3:2 &lt;
N 1 = 4. And (evict; 4; eo) 62 Expired because
ExpPrs(eo; b1) = Pw2W Prs(eo; w) b1(w) = 1 6&lt; N 4 = 1.
f(eTvhiecretf;o4r;ee,o)g( 5=)22 Clos=est absoClultoes(esBta;bfsobl1ute ( Bnu; llF; MeoEg))
f(evict; 4; eo)g = Closest absolute ( B; fb2g) f(evict; 4; eo)g.</p>
        <p>B, hence, Closest absolute ( B; fb2g) = b2. ( 5)22 is thus
b2 2</p>
        <p>evict; eo.</p>
        <p>Recall that Eng(eo; w; evict) equals 0.75 if w
`. And recall that
`, but 0.5 if
ba;z(w) = Exg(z; w; a)bz(w) + Eng(z; w; a)bSaE;z(w):</p>
      </sec>
      <sec id="sec-6-3">
        <title>So, for instance,</title>
        <p>(b2)evict;eo(eo`)
= Exg(eo; eo`; evict)(b2)eo(eo`)
+ Eng(eo; eo`; evict)beSvEict;eo(eo`)</p>
        <p>SE
= 0:25(b2)eo(eo`) + 0:75(b2)evict;eo(eo`)</p>
        <p>At N = 6, Expired = f(null; 1; eo); (evict; 4; eo)g
because (null; 1; eo) had already expired at N = 5, and
ExpPrs(eo; (b2)evict;eo) = 1 &lt; N 4 = 2.</p>
        <p>Therefore, ( 6)22 = Closest absolute ( B; F ME ) fg =
Closest absolute ( B; f(b2)evict;eog).
7</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>CONCLUDING REMARKS</title>
      <p>I believe that it is important to have the facility to reason about both
(exogenous) events and (endogenous) actions; I am definitely not the
first to propose a framework with both notions [28, 33, 9, 6, e.g.].
The framework also has two belief bases, one to represent fixed,
background beliefs and one to accommodate defeasible information
(observations which have become ‘stale’, but not necessarily false).</p>
      <p>Inherent to the framework is that the agent’s knowledge may be
incomplete. There is much work on dealing with ignorance or missing
information [14, 16, 37, 38, 21, 30, e.g.].</p>
      <p>What makes the proposed framework potentially significant is its
generality, applicable to many domains and agent-designer
requirements. I want to amplify the point that the belief change operations
used in this paper are only suggestions and used due to personal
familiarity with them – the researcher / agent designer is given the
flexibility to suggest or design their own operators to suit their needs.</p>
      <p>Another feature of this framework which potentially adds to its
significance, is the nature of the observation stream (with expiring
observations) and how it interacts with the dual belief base approach.
We saw the rich potential of patterns of cognition, which can be
simultaneously confusing and empowering to the framework user.
However, some of the confusion was cleared up in Section 5. My
feeling is that this observation-stream-dual-belief-base system holds
much potential for investigating deep questions in how to model
continual observation and cognition within a belief change setting.</p>
      <p>One line of research that should be made is to generalize the
dualbelief-base approach: What would happen if three, four, more belief
bases are employed, each accommodating a different degree of
entrenchment of given and received information? Would such a
generalization reduce to the known systems of belief change (which
include notions of preference, plausibility, entrenchment, etc.)?</p>
      <p>Closely related to the discussion in the previous paragraph is that
keeping the belief base B fixed is quite a strong stance. In reality,
only the most stubborn people will never change their core views
even a little bit. Such stubbornness indicates an inability to grow,
that is, an inability to improve one’s reasoning and behaviour. In the
current framework, the plasticity of the assumptions, although
important for accommodating and aggregating recent observations, are
always dominated by B. In future versions, I would like to make B
more amenable to learning, while minding sound principles of belief
change in logic and cognitive psychology.</p>
      <p>Perhaps one of the most difficult aspects of using the proposed
framework is to specify the persistence function Prs( ). However,
the specification is made easier by the property of the operational
semantics that: expired observations keep on having an influence of
the agent’s reasoning until (if) they are ‘overridden’ by the process
of refreshing set F . This means that the agent designer should rather
err by specifying less persistence of observations when s/he is
uncertain about the period to specify. In other words, the agent designer is
advised to specify the longest period an observation is guaranteed to
persist.</p>
      <p>I also found it challenging to thinking about how to model how
endogenous evict is, for the different worlds and observations. For
instance, if Eng (eo; w; evict) = 0:75 when w `, should it
constrain the values that Eng (`; w; evict) can take? And if it is
impossible for the elves to evict while they are outside the forest, can
Eng (z; w; evict) be greater than 0 if w :e? There seem to be
several inter-related issues in the modeling of endogeny, including
action executability, perceivability and consistency among related
observations. I did not focus on these issues here.</p>
      <p>
        It would be straightforward to add a utility function (e.g., a
POMDP reward function) to the environment model M. Existing
planning and decision-making algorithms and methods can then be
used together with the proposed framework. Instead of using the
POMDP state estimation function during POMDP planning, for
instance, a more general belief change operator ( ) could be used in
planning under uncertainty, where the operator’s definition depends
on the elements of the proposed framework. Little work on planning
and decision-making with underspecified knowledge exists [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ], [36,
Sec. 5].
      </p>
      <p>
        The fundamental distinction between focusing and belief revision
when dealing with generic knowledge has been made by Dubois and
Prade [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]: “Revision amounts to modifying the generic knowledge
when receiving new pieces of generic knowledge (or the factual
evidence when obtaining more factual information), while focusing is
just applying the generic knowledge to the reference class of
situations which exactly corresponds to all the available evidence
gathered on the case under consideration.” The distinction seems very
relevant to general systems for knowledge management in dynamic
environments.
      </p>
      <p>This paper touches on several aspects of computational
reasoning, including stochasticity, imprecision/ignorance, knowledge
entrenchment, default knowledge, physical change and belief update,
new evidence and belief revision, and the persistence of evidence.
Except for evidence persistence, there are probably hundreds of
papers and article on combinations of these aspects. I could not yet
find any work dealing with the persistence of veracity of new
evidence/observations, as presented in the present paper. Besides the
work already cited in this paper, the following may be used as a
bibliography to better place the present work in context, and to point
to methods, approaches and techniques not covered in the proposed
framework, which could possibly be added to it.</p>
      <p>
        Probabilistic logics for reasoning with defaults and for belief
change or learning [
        <xref ref-type="bibr" rid="ref15 ref24">15, 24</xref>
        ].
      </p>
      <p>
        Nonmonotonic reasoning systems with optimum entropy
inference as central concept [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">4, 2, 3</xref>
        ].
      </p>
      <p>
        Dynamic epistemic logics for reasoning about probabilities [
        <xref ref-type="bibr" rid="ref32 ref35">35,
32</xref>
        ].
      </p>
    </sec>
    <sec id="sec-8">
      <title>ACKNOWLEDGEMENTS</title>
      <p>I would like to thank Edgar Jembere for his comments on a draft of
this paper.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>K.</given-names>
            <surname>Astro</surname>
          </string-name>
          <article-title>¨m, 'Optimal control of Markov decision processes with incomplete state estimation'</article-title>
          ,
          <source>Journal of Mathematical Analysis and Applications</source>
          ,
          <volume>10</volume>
          ,
          <fpage>174</fpage>
          -
          <lpage>205</lpage>
          , (
          <year>1965</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>C.</given-names>
            <surname>Beierle</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          , '
          <article-title>On the modelling of an agent's epistemic state and its dynamic changes'</article-title>
          ,
          <source>Electronic Communications of the European Association of Software Science and Technology</source>
          ,
          <volume>12</volume>
          , (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Beierle</surname>
          </string-name>
          and
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          , '
          <article-title>Towards an agent model for belief management'</article-title>
          ,
          <source>in Advances in Multiagent Systems, Robotics and Cybernetics: Theory and Practice</source>
          . (Volume III), eds., G. Lasker and
          <string-name>
            <given-names>J.</given-names>
            <surname>Pfalzgraf</surname>
          </string-name>
          ,
          <string-name>
            <surname>IIAS</surname>
          </string-name>
          , Tecumseh, Canada, (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Bourne</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Parsons</surname>
          </string-name>
          , '
          <article-title>Extending the maximum entropy approach to variable strength defaults'</article-title>
          , Ann. Math. Artif. Intell.,
          <volume>39</volume>
          (
          <issue>1-2</issue>
          ),
          <fpage>123</fpage>
          -
          <lpage>146</lpage>
          , (
          <year>2003</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>I. Csisza</surname>
          </string-name>
          ´r, '
          <article-title>I-divergence geometry of probability distributions and minimization problems'</article-title>
          ,
          <source>Annals of Probability</source>
          ,
          <volume>3</volume>
          ,
          <fpage>146</fpage>
          -
          <lpage>158</lpage>
          , (
          <year>1975</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Dupin de Saint-Cyr</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Lang</surname>
          </string-name>
          , '
          <article-title>Belief extrapolation (or how to reason about observations and unpredicted change)', Artif</article-title>
          . Intell.,
          <volume>175</volume>
          , (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Doder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Markovic´</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z. Ognjanovic´</surname>
          </string-name>
          , A. Perovic´, and
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Rasˇkovic´, A Probabilistic Temporal Logic That Can Model Reasoning about Evidence</article-title>
          ,
          <fpage>9</fpage>
          -
          <lpage>24</lpage>
          , Lecture Notes in Computer Science, Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dubois</surname>
          </string-name>
          and
          <string-name>
            <given-names>H.</given-names>
            <surname>Prade</surname>
          </string-name>
          ,
          <article-title>Focusing vs. belief revision: A fundamental distinction when dealing with generic knowledge</article-title>
          ,
          <fpage>96</fpage>
          -
          <lpage>107</lpage>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ferrein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Fritz</surname>
          </string-name>
          , and G. Lakemeyer, '
          <article-title>On-line decision-theoretic Golog for unpredictable domains'</article-title>
          , in KI 2004:
          <article-title>Advances in Artif</article-title>
          . Intell., volume
          <volume>238</volume>
          /
          <year>2004</year>
          ,
          <fpage>322</fpage>
          -
          <lpage>336</lpage>
          , Springer Verlag, Berlin / Heidelberg, (
          <year>2004</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Gabbay</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Hodkinson</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          ,
          <source>Temporal Logic: Mathematical Foundations and Computational Aspects</source>
          , volume
          <volume>1</volume>
          ,
          <string-name>
            <surname>Clarendon</surname>
            <given-names>Press</given-names>
          </string-name>
          , Oxford,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Gabbay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reynolds</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Finger</surname>
          </string-name>
          ,
          <source>Temporal Logic: Mathematical Foundations and Computational Aspects</source>
          , volume
          <volume>2</volume>
          , Oxford University Press, Oxford,
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ga</surname>
          </string-name>
          <article-title>¨rdenfors, Knowledge in Flux: Modeling the Dynamics of Epistemic States</article-title>
          , MIT Press, Massachusetts/England,
          <year>1988</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ga</surname>
          </string-name>
          <article-title>¨rdenfors, Belief Revision</article-title>
          , volume
          <volume>29</volume>
          of Cambridge Tracts in Theoretical Computer Science, Cambridge University Press, Massachusetts/England,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H.</given-names>
            <surname>Geffner</surname>
          </string-name>
          and
          <string-name>
            <given-names>B.</given-names>
            <surname>Bonet</surname>
          </string-name>
          , '
          <article-title>High-level planning and control with incomplete information using POMDPs'</article-title>
          ,
          <source>in Proceedings of the Fall AAAI Symposium on Cognitive Robotics</source>
          , pp.
          <fpage>113</fpage>
          -
          <lpage>120</lpage>
          , Seattle, WA, (
          <year>1998</year>
          ). AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Goldszmidt</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Pearl</surname>
          </string-name>
          , '
          <article-title>Qualitative probabilities for default reasoning, belief revision, and causal modeling'</article-title>
          ,
          <source>Artificial Intelligence</source>
          ,
          <volume>84</volume>
          ,
          <fpage>57</fpage>
          -
          <lpage>112</lpage>
          , (
          <year>1996</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Grove</surname>
          </string-name>
          and
          <string-name>
            <given-names>J.</given-names>
            <surname>Halpern</surname>
          </string-name>
          , '
          <article-title>Updating sets of probabilities'</article-title>
          ,
          <source>in Proceedings of the Fourteenth Conf. on Uncertainty in Artif. Intell</source>
          .,
          <source>UAI'98</source>
          , pp.
          <fpage>173</fpage>
          -
          <lpage>182</lpage>
          , San Francisco, CA, USA, (
          <year>1998</year>
          ). Morgan Kaufmann.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hansson</surname>
          </string-name>
          ,
          <article-title>A textbook of belief dynamics: theory change and database updating</article-title>
          , Kluwer Academic, Dortrecht, The Netherlands,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>E.</given-names>
            <surname>Jaynes</surname>
          </string-name>
          , '
          <article-title>Where do we stand on maximum entropy?'</article-title>
          , in The Maximum Entropy Formalism,
          <fpage>15</fpage>
          -
          <lpage>118</lpage>
          , MIT Press, (
          <year>1978</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>H.</given-names>
            <surname>Katsuno</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Mendelzon</surname>
          </string-name>
          , '
          <article-title>On the difference between updating a knowledge base and revising it'</article-title>
          ,
          <source>in Proceedings of the Second Intl. Conf. on Principles of Knowledge Representation and Reasoning</source>
          , pp.
          <fpage>387</fpage>
          -
          <lpage>394</lpage>
          , (
          <year>1991</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          ,
          <article-title>'Characterizing the principle of minimum crossentropy within a conditional-logical framework', Artif</article-title>
          . Intell.,
          <volume>98</volume>
          (
          <issue>12</issue>
          ),
          <fpage>169</fpage>
          -
          <lpage>208</lpage>
          , (
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          , '
          <article-title>Linking iterated belief change operations to nonmonotonic reasoning'</article-title>
          ,
          <source>in Proceedings of the Eleventh Intl. Conf. on Principles of Knowledge Representation and Reasoning</source>
          , pp.
          <fpage>166</fpage>
          -
          <lpage>176</lpage>
          , Menlo Park, CA, (
          <year>2008</year>
          ). AAAI Press.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kullback</surname>
          </string-name>
          ,
          <source>Information theory and statistics</source>
          , volume
          <volume>1</volume>
          ,
          <string-name>
            <surname>Dover</surname>
          </string-name>
          , New York, 2nd edn.,
          <year>1968</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>W.</given-names>
            <surname>Lovejoy</surname>
          </string-name>
          , '
          <article-title>A survey of algorithmic methods for partially observed Markov decision processes'</article-title>
          ,
          <source>Annals of Operations Research</source>
          ,
          <volume>28</volume>
          ,
          <fpage>47</fpage>
          -
          <lpage>66</lpage>
          , (
          <year>1991</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>T.</given-names>
            <surname>Lukasiewicz</surname>
          </string-name>
          , '
          <article-title>Nonmonotonic probabilistic logics under variablestrength inheritance with overriding: Complexity, algorithms</article-title>
          , and implementation',
          <source>International Journal of Approximate Reasoning</source>
          ,
          <volume>44</volume>
          (
          <issue>3</issue>
          ),
          <fpage>301</fpage>
          -
          <lpage>321</lpage>
          , (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>G.</given-names>
            <surname>Monahan</surname>
          </string-name>
          , '
          <article-title>A survey of partially observable Markov decision processes: Theory, models</article-title>
          , and algorithms', Management Science,
          <volume>28</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          , (
          <year>1982</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26] J. Paris,
          <source>The Uncertain Reasoner's Companion: A Mathematical Perspective</source>
          , Cambridge University Press, Cambridge,
          <year>1994</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>J.</given-names>
            <surname>Paris</surname>
          </string-name>
          and A. Vencovsk, '
          <article-title>In defense of the maximum entropy inference process', Intl</article-title>
          .
          <source>Journal of Approximate Reasoning</source>
          ,
          <volume>17</volume>
          (
          <issue>1</issue>
          ),
          <fpage>77</fpage>
          -
          <lpage>103</lpage>
          , (
          <year>1997</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>D.</given-names>
            <surname>Poole</surname>
          </string-name>
          , '
          <article-title>Decision theory, the situation calculus and conditional plans'</article-title>
          ,
          <source>Linko¨ping Electronic Articles in Computer and Information Science</source>
          ,
          <volume>8</volume>
          (
          <issue>3</issue>
          ), (
          <year>1998</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>G.</given-names>
            <surname>Rens</surname>
          </string-name>
          , '
          <article-title>On stochastic belief revision and update and their combination'</article-title>
          ,
          <source>in Proceedings of the Sixteenth Intl. Workshop on Non-Monotonic Reasoning</source>
          (NMR), eds.,
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Wassermann</surname>
          </string-name>
          , pp.
          <fpage>123</fpage>
          -
          <lpage>132</lpage>
          . Technical University of Dortmund, (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>G.</given-names>
            <surname>Rens</surname>
          </string-name>
          , T. Meyer, and G. Casini, '
          <article-title>Revising incompletely specified convex probabilistic belief bases'</article-title>
          ,
          <source>in Proceedings of the Sixteenth Intl. Workshop on Non-Monotonic Reasoning</source>
          (NMR), eds.,
          <string-name>
            <given-names>G.</given-names>
            <surname>Kern-Isberner</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Wassermann</surname>
          </string-name>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>142</lpage>
          . Technical University of Dortmund, (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>G.</given-names>
            <surname>Rens</surname>
          </string-name>
          , T. Meyer, and G. Lakemeyer, '
          <article-title>A modal logic for the decisiontheoretic projection problem'</article-title>
          ,
          <source>in Proceedings of the Seventh Intl. Conf. on Agents and Artif. Intell. (ICAART)</source>
          , Revised Selected Papers, eds.,
          <string-name>
            <given-names>B.</given-names>
            <surname>Duval</surname>
          </string-name>
          , J. Van den Herik, S. Loiseau, and
          <string-name>
            <given-names>J.</given-names>
            <surname>Filipe</surname>
          </string-name>
          , LNAI, pp.
          <fpage>3</fpage>
          -
          <lpage>19</lpage>
          . Springer Verlaag, (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>J.</given-names>
            <surname>Sack</surname>
          </string-name>
          , '
          <article-title>Extending probabilistic dynamic epistemic logic'</article-title>
          , Synthese,
          <volume>169</volume>
          ,
          <fpage>124</fpage>
          -
          <lpage>257</lpage>
          , (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>S.</given-names>
            <surname>Shapiro</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Pagnucco</surname>
          </string-name>
          , '
          <article-title>Iterated belief change and exogenous actions in the situation calculus'</article-title>
          ,
          <source>in Proceedings of the Sixteenth European Conf. on Artif. Intell. (ECAI-04)</source>
          , pp.
          <fpage>878</fpage>
          -
          <lpage>882</lpage>
          , Amsterdam, (
          <year>2004</year>
          ). IOS Press.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>J.</given-names>
            <surname>Shore</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Johnson</surname>
          </string-name>
          , '
          <article-title>Axiomatic derivation of the principle of maximum entropy and the principle of minimum cross-entropy', Information Theory</article-title>
          , IEEE Transactions on,
          <volume>26</volume>
          (
          <issue>1</issue>
          ),
          <fpage>26</fpage>
          -
          <lpage>37</lpage>
          , (
          <year>Jan 1980</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <surname>J. Van Benthem</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gerbrandy</surname>
            , and
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Kooi</surname>
          </string-name>
          , '
          <article-title>Dynamic update with probabilities'</article-title>
          ,
          <source>Studia Logica</source>
          ,
          <volume>93</volume>
          (
          <issue>1</issue>
          ),
          <fpage>67</fpage>
          -
          <lpage>96</lpage>
          , (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>F.</given-names>
            <surname>Voorbraak</surname>
          </string-name>
          , 'Partial Probability:
          <article-title>Theory and Applications'</article-title>
          ,
          <source>in Proceedings of the First Intl. Symposium on Imprecise Probabilities and Their Applications</source>
          , pp.
          <fpage>360</fpage>
          -
          <lpage>368</lpage>
          , (
          <year>1999</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>F.</given-names>
            <surname>Voorbraak</surname>
          </string-name>
          , '
          <article-title>Probabilistic belief change: Expansion, conditioning and constraining'</article-title>
          ,
          <source>in Proceedings of the Fifteenth Conf. on Uncertainty in Artif. Intell</source>
          .,
          <source>UAI'99</source>
          , pp.
          <fpage>655</fpage>
          -
          <lpage>662</lpage>
          , San Francisco, CA, USA, (
          <year>1999</year>
          ). Morgan Kaufmann Publishers Inc.
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yue</surname>
          </string-name>
          and W. Liu, '
          <article-title>Revising imprecise probabilistic beliefs in the framework of probabilistic logic programming</article-title>
          .',
          <source>in Proceedings of the Twenty-third AAAI Conf. on Artif. Intell. (AAAI-08)</source>
          , pp.
          <fpage>590</fpage>
          -
          <lpage>596</lpage>
          , (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>