<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Course of Action Planning Ontology</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Timothy P. Darr</string-name>
          <email>tdarr@kbsi.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ph. D.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Perakath Benjamin</string-name>
          <email>pbenjamin@kbsi.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ph. D.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Richard Mayer</string-name>
          <email>rmayer@kbsi.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ph.D.</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>N00014-09-C-0334. Timothy P. Darr is a Research Scientist at Knowledge Based Systems Incorporated, College Station</institution>
          ,
          <addr-line>TX 77840 USA (</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2009</year>
      </pub-date>
      <abstract>
        <p>² T his pape r desc ribes an ontology to support course of action (C O A) planning that provides an extensible framewor k for modeling C O A plans consistent with A r my and M arine Corp doctrine. T his ontology is str uctur ed into a core ontology that includes definitions of common C O A planning concepts (activities, phases, outcomes, measu res-of-perfor mance, measur es-of-effectiveness, etc.), and multiple domain-specific ontologies that extend the cor e ontology for specific types of plans (stability operations, counterinsurgency planning, infor mation operations planning, etc.). A p refer ence relation between desc riptions of the " plan state " using measures-of-effectiveness is introduced to allow subject-matter experts to specify a r an k ing ove r plan activities or phases from a specific H uman Social C ultu re B ehavior ( HS C B) perspective. T hese prefe rences can be used in the planning p rocess to identify or pr une blac k holes and blind alleys.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>I ndex Terms² Course of A ction Planning, I nfor mation
O pe rations, C ounte rinsurgency Planning, Utility T heory,
Preferential Reasoning, Q ualitative Reasoning.</p>
    </sec>
    <sec id="sec-2">
      <title>I. INTRODUCTION</title>
      <p>TAction ontology (COA-Ontology) that can be used to</p>
      <p>
        HIS paper presents the initial design of a
Course-ofsupport course of action (COA) planning. This ontology
applies to the COA planning processes defined for the United
States Army and Marine Corps for multiple domains, to
include stability operations planning [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], counterinsurgency
planning [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and information operations planning [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The core
ontology includes definitions of the common concepts and
properties for defining COA plans, including: COAs, COA
activities, COA phases, measures-of-performance (MOP) and
measures-of-effectiveness (MOE).
      </p>
      <p>The COA-Ontology consists of multiple sub-ontologies,
each of which contains a small number of concepts and
properties and can easily be integrated into other ontologies.
For example, we have defined a measure-of-effectiveness
(MOE) ontology that can be imported as a standalone
ontology into other ontologies for the purpose of providing a
common definition for representing MOEs. The urban
Counterinsurgency (COIN) ontology is an example of how to
extend the core ontology to define activities, MOP, MOE and
phases that are tailored to a specific domain.</p>
      <p>Qualitative MOPs and MOEs are used to describe plan
states and outcomes. It is often the case that objectives are
defined not in terms of fixed numeric quantities or ranges, but
rather in qualitative terms such as: "reduce the number of
attacks against coalition forces", or "increase the level of
activity in the central market during daylight hours."</p>
      <p>The Deep Maroon1 ontology currently supports two types of
reasoning: (i) utility-theoretic preferential or preference
reasoning and (ii) goal-directed forward and backward
chaining reasoning.</p>
      <p>Preference reasoning provides a way to rank-order states
from the perspective of a given community group
(counterinsurgents, insurgent group, religious or ethnic group,
etc.). An inference algorithm can use these preferences to
reason about assessment of how a given activity will be
perceived and can assist a planner in the identification of black
holes or blind alleys2.</p>
      <p>Goal-directed forward and backward chaining reasoning
provides a way to reason about the desired trajectory of the
plan over time (forward chaining), and given an end state,
determine a set of starting states that would result in that end
state (backward chaining). In forward chaining, the sequence
of activities that are available at each plan state can be
determined by matching activity preconditions with the
current state and asserting the new state that results from the
application of the activity. In backward chaining, the possible
states that can achieve a given outcome are determined,
followed by the activities that can achieve that state.
Interleaving forward and backward chaining with
preferencebased filtering helps to mitigate the complexity of developing
and analyzing realistic plans3.</p>
      <sec id="sec-2-1">
        <title>A. Ontology D esign Goals</title>
        <p>
          The design of the COA-Ontology strives to achieve
extensibility, flexibility and reuse. Our experience is that
many ontologies are monolithic and cumbersome; making
them difficult to understand and difficult to reuse. An example
of what the authors believe is a concise, well-defined ontology
1 The name coined for this system by analogy to the DARPA Deep Green
force-on-force planning and execution management initiative [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] and the IBM
Deep Blue chess system.
        </p>
        <p>2 In the context of a COA plan, a black hole is a state that once you get
into, you can never get out of. An example of a black hole is an activity that
leads to an inflammatory situation such as civil war or increased intra-militia
violence. A blind alley is a state that is unproductive in that there is no
feasible next state or no path to a goal state. Unfortunately, in the COIN
context blind alleys often turn into black holes as you may not be able to
retrace to an earlier state.</p>
        <p>
          3 How the complexity is mitigated in this way is beyond the scope of this
paper.
is the W3C geospatial ontology [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]. This ontology consists of
only two concepts (geo:SpatialThing, geo:Point), and five
properties (geo:alt, geo:lat, geo:long, geo:lat_long,  
geo:location): the minimal amount of ontological
information to represent a geospatial region. From this
ontology, more specific geospatial regions such as polygons,
ellipses, and points can be defined by extending this geospatial
ontology.
        </p>
        <p>The COA-Ontology design principles are as follows4
x Ontologies should have a well defined purpose and
support a well-defined set of use cases;
x Ontologies should include the minimal number of
concepts and properties to support the purpose;
x Ontologies should strive to be more than a simple
taxonomy of domain concepts; and
x When possible, allow for importing / exporting
concepts and properties from / to other ontologies.</p>
        <p>We realize that the design principles described above can be
very subjective, but we believe that they are useful as a
starting point for ontology design.</p>
        <p>The purpose of the COA-Ontology is as follows:
x To represent COA plans that can be incorporated into
other tools / applications, or to be used as a
communication medium for the plans across systems.
x To represent concepts in the Human Social Culture
Behavior (HSCB) modeling domain so that COAs may
be assessed in a reusable, computer-process-able format
from the perspective of multiple interested
communities.</p>
      </sec>
      <sec id="sec-2-2">
        <title>B. Paper Organization</title>
        <p>The remainder of this paper is organized as follows. Section
II defines the COA planning problem context. Section III
describes the structure of the COA-Ontology family of
ontologies. Section IV outlines the representation of
measuresof-performance and measures-of-effectiveness. Section V
describes the representation of COAs. Section VI describes the
representation of preferences. Section VII describes inference
support in COA-Ontology. Section VIII outlines conclusions
and opportunities for future work.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>II. PROBLEM CONTEXT</title>
      <p>This section describes the context that the COA-Ontology
supports. Fig. 1 shows a simplified example of a stability
operations COA plan. This figure shows one of three possible
COAs that are proposed to achieve a commander's objective5.
This COA consists of three phases: establish security,
establish civil control and restore essential services. Each
phase is terminated by an outcome that serves as a milestone
for measuring progress of the plan. Each phase contains a
sequence of activities that are performed to achieve the
endphase outcomes. The activities can be sequential, as shown in
the establish security and restore essential services phases; or</p>
    </sec>
    <sec id="sec-4">
      <title>4 Similar to the principles defined in [8].</title>
      <p>5 Per doctrine, three COAs are typically presented to the commander for
approval.
branch-and-sequence as shown in the establish civil control
phase6.</p>
      <p>The forward chaining reasoning supported by the
COAOntology can be used to reason from the initial state
represented by the candidate COA on the left-hand side, to the
activities that are possible at that state, to intermediate states
that are achieved by each activity, to an end-phase outcome.
The same reasoning is possible treating each end-phase
outcome as an initial state. The backward chaining reasoning
supported by the COA-Ontology can be used to determine
what activities can achieve the end-phase outcome, backward
through the states that enable the activities back to the initial
state on the left-hand side.</p>
      <p>Fig. 2 illustrates the identification of black holes and blind
alleys in a COA plan. Using the preference knowledge
specified by an SME for a particular community segment, an
inference engine can identify activities and outcomes as
infeasible or uncertain, respectively. This has the potential to
significantly reduce the search space as black holes and blind
alleys are pruned from consideration or identified for further
investigation.</p>
      <p>This section describes the structure of the COA-Ontology.
Fig. 3 shows the structure of the core ontology and an
extension of the core ontology to support urban COIN COA
planning. This organization allows users to import or use only
those elements that are required for a given application,
promoting flexibility and reuse.</p>
      <p>6 It is also possible to have concurrent activities, though not shown in the
figure.</p>
      <p>The solid arrows represent ontology imports within the core
and urban coin ontologies. The dotted arrows across the core /
urban COIN boundary represents ontology imports from the
urban COIN ontology to the core ontology.</p>
      <p>The core ontology consists of six sub-ontologies contained
in six OWL files that define the most general concepts and
properties. The core ontology is completely self-contained and
can be used as the basis for more specific COA planning
ontologies, promoting extensibility. The urban COIN ontology
shown at the right of the figure is one such extension of the
core ontology.</p>
      <p>The common ontology (common.owl) contains definitions
of general concepts and properties that are common to COA
planning. The common ontology includes four classes:
x IO-­Thing - a subclass of owl:Thing that is used to
attach properties and relationships that are common to
all classes in the ontology.
x COA-­Variable - a subclass of IO-­Thing that represents
a variable that can describe some feature of a domain.
A measure-of-performance or measure-of-effectiveness
is a subclass of COA-­Variable. This class has two
properties: hasValue and hasValueDirection. The
hasValue property is the actual value of the variable
and can be one of the standard types (xsd:integer;
xsd:float, etc.) as well as a qualitative value (see
below).
x</p>
      <p>Qualitative-­Direction - a subclass of owl:Thing that
represents a qualitative description (increasing,
decreasing, stable) of the direction or trajectory of an
COA variable.
x Qualitative-­Values - subclass of owl:Thing that
represents qualitative values (HIGH, MEDIUM, LOW,
etc.).</p>
      <p>
        The measures-of-performance ontology
(measures-ofperformance.owl) contains the definition of measures of
performance. According to COIN doctrine [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], a measure of
performance is defined as "a criterion to assess friendly
actions that is tied to measuring task accomplishment." MOPs
in the COA-ontology are effectively state variables that are
used to define the pre- and post-conditions for an activity and
to be used as inputs to the calculation of MOEs. The MOP
ontology includes a single class:
x
      </p>
      <p>Measure-­of-­Performance - a subclass of COA-­
Variable that represents a measure of performance.
This class has the property hasTimeStamp to indicate
the time at which the measurement was collected.</p>
      <p>
        The measures-of-effectiveness ontology
(measures-ofeffectiveness.owl) contains the definition of measures of
effectiveness. According to COIN doctrine [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], a measure of
effectiveness is defined as "a criterion used to assess changes
in system behavior, capability, or operational environment that
is tied to measuring the attainment of an end state,
achievement of an objective, or creation of an effect." MOEs
in the COA-ontology define the commander's objective
(endof-COA outcome), objectives to be achieved at the end of each
COA phase, and objectives to be achieved after each COA
activity is applied at a given state. The MOEs depend on a set
of MOPs. The MOE ontology includes two classes:
x
      </p>
      <p>Measure-­of-­Effectiveness - a subclass of COA-­
Variable that represents a measure-of-effectiveness.
This class has the property influencingMOP that
defines the set of MOPs that influence the MOE. An
MOE can be views as a function that takes as input a set
of MOPs and generates a measure. The
influencingMOP defines the arguments to the function.
x COA-­Outcome - a subclass of IO-­Thing that represents
an objective or outcome. This class has the property
hasMOE that defines the MOEs that describe the
outcome.</p>
      <p>The activities ontology (activities.owl) contains the
definition of a COA activity. The activities ontology includes
a single class:
x</p>
      <p>COA-­Activity - a subclass of IO-­Thing that represents
an activity within a COA phase. This class has four
properties:</p>
      <p>preconditionMOP, postconditionMOP,
previousActivity, subsequentActivity, and
hasActivityOutcome. The preconditionMOP and
postconditionMOP properties are used to define the
precondition MOP for applying the activity and the
state that results from applying the activity,
respectively. The previousActivity and
subsequentActivity properties are used to define a
sequence of activities to perform within a COA phase.
The activityOutcome property is the outcome that
results from applying the activity.</p>
      <p>The course-of-action ontology (COA.owl) contains the
definition of a COA. The COA ontology includes two classes:
x</p>
      <p>Course-­of-­Action - a subclass of IO-­Thing that
represents a COA. This class has two properties:
hasPhases and hasOutcome. The hasPhases property
defines the phases within the COA. The hasOutcome
property defines the commander's objective for the
COA.</p>
      <p>COA-­Phase - a subclass of IO-­Thing that represents a
COA phase. This class has four properties:
and
hasNextPhase, hasPrevPhase, hasActivities
hasOutcome. The hasNextPhase and hasPrevPhase
properties are used to define a sequence of phases
within a COA. The hasActivities property defines the
x
activities within the COA phase. The hasOutcome
property defines the outcome of the COA phase.</p>
      <p>The preferences ontology (preferences.owl) contains the
definition of preferences over COA outcomes. A preference in
this context is a relation between two outcomes in which one
of the outcomes is preferred to the other outcome, given the
perspective of a specific human social culture behavior
(HSCB) perspective. These preferences are typically asserted
by an SME while role playing a specific HSCB perspective or
community group. For example, in an agricultural community
in which there is little or no electricity, a COA whose outcome
involves restoration of economic self-sufficiency via the
activity of building or restoring a canal system for crop
irrigation, will be preferred to a COA in which the same
outcome is achieved via the activity of providing electrical
power to the local market. The preference ontology contains a
single class:
x</p>
      <p>Preference-­Relation - a subclass of IO-­Thing that
represents the pairwise preference between two
outcomes from the perspective of a specific community
group. This class has two properties: lessPreferred
and morePreferred. The lessPreferred property
refers to the outcome that is less preferred from the
perspective of the community group. The
morePreferred property refers to the outcome that is
more preferred from the perspective of the community
group.</p>
      <p>
        KBSI is currently developing capabilities to reason about
preferences of this kind using the application of Imprecisely
Specified Multi-Attribute Utility Theory (ISMAUT) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
Preferences are used to model HSCB perspectives for the
purpose of supporting the decision maker(s) and COA
planner(s) in COA development, war gaming, comparison and
decision making.
      </p>
      <p>IV. MEASURES-OF-PERFORMANCE AND
MEASURES-OF</p>
      <p>EFFECTIVENESS</p>
      <p>This section illustrates the relationship between MOPs and
MOEs in more detail, as shown in Fig. 4. The namespaces for
each of the ontologies are defined in the lower left of the
diagram. An MOP has a unique timestamp and a value with a
qualitative direction. An MOE is a specialization of an MOP
with a set of MOPs that represent the arguments to a function
that calculates the value of the MOE, given the values of the
influencing MOPs. An outcome is described by one or more
MOEs.</p>
      <sec id="sec-4-1">
        <title>A. Example - Urban C OIN</title>
        <p>
          Fig. 5. shows example MOPs and MOEs from the COIN
domain [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. In this example, there is a single MOP, force
density; two MOEs, establish presence and increase level of
security; and a single outcome, increase level of security.
Force density is a measure of the amount of force in a given
area. The establish presence MOE is a qualitative measure of
the amount of presence of U.S. and Host Nation (HN) forces
in a given area. This could range from military patrols (U.S.
only or U.S. and HN) or the establishment of police stations or
outposts. The reduce reaction time is a statistical measure of
the amount of time it takes to respond to a significant event in
the area of interest (AOI). Each of these MOEs are a function
of the force density, as well as other "state variables" (not
shown) contained in an Intelligence Preparation of the
Battlefield (IPB) such as the AOI, force structure and human
terrain in the AOI. The µ increase level of security¶ outcome is
a qualitative objective that measures whether or not there was
an increase in the level of security in a given area. This
objective is described by the MOEs, establish presence and
reduce reaction time.
        </p>
        <p>This section illustrates the relationship between COAs,
COA phases, COA activities, outcomes and MOPs, as shown
in Fig. 6. The namespaces for each of the ontologies are
defined in the lower left of the figure. A COA consists of one
or more phases. Each phase has an outcome and consists of a
sequence of activities and may have a previous and next
phase. Each activity has MOP pre- and post-conditions and an
outcome. The precondition MOPs must be satisfied in order
for the activity to be applied. Whenever the activity is
performed, the postcondition MOPs are -determined.</p>
      </sec>
      <sec id="sec-4-2">
        <title>A. Example - Urban C OIN</title>
        <p>
          Fig. 7 shows an example of a partial COA from the COIN
domain [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Many of the details are missing, particularly the
precondition and postcondition MOPs as described previously.
In this example, there is one COA, two COA phases (one with
an outcome shown) and three COA activities. The COA,
labeled COA-1 has an establish security phase and an restore
essential services phase, consistent with COIN and stability
operations planning [
          <xref ref-type="bibr" rid="ref1 ref2">1,2</xref>
          ]. The establish security phase has an
increase level of security outcome as described in the previous
section. The establish security phase consists of activities for
establishing access points, performing a census and
establishing barriers. In this particular COA, the establish
access points activity is succeeded by the perform census and
establish barriers activities, which can be performed in
parallel.
        </p>
        <p>This section illustrates the relationship between preferences
and outcomes, as shown in Fig. 8. A preference is a pairwise
relationship between two outcomes, one of which is preferred
to the other assuming a specific HSCB perspective.</p>
        <p>This section describes the inference supported by the
COAontology.</p>
      </sec>
      <sec id="sec-4-3">
        <title>A. Class Subsumption</title>
        <p>
          As much as possible, the COA-Ontology utilizes class
subsumption via Description Logics (DL) based class
definitions [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. This section illustrates the COA ontology
using an example from the COIN domain [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>Fig. 9 shows a subsumption axiom for the MOP concept.
This axiom states that every MOP is a concept such that there
exists a hasValue relationship with an integer, float, boolean
or qualitative value, and there exists a hasValueDirection
relationship with a Qualitative-­Values concept, and there is
a hasTimeStamp relationship with an integer.</p>
        <p>׊ ݔ  ׌ ݕ , ݖ  ݄ ܽ ݏ ܸ ܽ ݈ ݑ ݁ ሺ ݔ , ݕ ሻ
ר ൫ ݔ ݏ ݀ : ݅ ݊ ݐ ݁ ݃ ݁ ݎ ሺ ݕ ሻ ש ݔ ݏ ݀ : ݂ ݈ ݋ ܽ ݐ ሺ ݕ ሻ
ש ݔ ݏ ݀ : ܾ ݋ ݋ ݈ ݁ ܽ ݊ ሺ ݕ ሻ
ש ܿ ݋ ݉ ݉ ݋ ݊ : ܳ ݑ ܽ ݈ ݅ ݐ ܽ ݐ ݅ ݒ ݁ െ ܸ ܽ ݈ ݑ ݁ ݏ ሺ ݕ ሻ ൯
ר ܿ ݋ ݉ ݉ ݋ ݊ : ݄ ܽ ݏ ܸ ܽ ݈ ݑ ݁ ܦ ݅ ݎ ݁ ܿ ݐ ݅ ݋ ݊ ሺ ݔ , ݖ ሻ
ר ܿ ݋ ݉ ݉ ݋ ݊ : ܳ ݑ ܽ ݈ ݅ ݐ ܽ ݐ ݅ ݒ ݁ െ ܸ ܽ ݈ ݑ ݁ ݏ ሺ ݖ ሻ
ר ݄ ܽ ݏ ܶ ݅ ݉ ݁ ܵ ݐ ܽ ݉ ݌ ሺ ݔ ሻ
՜ ݎ ݀ ݂ ݏ : ݏ ݑ ܾ ܿ ݈ ܽ ݏ ݏ ܱ ݂ (ݔ , ܯ ݁ ܽ ݏ ݑ ݎ ݁ െ ݋ ݂
െ ܲ ݁ ݎ ݂ ݋ ݎ ݉ ܽ ݊ ܿ ݁ )
Fig. 9. MOP Subsumption Axiom</p>
      </sec>
      <sec id="sec-4-4">
        <title>B. Preferential Dominance</title>
        <p>The concept of preferential dominance is important to
reasoning about preferences as it allows outcomes to be
pruned very efficiently, thereby reducing the computational
complexity, at very little cost, of searching through potentially
very large outcome spaces. Intuitively, one outcome
LDWHGQRVP RDQWHUK FWRXHP IL LW LV EHWU³´ WDQK HWK RWHUK
outcome along all the values of the variables that describe the
outcomes.</p>
        <p>Fig 10 shows the axiom for value dominance. The
atLeastAsGoodAs relation is the greater-than-or-equal-to
operator for numeric values. For qualitative values, this
relation LQGHIDSURWO\[P+,*´³YXLV
WD³ OHDVW VD RJG DV´ D 0(',8 DOY ue if the objective is to
maximize the value of that variable.
׊ ݉ ݋ ݁ 1, ݉ ݋ ݁ 2, ݒ 1, ݒ 2  (ݎ ݀ ݂ ݏ : ݏ ݑ ܾ ܥ ݈ ܽ ݏ ݏ ܱ ݂ ሺ ݉ ݋ ݁ 1, ܯ ݁ ܽ ݏ ݑ ݎ ݁
െ ݋ ݂ െ ܧ ݂ ݂ ݁ ܿ ݐ ݅ ݒ ݁ ݊ ݁ ݏ ݏ ሻ
ר ݎ ݀ ݂ ݏ : ݏ ݑ ܾ ܿ ݈ ܽ ݏ ݏ ܱ ݂ ሺ ݉ ݋ ݁ 2, ܯ ݁ ܽ ݏ ݑ ݎ ݁ െ ݋ ݂
െ ܧ ݂ ݁ ݁ ܿ ݐ ݅ ݒ ݁ ݊ ݁ ݏ ݏ ሻ ר ݄ ܽ ݏ ܸ ܽ ݈ ݑ ݁ ሺ ݉ ݋ ݁ 1, ݒ 1ሻ
ר ݄ ܽ ݏ ܸ ܽ ݈ ݑ ݁ ሺ ݉ ݋ ݁ 2, ݒ 2ሻ
ר ܽ ݐ ܮ ݁ ܽ ݏ ݐ ܣ ݏ ܩ ݋ ݋ ݀ ܣ ݏ ሺ ݒ 1, ݒ 2ሻ
՜ ݒ ܽ ݈ ݑ ݁ ܦ ݋ ݉ ݅ ݊ ܽ ݐ ݁ ݏ ሺ ݉ ݋ ݁ 1, ݉ ݋ ݁ 2ሻ )
Fig. 10. Value Dominance Axiom
Fig 11 shows the axiom for outcome dominance. Dominance
is defined for MOEs first and then outcome dominance is
derived by reasoning over the set of MOEs that describe each
outcome. One outcome dominates another outcome if each
MOE that describes the first outcome dominates the second
outcome. Any outcome that is dominated by another outcome
can be pruned from a search space, since it would never be
chosen as there is a better or more preferred outcome.
׊ ݋ 1, ݋ 2, ݉ ݋ ݁ 1, ݉ ݋ ݁ 2  (ݎ ݀ ݂ ݏ : ݏ ݑ ܾ ܥ ݈ ܽ ݏ ݏ ܱ ݂ (݋ 1, ܥ ܱ ܣ
െ ܱ ݑ ݐ ܿ ݋ ݉ ݁ ) ר ݎ ݀ ݂ ݏ : ݏ ݑ ܾ ܥ ݈ ܽ ݏ ݏ ܱ ݂ (݋ 2, ܥ ܱ ܣ
െ ܱ ݑ ݐ ܿ ݋ ݉ ݁ ) ר ݄ ܽ ݏ ܯ ݋ ݁ (݋ 1, ݉ ݋ ݁ 1)
ר ݄ ܽ ݏ ܯ ݋ ݁ (݋ 2, ݉ ݋ ݁ 2)
ר ݒ ܽ ݈ ݑ ݁ ܦ ݋ ݉ ݅ ݊ ܽ ݐ ݁ ݏ (݉ ݋ ݁ 1, ݉ ݋ ݁ 2)
՜ ݋ ݑ ݐ ܿ ݋ ݉ ݁ ܦ ݋ ݉ ݅ ݊ ܽ ݐ ݁ ݏ (݋ 1, ݋ 2)
Fig 11. Outcome Dominance</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>VIII. CONCLUSION AND FUTURE WORK</title>
      <p>This paper has described the preliminary design and of an
ontology for describing COAs and some inference rules for
using preferential reasoning to prune the space of possible
outcomes. KBSI is in the process of implementing this</p>
    </sec>
    <sec id="sec-6">
      <title>COA ontology, including</title>
      <p>contacting the authors.
ontology and using it to reason about COAs that have
significant HSCB characteristics. More information about the
access, can be obtained by</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>[1] "Stability Operations", Headquarters of the Department of the Army</article-title>
          , Field Manual No.
          <fpage>3</fpage>
          -
          <lpage>07</lpage>
          (
          <issue>FM 3-07</issue>
          ),
          <year>October 2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <article-title>[2] "Counterinsurgency", Headquarters Department of the Army</article-title>
          , Field Manual No.
          <fpage>3</fpage>
          -
          <lpage>24</lpage>
          (
          <issue>FM 3-07</issue>
          ),
          <article-title>Headquarters Marine Corps Combat Development Command Department of the Navy</article-title>
          ,
          <source>Marine Corps Warfighting Publication No. 3-33.5 (MCWP 3-33.5)</source>
          ,
          <year>December 2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>[3] "Information Operations: Doctrine, Tactics, Techniques and Procedures"</article-title>
          , Headquarters Department of the Army, Field Manual No.
          <fpage>3</fpage>
          -
          <lpage>13</lpage>
          (
          <issue>FM 3-13</issue>
          ),
          <year>November 2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4] http://www.w3.org/
          <year>2003</year>
          /01/geo/wgs84_pos
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>White</surname>
            ,
            <given-names>C. C.</given-names>
          </string-name>
          , et al $³
          <article-title>OGHPR IR PWXLO - attribute decision making and trade-IR KLWJHZ LPQURDWGH UGXHQ ´L\WXUDQFH (</article-title>
          ,
          <source>DWQULVFR7 on Systems, Man and Cybernetics</source>
          , Vol. SMC-
          <volume>14</volume>
          , No.
          <issue>2</issue>
          , pp.
          <fpage>223</fpage>
          -
          <lpage>229</lpage>
          ,
          <year>1984</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>The</given-names>
            <surname>Description Logic Handbook: Theory</surname>
          </string-name>
          , Implementation, and
          <string-name>
            <surname>Applications</surname>
            , Baader,
            <given-names>F.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Calvanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D. L.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Patel-Schneider P. F</surname>
          </string-name>
          . (eds), Cambridge University Press,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Surdu</surname>
            ,
            <given-names>J. R.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Kittka</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <article-title>The Deep Green concept</article-title>
          .
          <source>In Proceedings of the 2008 Spring Simulation Multiconference (Ottawa, Canada, April 14 - 17</source>
          ,
          <year>2008</year>
          ). Spring Simulation Multiconference. The Society for Computer Simulation International, San Diego, CA,
          <fpage>623</fpage>
          -
          <lpage>631</lpage>
          .
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>[8] XUEH*57GDURZ³VHOSLQFRIU WKHQVLJG IRVLHJQOWGX RIU GJHVKOLD´UNZQR International Journal of Human-Computer Studies</article-title>
          , Vol.
          <volume>43</volume>
          ,
          <string-name>
            <surname>Issues</surname>
          </string-name>
          4-
          <issue>5</issue>
          ,
          <year>November 1995</year>
          , pp.
          <fpage>907</fpage>
          -
          <lpage>928</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>