<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards a Forensic Event Ontology to Assist Video Surveillance-based Vandalism Detection?</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Faranak Sobhani</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Umberto Straccia</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>ISTI-CNR</institution>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Queen Mary University of London</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The detection and representation of events is a critical element in automated surveillance systems. We present here an ontology for representing complex semantic events to assist video surveillance-based vandalism detection. The ontology contains the de nition of a rich and articulated event vocabulary that is aimed at aiding forensic analysis to objectively identify and represent complex events. Our ontology has then been applied in the context of London Riots, which took place in 2011. We report also on the experiments conducted to support the classi cation of complex criminal events from video data.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>In the context of vandalism and terrorist activities, video surveillance forms an
integral part of any incident investigation and, thus, there is a critical need
for developing an \automated video surveillance system" with the capability of
detecting complex events to aid the forensic investigators in solving the criminal
cases. As an example, in the aftermath of the London riots in August 2011
police had to scour through more than 200,000 hours of CCTV videos to identify
suspects. Around 5,000 o enders were found by trawling through the footage,
after a process that took more than ve months.</p>
      <p>
        With the aim to develop an open and expandable video analysis
framework equipped with tools for analysing, recognising, extracting and classifying
events in video, which can be used for searching during investigations with
unpredictable characteristics, or exploring normative (or abnormal) behaviours,
several e orts for standardising event representation from surveillance footage
have been made [
        <xref ref-type="bibr" rid="ref10 ref11 ref22 ref23 ref28 ref30 ref9">9,10,11,22,23,28,30,37</xref>
        ].
      </p>
      <p>While various approaches have relied on o ering foundational support for the
domain ontology extension, to the best of our knowledge, a systematic ontology
for standardising the event vocabulary for forensic analysis and an application
of it has not been presented in the literature so far.</p>
      <p>
        In this paper, we present an OWL 2 [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ] ontology for the semantic retrieval of
complex events to aid video surveillance-based vandalism detection. Speci cally,
the ontology is a derivative of the DOLCE foundational ontology [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] aimed to
? This work was partially funded by the European Union's Seventh Framework
Programme, grant agreement number 607480 (LASIE IP project).
represent events that forensic analysts commonly encounter to aid in the
investigation of criminal activities. The systematic categorisation of a large number
of events aligned with the philosophical and linguistic theories enables the
ontology for interoperability between surveillance systems. We also report on the
experiments we conducted with the developed ontology to support the (semi-)
automatic classi cation of complex criminal events from semantically annotated
video data.
      </p>
      <p>
        Our work signi cantly extends the preliminary works [
        <xref ref-type="bibr" rid="ref12 ref31">12,31</xref>
        ]. The work [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]
is an embryonal work investigating about the use of an ontology for automated
visual surveillance systems, which then has been then further developed in [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ].
While our work shares with [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] some basic principles in the development of the
ontology, here the level of details is now higher (e.g., the Endurant class (see
Section 3.2) and its sub-classes have not been addressed in [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]) and various
ontological errors have been revised. Additionally, and more importantly, in our
work experiments have been conducted for criminal event classi cation based
on London 2011 riots videos. Furthermore, but less related, is [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ] in which the
technical challenges facing researchers in developing computer vision techniques
to process street-scene videos are addressed. The work focusses on standard
image processing methods and does not deal with ontologies in any way.
      </p>
      <p>The remainder of the paper is organised as follows. Related work is addressed
in Section 2. Section 3 presents a detailed description of the forensic ontology
about complex criminal events. In Section 4 we discuss how to use the ontology
to assist video surveillance-based vandalism detection. In Section 5 we conduct
some experiments with our ontology based on CCTV footage of London riots
from 2011, and nally, Section 6 concludes.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        In [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ], the Event Recognition Language (ERL) is presented, which can describe
hierarchical representation of complex spatiotemporal and logical events. The
proposed event structure consists of primitive, single-thread, and multi thread
events. Another event representation ontology, called CASEE , is based on
natural language representation and is proposed in [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and then extended in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
Subsequently, in [
        <xref ref-type="bibr" rid="ref22 ref9">9,22</xref>
        ] a Video Event Representation Language (VERL) was
proposed for describing an ontology of events and the companion language called
Video Event Markup Language (VEML), which is a representation language for
describing events in video sequences based on OWL [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ], event detection
is performed using a set of rules using the SWRL language [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ].
      </p>
      <p>
        The Event Model E [37] has been developed based on an analysis and
abstraction of events in various domains such as research publications, personal
media [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], meetings [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], enterprise collaboration [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] and sports [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. The
framework provides a generic structure for the de nition of events and is extensible to
the requirements ontology of events in the most di erent concrete applications
and domains.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] a formal model of events is presented, called Event-Model-F. The
model is based on the foundational ontology DOLCE+DnS Ultralite (DUL) and
provides comprehensive support to represent time and space, objects and
persons, as well as mereological, casual, and correlative relationships between events.
In addition, the Event-Model-F provides a exible means for event composition,
modelling event causality and event correlation, and representing di erent
interpretations of the same event. The Event-Model-F is developed following the
pattern-oriented approach of DUL, is modularised in di erent ontologies, and
can be easily extended by domain speci c ontologies.
      </p>
      <p>While the above-mentioned approaches essentially provide frameworks for
the representation of events, none of them address the problem of formalising
forensic events in terms of a standard representation language such as OWL 23
and, importantly, none have been applied and tested so far in a real use case,
which are the topics of the following sections.
3</p>
    </sec>
    <sec id="sec-3">
      <title>A Forensic Event Ontology</title>
      <p>In the following, we present an OWL 2 ontology to support to some extent the
semantic retrieval of complex events to aid automatic or semi-automatic video
surveillance-based vandalism detection. The idea is to develop an ontology that
not only conveys a shared vocabulary, but some inferences based on it may assist
a human being to support the video analysis by hinting to videos that may be
more relevant than others in the detection of criminal events.
3.1</p>
      <sec id="sec-3-1">
        <title>The Role of a Foundation Ontology</title>
        <p>
          To facilitate the elimination of the terminological ambiguity and the
understanding and interoperability among people and machines [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ], it is common
practice to consider a so-called foundational ontology. Let us note that several
e orts have been taken by researchers in de ning the foundational ontologies,
such as BFO,4 SUMO,5 UFO6 and DOLCE,7 to name a few. As DOLCE
ontology o ers a cognitive bias with the ontological categories underlying natural
language and human common sense, the same is selected for our proposed
extension. We recall that the DOLCE foundational ontology encompasses Endurant
and Perdurant entities. Endurant entities are ever-present at any time as opposed
to Perdurant entities that extended in time by accumulating di erent temporal
parts. A more thorough explanation on the DOLCE events conceptualisation
can be found e.g. in [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
3 We recall that the relationship to our previous work [
          <xref ref-type="bibr" rid="ref12 ref31 ref32">12,31,32</xref>
          ] has been addressed
in the introductory section.
4 http://ifomis.uni-saarland.de/bfo/
5 http://www.adampease.org/OP/
6 https://oxygen.informatik.tu-cottbus.de/drupal7/ufo/
7 http://www.loa.istc.cnr.it/old/Papers/DOLCE2.1-FOL.pdf
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>A Forensic Complex Event Ontology</title>
        <p>
          Our complex event classes extend DOLCE's Perdurant class. To assign the action
classes into respective categories, we follow a four-way classi cation of
actionverbs: namely, into State, Process, Achievement and Accomplishment using event
properties such as telic, stage and cumulative (see [
          <xref ref-type="bibr" rid="ref27">27,35,36</xref>
          ]). The distinction
between these concepts are derived from the event properties as illustrated in
Table 1, which we summarise below.
{ State [-telic,-stage] This action category represents a long, non-dynamic
event in which every instance is the same: there cannot be any distinction
made between the stages. States are cumulative and homogenous in nature.
{ Process [-telic, +stage] The action category, like State, is atelic, but unlike
State, the action undertaken are dynamic. The actions appear progressively
and thus can be split into a set of stages for analysis.
{ Accomplishment [+telic, +stage] Accomplishments are telic and
cumulative activities, and thus behave di erently from both State and Process.
The performed action can be analysed in stages and in this way, they are
similar to Process. Intuitively, an accomplishment is an activity which moves
toward a nishing point as it has variously been called in the literature.
Accomplishment is also cumulative activity.
{ Achievement [+telic, -stage] Achievements are similar to Accomplishment
in their telicity. They are also not cumulative with respect to contiguous
events. Achievements do not go on or progress, because they are near
instantaneous, and are over as soon as they have begun.
        </p>
        <p>Forensic Perdurant Entities. Perdurant entities extend in time by
accumulating di erent temporal parts and some of their proper temporal parts may be
not present. To this end, Perdurant entities are divided into the classes Event and
Stative, classi ed according to their temporal characteristics.</p>
        <p>The axiom sets below provide a subset of our formal extension of the Perdurant
vocabulary.</p>
        <sec id="sec-3-2-1">
          <title>Perdurant v SpatioTemporalParticular</title>
        </sec>
        <sec id="sec-3-2-2">
          <title>Perdurant v 9participant:Endurant</title>
        </sec>
        <sec id="sec-3-2-3">
          <title>Fighting v 9participant:GroupOfPeople</title>
        </sec>
        <sec id="sec-3-2-4">
          <title>Perdurant v :Endurant</title>
        </sec>
        <sec id="sec-3-2-5">
          <title>Kicking v :Vehicle</title>
        </sec>
        <sec id="sec-3-2-6">
          <title>State v Stative</title>
        </sec>
        <sec id="sec-3-2-7">
          <title>MetaLevelEvent v State</title>
        </sec>
        <sec id="sec-3-2-8">
          <title>Accusing v MetaLevelEvent</title>
        </sec>
        <sec id="sec-3-2-9">
          <title>Believing v MetaLevelEvent</title>
        </sec>
        <sec id="sec-3-2-10">
          <title>PsycologicalAggression v State</title>
        </sec>
        <sec id="sec-3-2-11">
          <title>Blaming v PsycologicalAggression</title>
        </sec>
        <sec id="sec-3-2-12">
          <title>Bullying v PsycologicalAggression Process v Stative Action v Process Gesture v Process</title>
          <p>Event</p>
          <p>Stative
Achievement</p>
          <p>Accomplishment
Saying
Seeing</p>
          <p>Physical
Aggression</p>
          <p>Process</p>
          <p>Action
Gesture</p>
          <p>State
MataLevel</p>
          <p>Event
Psycological
Aggression
An excerpt of the forensic ontology is shown in Figure 1.</p>
          <p>The concept State o ers representation for MetaLevelEvent which
encompasses abstract human events such as Accusing, Believing and Liking among
others. As previously stated, the concept State represents a collection of events which
are exhibited by a human that is time-consuming, non-dynamic, cumulative and
homogenous. The other sub-class of State is PsychologicalAggression which
characterises the human actions such as Blaming, Decrying, Harassing and so forth.
The concept Process includes several human action categories that represent
dynamic events which can be split into several intermediate stages for analysis.
For the purposes of clarity, the concept Process o ers three sub-concepts namely
Action, Gesture and PhysicalAggression. The Action class incorporates di erent
event such as Dancing, Greeting, Hugging among other concepts de ned. The
concept Gesture formalises the di erent interest points related to human gestures. In
order to eliminate the ambiguity traditionally present in human gestures across
cross-cultural impact, the action performed during the gesture is captured and
represented in the ontology and, thus, enabling the removal of subjectivity from
the concept de nition. The nal sub-class of the Process class includes the
concept PhysicalAggression and formalises human con icting actions.</p>
          <p>By and large, the human actions categorised into State and Process represent
the microscopic movements of humans.</p>
          <p>From the automatic surveillance viewpoint, these microscopic events may be
extracted from media items. In contrast, the event representations formalised
by means of the concepts Achievement and Accomplishment o er a rich
combination of human events that allow for the construction of complex events with
or without the combination of microscopic features. For instance, the concept
Unlawful
Entry</p>
          <p>Attempted</p>
          <p>ForcibleEntry
Forcible
Entry
Entering
Property
Cyber
stalking
Cyber
mobbing</p>
          <p>TheftOf</p>
          <p>Information
TheftOf
Identity</p>
          <p>TheftOf
Password
Vandalism
Molotov
Throwing</p>
          <p>Gun
Shot
Damage
Structure
Damage</p>
          <p>Apartment
Cyber
Bullying
Cyber
Crime</p>
          <p>Phishing</p>
          <p>Blackmail
Cyber
Threat</p>
          <p>Botnet
Malware</p>
          <p>Hacking
hierarchy for Vandalism is illustrated in Figure 2, while the concept hierarchy for
CyberCrime is shown in Figure 3 instead.</p>
          <p>
            Forensic Endurant Entities. DOLCE is based on fundamental distinction
among Endurant and Perdurant entities. The di erence between Endurant and
Perdurant entities is related to their behaviour in time. Endurant are wholly
present at any time they are present. Philosophers believe that endurant are
entities that are in time while lacking, however, temporal parts [
            <xref ref-type="bibr" rid="ref19">19</xref>
            ]. Therefore,
the proposed vocabulary structure of all possible forensic entities also extends
on Endurants entities.
Social
Object
Agentive
Physical
Object
NonAgentive
PhysicalObject
          </p>
          <p>Material
Artifact</p>
          <p>Axiom set (1) describes a subset formalization of the Endurant vocabulary
and an excerpt of the forensic extension of the ontology structure shown in
Figure 4.</p>
        </sec>
        <sec id="sec-3-2-13">
          <title>Endurant v SpatioTemporalParticular</title>
        </sec>
        <sec id="sec-3-2-14">
          <title>Endurant v 9participantIn:Perdurant</title>
          <p>participantIn = participant</p>
        </sec>
        <sec id="sec-3-2-15">
          <title>NonPhysicalEndurant v Endurant</title>
        </sec>
        <sec id="sec-3-2-16">
          <title>PhysicalEndurant v Endurant</title>
        </sec>
        <sec id="sec-3-2-17">
          <title>ArbitrarySum v Endurant :</title>
          <p>(1)
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Assisting Video Surveillance-based Vandalism</title>
    </sec>
    <sec id="sec-5">
      <title>Detection</title>
      <p>We next show how the so far developed ontology is expected to be used to assist
video surveillance-based vandalism detection.
4.1</p>
      <sec id="sec-5-1">
        <title>Annotating Media Objects, viz. Surveillance Videos</title>
        <p>Given surveillance videos and any media in general, we need a method to
annotate them by using the terminology provided by our ontology. This gives rise to
a set of facts that, together with the inferred facts, may support a more e ective
automatic or, more likely, semi-automatic retrieval of relevant information, such
as e.g. vandalic acts. Speci cally, the inferred information may suggest a user
look at some e.g. video sequences or video still images, rather than to others
rst.</p>
        <p>
          The general model we are inspired on is based on [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. Conceptually,
according to [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], a media object o (e.g., an image region, a video sequence, a piece
of text, etc.) is annotated with one (or more) entities t of the ontology (see
e.g. Figure 5).
        </p>
        <p>For instance, stating that an image object o is about a DamageVehicle can be
represented conceptually via the DL expression</p>
        <p>
          (9isAbout:DamageVehicle)(o) :
As speci ed in [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ], such an annotation may come manually from a user or, if,
available, from an image classi er. In the latter case, it may annotate the image
automatically, or, semi-automatically by suggesting to a human annotator, which
are the most relevant entities of the ontology that may be used for a speci c
media object o. Note, however, that, the above methodology just illustrates the
concept. In our case, for the sake of ease the annotation, we may not enforce
the use of the object property isAbout (see Example 2 later on). Generally, we
will annotate a Resource with Perdurants and Endurants: thus, if an image is
annotated with e.g. a perdurant that is a damaged vehicle, then this means that
the image is about a damaged vehicle.
        </p>
        <p>We recall that Resources (and Sources) are modelled as follows:</p>
        <sec id="sec-5-1-1">
          <title>Source v Endurant u 9has:Resource</title>
          <p>u9hasCameraId:string
u9hasLatitude:string
u9hasLongitute:string
u9hasLocationName:string</p>
        </sec>
        <sec id="sec-5-1-2">
          <title>Resource v Endurant u 9has:Perdurant</title>
          <p>has = isFrom
has has v has :
Note that in the last role inclusion axiom, is role composition and, thus,
has has v has dictates that the property has is transitive, while with has =
isFrom we say that isFrom is the inverse of has. Therefore, isFrom is transitive
as well.</p>
          <p>The following example illustrates the mechanism of image of annotation
together with a meaningful inference.</p>
          <p>Example 1. Consider the following DL axioms resulting from annotating images
of a video (video6) registered by a camera (cameraC004):
participateIn(personA; throwing5) ; Throwing(throwing5)</p>
        </sec>
        <sec id="sec-5-1-3">
          <title>NaturalPerson(personA) ; Throwing v ActivePhysicalAggression</title>
        </sec>
        <sec id="sec-5-1-4">
          <title>ActivePhysicalAggression v PhysicalAggression ; PhysicalAggression v Process</title>
          <p>isFrom(throwing5; endurant6) ; Resource(endurant6)
hasVideoId(endurant6; video6) ; Source(endurant7)
hasCameraId(endurant7; cameraC004) ; has(endurant7; endurant6) :
Now, as isFrom is transitive, we may infer:
Then, it is not di cult to see that we
nally infer
isFrom(throwing5; endurant7) :</p>
        </sec>
        <sec id="sec-5-1-5">
          <title>9paticipateIn:(PhysicalAggression u</title>
        </sec>
        <sec id="sec-5-1-6">
          <title>9isFrom:(Source u 9hasCameraID:fcameraC004g)) (PersonA) ;</title>
          <p>which can be read as:
\A person (PersonA) participated in a physical aggression that has been
registered by camera C004".
4.2</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>Modelling GCIs for Vandalism Event Detection</title>
        <p>As we are focusing on forensic domain and dealing with variety of concepts
aiming at aiding forensic analysis, to objectively identify and represent complex
events, we next show that a (manually build) General Concept Inclusion (GCI)
axiom may help to classify high-level events in terms of a composition of some
lower level events. The following are such GCI examples:</p>
        <sec id="sec-5-2-1">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-2-2">
          <title>9participant:(Vehicle u</title>
        </sec>
        <sec id="sec-5-2-3">
          <title>9participantIn:(BreakingDoor t BreakingWindows)) v DamageVehicle :</title>
          <p>\If an event involves a vehicle that is subject of a breaking door or
breaking windows then the event is about a damaged vehicle" (see
Figure 6).</p>
        </sec>
      </sec>
      <sec id="sec-5-3">
        <title>DamageStructure:</title>
        <sec id="sec-5-3-1">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-3-2">
          <title>9participant:(Structure u</title>
        </sec>
        <sec id="sec-5-3-3">
          <title>9participantIin:Kicking) v DamageStructure :</title>
          <p>\If an event involves a structure that is subject of kicking, then the
event is about a damaged structure" (see Figure 6).</p>
          <p>The following example illustrates the use of such GCIs.</p>
          <p>Example 2. Suppose we have an image classi er that is able to provide us with
the following facts. Speci cally, assume it is able to identify vehicles and breaking
windows:</p>
          <p>participant(Perdurant2; Endurant1); Vehicle(Endurant1) ; BreakingWindows(Perdurant2) :
From these facts and the GCI about DamageVehicle, we may infer that the
image is about a damaged vehicle, i.e. we may infer</p>
          <p>DamageVehicle(Perdurant2) :
The following set of GCIs illustrates instead how one may have multiple GCIs
to classify a single event, such as those for Vandalism (see, e.g. Figure 7).8
Note that in the example above, we assume that events (perdurant) may be
complex in the sense that they may compose by multiple sub-events (parts). So,
e.g. in the last GCI, we roughly state
8 Recall that all these GCIs provide su cient conditions to be an instance of Vandalism,
but no necessary condition.</p>
          <p>\If a (complex) event involves both throwing and an explosion (two
subevents) then the event is about vandalism".</p>
          <p>Following our previous examples, we next are going to formulate another kind of
background knowledge. Our main focus in this example is on recognizing
highlevel events, which occur in the same location (same street in our modelling).
In order to model this scenario, we may use the Semantic Web Rule Language
(SWRL) to model the locatedSameAs role and then use it in GCIs. The SWRL
rule is:
\Two perdurants that occur in the same street occur in the same place.</p>
          <p>Perdurant(?p1); Perdurant(?p2); hasLocationName(?p1; ?l1);</p>
          <p>hasLocationName(?p2; ?l2); SameAs(?l1; ?l2) ! locatedSameAs(?p1; ?p2) :</p>
          <p>The following axioms illustrate how to use the previously de ned relation (few
examples captured from our data set by these rules are illustrated in Figure 8).</p>
        </sec>
        <sec id="sec-5-3-4">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-3-5">
          <title>9part:(Crowding u 9locatedSameAs:Explosion) v Vandalism</title>
        </sec>
        <sec id="sec-5-3-6">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-3-7">
          <title>9part:(Crowding u 9locatedSameAs:DamageStructure) v Vandalism</title>
        </sec>
        <sec id="sec-5-3-8">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-3-9">
          <title>9part:(Crowding u 9locatedSameAs:Throwing) v Vandalism</title>
        </sec>
        <sec id="sec-5-3-10">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-5-3-11">
          <title>9part:(DamageStructure u 9locatedSameAs:Throwing) v Vandalism :</title>
          <p>5</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Experiments</title>
      <p>We conducted two experiments with our ontology, which we are going to describe
in the following.9</p>
      <p>In the rst case, we evaluated the classi cation e ectiveness of manually built
GCIs to identify crime events, while in the second case we drop the
manualbuilt GCIs and, try to learn such GCIs instead automatically from examples
and compare their e ectiveness with respect to the manually built ones.
9 The ontologies used in the experiments and experimental results can be found at
http://www.umbertostraccia.it/cs/ftp/ForensicOntology.zip.
5.1</p>
      <sec id="sec-6-1">
        <title>Classi cation via Manually Built GCIs</title>
        <p>Roughly, we have considered a number of crime videos, annotated them manually
and then checked whether the manually built GCIs, as described in Section 4.2,
were able to determine crime events correctly.</p>
        <p>Setup. Speci cally, we considered our ontology and around 3.07 TB of video
data about the London riot 2011,10 of which 929 (GB) is in a non-proprietary
format. We considered 140 videos (however, the videos cannot be made publicly
available). Within these videos, all the available CCTV cameras (35 CCTV)
along with their features such as latitude, longitude, start time, end time and
street name, have been annotated manually according to our methodology
described in Section 4 and included into our ontology. We have also calculated all
the geographic distances between each camera. The resulting ontology contains
1800 created individuals of which, e.g. 106 are of type Event.</p>
        <p>Vandalism (13; 57) Riot (4; 21) AbnormalBehavior (2; 80)
Crowding (1; 64) DamageStructure (3; 9) DamageVehicle (3; 16)</p>
        <p>Throwing (1; 30)
Then, we considered criminal events occurring in the videos (speci cally, we
focused on vandalic events). For each class of events, we manually built one or
more GCIs, as illustrated in Section 4.2. The list of crime events considered is
reported in Table 2. In it, the rst number in parenthesis reports the number of
GCIs we built for each of them, while the second number indicates the number
of event instances (individuals) we created during the manual video
annotation process. So, for instance, for the event DamageStructure we have built 3
classi cation GCIs and we have created 9 instances of DamageStructure during
the manual video annotation process. For further clari cation, the 3 GCIs for
DamageStructure are</p>
        <sec id="sec-6-1-1">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-6-1-2">
          <title>9participant:(Structure u</title>
        </sec>
        <sec id="sec-6-1-3">
          <title>9participantIin:Kicking) v DamageStructure</title>
        </sec>
        <sec id="sec-6-1-4">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-6-1-5">
          <title>9participant:(Structure u</title>
        </sec>
        <sec id="sec-6-1-6">
          <title>9participantIin:Beating) v DamageStructure</title>
        </sec>
        <sec id="sec-6-1-7">
          <title>Perdurant u</title>
        </sec>
        <sec id="sec-6-1-8">
          <title>9participant:(Structure u</title>
        </sec>
        <sec id="sec-6-1-9">
          <title>9participantIin:BreakingWindows) v DamageStructure ;</title>
          <p>while, e.g., an instance of DamageStructure is the individual Kicking1, whose
related information excerpt is:</p>
          <p>Kicking(Kicking1); isFrom(Kicking1; 2bdf); Resource(2bdf); isFrom(2bdf; C004);
has(2bdf; pr11); part(pr11; Kicking1); part(pr11; BreackingWindows3);</p>
          <p>BreackingWindows(BreackingWindows3); : : :
As a matter of general information, the global metric statistics of the so built
ontology is reported in Table 3.
10 These are part of the EU funded project LASIE \Large Scale Information
Exploitation of Forensic Data", http://www.lasie-project.eu.
Evaluation. Let O be the built ontology from which we drop axioms stating
explicitly that an individual is an instance of a crime event listed in Table 2.
Please note that without the GCIs none of the crime events instances in O can
be inferred to be instances of the crime events in Table 2.11 Now, on O we run
an OWL 2 reasoner that determines the instances of all crime event classes in
the ontology.</p>
          <p>To determine the classi cation e ectiveness of the GCIs, we compute the
so-called micro/macro averages of precision, recall and F1-score w.r.t. inferred
data. The evaluation result of the rst test is shown in Table 4.
In the second experiment, we apply a concept learning approach to replace the
manually built GCIs describing the crime events listed in Table 2. To this end,
11 Roughly, crime events are subclasses of the Event class, while crime event instances
are instances of the class Stative (see Figure 3).
the DL-Learner12 system was used to learn descriptions of the criminal events
in Table 2, based on existing instances of these classes.</p>
          <p>
            Setup. Let now O be the ontology as in Section 5.1, but from which we
also drop additionally the manually created GCIs for the crime event listed in
Table 2. On it we used the CELOE algorithm [
            <xref ref-type="bibr" rid="ref15 ref6">6,15</xref>
            ] with its default settings to
generate suggestion de nitions (inclusion axioms) for each target class C.
          </p>
          <p>
            Speci cally, we used a K-fold cross style validation method [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ], which divides
the available crime event instances into a K disjoint subsets. That is, we split
each target class C into K disjoint subsets C1; : : : ; CK of equal size. In our
experiment, K is the number of instances of C and, thus, each Ci has size one
For each Ci, the training set is (SK
          </p>
          <p>i=1 Ci) n Ci and is denoted as T rainseti. Then,
for each Ci we run CELOE on the training set T rainseti and generated at most
10 class expressions of the form Dj v Ci, out of which we have chosen the best
solution (denoted DCi v Ci or GCIi). If the best solution is not unique, we
select the rst listed one.</p>
          <p>The best-selected GCIs found by CELOE for each of the target classes in
Table 2 are:</p>
        </sec>
        <sec id="sec-6-1-10">
          <title>PhysicalAggression u 9immediateRelation:Structure v DamageStructure</title>
        </sec>
        <sec id="sec-6-1-11">
          <title>9immediateRelation:Vehicle v DamageVehicle</title>
        </sec>
        <sec id="sec-6-1-12">
          <title>9immediateRelation:Vandalism v AbnormalBehavior</title>
        </sec>
        <sec id="sec-6-1-13">
          <title>9immediateRelation:Arm v Throwing</title>
        </sec>
        <sec id="sec-6-1-14">
          <title>9immediateRelation:Group v Crowding :</title>
          <p>With the help of a reasoner, we then infer all instances in O, that are not in
T rainseti, that are instances of the selected DCi and consider them as our result
set (denoted Resultseti).</p>
          <p>Evaluation. To determine the classi cation e ectiveness of the learned GCIs,
i.e. of GCIi, average precision, recall and F1 measures across the folds are
computed. The evaluation results of the second test are shown in Table 5.
Discussion. The results are generally promising. In the manually built GCI case,
precision and F1 are reasonably good, though in one case (Riot) the recall and,
thus, F1 is not satisfactory. For the learned GCI case, the individual measures
are generally comparable to the manual ones.
12 http://dl-learner.org/</p>
          <p>Given that the learned GCIs are completely di erent than the manually built
ones, it is surprising that both sets perform more or less the same. However,
please note that DL-Learner was neither able to learn a GCI for Vandalism nor
for Riot. This fact is re ected in the generally worse micro/macro precision, recall
and F1 measures.</p>
          <p>Eventually, we also merged the manually built GCIs and the learned ones
together and tested them as in Section 5.1. The results in Table 6 show, however,
that globally their e ectiveness is as for the manual case (and does not improve).
In this work, we have proposed an extensive ontology for representing complex
criminal events. The proposed ontology focuses on events that are often required
by forensic analysts. In this context, the Perdurant, as de ned in the DOLCE
ontology as an occurrence in time, and the Endurant, de ned in the DOLCE
ontology as contentious in time, have both been extended to represent all forensics
entities together with meaningful entities for video surveillance-based vandalism
detection. The aim of the built ontology is to support the interoperability of the
automated surveillance system.</p>
          <p>To classify high-level events in terms of the composition of lower level events
we focused on both manually built and automatically learned GCIs and have
compared the evaluation results of both experiments. The results are generally
promising and the e ectiveness of machine derived de nitions for high-level crime
events is encouraging though needs further development.</p>
          <p>
            In the future, we intend to deal with vague or imprecise knowledge and we
would like to work on the problem of automatically learn fuzzy concept
description [
            <xref ref-type="bibr" rid="ref16 ref17 ref18 ref33 ref34 ref4 ref5">4,5,16,17,18,33,34</xref>
            ] as most of the involved entities are fuzzy by nature.
30th Annual ACM Symposium on Applied Computing (SAC-15). pp. 345{352.
          </p>
          <p>ACM, Salamanca, Spain (2015)
35. Vendler, Z.: Verbs and times. The Philosophical Review 62(2), 143{160 (1957)
36. Vendler, Z. (ed.): Linguistics in Philosophy. G - Reference, Information and
Interdisciplinary Subjects Series, Cornell University Press (1967), https://books.
google.co.uk/books?id=OR1DAAAAIAAJ
37. Westermann, U., Jain, R.: Toward a common event model for multimedia
applications. IEEE Multimedia 14(1) (2007)</p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Appan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sundaram</surname>
          </string-name>
          , H.:
          <article-title>Networked multimedia event exploration</article-title>
          .
          <source>In: Proceedings of the 12th Annual ACM International Conference on Multimedia</source>
          . pp.
          <volume>40</volume>
          {
          <fpage>47</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Calvanese</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patel-Schneider</surname>
            ,
            <given-names>P.F</given-names>
          </string-name>
          . (eds.):
          <article-title>The Description Logic Handbook: Theory, Implementation, and Applications</article-title>
          . Cambridge University Press (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Baader</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sattler</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Description logics</article-title>
          . In: Staab,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Studer</surname>
          </string-name>
          ,
          <string-name>
            <surname>R</surname>
          </string-name>
          . (eds.) Handbook on Ontologies, pp.
          <volume>21</volume>
          {
          <fpage>43</fpage>
          .
          <source>International Handbooks on Information Systems</source>
          , Springer Verlag (
          <year>2009</year>
          ), https://doi.org/10.1007/ 978-3-
          <fpage>540</fpage>
          -92673-
          <issue>3</issue>
          _
          <fpage>1</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bobillo</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Fuzzy ontology representation using OWL 2</article-title>
          .
          <source>International Journal of Approximate Reasoning</source>
          <volume>52</volume>
          ,
          <issue>1073</issue>
          {
          <fpage>1094</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bobillo</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>The fuzzy ontology reasoner fuzzyDL</article-title>
          .
          <source>Knowledge-Based Systems 95</source>
          ,
          <fpage>12</fpage>
          {
          <fpage>34</fpage>
          (
          <year>2016</year>
          ), http://www.sciencedirect.com/science/article/ pii/S0950705115004621
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6. Buhmann, L.,
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Westphal</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>DL-Learner framework for inductive learning on the semantic web</article-title>
          .
          <source>Web Semantics: Science, Services and Agents on the World Wide Web</source>
          <volume>39</volume>
          ,
          <issue>15</issue>
          {
          <fpage>24</fpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Casati</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varzi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : Events. In: Zalta,
          <string-name>
            <surname>E.N.</surname>
          </string-name>
          <article-title>(ed</article-title>
          .)
          <source>The Stanford Encyclopedia of Philosophy. Winter</source>
          <year>2015</year>
          edn. (
          <year>2015</year>
          ), http://plato.stanford.edu/archives/ win2015/entries/events/
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Forman</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scholz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Apples-to-apples in cross-validation studies: pitfalls in classi er performance measurement</article-title>
          .
          <source>ACM SIGKDD Explorations Newsletter</source>
          <volume>12</volume>
          (
          <issue>1</issue>
          ),
          <volume>49</volume>
          {
          <fpage>57</fpage>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Francois</surname>
            ,
            <given-names>A.R.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nevatia</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hobbs</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bolles</surname>
          </string-name>
          , R.C.
          <article-title>: VERL: an ontology framework for representing and annotating video events</article-title>
          .
          <source>IEEE MultiMedia 12(4)</source>
          ,
          <volume>76</volume>
          {
          <fpage>86</fpage>
          (
          <year>2005</year>
          ), https://doi.org/10.1109/MMUL.
          <year>2005</year>
          .87
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Hakeem</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sha</surname>
            <given-names>que</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            ,
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.:</surname>
          </string-name>
          <article-title>An object-based video coding framework for video sequences obtained from static cameras</article-title>
          .
          <source>In: Proceedings of the 13th Annual ACM International Conference on Multimedia</source>
          . pp.
          <volume>608</volume>
          {
          <fpage>617</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Hakeem</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sheikh</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>CASEE: A hierarchical event representation for the analysis of videos</article-title>
          .
          <source>In: Proceedings of the 19th National Conference on Arti cial Intelligence</source>
          . pp.
          <volume>263</volume>
          {
          <fpage>268</fpage>
          . AAAI Press (
          <year>2004</year>
          ), http://dl.acm.org/citation.cfm? id=
          <volume>1597148</volume>
          .
          <fpage>1597192</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Henderson</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blasi</surname>
            ,
            <given-names>S.G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sobhani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          , ,
          <string-name>
            <surname>Izquierdo</surname>
          </string-name>
          , E.:
          <article-title>On the impurity of streetscene video footage</article-title>
          .
          <source>In: 6th International Conference on Imaging for Crime Prevention and Detection (ICDP-15)</source>
          .
          <source>The Institution of Engineering and Technology (IET)</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Jain</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Experiential meeting system</article-title>
          .
          <source>In: Proceedings of the 2003 ACM SIGMM Workshop on Experiential Telepresence</source>
          . pp.
          <volume>1</volume>
          {
          <fpage>12</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2003</year>
          ), http://doi.acm.
          <source>org/10</source>
          .1145/982484.982486
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Podlaseck</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pingali</surname>
          </string-name>
          , G.:
          <article-title>Personal chronicling tools for enhancing information archival and collaboration in enterprises</article-title>
          .
          <source>In: Proceedings of the the 1st ACM Workshop on Continuous Archival and Retrieval of Personal Experiences</source>
          . pp.
          <volume>56</volume>
          {
          <fpage>65</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2004</year>
          ), http://doi.acm.
          <source>org/10</source>
          .1145/1026653.1026662
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Lehmann</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Auer</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , Buhmann, L.,
          <string-name>
            <surname>Tramp</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Class expression learning for ontology engineering</article-title>
          .
          <source>Web Semantics: Science, Services and Agents on the World Wide Web</source>
          <volume>9</volume>
          (
          <issue>1</issue>
          ),
          <volume>71</volume>
          {
          <fpage>81</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Lisi</surname>
            ,
            <given-names>F.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>A logic-based computational method for the automated induction of fuzzy ontology axioms</article-title>
          .
          <source>Fundamenta Informaticae</source>
          <volume>124</volume>
          (
          <issue>4</issue>
          ),
          <volume>503</volume>
          {
          <fpage>519</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Lisi</surname>
            ,
            <given-names>F.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>A foil-like method for learning under incompleteness and vagueness</article-title>
          .
          <source>In: 23rd International Conference on Inductive Logic Programming. Lecture Notes in Arti cial Intelligence</source>
          , vol.
          <volume>8812</volume>
          , pp.
          <volume>123</volume>
          {
          <fpage>139</fpage>
          . Springer Verlag, Berlin (
          <year>2014</year>
          ), revised Selected Papers
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Lukasiewicz</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Managing uncertainty and vagueness in description logics for the semantic web</article-title>
          .
          <source>Journal of Web Semantics</source>
          <volume>6</volume>
          ,
          <issue>291</issue>
          {
          <fpage>308</fpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Masolo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Borgo</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gangemi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guarino</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oltramari</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Wonderweb Deliverable D18, Ontology Library ( nal)</article-title>
          .
          <source>ICT project 33052</source>
          (
          <year>2003</year>
          ), http:// wonderweb.man.ac.uk/deliverables/documents/D18.pdf
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Meghini</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sebastiani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>A model of multimedia information retrieval</article-title>
          .
          <source>Journal of the ACM</source>
          <volume>48</volume>
          (
          <issue>5</issue>
          ),
          <volume>909</volume>
          {
          <fpage>970</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Motik</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grau</surname>
            ,
            <given-names>B.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horrocks</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fokoue</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lutz</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , et al.:
          <article-title>OWL 2 Web Ontology Language Pro les</article-title>
          .
          <source>W3C recommendation 27</source>
          ,
          <issue>61</issue>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Nevatia</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hobbs</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bolles</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>An ontology for video event representation</article-title>
          .
          <source>In: IEEE Conference on Computer Vision and Pattern Recognition Workshops</source>
          . p.
          <fpage>119</fpage>
          . IEEE Computer Society (
          <year>2004</year>
          ), https://doi.org/10.1109/CVPR.
          <year>2004</year>
          .301
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Nevatia</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hongeng</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Hierarchical language-based representation of events in video streams</article-title>
          .
          <source>In: IEEE Conference on Computer Vision and Pattern Recognition</source>
          . p.
          <fpage>39</fpage>
          . IEEE Computer Society (
          <year>2003</year>
          ), https://doi.org/10.1109/ CVPRW.
          <year>2003</year>
          .10038
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>OWL</surname>
            ,
            <given-names>S.A.S.W.R.L.C.</given-names>
          </string-name>
          , RuleML: https://www.w3.org/Submission/SWRL/.
          <source>W3C</source>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <article-title>OWL 2 Web Ontology Language Document Overview</article-title>
          : http://www.w3.org/TR/ 2009/REC-owl2
          <string-name>
            <surname>-</surname>
          </string-name>
          overview-20091027/. W3C (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Pingali</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Opalach</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jean</surname>
            ,
            <given-names>Y.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carlbom</surname>
            ,
            <given-names>I.B.</given-names>
          </string-name>
          :
          <article-title>Instantly indexed multimedia databases of real world events</article-title>
          .
          <source>IEEE Transactions on Multimedia</source>
          <volume>4</volume>
          (
          <issue>2</issue>
          ),
          <volume>269</volume>
          {
          <fpage>282</fpage>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>Rothstein</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>Chapter 1: Verb classes and aspectual classi cation</article-title>
          .
          <source>In: Structuring Events: A Study in the Semantics of Lexical Aspect</source>
          , pp.
          <volume>1</volume>
          {
          <fpage>35</fpage>
          . Wiley Online Library (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Scherp</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Franz</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Saatho</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Staab</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>F{A model of events based on the foundational ontology DOLCE+DnS ultralight</article-title>
          .
          <source>In: Proceedings of the Fifth International Conference on Knowledge Capture</source>
          . pp.
          <volume>137</volume>
          {
          <fpage>144</fpage>
          .
          <string-name>
            <surname>ACM</surname>
          </string-name>
          (
          <year>2009</year>
          ), http: //doi.acm.
          <source>org/10</source>
          .1145/1597735.1597760
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <surname>Schmidt-Schau</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smolka</surname>
          </string-name>
          , G.:
          <article-title>Attributive concept descriptions with complements</article-title>
          .
          <source>Arti cial Intelligence</source>
          <volume>48</volume>
          ,
          <issue>1</issue>
          {
          <fpage>26</fpage>
          (
          <year>1991</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Snidaro</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Belluz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Foresti</surname>
            ,
            <given-names>G.L.</given-names>
          </string-name>
          :
          <article-title>Representing and recognizing complex events in surveillance applications</article-title>
          .
          <source>In: Fourth IEEE International Conference on Advanced Video and Signal Based Surveillance</source>
          . pp.
          <volume>493</volume>
          {
          <fpage>498</fpage>
          . IEEE Computer Society (
          <year>2007</year>
          ), https://doi.org/10.1109/AVSS.
          <year>2007</year>
          .4425360
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Sobhani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chandramouli</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Izquierdo</surname>
          </string-name>
          , E.:
          <article-title>Formal representation of events in a surveillance domain ontology</article-title>
          .
          <source>In: 2016 IEEE International Conference on Image Processing</source>
          . pp.
          <volume>913</volume>
          {
          <fpage>917</fpage>
          . IEEE Computer Society (
          <year>2016</year>
          ), https://doi. org/10.1109/ICIP.
          <year>2016</year>
          .7532490
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Sobhani</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kahar</surname>
            ,
            <given-names>N.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          :
          <article-title>An ontology framework for automated visual surveillance system</article-title>
          .
          <source>In: 13th International Workshop on Content-Based Multimedia Indexing</source>
          . pp.
          <volume>1</volume>
          {
          <issue>7</issue>
          .
          <string-name>
            <given-names>IEEE</given-names>
            <surname>Computer</surname>
          </string-name>
          <article-title>Society (</article-title>
          <year>2015</year>
          ), https://doi.org/10.1109/ CBMI.
          <year>2015</year>
          .7153628
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Foundations of Fuzzy Logic and Semantic Web Languages</article-title>
          .
          <source>CRC Studies in Informatics Series</source>
          , Chapman &amp; Hall (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Straccia</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mucci</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>pFOIL-DL: Learning (fuzzy) EL concept descriptions from crisp OWL data using a probabilistic ensemble estimation</article-title>
          .
          <source>In: Proceedings of the</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>