<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Mobile objects and sensors within a video surveillance system: Spatio-temporal model and queries</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dana Codreanu</string-name>
          <email>codreanu@irit.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ana-Maria Manzat</string-name>
          <email>manzat@irit.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Florence Sedes</string-name>
          <email>sedes@irit.fr</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Université de Toulouse - IRIT - UMR</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>118 Route de Narbonne</institution>
          ,
          <addr-line>31062 Toulouse Cedex 9</addr-line>
          ,
          <country country="FR">France</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The videos recorded by video surveillance systems represent a key element in a police inquiry. Based on a spatio-temporal query specified by a victim, (e.g., the trajectory of the victim before and after the aggression) the human operators select the cameras that could contain relevant information and analyse the corresponding video contents. This task becomes cumbersome because of the huge volume of video contents and the cameras' mobility. This paper presents an approach, which assists the operator in his task and reduces the research space. We propose to model the cameras' network (fixed and mobile cameras) on top of the city's transportation network. We consider the video surveillance system as a multilayer geographic information system, where the cameras are situated into a distinct layer, which is added on top of the other layers (e.g., roads, transport) and is related to them by the location. The model is implemented in a spatio-temporal database. Our final goal is that based on a spatio-temporal query to automatically extract the list of cameras (fixed and mobile) concerned by the query. We propose to include this automatically computed relative position of the cameras as an extension of the standard ISO 22311.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>The number of video surveillance cameras increases in
public and private areas (e.g., in train and metro stations,
onboard of buses and trains, inside commercial areas, inside
enterprises buildings). For example, some estimations show that
there are more than 400000 cameras in London and that only the
RATP also known as Régie Autonome des Transports Parisiens
(English: Autonomous Operator of Parisian Transports)
surveillance system comprises around 9000 cameras in Paris. In
these conditions, any person that lives and walks in those two big
European capitals is likely to be captured many times during a day
(up to 300 times in London) by several video surveillance systems
(e.g., the traffic surveillance cameras, the cameras in the subway,
and the cameras of a commercial centre). The only markers</p>
      <p>This standard does not consider the video surveillance
cameras’ mobility or their fields’ of view modification. In fact, at
the beginnings of video surveillance systems the cameras were
placed in fixed locations in order to monitor indoor and outdoor
places. With the improvements in the hardware and software
technologies, on-board cameras are more and more employed in
mobile vehicles (e.g., buses, police cars). This cameras’ mobility
makes the task of security agents even more difficult in the
context of an inquiry, when they have to analyse a huge amount of
video contents and to have supplementary knowledge on the
system’s characteristics (e.g., the bus timetables, the city transport
plan) in order to select the most appropriate video contents.</p>
      <p>In this context, our goal is to provide users with tools that
could assist them in their research and reduce the research space.
In order to achieve this objective, in this article, we propose an
extension of the ISO 22311 standard in order to take into account
the cameras’ mobility. We consider the video surveillance system
as a multilayer geographic information system, where the cameras
are situated on a distinct layer, which is added on top of the other
layers (e.g., roads, transport) through the location. We
implemented our solution using a spatial database in order to
select the cameras that might have acquired video contents
corresponding to a user’s spatio-temporal query.</p>
      <p>The remainder of this paper is organized as follows. After a
review of related work concerning the three aspects addressed in
this paper, video surveillance systems, standard ISO 22311 and
mobile objects modelling in the Section 2, Section 3 presents our
multilayer modelling approach. This model is implemented using
a spatio-temporal database. Some queries that can be answered
based on this database are presented in Section 4. Finally, Section
5 concludes and discusses possible future research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. STATE OF THE ART</title>
    </sec>
    <sec id="sec-3">
      <title>2.1 Video Surveillance Systems</title>
      <p>The generic schema of a video surveillance system is
illustrated in Figure 1. The content is captured and stored in a
distributed manner and analysed in a control centre by human
operators that watch a certain number of screens displayed in a
matrix (the Video Wall in Figure 1).</p>
      <p>
        There is a big diversity of cameras and sensors that constitute
the acquisition part of surveillance systems and a heterogeneity of
their installation contexts (e.g., on the halls or platforms of
railway or metro stations, on-board of trains and buses, on the
streets, in commercial centres or office buildings). Therefore, we
have fixed and mobile cameras having different technical
characteristics (most of the time dynamic) (see Figure 2 for an
example of such cameras) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]:




      </p>
      <sec id="sec-3-1">
        <title>Camera type: optical, thermal, infrared</title>
        <p>Sensor type and dimension: CMOS, CCD
Transmission type: analogous/ IP
Angle of view (horizontal and vertical), focal
distance, pan-tilt-zoom, field of view orientation,
visible distance etc.</p>
        <p>We started by analysing the way a query is processed in a
video surveillance system today. When a person (victim of an
aggression for example) files a complaint, he is asked to fill a
form describing the elements that could help the investigators to
find the relevant video segment (the Figure 3 illustrates an
example of such form). Based on the spatial and the temporal
aspects of the query, the surveillance operator uses his own
knowledge concerning the spatial disposal of the cameras’
network in order to select the most relevant video contents. Then
he analyses these contents by playing them on the different
screens that he has in front of him. The monitors themselves show
no spatial relationship of any kind, only the numbering of the
cameras is in a somewhat logical order.</p>
        <p>Therefore, the operators’ tasks become cumbersome taking
into consideration the huge volume of video contents to be
analysed, the mobility and the different characteristics of cameras.
Moreover, in the current systems, most of the stored contents is
not exploitable because of the recording’s low quality. This lack
of quality is often caused by inappropriate installation of cameras,
bad shooting, bad illumination conditions etc. The operator has no
a priori knowledge on the quality of the video contents and thus
he loses time by visualizing the low quality contents also.</p>
        <p>
          The video surveillance domain has seen a big number of
commercial systems developed [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. In the research area, many
projects were developed as well: CROMATICA [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ],
CARETAKER2 [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], VANAHEIM3 for the indoor static video
surveillance, and SURTRAIN [
          <xref ref-type="bibr" rid="ref21">20</xref>
          ], BOSS4 [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ],
PROTECTRAIL5 projects for the on-board mobile surveillance.
All these heterogeneous projects concentrate on the development
of the system’s physical architecture and of better detection
algorithms in order to obtain a fully automatic system [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ], [
          <xref ref-type="bibr" rid="ref25">24</xref>
          ].
        </p>
        <p>
          We can summarize by saying that there is a growing concern
in the research and industrial environments for developing
algorithms for video content analysis (VCA) in order to
automatically index content and detect objects (e.g., abandoned
packets or luggage) and events (e.g., intrusions, people or vehicles
going the wrong way) [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] or to draw operators’ attention to
events of interest (e.g., alarms). However, solutions for assistance
to a posteriori investigation are at a lesser stage of maturity, and to
date most of the data remain unexploited.
        </p>
        <p>In this article, we are going to address also the lack of
interoperability between different surveillance systems. In the
context of an inquiry, the police might need to analyse data from
different sources (systems), so it is important that the different
outputs of the systems to be interoperable. As a consequence, the
big actors of the domain started to unify efforts in order to
standardize the structure of folders and of metadata files generated
by video surveillance systems. A result of these efforts is
represented by the ISO 22311 standard that proposes a structure
2 http://cordis.europa.eu/ist/kct/caretaker_synopsis.htm</p>
        <sec id="sec-3-1-1">
          <title>3 http://www.vanaheim-project.eu/</title>
        </sec>
        <sec id="sec-3-1-2">
          <title>4 http://celtic-boss.mik.bme.hu/</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>5 http://www.protectrail.eu/</title>
          <p>for the data issued from video surveillance systems and the
metadata needed to exploit that data.</p>
          <p>In the following, we are going to present the ISO 22311
standard, especially the part concerning the description of the
cameras characteristics and mobility. We are going to highlight
the interesting elements which relate to our research.
2.2 Standard ISO 22311</p>
          <p>The Standard ISO 22311 defines an interoperability format
for the data generated by video surveillance systems and for the
metadata needed to exploit these huge volumes of data.</p>
          <p>The audio visual packages (containing audio, video or
metadata files) have to be structured hierarchically (in files,
folders and groups of folders) according to time intervals in
Coordinated Universal Time (UTC). For each group of folders it
is mandatory for the system to provide a XML description of the
source(s) (e.g., cameras, GPS, video analysis tools), codec(s), file
formats and a temporal index enabling an easy access to the
content.</p>
          <p>The current technologies and processing power enable the
analysis of video content and the extraction of metadata
describing objects, events, scenes etc. This analysis depends on
the acquisition context (e.g., the position of the camera, the image
quality, the type of sensors). Therefore, the standard distinguishes
between the systems, those that can generate such metadata (i.e.,
level 2 systems) and provides a general structure and dictionary
for describing sensors and events (i.e., metadata).</p>
          <p>As in this paper we are going to address the problem of
cameras’ geo-localization we present the schema for the sensors
description in Figure 4.</p>
          <p>Each camera has an absolute location (GPS coordinates) as
more and more of the installed cameras have an embedded GPS
transmitter. But, there are many cases when the GPS is not
enough because: (1) we need to model the position of the camera
with regards to the video surveillance system and not to the world;
(2) in some situations, for example in indoor environments, the
GPS positions do not provide a good precision.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>In the context of a video surveillance system:</title>
        <p> The mobile cameras are embedded in buses, train, police
cars;
 The movement of these vehicles is constrained by a road
network and a transportation network.</p>
        <p>By analysing the standard, we can notice that it defines a
relative position for a camera that is today a simple link to an
image (the plan of the network of cameras or of a building). This
kind of location is not easily exploitable. Furthermore, the
standard does not consider the video surveillance cameras’
mobility. In order to overcome these issues, we propose to extend
this standard through a multilayer modelling approach, where the
network of cameras is put on top of a transportation network.</p>
        <p>In the following, we present a state of the art of the mobile
objects modelling as the cameras’ mobility management
represents the main focus of this paper.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>2.3 Mobile Objects Modelling</title>
      <p>
        With the technology’s evolution, the mobility became very
important in the context of video surveillance systems. Not only
the objects (e.g., persons, cars) are moving in the monitored
scene, but also the surveillance cameras are moving. The great
majority of the research papers concerning the mobile objects in
the video surveillance domain concentrate on the video content
analysis in order to detect and track the objects, to interpret their
behaviour and to understand the visual events of the monitored
scene [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Thus, the mobility of the cameras is not exploited.
      </p>
      <p>
        In the field of moving objects, a mobile object means the
continuous evolution of any object over the time, in terms of
position and dimension [
        <xref ref-type="bibr" rid="ref22">21</xref>
        ]. This movement of the mobile
objects can be effectuated in an unconstrained environment [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]
(e.g., for hurricanes, fires) or in a constrained environment [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]
(e.g., cars move on road and transportation networks).
      </p>
      <p>
        In the video surveillance domain, the objects are moving in a
constrained environment, mainly by the road network. This
environment is represented as a graph-based model [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], [
        <xref ref-type="bibr" rid="ref26">25</xref>
        ],
where the vertices are junctions and the edges are the roads
between the two junctions. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] considers also the connectivity at
each junction in order to represent the road network. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] extends
the model proposed by [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] in order to consider the predefined
trajectories that some objects could have (e.g., buses). [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
proposes a mobile object data model where they consider the road
and rail networks. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] takes into account the transport network in
a city as a graph and they add to each graph vertex the transport
modes available (i.e., pedestrian, auto, urban rail, metro, bus).
      </p>
      <p>
        In the management of mobile objects, a major issue is the
storage of the objects’ spatio-temporal positions. Several
strategies can be considered: using the spatio-temporal data types
defined by [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] (e.g., moving points, moving lines, moving
regions), or using the dynamic attributes [
        <xref ref-type="bibr" rid="ref24">23</xref>
        ] (e.g., motion vector)
which enables to limit the size of the data that has to be stored and
queried.
      </p>
      <p>As far as we know, the video content’s mobility is not taken
into account in the video surveillance domain. In this article, we
want to exploit the advances in the field of mobile objects and
apply them in the video surveillance domain in order to consider
the mobile aspect of surveillance cameras.</p>
    </sec>
    <sec id="sec-5">
      <title>3. Extension of the Standard 22311 for the management of cameras mobility</title>
      <p>As you could see in Section 2.2, the Standard 22311,
defines a fix position of video surveillance camera, through the
GPS coordinates and a link to an image containing the plan of the
network. In order to overcome this issue, we propose to compute a
relative position with regards to a map which will enable us to:
 Model the distances between the cameras and select the
relevant cameras for a certain trajectory;
 Model the connections between the cameras ( e.g., possible
path between camera1 and camera2 but not between
camera2 and camera3 );
 Model trajectories for mobile cameras;
 Model the fields of view and the maximum detection
distances of fixed and mobile cameras.</p>
      <p>
        In order to achieve this goal we took our inspiration from the
domain of GIS (Geographical Information Systems) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and
mobile objects modelling.
      </p>
      <p>By considering the video surveillance system as a GIS we
benefit from the separation between the conceptual layers. Thus at
any time, a new layer can be added without modifying the existing
layers.</p>
      <p>In our approach, we propose a four layer model: (1) Road
network, (2) Transportation network, (3) Objects and (4) Cameras
network. The Figure 5 illustrates the UML model for the first
three layers.</p>
      <p>The “Road network” layer, presented in blue in Figure 5, is
based on the graph modelling approach well-known in the
literature. The road network is considered as an undirected graph
G= (V, E), with V a set of vertices and E a set of edges defined
according to the granularity level that we want to consider (for a
big boulevard of a European capital for example we can consider
each segment of the road, each segment between two intersections
or the entire boulevard). Each vertex has an identifier and a 2D
position. Each edge is determined by two vertices.</p>
      <p>
        The “Transportation network”, presented in yellow in
Figure 5, is also based on a graph model. At this level, the vertices
of the transportation network are intersections between roads, and
bus stations. Each transportation vertex has a position with
regards to a road segment. Ordered sequences of transportation
vertices constitute sections, which form lines (e.g., bus lines). The
advantage of our approach with regards to the ones proposed in
the state of the art [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] is that we have two independent graphs that
are connected to each other through the positions of transportation
vertices. That way if the buses stations are modified or new buses
lines are introduced we do not have to recompute the underlying
road graph.
      </p>
      <p>The “Objects” layer, presented in red in Figure 5, models
the positions of fixed and mobile objects with regards to the
underlying layers.</p>
      <p>
        The Fixed Object has a position on a road segment. Its
position is defined as a distance from each end of the segment.
For this kind of objects, we adopt the same localisation as the one
proposed by [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>
        In the case of Mobile Objects (e.g., buses, police cars,
persons), the position changes in time. Each object will
periodically transmit its position using different strategies (e.g.,
each Δt seconds, each time the object is changing the segment,
when the object's position predicted by the motion vector deviates
from the real position by more than a threshold [
        <xref ref-type="bibr" rid="ref24">23</xref>
        ]) that are out
of the scope of this article. We suppose that we periodically
receive updates containing time-stamped GPS points that we
transform into a relative position with regards to the road network
(i.e., the segments). We use this information to reconstitute
object’s trajectory.
      </p>
      <p>We distinguish two types of mobile objects: objects that
move freely within the road and transportation networks (e.g., car,
person) and objects of which trajectories are constrained by a
“line” (e.g., buses).</p>
      <p>On top of all these layers, we model a video surveillance
cameras’ network. A simplified schema of this model is illustrated
in Figure 6.</p>
      <p>The cameras’ network is composed of fixed and mobile
cameras. The fixed cameras have a 2D position that is given at
installation time. The mobile cameras are associated with mobile
objects (e.g., buses) and their trajectory is the same as the object’s
one.</p>
      <p>The new generation of digital surveillance cameras has
embedded GPS transmitters and even compasses. The
technologies developed around these cameras make it possible to
automatically extract information from the camera related to its
orientation, pan, tilt, zoom, focal distance, compression
parameters etc.</p>
      <p>
        Based on all these elements it is possible to model the field
of view for each camera and track its modifications in time. The
field of view is computed based on four parameters [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]: the 2D
position, the viewable angle, the orientation and the visible
distance. A schema of a 2D field of view proposed by [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] is
shown in Figure 7.
      </p>
      <p>In order to select the most appropriate attributes to describe
a video surveillance camera, we studied the sensor description
proposed by the ISO 22311 standard, SensorGML6, KML7. We
separated the identified camera’s properties in two categories:
properties that could be modified over the time, and fixed
characteristics.</p>
      <p>Thus, the extension of the standard ISO 22311 is realised at
three levels:
 Taking into account the road and transportation networks as
a graph and not as an image;
 Taking into account the camera’s relative position and its
mobility on the networks;
 Taking into account the camera’s characteristics change
over the time.</p>
      <p>Our model is implemented in a spatio-temporal database
that can be queried by users in order to retrieve the relevant
cameras for a given trajectory. The originality of our research
work is given by:
 the fact that it combines different spatio-temporal
information (e.g., road network, transportation network,
objects’ positions) and computation (e.g., trajectories, field
of view) within the same database;
 the twofold mobility, of the target objects and of the
cameras.</p>
      <p>In the next section we present the general architecture of the
tool that could assist the video surveillance operators in their
research based on our spatio-temporal database and some
examples of queries.</p>
    </sec>
    <sec id="sec-6">
      <title>4. Spatiotemporal database and queries</title>
      <p>Based on the presented model, our goal is to automatically
select the cameras (fixed and mobile) that could contain relevant
video content with regards to the user query (their field of view
intersected the query trajectory).</p>
      <sec id="sec-6-1">
        <title>6 http://www.opengeospatial.org/standards/sensorml</title>
      </sec>
      <sec id="sec-6-2">
        <title>7 http://www.opengeospatial.org/standards/kml</title>
        <p>More precisely, the idea is to compare a spatio-temporal
query of the user (e.g., Rivoli Street from Louvre to Metro
Chatelet the 14th of July between 10h and 14h) with the
trajectories stored in our database and, for a better precision, with
the cameras fields of view. The Figure 8 illustrates the generic
architecture of a system based on our spatio-temporal database for
assisting the video surveillance in their research.</p>
        <p>From the Figure 8 it is easy to observe that there are two
main questions when developing such system: How to query the
system? and How to update the system?. As explained in the
previous section our work addresses only the querying aspect that
we are going to describe in the following.</p>
        <p>First, a Query Interpreter module will transform the user
query (e.g, Rivoli Street from Louvre to Metro Chatelet the 14th of
July between 10h and 14h) in a spatio-temporal query. By
spatiotemporal query we understand a sequence of road segments and a
time interval that will be further transformed in a SQL query, by
the SQL Query Generator module. The SQL query is executed on
the database having as a result a list of cameras. Based on some
image quality parameters a score per camera can be computed and
the initial list can then be ranked according to this relevance
score.</p>
        <p>In the following we present two examples of
spatiotemporal queries executed on our database implemented in Oracle
Spatial 8:
</p>
        <p>The first selects the fixed cameras of which geometry
(field of view) intersects the geometry of the Rivoli street;
SELECT IdCamera
FROM FixedCamera
WHERE SDO_RELATE(
camera_geom,
(SELECT street_geom</p>
        <p>FROM Road</p>
        <p>WHERE Name ='Rivoli‘ ),
'mask=OVERLAPBDYDISJOINT querytype=WINDOW'
)='TRUE';
8http://www.oracle.com/fr/products/database/options/spatial/index
.html</p>
        <p>The second selects the mobile cameras that are associated
with the buses that crossed the street within the given time
interval.</p>
        <p>Another perspective of our work is the improvement of the
resulted cameras list by re-ranking it based on cameras’
characteristics (e.g., image quality, visible distance).</p>
        <p>LET TimePeriod = Timestamp(hour(2013,1,14,10),
hour(2013,1,14, 12));
SELECT ObjetID</p>
        <p>FROM ConstrainedObject
WHERE Type.MobileObject= “Bus” AND
TimePeriod.ConstrainedObject (atperiods (Timestamp,
TimePeriod));
SELECT DISTINCT IdMobileCamera</p>
        <p>FROM ConstrainedObject, FreeObject, MobileCamera
WHERE Intersect (MobileCamera.geom,
ConstrainedObject.geom) AND Intersect
(MobileCamera.geom, FreeObject.geom);</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. CONCLUSION</title>
      <p>In this paper, we presented a spatio-temporal modelling
approach of fixed and mobile cameras within a common
transportation network. Taking our inspiration from the multilayer
representation of the geographical information systems, we model
spatial information about the road and transportation
infrastructures and mobile objects’ trajectories in four
independent layers: (1) Road network, (2) Transportation
network, (3) Objects and (4) Cameras network.</p>
      <p>Based on this modelling approach we also proposed a
generic architecture for a system that could assist the video
surveillance operators in their research. Starting from a sequence
of trajectory segments and a temporal interval, such system
generates the list of cameras that could contain relevant
information concerning the query (that “saw” the query’s
trajectory).</p>
      <p>The need of such assisting tools was identified within the
French National Project METHODEO. Among the project’s
partners, we mention the French National Police, Thales and the
RATP also known as Régie Autonome des Transports Parisiens
(English: Autonomous Operator of Parisian Transports). Our
approach has been validated and will be evaluated within the
project.</p>
      <p>Obviously, many questions are still left with no answer
giving way to a large number of perspectives. We will present
several of them in the following.</p>
      <p>For now, our model considers only outdoor transportation
and surveillance networks. We plan to extend our model to indoor
spaces also in order to model cameras inside train or subway
stations for example.</p>
      <p>Our work is situated in the context of the a posteriori
research in the case of a police inquiry. We would like to extend
this context in the future in order to be able to process real time
queries or to predict trajectories based on some statistics realized
based on the stored data (e.g., average speed on some road
segments).</p>
    </sec>
    <sec id="sec-8">
      <title>6. ACKNOWLEDGMENTS</title>
      <p>This work has been supported by the ANR CSOSG-National
Security (French National Research Agency) project
METHODEO.
7
58
8
59</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Ay</surname>
            <given-names>S. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zimmermann</surname>
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kim</surname>
            <given-names>S.O.. Relevance</given-names>
          </string-name>
          <article-title>Ranking in Georeferenced Video Search</article-title>
          .
          <source>In Multimedia Systems Journal</source>
          ,
          <volume>16</volume>
          , 2 (March
          <year>2010</year>
          ), Springer,
          <fpage>105</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Booth</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sistla</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolfson</surname>
            <given-names>O.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Cruz</surname>
            <given-names>I. F.</given-names>
          </string-name>
          <article-title>A data model for trip planning in multimodal transportation systems</article-title>
          .
          <source>In Proceedings of the 12th International Conference on Extending Database Technology: Advances in Database Technology, ACM</source>
          ,
          <year>2009</year>
          ,
          <fpage>994</fpage>
          -
          <lpage>1005</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>CARETAKER</given-names>
            <surname>Consortium</surname>
          </string-name>
          .
          <article-title>Caretaker puts knowledge to good use</article-title>
          .
          <source>In European Public Transport Magazine</source>
          .
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Chang K.-T.</surname>
          </string-name>
          Introduction to Geographic Information Systems,
          <string-name>
            <surname>McGraw-Hill Higher</surname>
            <given-names>Education</given-names>
          </string-name>
          ,
          <year>2006</year>
          , 450 pages
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Deparis</surname>
            <given-names>J.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Velastin</surname>
            <given-names>S.A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Davies A.C.</surname>
          </string-name>
          <article-title>Cromatica project</article-title>
          .
          <source>In Advanced Video-Based Surveillance Systems</source>
          , The Springer International Series in Engineering and Computer Science. volume
          <volume>488</volume>
          ,
          <year>1999</year>
          ,
          <fpage>203</fpage>
          -
          <lpage>212</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Dingin Z.</surname>
          </string-name>
          , and
          <article-title>Deng K. Collecting and managing networkmatched trajectories of moving objects in databases</article-title>
          .
          <source>In Proceedings of the 22nd international conference on Database and expert systems applications, LNCS</source>
          Volume
          <volume>6860</volume>
          ,
          <year>2011</year>
          , pp
          <fpage>270</fpage>
          -
          <issue>279</issue>
          <year>2011</year>
          ,
          <fpage>270</fpage>
          -
          <lpage>279</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>El</surname>
            <given-names>bouziri A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boulmakoul</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Laurini R. Mobile</surname>
          </string-name>
          <article-title>Object and Real Time Information System Modeling for Urban Environment</article-title>
          ,
          <source>In Proceedings of the 26th Urban and Regional Data Management Symposium</source>
          ,
          <year>2007</year>
          ,
          <fpage>403</fpage>
          -
          <lpage>413</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>GDANSK</surname>
          </string-name>
          ,
          <source>KU. Deliverable 2</source>
          .1
          <article-title>- Review of existing smart video surveillance systems capable of being integrated with ADDPRIV</article-title>
          .
          <source>ADDPRIV consortium</source>
          ,
          <year>2011</year>
          , www.addpriv.eu
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Güting</surname>
            <given-names>R.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Almeida</surname>
          </string-name>
          , V.T., and Ding Z.
          <article-title>Modeling and Querying Moving Objects in Networks</article-title>
          .
          <source>In VLDB Journal</source>
          ,
          <volume>15</volume>
          ,
          <issue>2</issue>
          (
          <year>2006</year>
          ),
          <fpage>165</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Joshi</surname>
            <given-names>K. A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Thakore</surname>
            <given-names>D. G.</given-names>
          </string-name>
          <article-title>A Survey on Moving Object Detection and Tracking in Video Surveillance System</article-title>
          , in
          <source>International Journal of Soft Computing and Engineering</source>
          ,
          <volume>2</volume>
          , 3
          <issue>(</issue>
          <year>July 2012</year>
          ),
          <fpage>44</fpage>
          -
          <lpage>48</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Kim</surname>
            <given-names>I.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            <given-names>H. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yi</surname>
            <given-names>K. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Choi</surname>
            <given-names>J. Y.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Kong S. G. Intelligent Visual Surveillance - A Survey</surname>
          </string-name>
          . In
          <source>International Journal of Control Automation and Systems</source>
          ,
          <volume>8</volume>
          , 5 (april
          <year>2010</year>
          ),
          <fpage>926</fpage>
          -
          <lpage>939</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Lakshmi</surname>
            <given-names>Devasena C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Revathí</surname>
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Hemalatha</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Video Surveillance</surname>
          </string-name>
          Systems-A Survey, in
          <source>International Journal of Computer Science Issues</source>
          ,
          <volume>8</volume>
          , 4
          <issue>(</issue>
          <year>July 2011</year>
          ),
          <fpage>635</fpage>
          -
          <lpage>642</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Lamy-Bergot</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ambellouis</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khoudour</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanz</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Malouch</surname>
            <given-names>N.</given-names>
          </string-name>
          ,.
          <string-name>
            <surname>Hocquard</surname>
            <given-names>A</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bruyelle</surname>
            <given-names>J-L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petit</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cappa</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barro</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Villalta</surname>
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jeney</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <article-title>and Egedy K. Transport system architecture for on board wireless secured a/v surveillance and sensing</article-title>
          .
          <source>In Proceedings of the 9th international conference on Intelligent Transport Systems Telecommunications, IEEE</source>
          ,
          <year>2009</year>
          ,
          <fpage>564</fpage>
          -
          <lpage>568</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Le</given-names>
            <surname>Barz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            and
            <surname>Lamarque</surname>
          </string-name>
          ,
          <string-name>
            <surname>T.</surname>
          </string-name>
          <article-title>Video Surveillance Cameras</article-title>
          . In Jean-Yves Dufour, editor,
          <source>Intelligent Video Surveillance Systems, chapter 4</source>
          , Wiley, novembre
          <year>2012</year>
          .
          <volume>33</volume>
          -
          <fpage>46</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Liu</surname>
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            <given-names>J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ding Z. Effective</surname>
          </string-name>
          <article-title>Mapmatching on the Most Simplified Road Network</article-title>
          .
          <source>In Proceedings of the 20th International Conference on Advances in Geographic Information Systems (SIGSPATIAL '12)</source>
          . ACM,
          <year>2012</year>
          ,
          <fpage>609</fpage>
          -
          <lpage>612</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Marraud</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Cepas</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Reithler</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <article-title>Semantic browsing of video surveillance databases through Online Generic Indexing</article-title>
          ,
          <source>in Proceeding of the 3rd ACM/IEEE International Conference on Distributed Smart Cameras</source>
          ,
          <year>2009</year>
          , .1-
          <fpage>8</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>McKenney</surname>
            <given-names>M,</given-names>
          </string-name>
          and
          <string-name>
            <surname>Schneider</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <article-title>Spatial Partition Graphs: A Graph Theoretic Model of Maps</article-title>
          .
          <source>In Proceedingsof the 10th Int. Symp. on Spatial and Temporal Databases, LNCS 4605</source>
          , Springer,
          <year>2007</year>
          ,
          <fpage>167</fpage>
          -
          <lpage>184</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Parent</surname>
            <given-names>Ch.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spaccapietra</surname>
            <given-names>S.</given-names>
          </string-name>
          , and Zimányi E.
          <article-title>Conceptual Modeling for Traditional and Spatio-Temporal Applications: The MADS Approach</article-title>
          . Springer-Verlag New York,
          <year>2006</year>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Popa</surname>
            <given-names>I.S.</given-names>
          </string-name>
          , and
          <article-title>Zeitouni K. Modeling and Querying Mobile Location Sensor Data</article-title>
          ,
          <source>In Proceedings of the 4th Int.</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>Conference on Advanced Geographic Information Systems</source>
          , Applications, and Services,
          <string-name>
            <surname>IARIA</surname>
          </string-name>
          ,
          <year>2012</year>
          ,
          <fpage>222</fpage>
          -
          <lpage>231</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Priam</surname>
            <given-names>Q.-C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lapeyronnie</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baudry</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lucat</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sayd</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ambellouis</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sodoyer</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Flancquart</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barcelo</surname>
            <given-names>A</given-names>
          </string-name>
          .-C.,
          <string-name>
            <surname>Heer</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ganansia</surname>
            <given-names>F.</given-names>
          </string-name>
          , and Delcourt V.
          <article-title>Audio-video surveillance system for public transportation</article-title>
          .
          <source>In Proceeding of the 2nd international conference on Image Processing Theory Tools and Applications</source>
          , IEEE,
          <year>2010</year>
          ,
          <fpage>47</fpage>
          -
          <lpage>53</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Schneider</surname>
            <given-names>M. Moving</given-names>
          </string-name>
          <article-title>Objects in Databases and GIS: Stateof-the-Art and Open Problems</article-title>
          , in Research Trends in Geographic Information Science. Springer-Verlag,
          <year>2009</year>
          ,
          <fpage>169</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Sedes</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sulzer</surname>
            <given-names>J.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marraud</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Mulat</given-names>
            <surname>Ch</surname>
          </string-name>
          ., and
          <string-name>
            <surname>Cepas</surname>
            <given-names>B..</given-names>
          </string-name>
          <article-title>A Posteriori Analysis for Investigative Purposes</article-title>
          . In JeanYves Dufour, editor,
          <source>Intelligent Video Surveillance Systems, chapter 3</source>
          , Wiley, novembre
          <year>2012</year>
          ,
          <fpage>33</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Sistla</surname>
            <given-names>A.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wolfson</surname>
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chamberlain</surname>
            <given-names>S.</given-names>
          </string-name>
          , and Dao S.
          <article-title>Querying the Uncertain Posi-tion of Moving Objects</article-title>
          .
          <source>In: Temporal Databases: Research and Practice, LNCS 1399</source>
          , Springer,
          <year>1998</year>
          ,
          <fpage>310</fpage>
          -
          <lpage>337</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Soro</surname>
            <given-names>S.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Heinzelman W. A</surname>
          </string-name>
          <article-title>Survey of Visual Sensor Networks</article-title>
          ,
          <source>in Advances in Multimedia</source>
          , vol.
          <year>2009</year>
          ,
          <year>2009</year>
          ,
          <fpage>1</fpage>
          -
          <lpage>22</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Xu</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ding</surname>
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sun</surname>
            <given-names>X.</given-names>
          </string-name>
          , and Liu C.
          <article-title>Traffic aware route planning in dynamic road networks</article-title>
          .
          <source>In Proceedings of the 17th international conference on Database Systems for Advanced Applications, LNCS 7238</source>
          ,
          <year>2012</year>
          ,
          <fpage>576</fpage>
          -
          <lpage>591</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>