<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SLAM method: reconstruction and modeling of environment with moving ob jects using an RGBD camera</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Korney Tertychny</string-name>
          <email>tertychny@volsu.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dmitry Krivolapov</string-name>
          <email>krivolapov@volsu.ru</email>
          <email>krivolapov@volsu.ru Alexander Khoperskov Volgograd State University Volgograd, Russia khoperskov@volsu.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sergey Karpov</string-name>
          <email>karpov.sergey@volsu.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Volgograd State University</institution>
          ,
          <addr-line>Volgograd</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <abstract>
        <p>We have analyzed high and low dynamic environment problems of Simultaneous Localization and Mapping (SLAM) methods, which we use for automatic 3D models construction. Constructed models are convenient for developing lifelike VR (virtual reality) systems of real places. Our software provides preliminary processing of data before transferring it to the main SLAM method. Developed software allows utilizing various computer vision algorithms or arti cial neural networks for recognition of objects which may change their location. Microsoft Kinect sensor allows easily recognize movable objects like human while testing.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Navigation systems development is one of the most important problem in robotic engineering. To make decisions
on further action a robot needs information about environment, surrounding objects, and its localization. There
are a lot of various navigation methods based on odometry [
        <xref ref-type="bibr" rid="ref10">13</xref>
        ], inertial navigation, magnetometer, active labels
(RFIDS, GPS) [
        <xref ref-type="bibr" rid="ref4">7</xref>
        ], label and map matching. Simultaneous Localization and Mapping (SLAM) approach is one
of the most promising way of navigation [
        <xref ref-type="bibr" rid="ref4">6, 7</xref>
        ]. Modern SLAM solutions provide mapping and localization in an
unknown environment [
        <xref ref-type="bibr" rid="ref1">2</xref>
        ]. Some of them can be used for updating the map which has been made before. SLAM
approach can be used in mobile autonomous systems like robots, cars, etc. SLAM approach is the only method
which provides its processing without any external information sources or the preparation of the environment
(for example, placing labels) [
        <xref ref-type="bibr" rid="ref7">10</xref>
        ]. SLAM is the general methodology for solving two problems [
        <xref ref-type="bibr" rid="ref15 ref3 ref6">4, 9, 18</xref>
        ]: 1)
environment mapping and 3D model construction; 2) localization using generated map and trajectory processing.
      </p>
      <p>
        E cacy of SLAM solution determines navigation and environment reconstruction accuracy [
        <xref ref-type="bibr" rid="ref10 ref10">13, 13</xref>
        ]. SLAM
is one of the general directions in computer vision and robotic engineering for the last than 25 years. It appears
that, all animals and human beings use this principle of navigation.
      </p>
      <p>
        However, there are some problems in modern SLAM solutions [
        <xref ref-type="bibr" rid="ref4">6, 7</xref>
        ]. One of them is dynamic environment
processing, where objects can move both in sight or out of sight. The problem becomes even more serious, since
a change of the position of objects may lead to a complete disorientation of the robotic system. Our research
is focused on the algorithm for removing moving objects to increase accuracy of 3D reconstruction of a real
environment using SLAM approach development.
      </p>
      <p>
        There are four types of robots [
        <xref ref-type="bibr" rid="ref10">13</xref>
        ]: indoor, outdoor, underwater and airborne robots. Our solution is applicable
with all four types, but the main relevance are indoor and outdoor robots. We have applied and tested it with
indoor robots due to technical features of our equipment.
      </p>
      <p>
        Figure 1 presents a common operation SLAM algorithm, which uses RGBD sensor. The output data of
a RGBD camera is an RGB image and a depthmap. In the depthmap image pixels contain data about the
distance between the sensor and the object. RGB image and the depthmap are transfered to the SLAM
frontend from a RGBD sensor. A feature matching algorithm processes the RGB image. Selected features are used
for a localization (See yellow and green points on \Feature matching" panel Figure 1). ORB, SURF, SIFT are
examples of such algorithms [
        <xref ref-type="bibr" rid="ref12">6, 15</xref>
        ]. Then the image with the selected features and the depthmap are resized,
cut and calibrated. Calibration, cutting and resizing are required because the RGB and the depthmap may
have di erent resolution, view angle and these cameras are located in some distance from each other [
        <xref ref-type="bibr" rid="ref2">3</xref>
        ]. Some
SLAM solutions combine RGB, depthmap and features information into a single structure. Then the data is
transfered to SLAM back-end, where the localization of the image, the camera and the features are provided.
Afterwards SLAM solution can make and locate point cloud on the 3D scene. Also it can make 2D maps, a
camera trajectory, and a feature map (features and their location).
      </p>
      <p>
        Usually SLAM solutions based on graph-based representation may be subdivided into two parts: the front-end
and the back-end. Using sensor data the front-end provides preliminary solution as a feature map. The back-end
uses the data from the front-end and a probabilistic approach to make solution with maximum probability. As
a rule, such approach is faster and more reliable than lter{based approaches due to better ability to cope with
SLAM nonlinearity [
        <xref ref-type="bibr" rid="ref10 ref3">4, 13</xref>
        ].
      </p>
      <p>
        In the classical SLAM scheme, when a SLAM algorithm starts working there is no knowledge on surrounding
environment, no preliminary map or labels [
        <xref ref-type="bibr" rid="ref3">4</xref>
        ]. Thus, the solution is only the result of sensor measurements.
      </p>
      <p>
        The other peculiarity is that the environment, where the robot works, must be static. It means that objects
seen and memorized by the robot must not change their positions and properties [
        <xref ref-type="bibr" rid="ref3">4</xref>
        ]. The lighting conditions
should be taken into the account depending on used equipment. All of them is hardly possible in a real
environment. We can highlight several important SLAM problems, which appear in the real world: dynamic or
changeable environment [
        <xref ref-type="bibr" rid="ref15 ref6">9, 18</xref>
        ] and accumulative mistakes eliminating (See discussion in works [
        <xref ref-type="bibr" rid="ref10 ref14 ref15 ref2 ref5 ref6 ref9">3, 13, 9, 18, 8, 12, 1, 17</xref>
        ]).
In the current work we focus on the rst problem.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>The dynamic environment problem</title>
      <p>
        Emergence of moving and movable objects in an environment which is surveyed by a robot is a signi cant problem.
We call high-dynamic environment, if objects move in sight, and low-dynamic environment, if objects move out
of sight [
        <xref ref-type="bibr" rid="ref15 ref6">9, 18</xref>
        ]. The presence of such objects may cause errors in the resulting 3D model, or localizing failure.
Any objects in a dynamic environment can be divided into objects moving in sight or out of sight. Some SLAM
solutions, such as DP-SLAM and RTAB-Map, may cope with small out of sight changes [
        <xref ref-type="bibr" rid="ref11">14, 5</xref>
        ]. DP-SLAM can
include an overlapping frames, and after some time passes it can replace wrong frames with overlapping ones.
RTAB-Map overwrites voxels, but the object don't disappear while all voxels are overwritten. Actually some
parts of the object will never eliminate. It looks like transparent object (See Figure 2). It is important that such
objects can contain some features, which cause localization failure or errors.
      </p>
      <p>We can see the point cloud made from the frames set. Half of them contain a human being who left the frame.
In that case some features are located at the defunct object. So there are several errors, but the method still
works because greater amount of features are on static objects. Also the model is extended with the parts which
have been invisible before the human being has gone. The place which was invisible looks like a shadow on the
oor.</p>
      <p>
        If there are too many features on a defunct or replaced object, the robot can't understand that it is the same
place. It may cause unexpected behavior [
        <xref ref-type="bibr" rid="ref10 ref6">13, 9</xref>
        ]. Any dramatic global changes may cause such problems too.
      </p>
      <p>In a high dynamic environment sensors provide erroneous data on environment with moving objects. If moving
objects don't carry features, modern SLAM solutions don't have any problems. Nevertheless such objects leave
traces in point cloud (See Figure 3). All point cloud images in this article are rotated relative to their RGB and
depth analogs for better visibility.</p>
      <p>There are a lot of traces and features (yellow dots) left by a moving object. We can see signi cant amount of
features, which are not placed on static objects. Figures 4 and 5 shows how features are located on the image
with a human being. A lighter space on the RGBD image shows where the distance was correctly measured.
Only points with correct distance are placed to the point cloud.</p>
      <p>It is easy to observe that on the RGB image there are approximately 30% of features located on the human
being. The transition to the RGBD image shows that the human being has already occupied more than 50%
of features. This is caused by the fact that all the features without correct distance (not located on the RGBD
image) are discarded.</p>
      <p>A human being is usually separated form any environment and the distance to him can be correctly measured.
Also a human being is a good source of features. When several human beings appear in the frame incorrect work
of SLAM solutions is practically guaranteed. On the RGB image di erent circle colors and sizes mean reliability
level and location accuracy of features. The smaller the radius, the more accurate they are determined.</p>
      <p>It should be mentioned that in the current work we consider human as an object which can change its position,
since it is easy to locate a human being using Kinect. Thus all the argumentation above is valid for di erent
movable and moving objects, such as animals, cars, papers, chairs etc. All of those objects are not a reliable
features source.</p>
      <p>
        The solutions for the dynamic environment problem have been already o ered. Authors [
        <xref ref-type="bibr" rid="ref6">9</xref>
        ] propose to make
static and dynamic maps for high-dynamic environment separately. The localization is provided by using a static
map. Not real objects but only moving elements of the frame are determined. If a pixel has changed its color, it
is considered as a moving object and removed from the static map. This method does not work in a low-dynamic
environment with objects which can change their position out of sight.
      </p>
      <p>
        In Ref.[
        <xref ref-type="bibr" rid="ref15">18</xref>
        ] a framework solving the problem with a low-dynamic environment has been proposed. In their
work an environment changes between sessions. Their algorithm nds outdated frames from last session and
replace them with fresher ones. The algorithm does not treat moving objects in frame leading to the instability
when moving object has too many features. Both of these solutions solve the problem in the back-end part.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Modi cated method</title>
      <p>
        We propose a solution which can work both with a high and low-dynamic environment in the front-end. It
processes data before transfering it to the main SLAM method. So a di erent back-end can be used in order
to transform input data. Our solution removes movable objects from data stream. Our algorithm can allocate
objects which don't move but can move unlike the solution presented in Ref.[
        <xref ref-type="bibr" rid="ref6">9</xref>
        ].
      </p>
      <p>We establish human beings as removing objects for our algorithm in this article. This type of an object has
been chosen because it is easy to locate people using Kinect sensor and Kinect SDK, which provide a body frame
stream. We use a body frame to remove human being image from data streams. We overwrite a distance as
incorrect (less than minimum) in pixels where a human location is speci ed. It is displayed with a red colored
human silhouette on Figure 6. On Figure 6 the red color was chosen to make images more visible.</p>
      <p>After processing our algorithm transfers RGB and depthmap frames to feature matching algorithms and then
to SLAM back-end. An examples of solutions which uses an RGB, depthmap and feature map are RTAB-Map
(Real-Time Appearance-Based Mapping), and DP-SLAM (Distributed Particle).</p>
      <p>After several tests we noticed that sometimes some pixels, which belong to removed objects, stay away of
overwritten part. We determined that there are 0 to 3 pixels around an overwritten space. Figure 7 shows the
sequence of updated SLAM method operations. Our algorithm processes data from any RGBD sensor overwriting
unwelcome objects on the depth frames, and transfers data to the main SLAM solution. It should be mentioned
that in the \Object rejection" stage we can remove any unwelcome objects, which we can allocate by computer
vision algorithms or arti cial neural networks.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Results and discussion</title>
      <p>In the current work the algorithm which provides preliminary data processing to remove objects which may
cause errors in simultaneous localization and mapping is described. The examples of such objects are people,
cars, animals and other movable or moving objects. We test developed software using a Kinect 2 sensor, and
RTAB-Map as a SLAM solution. We take out a human as an sample of moving objects from data provided
by sensors, since Microsoft Kinect SDK provides simple and convenient work with human contours. We use an
output point cloud for our VR (virtual reality) system. It provides virtual tours to reconstructed real places.</p>
      <p>Figure 9 shows the output point cloud of our hardware and software system. Our algorithm creates the point
cloud with no erroneous data caused by moving objects. If there is some place hidden by removing objects, there
will be empty space in the point cloud. After it becomes visible, system will add information.</p>
      <p>If removing objects occupy a signi cant part of the eld of vision so the SLAM solution can't localize, it
may show an error and continue localizing after the object moves out. There is no problem if there are enough
features outside a moving object. As it was before, yellow and green dots on Figure 9 indicate features. SLAM
solutions such as RTAB-Map use features for localization and place memorization.</p>
      <p>While testing we have found out that there may be some pixels which are not overwritten by our algorithm
around the removing object. So we have to modify our solution to overwrite at least 3 more pixels around
allocated objects.</p>
      <p>There is an empty space that looks like a shadow which is made by human. Standard SLAM solution doesn't
add points there (See Figure 8). Without moving of the camera, the algorithm continues to assume that the
cloud of points is complete and true and the empty place is out of view. The key feature of our algorithm is
an ability to create a list of unwelcome objects, that are not desirable to transmit to SLAM input. We can
use di erent computer vision algorithms or arti cial neural networks to recognize objects that can change their
position. Then such objects can be removed, as it has been shown in this article using a human being as an
example. Our algorithm has an ability to remove any recognizable object. In addition to the modular principle,
it is possible to work with the method in combination with various solutions that implement the SLAM approach.</p>
      <p>The tests of the developed algorithm give satisfactory results completely excluding undesirable objects from
the input data of the main SLAM algorithm. SLAM solution without using the proposed algorithm results shown
on Figure 8. It is seen that in the cloud of points there is a large number of features, which are not tied to real
objects. Such features are not a reliable source of information about an environment. Figure 9 shows the cloud of
points obtained with the proposed algorithm. It does not contain any erroneous points which always appear due
to the presence of a moving object. Tests show that it is enough to remove an object image from the depthmap,
because SLAM solutions use only the features with correctly measured distance for localization and mapping.
Furthermore, removing one object like a human doesn't cause performance decline using medium-power personal
computer.</p>
      <p>Acknowledgements
Work has been supported by the Ministry of Education and Science of the Russian Federation (government task
No. 2.852.2017/4.6).</p>
      <p>[1] Bakkay, M.C., Arafan, M., Zagrouba, E.: Dense 3D SLAM in dynamic scenes using kinect. Lecture
Notes in Computer Science (including subseries Lecture Notes in Arti cial Intelligence and Lecture
Notes in Bioinformatics) 9117, pp. 121-129 (2015)</p>
      <p>Scale
pp</p>
      <p>with FAB-MAP
1100-1123 (2010).
[5] DP Developers, https://github.com/jordant0/DP-SLAM. Last accessed 27 March 2018
[6] Durrant-Whyte, H., Bailey, T.: Simultaneous localization and mapping: Part I. IEEE Robotics and</p>
      <p>Automation Magazine 13(2), pp. 99-110 (2006) https://doi.org/10.1109/MRA.2006.1638022</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Cao</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhuang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
          </string-name>
          , H.,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Robust Place Recognition and Loop Closing in LaserBased SLAM for UGVs in Urban Environments</article-title>
          .
          <source>IEEE Sensors Journal</source>
          <volume>18</volume>
          (
          <issue>10</issue>
          ),
          <fpage>4242</fpage>
          -
          <lpage>4252</lpage>
          (
          <year>2018</year>
          ) https://doi.org/10.1109/JSEN.
          <year>2018</year>
          .2815956
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Civera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bueno</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davison</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montiel</surname>
          </string-name>
          , J.:
          <article-title>Camera self-calibration for sequential Bayesian structure from motion</article-title>
          .
          <source>IEEE International Conference on Robotics and Automation</source>
          <year>2009</year>
          , pp.
          <fpage>403</fpage>
          -
          <lpage>408</lpage>
          . IEEE (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Cummins</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Newman</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Appearance-only SLAM 2.0</article-title>
          . The International Journal of Robotics Research, https://doi.org/10.1177/0278364910385483 at Large Vol.
          <volume>30</volume>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Durrant-Whyte</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bailey</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Simultaneous localization and mapping: Part II IEEE Robotics</article-title>
          and
          <source>Automation Magazine</source>
          <volume>13</volume>
          (
          <issue>3</issue>
          ), pp.
          <fpage>108</fpage>
          -
          <lpage>117</lpage>
          (
          <year>2006</year>
          ) https://doi.org/10.1109/MRA.
          <year>2006</year>
          .1678144
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Endres</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sturm</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cremers</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Burgard</surname>
            , W.: 3-
            <given-names>D</given-names>
          </string-name>
          <string-name>
            <surname>Mapping</surname>
          </string-name>
          with
          <article-title>an RGBD camera</article-title>
          .
          <source>In IEEE Transactions on Robotics</source>
          <year>2014</year>
          , Vol
          <volume>30</volume>
          (
          <issue>1</issue>
          ), pp.
          <fpage>177</fpage>
          -
          <lpage>187</lpage>
          . IEEE (
          <year>2014</year>
          ) https://doi.org/10.1109/TRO.
          <year>2013</year>
          .2279412
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Hahnel</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Triebel</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Burgard</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thrun</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <article-title>: in dynamic environments</article-title>
          .
          <source>Robotics Automation</source>
          , https://doi.org/10.1109/ROBOT.
          <year>2003</year>
          .
          <article-title>1241816 Map building with mobile robots 2003</article-title>
          , pp.
          <fpage>1557</fpage>
          -
          <lpage>1563</lpage>
          . IEEE (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cho</surname>
            ,
            <given-names>Y.K.</given-names>
          </string-name>
          :
          <article-title>SLAM-driven robotic mapping and registration of 3D point clouds</article-title>
          .
          <source>Automation in Construction 89</source>
          ,
          <fpage>38</fpage>
          -
          <lpage>48</lpage>
          (
          <year>2018</year>
          ) https://doi.org/10.1016/j.autcon.
          <year>2018</year>
          .
          <volume>01</volume>
          .009
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Lachat</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Macher</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Landes</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grussenmeyer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Assessment and calibration of a RGB-D camera (Kinect v2 Sensor) towards a potential use for close-range 3D modeling</article-title>
          .
          <source>Remote Sensing</source>
          <volume>7</volume>
          (
          <issue>10</issue>
          ),
          <fpage>13070</fpage>
          -
          <lpage>13097</lpage>
          (
          <year>2015</year>
          ) https://doi.org/10.3390/rs71013070
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Majdi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bakkay</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zagrouba</surname>
          </string-name>
          , E.:
          <article-title>3D modeling of indoor environments using Kinect sensor</article-title>
          .
          <source>In IEEE 2nd International Conference on Image Information</source>
          Processing pp,
          <year>2013</year>
          ,pp.
          <fpage>67</fpage>
          -
          <lpage>72</lpage>
          . IEEE (
          <year>2013</year>
          ) https://doi.org/10.1109/ICIIP.
          <year>2013</year>
          .6707557
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Nister</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naroditsky</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bergen</surname>
            ,
            <given-names>J.: Visual</given-names>
          </string-name>
          <string-name>
            <surname>Odometry</surname>
          </string-name>
          .
          <source>IEEE Computer Society Conference on Computer Vision and Pattern Recognition</source>
          <year>2004</year>
          , vol.
          <volume>1</volume>
          , pp.
          <fpage>652</fpage>
          -
          <lpage>659</lpage>
          . IEEE (
          <year>2004</year>
          ) https://doi.org/10.1109/CVPR.
          <year>2004</year>
          .1315094
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [14]
          <article-title>RTAB map Developers</article-title>
          , http://introlab.github.io/rtabmap/.
          <source>Last accessed 27 March 2018</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Rublee</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rabaud</surname>
            ,
            <given-names>V</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konolige</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bradski</surname>
          </string-name>
          , G.:
          <article-title>ORB: an e cient alternative to SIFT or SURF</article-title>
          .
          <source>IEEE International Conference on Computer Vision</source>
          <year>2011</year>
          , pp.
          <fpage>2564</fpage>
          -
          <lpage>2571</lpage>
          , IEEE (
          <year>2011</year>
          ) https://doi.org/10.1109/ICCV.
          <year>2011</year>
          .6126544
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Shao</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>3D Indoor Environment Modeling and Detection of Moving Object and Scene Understanding</article-title>
          ,
          <source>Transactions on Edutainment XIV</source>
          ,
          <fpage>40</fpage>
          -
          <lpage>55</lpage>
          , IEEE International Conference on. IEEE, pp.
          <fpage>2564</fpage>
          -
          <lpage>2571</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Walcott-Bryant</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaess</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johannsson</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leonard</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          :
          <article-title>Dynamic pose graph SLAM: Longterm mapping in low dynamic environments</article-title>
          .
          <source>IEEE International Conference on Intelligent Robots and Systems</source>
          vol.
          <volume>6385561</volume>
          , pp.
          <fpage>1871</fpage>
          -
          <lpage>1878</lpage>
          . IEEE (
          <year>2012</year>
          ) https://doi.org/10.1109/IROS.
          <year>2012</year>
          .6385561
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xiong</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>A framework for multisession RGBD SLAM in low dynamic workspace environment</article-title>
          .
          <source>CAAI Transactions on Intelligence Technology</source>
          <volume>1</volume>
          ,
          <fpage>90</fpage>
          -
          <lpage>103</lpage>
          . CAAI (
          <year>2016</year>
          ) https://doi.org/10.1016/j.trit.
          <year>2016</year>
          .
          <volume>03</volume>
          .009
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Microsoft kinect sensor and its e ect</article-title>
          .
          <source>IEEE Multimedia</source>
          <volume>19</volume>
          (
          <issue>2</issue>
          ), pp.
          <fpage>4</fpage>
          -
          <lpage>10</lpage>
          (
          <year>2012</year>
          ) https://doi.org/10.1109/MMUL.
          <year>2012</year>
          .24
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>