<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Automated Landslide Monitoring through a Low-Cost Stereo Vision System</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Mauro Antonello</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Fabio Gabrieli</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Simonetta Cola</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emanuele Menegatti</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Information Engineering (DEI)</institution>
          ,
          <addr-line>via Gradenigo 6/B, Padova</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Dip. di Ingegneria Civile Edile ed Ambientale (ICEA)</institution>
          ,
          <addr-line>via Ognissanti 39, Padova</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we introduce an inexpensive yet e cient photogrammetry system that takes advantage of state of art computer vision techniques to monitor large natural environments. Speci cally, our system provides a precise evaluation of the terrain ow in wide landslides through optical ow applied to 2D image sequences and a back-projection of the resulting motion gradients to a 3D model of the landslide. Providing such a wide 3D model is one of the key issues and is addressed relying to a wide baseline stereo vision system. To initialize the stereo vision system, we propose an e ective multiview calibration process.</p>
      </abstract>
      <kwd-group>
        <kwd>Photogrammetry</kwd>
        <kwd>Stereo Vision</kwd>
        <kwd>Multiview Calibration</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>In last few years, the signi cance of environmental monitoring for natural
hazards prevention and mitigation has been constantly growing. Responsiveness
requirements along with an increased amount of data coming from ambient
sensors led to the necessity of automated systems able to detect critical situations
and alert authorities.</p>
      <p>In this work we address the problem of landslide monitoring, that means
detecting the ow of landslide material, a motion that is very limited: usually
only few meters over several weeks. The slipping of the landslide material is often
monitored analyzing image sequences exploiting optical ow techniques. One of
the main limitations of such process is that the direction and intensity of the
ow are the projection in the camera plane of the real world ows. Therefore, in
order to obtain a correct estimation of the material motion, ow gradients must
be back-projected to the landslide 3D model. Unfortunately, this 3D model is
hard to obtain due to the wideness of the monitored area.</p>
      <p>
        Several works presented in the literature rely on expensive laser-scanner or
aerial photogrammetry systems. Di erently, our work propose an innovative
stereo vision system that requires only two cameras, which ultimately makes
it a low-cost alternative for large environmental 3D reconstruction. Stereo vision
is one of the most widely used techniques for outdoor 3D reconstruction [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
especially in low cost solutions. Stereo vision only requires two cameras and can
retrieve the distance of an area observed by both cameras from its slightly
different appearance in the two images (see g. 1). In our case, the large distance of
monitored areas and the presence of adverse natural conditions, like heavy wind
or bad weather, make the calibration of a stereo vision system an hard task.
In order to maintain a good reconstruction quality at farthest monitored areas,
the distance between monitoring cameras (baseline, see sec. 2) needs to be a lot
higher than what is commonly the case when calibrating a stereo camera pair.
Indeed, the common way to calibrate a stereo vision system and nd its extrinsic
patameters (i.e. cameras mutual position) has been proposed by [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and makes
use of a small pattern with known geometry; however, this pattern needs to be
observed by both camera during the calibration process making this technique
infeasible when cameras are too far from each other. In our work we deal this
issue through a robust multiview calibration system which lesser constraints in
terms of maximum baseline allow us to extend the distance between cameras
without loss of precision in the calibration process.
The main issue connected to the reconstruction of wide areas is the sensitivity
of the stereo matching with respect to the distance of the target area. In stereo
vision, the error in the detected distance z is related to the quantization error
in digital images [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and is computed by means of the following equation [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]:
z2
z = d bf ;
(1)
where b [m] is the baseline between left and right cameras, f [px] is the focal
length, z [m] is the target distance and d [px] is the quantization in the disparity
map (usually one pixel).
      </p>
      <p>Equation 1 shows that there is a quadratic dependency of the depth error
to the distance between camera and observed region. Since farthest areas of the
landslide are located at more than 700 m from the cameras, a wide baseline is
needed in order to keep low the depth error. In table 2 estimated depth
errors with respect to target distance and baseline are reported for our camera
installation (18M px and 30 mm focal length).</p>
      <p>
        distance
[m]
10
12
To correctly process a couple of stereo images and obtain observed area distances,
the roto-translation between the two stereo cameras (extrinsic parameters) needs
to be known. The calibration process is usually performed observing a pattern
with a known geometry [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] but this simple method is not applicable in our case:
the wide baseline and the terrain conformation do not allow the observation of
the calibration pattern from both cameras.
      </p>
      <p>
        We obtained a good estimation of the extrinsic parameters exploiting a
multiview calibration [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. This technique makes use of a large set of images of the
same area taken from several viewpoints. Visual correspondences between all
possible couples of images are matched in order to impose constraints on the
roto-translation between each viewpoint couple; this way it is possible to
perform calibration exploiting the features available in the framed scene without
the need of dedicated patterns. In our work we collected a large set of images of
the landslide, taken from a number of di erent viewpoints; we then added such
images to those acquired by the stereo camera pair. Exploiting the multi-view
calibration it is possible to obtain the mutual position between all couples of
views, including the extrinsic calibration of the stereo couple.
2.2
      </p>
      <p>
        Landslide 3D Reconstruction
Once obtained the extrinsic calibration of the stereo system, images taken from
the two cameras are recti ed (see g. 2) and processed by a stereo matching
algorithm called Semi Global Block Matching [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] in order to produce a disparity
map. From the disparity map we retrieve the distance of each observed point
and create a dense 3D point cloud representing the landslide reconstruction (see
g. 3).
      </p>
    </sec>
    <sec id="sec-2">
      <title>Results</title>
      <p>The 3D reconstruction of the landslide allows us to precisely evaluate sliding of
the ground. We detect particle ows in the image sequence using Normalized
Cross-Correlation and then back-projecting the 2D ow onto the 3D landslide
model, obtaining the motion ow of the rocks. The monitoring system proposed
in this paper is completely autonomous and scalable and it is designed to issue
an alert when the sliding e ect exceeds a given threshold.</p>
      <p>
        As a future work, we will employ a continuous camera calibration
algorithm [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] to prevent the system to lose its calibration. This way it will be
possible to obtain good performance over time, since the capability of the multiview
stereo calibration of providing good estimation of extrinsic parameters is strongly
dependent on the mutual position of the cameras.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Strecha</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Von Hansen</surname>
          </string-name>
          , W.,
          <string-name>
            <surname>Van Gool</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fua</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thoennessen</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>On benchmarking camera calibration and multi-view stereo for high resolution imagery</article-title>
          .
          <source>In: Computer Vision and Pattern Recognition</source>
          ,
          <year>2008</year>
          .
          <article-title>CVPR 2008</article-title>
          . IEEE Conference on. (
          <year>2008</year>
          ) 1{
          <fpage>8</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>A exible new technique for camera calibration</article-title>
          .
          <source>Pattern Analysis and Machine Intelligence</source>
          , IEEE Transactions on
          <volume>22</volume>
          (
          <issue>11</issue>
          ) (
          <year>2000</year>
          )
          <volume>1330</volume>
          {
          <fpage>1334</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <issue>3</issue>
          .
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chatterjee</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Quantization error analysis in stereo vision</article-title>
          .
          <source>In: Signals, Systems and Computers</source>
          ,
          <year>1992</year>
          . 1992 Conference Record of The Twenty-Sixth Asilomar Conference on. (
          <year>1992</year>
          )
          <volume>1037</volume>
          {
          <fpage>1041</fpage>
          vol.2
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Gallup</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frahm</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mordohai</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pollefeys</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Variable baseline/resolution stereo</article-title>
          .
          <source>In: Computer Vision and Pattern Recognition</source>
          ,
          <year>2008</year>
          .
          <article-title>CVPR 2008</article-title>
          . IEEE Conference on. (
          <year>2008</year>
          ) 1{
          <fpage>8</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Hiep</surname>
            ,
            <given-names>V.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keriven</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Labatut</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pons</surname>
            ,
            <given-names>J.P.</given-names>
          </string-name>
          :
          <article-title>Towards high-resolution large-scale multi-view stereo</article-title>
          .
          <source>In: Computer Vision and Pattern Recognition</source>
          ,
          <year>2009</year>
          .
          <article-title>CVPR 2009</article-title>
          . IEEE Conference on. (
          <year>2009</year>
          )
          <volume>1430</volume>
          {
          <fpage>1437</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Hirschmuller</surname>
          </string-name>
          , H.:
          <article-title>Stereo processing by semiglobal matching and mutual information</article-title>
          .
          <source>Pattern Analysis and Machine Intelligence</source>
          ,
          <source>IEEE Transactions on 30(2)</source>
          (
          <year>2008</year>
          )
          <volume>328</volume>
          {
          <fpage>341</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Dang</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , Ho mann, C.,
          <string-name>
            <surname>Stiller</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Continuous stereo self-calibration by camera parameter tracking</article-title>
          .
          <source>Image Processing, IEEE Transactions on 18(7)</source>
          (
          <year>2009</year>
          )
          <volume>1536</volume>
          {
          <fpage>1550</fpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>