<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Automated Landslide Monitoring through a Low-Cost Stereo Vision System</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Mauro</forename><surname>Antonello</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Information Engineering (DEI)</orgName>
								<address>
									<addrLine>via Gradenigo 6/B</addrLine>
									<settlement>Padova</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Fabio</forename><surname>Gabrieli</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Dip. di Ingegneria Civile Edile ed Ambientale (ICEA)</orgName>
								<address>
									<addrLine>via Ognissanti 39</addrLine>
									<settlement>Padova</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Simonetta</forename><surname>Cola</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Dip. di Ingegneria Civile Edile ed Ambientale (ICEA)</orgName>
								<address>
									<addrLine>via Ognissanti 39</addrLine>
									<settlement>Padova</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Emanuele</forename><surname>Menegatti</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Information Engineering (DEI)</orgName>
								<address>
									<addrLine>via Gradenigo 6/B</addrLine>
									<settlement>Padova</settlement>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Automated Landslide Monitoring through a Low-Cost Stereo Vision System</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">7BECC36674208E5CEE65CF6D04D89901</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-19T16:12+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Photogrammetry</term>
					<term>Stereo Vision</term>
					<term>Multiview Calibration</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper we introduce an inexpensive yet efficient photogrammetry system that takes advantage of state of art computer vision techniques to monitor large natural environments. Specifically, our system provides a precise evaluation of the terrain flow in wide landslides through optical flow applied to 2D image sequences and a back-projection of the resulting motion gradients to a 3D model of the landslide. Providing such a wide 3D model is one of the key issues and is addressed relying to a wide baseline stereo vision system. To initialize the stereo vision system, we propose an effective multiview calibration process.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>In last few years, the significance of environmental monitoring for natural hazards prevention and mitigation has been constantly growing. Responsiveness requirements along with an increased amount of data coming from ambient sensors led to the necessity of automated systems able to detect critical situations and alert authorities.</p><p>In this work we address the problem of landslide monitoring, that means detecting the flow of landslide material, a motion that is very limited: usually only few meters over several weeks. The slipping of the landslide material is often monitored analyzing image sequences exploiting optical flow techniques. One of the main limitations of such process is that the direction and intensity of the flow are the projection in the camera plane of the real world flows. Therefore, in order to obtain a correct estimation of the material motion, flow gradients must be back-projected to the landslide 3D model. Unfortunately, this 3D model is hard to obtain due to the wideness of the monitored area.</p><p>Several works presented in the literature rely on expensive laser-scanner or aerial photogrammetry systems. Differently, our work propose an innovative stereo vision system that requires only two cameras, which ultimately makes it a low-cost alternative for large environmental 3D reconstruction. Stereo vision is one of the most widely used techniques for outdoor 3D reconstruction <ref type="bibr" target="#b0">[1]</ref>, especially in low cost solutions. Stereo vision only requires two cameras and can retrieve the distance of an area observed by both cameras from its slightly different appearance in the two images (see fig. <ref type="figure" target="#fig_0">1</ref>). In our case, the large distance of monitored areas and the presence of adverse natural conditions, like heavy wind or bad weather, make the calibration of a stereo vision system an hard task. In order to maintain a good reconstruction quality at farthest monitored areas, the distance between monitoring cameras (baseline, see sec. 2) needs to be a lot higher than what is commonly the case when calibrating a stereo camera pair. Indeed, the common way to calibrate a stereo vision system and find its extrinsic patameters (i.e. cameras mutual position) has been proposed by <ref type="bibr" target="#b1">[2]</ref> and makes use of a small pattern with known geometry; however, this pattern needs to be observed by both camera during the calibration process making this technique infeasible when cameras are too far from each other. In our work we deal this issue through a robust multiview calibration system which lesser constraints in terms of maximum baseline allow us to extend the distance between cameras without loss of precision in the calibration process. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">System Overview</head><p>The main issue connected to the reconstruction of wide areas is the sensitivity of the stereo matching with respect to the distance of the target area. In stereo vision, the error in the detected distance z is related to the quantization error in digital images <ref type="bibr" target="#b2">[3]</ref> and is computed by means of the following equation <ref type="bibr" target="#b3">[4]</ref>:</p><formula xml:id="formula_0">z = d z 2 bf ,<label>(1)</label></formula><p>where b [m] is the baseline between left and right cameras, f [px] is the focal length, z [m] is the target distance and d [px] is the quantization in the disparity map (usually one pixel). Equation <ref type="formula" target="#formula_0">1</ref>shows that there is a quadratic dependency of the depth error to the distance between camera and observed region. Since farthest areas of the landslide are located at more than 700 m from the cameras, a wide baseline is needed in order to keep low the depth error. In table 2 estimated depth errors with respect to target distance and baseline are reported for our camera installation (18M px and 30 mm focal length).  <ref type="table">1</ref>. Estimated depth errors with respect to the target distance and cameras baseline.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Extrinsic Multiview Calibration</head><p>To correctly process a couple of stereo images and obtain observed area distances, the roto-translation between the two stereo cameras (extrinsic parameters) needs to be known. The calibration process is usually performed observing a pattern with a known geometry <ref type="bibr" target="#b1">[2]</ref> but this simple method is not applicable in our case: the wide baseline and the terrain conformation do not allow the observation of the calibration pattern from both cameras.</p><p>We obtained a good estimation of the extrinsic parameters exploiting a multiview calibration <ref type="bibr" target="#b4">[5]</ref>. This technique makes use of a large set of images of the same area taken from several viewpoints. Visual correspondences between all possible couples of images are matched in order to impose constraints on the roto-translation between each viewpoint couple; this way it is possible to perform calibration exploiting the features available in the framed scene without the need of dedicated patterns. In our work we collected a large set of images of the landslide, taken from a number of different viewpoints; we then added such images to those acquired by the stereo camera pair. Exploiting the multi-view calibration it is possible to obtain the mutual position between all couples of views, including the extrinsic calibration of the stereo couple.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Landslide 3D Reconstruction</head><p>Once obtained the extrinsic calibration of the stereo system, images taken from the two cameras are rectified (see fig. <ref type="figure" target="#fig_2">2</ref>) and processed by a stereo matching algorithm called Semi Global Block Matching <ref type="bibr" target="#b5">[6]</ref> in order to produce a disparity map. From the disparity map we retrieve the distance of each observed point and create a dense 3D point cloud representing the landslide reconstruction (see fig. <ref type="figure" target="#fig_3">3</ref>).   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Results</head><p>The 3D reconstruction of the landslide allows us to precisely evaluate sliding of the ground. We detect particle flows in the image sequence using Normalized Cross-Correlation and then back-projecting the 2D flow onto the 3D landslide model, obtaining the motion flow of the rocks. The monitoring system proposed in this paper is completely autonomous and scalable and it is designed to issue an alert when the sliding effect exceeds a given threshold.</p><p>As a future work, we will employ a continuous camera calibration algorithm <ref type="bibr" target="#b6">[7]</ref> to prevent the system to lose its calibration. This way it will be possible to obtain good performance over time, since the capability of the multiview stereo calibration of providing good estimation of extrinsic parameters is strongly dependent on the mutual position of the cameras.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. The distance Z of a point P is retrieved from the disparity d = XR − XT . The distance the left and right optical centers (OT and OR) is called baseline B.</figDesc><graphic coords="2,236.81,292.91,141.73,129.82" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>c</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 2 .</head><label>2</label><figDesc>Fig. 2. Images processed by the stereo matching algorithm are first rectified. After rectification all correspondent points are located on the same row so the matching algorithm can search left-right correspondences in the same row.</figDesc><graphic coords="4,140.91,175.64,155.61,103.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Fig. 3 .</head><label>3</label><figDesc>Fig. 3. Left-right disparity map and 3D point cloud reconstruction of the landslide. For image clarity the point cloud has been down-sampled so that only one point every 10cm is kept.</figDesc><graphic coords="4,134.77,454.79,155.61,103.74" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">On benchmarking camera calibration and multi-view stereo for high resolution imagery</title>
		<author>
			<persName><forename type="first">C</forename><surname>Strecha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Von Hansen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Van Gool</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Thoennessen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Vision and Pattern Recognition</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
	<note>IEEE Conference on</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">A flexible new technique for camera calibration</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">11</biblScope>
			<biblScope unit="page" from="1330" to="1334" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
	<note>Pattern Analysis and Machine Intelligence</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Quantization error analysis in stereo vision</title>
		<author>
			<persName><forename type="first">C</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chatterjee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conference Record of The Twenty-Sixth Asilomar Conference on</title>
				<imprint>
			<date type="published" when="1992">1992. 1992. 1992</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="1037" to="1041" />
		</imprint>
	</monogr>
	<note>Signals, Systems and Computers</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Variable baseline/resolution stereo</title>
		<author>
			<persName><forename type="first">D</forename><surname>Gallup</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Frahm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mordohai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pollefeys</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Vision and Pattern Recognition</title>
				<imprint>
			<date type="published" when="2008">2008. 2008</date>
			<biblScope unit="page" from="1" to="8" />
		</imprint>
	</monogr>
	<note>IEEE Conference on</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Towards high-resolution large-scale multi-view stereo</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">H</forename><surname>Hiep</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Keriven</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Labatut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Pons</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Vision and Pattern Recognition</title>
				<imprint>
			<date type="published" when="2009">2009. 2009</date>
			<biblScope unit="page" from="1430" to="1437" />
		</imprint>
	</monogr>
	<note>IEEE Conference on</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Stereo processing by semiglobal matching and mutual information. Pattern Analysis and Machine Intelligence</title>
		<author>
			<persName><forename type="first">H</forename><surname>Hirschmuller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="328" to="341" />
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Continuous stereo self-calibration by camera parameter tracking</title>
		<author>
			<persName><forename type="first">T</forename><surname>Dang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hoffmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stiller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="1536" to="1550" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
	<note>Image Processing</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
