<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Investigation of algorithms for generating surfaces of 3D models based on an unstructured point cloud</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>E.S.Glumova</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>A.D.Filinskikh glumova.ek@yandex.ru</string-name>
          <email>glumova.ek@yandex.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>alexfil@yandex.ru</string-name>
          <email>alexfil@yandex.ru</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Glumova Ekaterina S. - student of Nizhny Novgorod State Technical University n.a. R.E. Alekseev</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Nizhny Novgorod State Technical University n a. R.Е. Alekseev</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Methods of 3D object model creation on the basis of unstructured (sparse) cloud of points are considered in the paper. The issues of combining point cloud compaction methods and subsequent surface generation are described. The comparative analysis of generation surfaces algorithms for the purpose of revealing of more effective method using as input data the depth maps received from the sparse cloud of points is carried out. The comparison is made by qualitative, quantitative and temporal criteria. The optimal method of 3D object model creation on the basis of unstructured (sparse) cloud of points and depth map data is chosen. The mathematical description of the point cloud compaction method on the basis of stereo-matching with application of two-phase algorithm of species search and depth map extraction from Multi-View Stereo for Community Photo Collections source image set is provided. The implementation of the method in open-source software Regard3D is realized in practice.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Today, the development of computing and surface
restoration technologies allows to recreate 3D models of
objects with high accuracy and high quality. One of such
technologies is laser scanning. With the help of laser
scanners, it is possible to get the geometry of high
accuracy, but unfortunately the devices that allow to
achieve accuracy in hundredths of millimeters cost tens
and hundreds of millions of rubles. One of the types of
non-contact scanning of objects is photogrammetry. The
cost of equipment for obtaining geometric data about an
object is hundreds of times lower than the equipment that
uses laser technology, and the main load for obtaining
high-quality models falls on the software.</p>
      <p>
        3D objects models are widely used in the field of
parametric architecture [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the industry of computer video
games and animation [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], in the development of scenes for
VR applications, as well as in mobile development [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
The quality of the model plays an important role in any of
these areas. It is important to distribute computational
resources of software correctly.
      </p>
      <p>There are quite a lot of various software on the market
for processing images and obtaining 3D models by series
of images. There are both paid software, costing about one
hundred thousand rubles, and free open source software.
In both cases, different algorithms are used at all stages
from photo processing to obtaining a 3D model.</p>
      <p>
        One of the photogrammetry methods is the one of
building a 3D structure by a set of images - Structure from
Motion. The method feature is automatic determination of
camera internal parameters [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This method restores such
camera parameters as the extrinsic calibration (the
orientation and position of the camera) and the intrinsic
calibration (focal length, radial distortion of the lens).
      </p>
      <p>The first step of SfM realization is to detect and match
point features in the input images. Special points (term
vary in different sources) - to put it informally - "well
detectable" fragments of an image. These are points
(pixels) with a characteristic (special) neighborhood - i.e.
different from all neighboring points. Local features
examples can be corner tops, isolated point features,
contours, etc. The keypoints are described by descriptors
vectors of features computed on the basis of
intensity/gradients or other characteristics of the
neighborhood points.</p>
      <p>
        The most popular feature descriptors used in modern
image processing systems are given in [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. A-KAZE
(nonlinear diffusion filtering for detecting and describing
2D objects) is used to solve the problem of keypoint
detection.
      </p>
      <p>
        Then the camera position is assessed and a cloud of
low density points or sparse points is selected. Keypoints
in multiple images are matched using approximate nearest
neighbor and ‘tracks’, linking specific keypoints in a set of
pictures. Tracks comprising a minimum of two keypoints
and three images are used for point-cloud reconstruction,
with those which fail to meet these criteria being
automatically discarded [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. After that triangulation is
used to estimate points three-dimensional positions and
gradual reconstruction scene geometry fixed into a relative
coordinate system.
      </p>
      <p>
        An enhanced density point-cloud can be derived by
implementing the Multi-View Stereo (MVS) algorithm
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], based on depth maps, the Clustering Views for
MultiView Stereo (CMVS) [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], the Patch-based MVS algorithm
(PMVS2) [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], the Shadow-Aware Multi-View Stereo
Algorithm (SMVS) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], that combines stereo and
shapefrom-shading energies into a single optimization scheme.
The camera positions obtained from a sparse point cloud
are used here as input data. The result of this additional
processing is a significant increase in point density.
      </p>
      <p>The color and texture information is then transferred to
a point cloud, after which the final 3D model is rendered.</p>
      <p>Simplified process of obtaining 3D-model based on the
images is shown in Fig. 1.</p>
      <p>
        A stage of reception of surface generation on the basis
of the received unstructured cloud of points by a
3Dreconstruction method MVS (Multi-View Stereo) are
considered separately [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        MVS is based on reconstructing a depth map for each
view (image). Despite the large redundancy of the output
data, the method has proven to be well suited for restoring
the detailed geometry of sufficiently large scenes. Another
advantage of depth maps as an intermediate representation
is that the geometry is parameterized in its natural domain,
and per-view data (such as color) is directly available from
the images. The excessive redundancy in the depth maps
can cause problems; not so significant in terms of storage,
but in terms of computational power [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        SfM, which reconstructs the parameters of the
cameras;
MVS for establishing dense;
surface generation (meshing), which merges the MVS
geometry into a globally consistent, colored mesh.
(MVSCPC) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], that reconstructs depth maps for each
image. The depth map represents the two-dimensional
one-channel image containing the information about
distance from a sensor plane to scene objects [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>The method is based on the idea of selecting images
from the collection so that they match both per-view and
per-pixel level. Appropriate choice of views ensures
reliable matches even with strong differences in images.
The stereo matching algorithm takes as input sparse 3D
points reconstructed from
SfM
and iteratively grows
surfaces from these points. Optimizing for surface norms
with a photoconsistency measure significantly improves
the</p>
      <p>matching results. The depth map quality is also
assessed.</p>
      <p>Stereo</p>
      <p>
        matching is performed at each pixel by
optimizing for both depth and normal, starting from an
initial estimate provided by a sparse point cloud. During
stereo
optimization, poorly
matching
views can
be
discarded and new ones added according to the local view
selection criteria. The detour Pixels can be revised and
their depth updated if a more accurate match is found [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>MVSCPC provides depth map assessment for each
input image - each image serves as a reference view only
once, after which a two-level view selection algorithm is
implemented. At the image level, global view selection
determines for each reference view a set of good neighbor
images to use for stereo matching.</p>
      <p>Global view. For each reference view R, global view
selection seeks a set N of neighboring views that are good
candidates for stereo matching in terms of scene content,
appearance, and scale. In addition, the neighboring views
should</p>
      <p>provide sufficient parallax (a change in the
apparent position of an object relative to a distant
background, depending on the position of the observer)
with respect to R and each other in order to enable a stable
match. Here we describe a scoring function designed to
measure the quality of each candidate neighboring view
based on these desiderata.</p>
      <p>Since matches and sparse point cloud extracted in the
SfM phase are not sufficient indicators for accurate surface
reconstruction (as they
are extracted
based on the
similarity of only the scene content), another assessment
of image matches reliability was proposed.</p>
      <p>A global score   for each view 
within a candidate
over features shared with  is computed as:
neighborhood</p>
      <p>(which includes  ) as a weighted sum
  ( ) =</p>
      <p>( ) ∙   ( )
and the weight functions are described below.
where   is the set of feature points observed in view  ,
To encourage a good range of parallax
within a
product over all pairs of views in  :
neighborhood, the weight function   ( ) is defined as a
  ( ) =</p>
      <p>( ,   ,   )
  ,  ∈
 ≠  ,  ∈    ∩   
= min(   

where</p>
      <p>,    ,  
angle between the lines of sight from   and   to  .
2
, 1) and  is the
(1)
(2)
  ( ) =
2</p>
      <p>, 2 ≤ 
1,1 ≤  &lt; 2</p>
      <p>,  &lt; 1
angles below  
angle.</p>
      <p>The function    ,   ,</p>
      <p>downweights triangulation
, which is usuall set to 10 degrees. The
quadratic weight function serves to counteract the trend of
greater numbers of features in common with decreasing</p>
      <p>The weighting function   ( ) measures similarity in
resolution of images 
and 
at feature  . The diameter
  ( ) of a sphere centered at  whose projected diameter
in  equals the pixel spacing in  is computed to estimate
the 3D sampling rate of  in the vicinity of the feature  .</p>
      <p>Similarly,   ( ) is calculated for 
and the scale
weight   is defined based on the ratio  =   ( )
using</p>
      <p>This weight function prefers views with equal or
higher resolution than a reference view. Having defined a
global estimate of species  and neighbors  , one can find
the best</p>
      <p>
        of a given size (usually | | = 10) by the sum
of species estimates ∑ ∈   ( ). For efficiency, a "greedy
algorithm" [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] is used
      </p>
      <p>and grow the neighborhood
incrementally by iterative adding to  the highest scoring
view, taking into account the current 
(which initially
contains only  ).</p>
      <p>Rescaling</p>
      <p>Views. Although global view selection
algorithm
tries to
select
neighboring
views
with
compatible scale, some inconsistencies in scale are
unavoidable due to differences in resolution within the
collection of photos, which may negatively affect stereo
matching. There are methods to adapt the scale of all views
by filtering to a common, narrow range or global,
pixelbased view. The first method is used in this research to
avoid resizing of the matching window in different areas
of the depth map. This approach finds a view with the
approximately
lowest-resolution   
match that lower resolution, and then
∈</p>
      <p>relative to  , resamples  to
resamples higher resolution to match  .</p>
      <p>In particular, the assessment the resolution scale of a
view  relative  is based on their common features

 ( ) =</p>
      <p>1
|  ∩   |
Then   
simply equals</p>
      <p>( ) is less than the threshold value  (
which is close to the 5x5 of the reference window on a 3x3
window in the neighboring view with the lowest relative
scale), the reference view is rescaled so that, after rescaling



 ( ) =  .</p>
      <p>Then
all
neighboring
views
with
(which itself may have been changed in the previous step).</p>
      <p>( ) &gt; 2 to match the scale of the reference view
It is important that all modified versions of the images are
discarded when moving to the depth map computation for
the next reference view.</p>
      <p>Local View. Global view selection determines a set of
well suited candidates for a reference view and matches
their scale. Instead of using all of these views for stereo
matching at a specific location in the reference view, the
smallest set  ⊂</p>
      <p>of active views is selected (usually
computation of the depth map.
| | = 4). Using this subset naturally speeds up the
(4)
 ( ).
 = 1,
  ( )
(3)</p>
      <p>During stereo matching,  is iteratively updated using
a set of local view selection criteria designed to select
views that, given a current depth and normal pixel
estimates, are photometrically consistent and provide a
sufficiently wide range of observation directions. To
measure the photometric consistency, the mean-removed
normalized cross correlation (NCC) between pixels within
a window about the given pixel in  and the corresponding
window in V is used. If the NCC score is above a fixed
threshold, then  is a candidate for addition to  .</p>
      <p>
        You can measure the angular distribution by looking at
gaps of directions from which the given scene point (based
on the current depth estimation for the reference pixel) is
observed. In practice, the angular spread of the epipolar
line [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] is considered instead, obtained by projecting each
viewing ray passing through the reference point to the
reference view. When deciding whether to add view  to
the active set  , the local score is calculated as
  ( ) =   ( ) ⋅
  ( ,  ′)
      </p>
      <p>(5)
 ′∈
 
, 1) and  is the acute
where   ( ,  ′) = min (
angle between the pair of epipolar lines in the reference
view as described above. Accept</p>
      <p>Then the local view selection algorithm is performed
in the following way. Taking the initial depth of the pixel,
the view</p>
      <p>with the highest   ( ) value is found. If this
view has a sufficiently high NCC score (threshold 5 is
used), it is added to  ; otherwise, the view is rejected. The
process is repeated until either set  reaches the desired
size or the view</p>
      <p>remains undecided. During stereo
matching, the depth (and normal) are optimized, and a
view may be removed (and marked as rejected). Then a
replaced view is added. The algorithm completes as the
=10 degrees.
deflected views are never revised.</p>
    </sec>
    <sec id="sec-2">
      <title>4. Surface Generation</title>
      <p>After computing arrays containing the best matching
candidates for each image, you can move towards the step
of surface generation. Merging the individual depth maps
into a single polygonal surface is a labor intensive task.
The depth maps inherit information about the multi-scale
properties of the original images, which leads to vastly
different sampling rates of the research surfaces.</p>
      <p>
        Many approaches for depth maps fusion have been
proposed [
        <xref ref-type="bibr" rid="ref16 ref17 ref18 ref19 ref20">16-20</xref>
        ]. Among them FSSR (Floating Scale
Surface Reconstruction) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and SPSR (Screened Poisson
Surface Reconstruction) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] were considered as methods
of surface generation, as they provide high detail of the
reconstructed 3D model.
      </p>
      <p>FSSR is widely used as outdoor scene reconstruction,
when data is too sparse for a reliable reconstruction. In this
case the
method
does not hallucinate
geometry in
incomplete regions, requiring manual intervention, but
leaves in these areas holes (i.e. these areas have gaps).</p>
      <p>
        The approach draws upon a simple yet efficient
mathematical formulation to construct an implicit function
as the sum of compactly supported basis functions. The
implicit function has spatially continuous “floating” scale
and can be readily evaluated without any preprocessing.
The final surface is extracted as the zero-level set of the
implicit function. One of the key properties of the
approach is that it is virtually parameter-free even for
complex, mixed-scale datasets [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>The FSSR method combines all depth maps in one
large point cloud. At this stage, the scale value is attached
to each point, indicating the factual size of the surface area
in which the point was measured. This value is derived
from the size of the regions identified in the MVS phase.
Then FSSR tools calculate a multi-scale 3D surface.</p>
      <p>
        SPSR is an improvement of the approach that
considers surface reconstruction as a spatial Poisson
problem [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. The approach explicitly incorporates the
point as interpolation constraints. Unlike other methods of
image processing and geometry processing, the term
screening is defined for a sparse set of points rather than
for the whole area. These rare constraints, however, can be
effectively integrated. Since the modified linear system
retains the same finite-element discretization, the sparse
structure is unchanged and the system can still be resolved
using a multi-mesh approach.
      </p>
      <p>
        In addition, Poisson's surface reconstruction presents
several algorithmic improvements that together reduce the
time complexity of the solution to linear in the number of
points, thus enabling faster and better surface
reconstruction [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ].
5.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Algorithm Comparison</title>
      <p>Consider a combination of MVS-FSSR and
MVSSPSR approaches here in more detail.</p>
      <p>Implementation is studied on the example of 21 photos
of the statuette (Fig. 2) and freely distributed software
Regard3D and MeshLab.</p>
      <p>In Fig. 3. shows the detection of keypoints by
Regard3D. This image contains 14,486 keypoints.</p>
      <p>These key points are then matched to establish sparse
matches between images (Fig. 4). The image features
require the invariance to the image scaling, rotation, noise
and changes in illumination.
The results of the pairing are then combined and
unfolded into multiple views, creating functional tracks.</p>
      <p>The next step in SfM implementation is incremental
triangulation algorithm. It assesses the relative position of
a well-matched original pair of the image, and then all
tracks visible in both images are triangulated. The
matching next images are incrementally added to the
reconstruction until all the reconstructed views become
part of the scene. Parameters of lens distortion are
evaluated during the reconstruction. The performance of
the following algorithms is significantly improved by
removing distortions from the original images.</p>
      <p>In Regard3D's "ideology", this method is called New
Incremental. The result is a sparse point cloud (Fig. 5).
21 cameras (according to the number of uploaded
images) have been calibrated by the program, i.e. 3D
positions and parameters of all images have been found.
13 583 points has received that match not only the model,
but also some part of the environment. The calculation
time of a point cloud has made slightly less than 30 s.</p>
      <p>Further we will proceed to compression of the sparse
point cloud by MVS method. The result of point
compaction using the MVS method is shown in the figure.
6. The computation time was 40.42 minutes; 4 756 185
points were created. As you can see, the point cloud has
holes on the side of the figure (Fig. 6).</p>
      <p>The corresponding depth map was obtained using the
MeshLab program (Fig.7).</p>
      <p>Surface generation by FSSR and SPSR. In Fig. 8. the
results of calculations in the Regard3D command line are
presented, illustrating the iterative algorithm of finding the
best candidates for comparison described above.
In general, 21 reports were produced - according to the
number of uploaded images. You can find the views
recommended for comparison view, as well as the number
of optimized points, i.e. points that have updated the depth
map data and normal in accordance with the described
algorithm.</p>
      <p>Fig. 9 shows the result of FSSR method surface
construction. The calculation time was 17.02 min. The
final surface contains 1,369,758 points. The model also
contains small noises and has gaps.</p>
      <p>In the right picture, you can see that the model has a
big hole. This is due to the fact that a shadow falls on this
area in the original images. The lighting change is
interpreted by the program as a lack of data for point
reconstruction, because the shaded area is found in only
23 species out of 21, which was a rejection of its revision
and surface reconstruction.</p>
      <p>Fig. 10 shows the result of surface reconstruction using
the SPSR method. The calculation time was 1.22 min. The
final surface contains 301,497 points. The model also
contains little noise and has gaps.</p>
      <p>In the right picture, you can see that the model has an
even greater gap than the previous method.</p>
      <p>We will compare the obtained models by several
indicators (Table 1).</p>
    </sec>
    <sec id="sec-4">
      <title>6. Conclusion</title>
      <p>Two surface generation algorithms were considered
during the research: Screened Poisson Surface
Reconstruction point approach and Floating Scale Surface
Reconstruction approach. In connection with the method
of point compression, the considered algorithms showed
different temporal and quantitative results. The result of
comparison of final 3D-models generated by these
methods is shown, reduction of time expenses in SPSR
method does not give qualitative result. Model MVS
Screened Poisson Surface Reconstruction is a much less
dense mesh than model MVS - Floating Scale Surface
Reconstruction. On the basis of the received data it is
possible to draw a conclusion that for reception of
qualitative 3D-models on the basis of not structured point
cloud it is necessary to use the algorithm of generation of
the surface based on changing scale of images. The surface
generation algorithm based on a point approach can be
used for small collections of photos that do not contain
multiscale images. Reduction of computational power in
model preparation, as well as their small volume can be
used, for example, for low-polygonal modeling in the
mobile applications.</p>
      <p>In the future, it is planned to conduct a comparative
analysis of existing algorithms based on depth map data,
as well as approaches that take into account changes in
illumination in photographs.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Sosnina</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Filinskikh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lozhkina</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          :
          <article-title>Analysis of the virtual nontrivial forms models creation methods (in Russian)</article-title>
          .
          <source>Information technologies, Т. 25. (11)</source>
          , pp.
          <fpage>679</fpage>
          -
          <lpage>681</lpage>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Sosnina</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Filinskikh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Korotaeva</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          :
          <article-title>Comparison of the low-polygonal 3D model creation methods (in Russian)</article-title>
          .
          <source>Information technologies, Т</source>
          .
          <volume>23</volume>
          (
          <issue>8</issue>
          ), pp.
          <fpage>564</fpage>
          -
          <lpage>568</lpage>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Malysheva</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tomchinskaya</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Features of the low-polygon modeling and texturing in the mobile applications (in Russian)</article-title>
          .
          <source>CONFERENCE KOGRAF-2019, ISBN 978-5-502-01200-3</source>
          , p.
          <fpage>51</fpage>
          -
          <lpage>54</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Westoby</surname>
            ,
            <given-names>Matthew J.</given-names>
          </string-name>
          , et al.:
          <article-title>Structure-fromMotion'photogrammetry: A low-cost, effective tool for geoscience applications</article-title>
          .
          <source>Geomorphology</source>
          ,
          <volume>179</volume>
          , pp.
          <fpage>300</fpage>
          -
          <lpage>314</lpage>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Kublanov</surname>
            <given-names>V.</given-names>
          </string-name>
          <article-title>and others: Biomedical signals and images in digital healthcare: storage, processing and analysis: a training manual (in Russian)</article-title>
          , pp.
          <fpage>193</fpage>
          -
          <lpage>195</lpage>
          , (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Snavely</surname>
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Scene reconstruction and visualization from internet photo collections</article-title>
          . USA : University of Washington, (
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Fuhrmann</surname>
            , Simon,
            <given-names>Fabian</given-names>
          </string-name>
          <string-name>
            <surname>Langguth</surname>
          </string-name>
          , and Michael Goesele:
          <article-title>Mve-a multi-view reconstruction environment</article-title>
          .
          <source>GCH</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>[8] Clustering views for multiple stereo views (CMVS)</article-title>
          , https://www.di.ens.fr/cmvs/.
          <source>Last accessed 10 May</source>
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Furukawa</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponce</surname>
            <given-names>J</given-names>
          </string-name>
          .: Accurate, Dense, and
          <article-title>Robust Multi-View Stereopsis (PMVS)</article-title>
          .
          <source>IEEE Computer Society Conference on Computer Vision</source>
          and Pattern
          <string-name>
            <surname>Recognition</surname>
          </string-name>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Langguth</surname>
            <given-names>F.</given-names>
          </string-name>
          et al.:
          <article-title>Shading-aware multi-view stereo</article-title>
          .
          <source>European Conference on Computer Vision</source>
          , Springer, Cham, pp.
          <fpage>469</fpage>
          -
          <lpage>485</lpage>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Seitz</surname>
            <given-names>S. M.</given-names>
          </string-name>
          et al.:
          <article-title>A comparison and evaluation of multi-view stereo reconstruction algorithms</article-title>
          .
          <source>2006 IEEE computer society conference on computer vision and pattern recognition (CVPR'06)</source>
          ,
          <source>Т.1</source>
          , pp.
          <fpage>519</fpage>
          -
          <lpage>528</lpage>
          (
          <year>2006</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Goesele</surname>
            <given-names>M.</given-names>
          </string-name>
          et al.
          <article-title>: Multi-view stereo for community photo collections</article-title>
          .
          <source>2007 IEEE 11th International Conference on Computer Vision</source>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Voronin</surname>
            ,
            <given-names>V</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fisunov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marchuk</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Svirin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Petrov</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Restoration of the depth map based on the combined processing of a multi-channel image (in Russian)</article-title>
          .
          <source>Modern problems of science and education, 6</source>
          , (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <article-title>Greedy algorithms</article-title>
          , https://habr.com/ru/post/120343/. Last accessed
          <issue>26</issue>
          <year>June 2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <article-title>Basics of Stereo Vision</article-title>
          , https://habr.com/ru/post/130300/. Last accessed 26 May
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Curless</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levoy</surname>
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A volumetric method for building complex models from range images</article-title>
          .
          <source>Proceedings of the 23rd annual conference on Computer graphics and interactive techniques</source>
          , pp.
          <fpage>303</fpage>
          -
          <lpage>312</lpage>
          (
          <year>1996</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Fuhrmann</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goesele</surname>
            <given-names>M.:</given-names>
          </string-name>
          <article-title>Fusion of depth maps with multiple scales</article-title>
          .
          <source>ACM Transactions on Graphics (TOG)</source>
          ,
          <source>Т</source>
          .
          <volume>30</volume>
          (
          <issue>6</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Fuhrmann</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goesele</surname>
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Floating scale surface reconstruction</article-title>
          .
          <source>ACM Transactions on Graphics (ToG)</source>
          ,
          <source>Т</source>
          .
          <volume>33</volume>
          (
          <issue>4</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Kazhdan</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoppe</surname>
            <given-names>H.</given-names>
          </string-name>
          :
          <article-title>Screened poisson surface reconstruction</article-title>
          .
          <source>ACM Transactions on Graphics (ToG)</source>
          ,
          <source>Т</source>
          .
          <volume>32</volume>
          (
          <issue>3</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Kazhdan</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bolitho</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hoppe</surname>
            <given-names>H.</given-names>
          </string-name>
          :
          <article-title>Poisson surface reconstruction</article-title>
          .
          <source>Proceedings of the fourth Eurographics symposium on Geometry processing, Т. 7</source>
          , (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>