<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Technology of Selection and Recognition of Information Objects on Images of the Earth's Surface Based on Multi-Projection Analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stepan Bilan</string-name>
          <email>bstepan@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Hnatiienko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleh Ilarionov</string-name>
          <email>oilarionov@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hanna Krasovska</string-name>
          <email>annavkrasovska@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>Volodymyrska Street, 60, Kyiv, 01033</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <fpage>27</fpage>
      <lpage>28</lpage>
      <abstract>
        <p>The paper describes the technology for selecting and identifying objects on raster images of the earth's surface. A method for describing images using a set of projections obtained in different directions is presented. According to the proposed method, an initial scanning window with floating borders is formed. Projections are formed, the analysis of which made it possible to change the dimensions of the scanning window or to shift it in the appropriate direction until the object completely enters the window area. According to the generated projections, the geometric shape of the selected object is determined. In this case, the selection of edge pixels in the image is initially carried out, which makes it possible to improve the accuracy of identification. The method allows to increase the accuracy due to the formation of a larger number of projections in different directions. To do this, rotate the image itself at different angles is used. This approach allows you to determine the placement of the object and determine the presence of several objects on the window image in the case when in many projections the objects can be considered as one. The analysis of the density of distribution of ones pixels on the projections of the image of objects also made it possible to select several objects in the image. The use of threshold processing for the resulting projections made it possible to eliminate unnecessary areas that can be identified as a result of using the method. The method does not require large computational costs and is characterized by ease of analysis. Projection analysis allows feedback for effective correction of objects in the image and the image itself. Given the limited computing resources, the developed method is an attractive option for the implementation of systems for processing and recognizing satellite images. Image, object selection, earth image, Radon transform, binary image projection, image edges</p>
      </abstract>
      <kwd-group>
        <kwd>Multi-Projection</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <sec id="sec-1-1">
        <title>Signals that are not informational for us for other people can carry a lot of information.</title>
        <p>
          Currently, among all the tasks of image processing and recognition, one of the urgent tasks is the
analysis of aerial photographs of the earth's surface. The analysis of images of the earth's surface is used
in many areas of the national economy, such as: space, agriculture, military affairs, transport, geology,
meteorology and others. One of the tasks in the analysis of images of the earth's surface is the task of
selecting informative objects and their recognition. Particular attention is paid to improving the
performance of operations and achieving real-time. This is relevant for aircraft that fly over the ground
and carry out automatic analysis of the earth's surface. Various methods are used to solve this problem.
[
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1-4</xref>
          ]. A successful solution depends on the exact selection of the object in the image. Often the problem
is assigning image elements to an object they do not belong to.
        </p>
        <p>This distorts the result of recognition of the selected object in the image and leads to false solutions.
In addition, images of selected objects, which have the same top view, but at the same time, have</p>
        <p>2023 Copyright for this paper by its authors.
different heights of the object elements, have a significant impact on the recognition result. Existing
methods for automatic selection of objects on the image of the earth's surface do not give the desired
results. A particular challenge is to achieve high performance. Also, the solution of the problem depends
on the choice of characteristic features, a correctly formed set of which gives high image quality. Since
all objects in the image are viewed in a geometric projection from above, the recognition of such objects
is considered as the recognition of geometric shapes and pictures on a two-dimensional plane. Methods
aimed at solving such a problem use different approaches that do not use a single universal approach
for all forms of images. Different characteristic features are used that require different measuring
instruments and different computational models.</p>
        <p>In this paper, we solve the problem of analyzing images of photographs of the earth's surface based
on a multi-projection analysis of their binary representation. The multi-projection method allows you
to automate the process of selecting local objects and form a homogeneous set of characteristic features
that require unified algorithms for their processing.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Relative works</title>
      <p>
        The problem of selecting and recognizing objects in images from images of the earth's surface has
recently received much attention from specialists. Many works are devoted to the fractal analysis of
images of the earth's surface [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. However, this approach requires significant resources. The methods
mainly use reference images that are prepared in advance. In this situation, the resistance of informative
features to various kinds of distortions is analyzed. The universal quality index is used as an indicator
of durability [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Resistant to change are those areas for which the highest value of the universal quality
index is provided. In [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], the image is divided into homogeneous segments of different shapes, which
are determined by the characteristics of space and brightness, and the characteristics of the segments
are compared with the reference segments, taking into account the value of the decision rule. Such
methods are characterized by a large number of empirical parameters, which requires additional
processing time. There are methods that use the gradient representation of the original image based on
the Kirsch operator and automatically determine the clustering centers [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The use of the Kirsch
operator limits the area of the analyzed neighborhood, which reduces the accuracy. Common methods
that use artificial neural networks [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10 - 12</xref>
        ]. Such networks extract features and classify objects
simultaneously. However, they cannot classify objects with different orientations, and also have low
classification accuracy along with high performance. There are problems that are determined by the
imposition of bounding boxes on the image. Therefore, methods based on the pyramid of attention are
used [
        <xref ref-type="bibr" rid="ref13 ref14 ref15 ref3">3, 13 - 15</xref>
        ]. In such works, estimates are used, as well as a complex learning process, which affects
the accuracy and speed of operations. In recent years, methods have been used that use the Radon
transform and the hexagonal shape of the coverage to represent images [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. Such methods allow you
to select objects in the image with high accuracy based on the obtained Radon projections. However,
they have a limitation due to the fact that modern computing systems use an orthogonal representation
of graphic information. In this paper, we solve the problem of extracting informative objects based on
two projections using an orthogonal representation of bitmaps.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Selection of objects based on multi-projection analysis</title>
      <p>
        The exact selection of an object in a color image depends on the exact selection of its edges, which
determine its geometric shape. To select edge pixels in an image, many well-known methods can be
used [
        <xref ref-type="bibr" rid="ref17 ref18 ref19">17 - 19</xref>
        ], which give a different effect, but in general, edges are selected on all image objects.
      </p>
      <p>
        Since it is very difficult to immediately determine the location of objects in the image in automatic
mode, there is a need to search for methods without user intervention. One of these methods is the
method using the Radon transform [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. The Radon transform is used to solve many image processing
problems in which there is a need to pre-select individual image elements. For a quadrangular mosaic
(rectangular pixel shape), as a result of the Radon transformation, three projections are formed in the
direction 00, 450 and 900. To solve most problems where there is a selection of objects of such a number
of projections, it is enough. However, to improve the accuracy of the tasks being solved, six projections
can be formed, obtained in the directions 00, 300, 600, 900, 1200 and 1500. To obtain such projections
without aliasing distortion, pixels with a hexagonal mosaic are used (pixels have the shape of an
equilateral hexagon), which can be modeled on a rectangular mosaic [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. This made it possible to
represent the image in a hexagonal mosaic. This representation of images and its advantages are
described in the works [16. 20]. To select objects in images of the earth's surface, it is sufficient to
analyze two projections in the directions 00 and 900. However, this increases the time spent on solving
the problem. You can also use one projection, but at the same time rotate the image by a given angle.
The smaller the angle of rotation, the higher the accuracy of object selection. The selection of an object
on the image of the earth's surface is carried out using the following sequence of operational steps.
1. An image of the earth's surface using aircraft or probes are formed.
2. With the help of the selected operator, edge pixels are selected in the image.
3. The number of selected pixels in the image by applying thresholding are reduced.
4. The resulting image is converted into a binary one, where the selected pixels correspond to the
logical code "1", and the background pixels correspond to the logical code "0".
5. The resulting binary image is divided into regions of predetermined shapes and areas. As a rule,
they are divided into four rectangular areas that have a common vertex in the center of the image.
6. For each selected area of the image, projections are formed in the direction of 00 and 900.
7. The resulting projections are analyzed and, based on the analysis, a new shape of the selection
window is determined.
8. The sixth and seventh steps are repeated until there is one selected object in the image in the
window.
9. When the projections show that one object is selected in the generated window, the projections
are analyzed and the geometric shape of the object is determined. The geometric shape of an object
is determined by searching for the nearest reference set of projections.
      </p>
      <p>Projection analysis can show that parts of object images are selected in the formed rectangular area.
An example of such a situation in Figure 1 is shown.</p>
      <p>In this case, the system must decide in which direction to execute the algorithm. The selection
window can move in the direction of the object, part of which fell into the field of the original window,
or expand its border, which intersects the information object in the image. Both options are shown as
an example in Figure 2. Resizing or moving the window is carried out until the projections show that
the object is completely within the area of the analyzed window. In this case, the size of the object in
the image may exceed the size of the analyzed window. Therefore, an attractive option is the option of
expanding the boundaries (increasing the size of the window) towards the object, part of which fell into
the original window. After the object is located in the field of the selection window, the analysis of the
resulting projections shows which areas of the window need to be cut off. Thus, the window is first
enlarged, and then reduced to the size of a rectangle covering the selected object. In accordance with
this, the algorithm for generating the final window sizes consists of the following steps, which are
expressed by the graph diagram of the algorithm (Fig. 3).</p>
      <p>In fact, the method works in such a way that a feedback between projections and window sizes is
implemented. If there are situations when the projections do not show a separately selected object in
the image, i.e. projection overlays are present, then the image is rotated by a certain discrete angle. After
each discrete rotation, new projections are formed and analyzed to identify the object and determine its
geometric shape. Instead of rotation, it is possible to form different projections in parallel, which greatly
speeds up the image processing process. On Figure 4 are shown examples of projections obtained as a
result of rotating the image of the selected area. Figure 4 shows that with a rotation angle of 300, two
objects are selected, the geometric shape of which can be determined. It is also possible to remove a
small number of pixels on the projections, which will allow you to select a separate object in the total
set of selected pixels in the image.</p>
    </sec>
    <sec id="sec-4">
      <title>4. A method for recognizing the geometric shape of an object based on the analysis of projections of its binary image</title>
      <p>The formed projections of the selected object on the image allow you to determine its shape. Each
projection is a set of numbers, each representing one straight line in the direction of the projection. For
an n × m image, the projection can be given by the following model</p>
      <p>= 〈 1,  2, … ,   〉,
were   = ∑   ,
  – the value of the code of the j-th pixel in the corresponding i-th row of the image matrix.</p>
      <p>The code value of each pixel can be either 0 or 1. In this case, the number   р is equal to the number
of ones in the corresponding i-th line of the image, which has the size of a rectangle covering the
selected object. There are works that describe the forms of projections of various binary figures. For
example, the projections of a circle filled inside a contour are the same. But at the same time, there are
figures that can also give the same projections as circles. They may have voids inside the contour
(Figure 5). The same can be said about other geometric shapes. You can eliminate such problems by
using the outline representation of shapes (Figure 6). Especially if the outline is one pixel thick.</p>
      <p>The search for straight lines from the resulting projections is determined by a large number in the
sequence of numbers that determines the resulting projection. For example, if the projection is defined
by a set of numbers &lt;10, 2, 2, 2, 2, 2, 2, 2, 10&gt;, then it is safe to say that the edges of the object are
limited to two segments ten pixels long. If this is a horizontal projection, then the resulting projections
of the segments determine the top and bottom projections of objects. To determine the right and left
boundaries, analyze the vertical projection or rotate the image by 900. In this case, for a perfect
rectangle, the projection will be determined by the following set of numbers &lt;7, 2, 2, 2, 2, 2, 2, 2, 7&gt;.</p>
      <p>Inclined line segments are determined by the ratio of sets of numbers in several projections. For
example, there are two sets P1=&lt;1, 1, 1, 1, 1, 0, 0, 0, 0, 0&gt; and P2=&lt;0, 0, 0, 1, 1, 1, 1, 1, 0, 0 &gt;, then
you can define a segment, as shown in Figure 7.</p>
      <p>According to Figure 7 it can be seen that several line styles are possible according to the obtained
projections. Therefore, to improve accuracy, it is better to rotate the image and analyze those projections
that form the largest number on one of the projections. This means that the segment is perpendicular to
the projection that forms the maximum number. The angle of rotation is fixed, which determines the
angle of inclination of the segment. Because orthogonal tiling results in aliasing, the outline can be
represented as a few pixels thick. In this case, allowable thickness limits are set for each projection,
which can reduce the percentage of false description of the geometric shape, as well as reduce the time
spent on processing. Before and during operation, the system is trained using the generated numerical
sets that define projections and geometric shapes. For each geometric shape, two projections (two sets
of numbers) are formed, which are determined by the ratios of the numbers in each numerical set.</p>
      <p>However, the resulting initial numerical set is quite difficult to process. Therefore, to simplify the
processing of sets of numbers, thresholding is used, with the help of which the numbers in each set are
sequentially reduced until zero values appear. An example of such processing for the initial set of
numbers &lt;10, 2, 2, 2, 2, 5, 2, 2, 2, 3, 2, 2, 2, 2, 9&gt; is carried out in time steps, presented in Table 1.</p>
      <p>The table1shows that perpendicular to this projection there are three lines located in the first, sixth
and fifteenth row (column). If this projection is formed in the horizontal direction, then these line
segments are parallel to the horizontal axis. In this example, the number 3 can be regarded as noise and,
within the confidence interval, it is related to the contour thickness. The numbers in the far right column
of Table 1 indicate the number of pixels that were removed by thresholding. According to the results
of processing the numerical sequence of the vertical projection, segments of vertical lines are
determined. For an example of a vertical projection that forms an array of numbers &lt;7, 2, 3, 3, 3, 3, 3,
2, 6&gt;, the image in fig. 8 is shown.</p>
      <p>If we neglect the difference between the two extreme horizontal segments, then this figure can be
identified as a rectangle with a horizontal segment inside. For each geometric shape, you must assign
an identifier, as well as assign it to a specific class. The form shown in Figure 8 can be attributed to a
rectangle with a horizontal segment inside. If we consider real objects in the image, then, for example,
a bus can be represented as a rectangle in a top view, a car can be represented by a quadrilateral with
an internal quadrilateral that represents the roof, as well as additional internal quadrangles that represent
the glass windows in the car.</p>
      <p>The selection of objects by color arrangement does not always give the correct result, since a large
number of individual objects in the image can be selected, each of which must be analyzed separately,
which requires additional processing time, and it is also necessary to determine the geometric shape of
the object. It is also difficult to determine the height of an object located on the ground from the top
view, which does not make it possible to accurately identify the 3D geometric shape. However, it can
be clearly defined that it belongs to one of the classes of geometric shapes of objects.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Selection of a group of objects based on the analysis of projections of contour objects</title>
      <p>After selecting the edge pixels on the image, two projections are formed and analyzed. Previously,
analysis was considered for a single selected object. However, it is possible to single out a group of
objects at once by threshold analysis of the projections themselves. For selection, the analysis of
quantities, the values of which were selected experimentally, is used. Let us introduce the symbols and
formulas for determining the values of the case of horizontal projections:
d(i) - number of non-zero pixels in row i; μ - density test size;
δ(i) = (d(i-μ) + d(i-μ+1) +… d(i-1) + d(i) + d(i+1) +… + d(i+μ-1) + d(i+μ)) / μ –
density of the i-th element of the row (Figure 9, the density is shown in red);
θ - threshold for deleting a projection area (Figure 9, the threshold is shown in yellow, the
supposed areas are in red frames in the image);</p>
      <p>The allocation of informative areas is carried out according to the following rule</p>
      <sec id="sec-5-1">
        <title>For vertical projections, all quantities are determined similarly.</title>
        <p>It was empirically determined that the optimal threshold value  for deleting an informative area
is determined by the following formula.</p>
        <p>= (ℎ+ ),
4,64
were Θ - informative area deletion threshold: if the estimated area contains non-zero pixels less than
Θ, it is removed (Figure 10, the area in the upper left corner has been removed);
h - plot height; w – width.</p>
        <p>Since the projections of objects can overlap each other in one of the directions, the method does not
immediately unambiguously perform recognition and determine clear boundaries. However, a repeated
launch in certain areas refines the found boundaries and selects objects without losing information or,
conversely, storing extra elements that do not belong to the object (Figure 11).</p>
        <p>When processing image projections, simple threshold removal showed low efficiency, since areas
that were inside the object were removed or, conversely, single pixels remained, perceived by the model
as informative areas (Figure 12). Determining the density helps to reduce the “estimation” of single
pixels and, conversely, smooth empty pixels inside the informative area. And thresholding leaves only
the necessary objects on the image. This behavior is shown in previous figures.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>A system for recognizing objects on satellite images has been successfully implemented in the
work. The results of the experiments confirmed the effectiveness and advantages of the developed
method in comparison with other neural network approaches. One of the key advantages of the
implemented system is its speed. Compared to other approaches that require significant computational
resources for training and direct work, the developed method offers an effective solution that allows
you to increase the processing speed. This is important given the volume and complexity of such
images. As a result of the experiment, the threshold for deleting an informative area is determined if
the proposed area contains non-zero pixels less than a given value, which allows you to select several
objects at the same time in one image. The developed system for recognizing objects on satellite
images has demonstrated its efficiency and speed advantages. Given the limited computing resources,
the developed method is an attractive option for the implementation of systems for processing and
recognizing satellite images.</p>
      <p>In future studies, the authors plan to study and analyze images of the earth's surface to highlight and
recognize objects fixed at an angle of less than 900, as well as increase the number of projections, which
will significantly improve recognition accuracy.
7. References</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Fan</surname>
            <given-names>Q</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhuo</surname>
            <given-names>W</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tang C-K</surname>
            , Tai
            <given-names>Y-W</given-names>
          </string-name>
          ,
          <article-title>Few-shot object detection with attention-RPN and multirelation detector, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) IEEE</article-title>
          , (
          <year>2020</year>
          ), pp.
          <fpage>4013</fpage>
          -
          <lpage>4022</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Ye</surname>
            <given-names>Y</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            <given-names>H</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            <given-names>C</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hao</surname>
            <given-names>X</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            <given-names>Z</given-names>
          </string-name>
          ,
          <article-title>Sarpnet: Shape attention regional proposal network for lidar-based 3d object detection: Neurocomputing</article-title>
          ,No.
          <volume>379</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>53</fpage>
          -
          <lpage>63</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Hongchun</given-names>
            <surname>Yuan</surname>
          </string-name>
          , Hui Zhou, Zhenyu Cai, Shuo Zhang, Ruoyou Wu,
          <article-title>Dynamic Pyramid Attention Networks for multi-orientation object detection</article-title>
          .
          <source>Journal of Internet</source>
          Technology Vol.
          <volume>23</volume>
          No.
          <issue>1</issue>
          ,
          <year>January 2022</year>
          , pp.
          <fpage>81</fpage>
          -
          <lpage>90</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <surname>Faster</surname>
          </string-name>
          r-cnn:
          <article-title>Towards real-time object detection with region proposal networks</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis &amp; Machine Intelligence</source>
          , Vol.
          <volume>39</volume>
          , No.
          <issue>6</issue>
          , pp.
          <fpage>1137</fpage>
          -
          <lpage>1149</lpage>
          , June,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Tarshin</surname>
            <given-names>V. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sydorenko</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sotnikov</surname>
            <given-names>O. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pevtsov</surname>
            <given-names>G. V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Megelbey</surname>
            <given-names>G. V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lupandin</surname>
            <given-names>V. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kovalchuk</surname>
            <given-names>V. A.</given-names>
          </string-name>
          (
          <year>2015</year>
          ).
          <article-title>The method of selection of informative sections of images based on the theory of fractal analysis</article-title>
          .
          <source>Patent of Ukraine for utility model No. 100560, BI No. 14. from 27.07</source>
          .
          <year>2015</year>
          p.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>[6] Fractal analysis of processes, structures and signals</article-title>
          . Collective monograph / Under 5 ed. R.E. Pashchenko // Kharkiv:
          <source>HOOO "NEO "Ecoperspektiva"</source>
          ,
          <year>2006</year>
          . - 348 p.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Wang</surname>
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bovik</surname>
            <given-names>А</given-names>
          </string-name>
          . С., Lu L.
          <article-title>Why is image quality assessment so difficult? //</article-title>
          <source>Proc. IEEE Int. Conf. Acoustics, Speech, and Signal Processing</source>
          ,
          <year>2002</year>
          . V. 4. P.
          <volume>3313</volume>
          -
          <fpage>3316</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Orda</surname>
            <given-names>M. V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abramov</surname>
            <given-names>S. V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Poltorak</surname>
            <given-names>M. F.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>A method of automated detection of planar objects and informative areas on halftone raster images of the earth's surface. Patent of Ukraine for a utility model No. 121816, Byul</article-title>
          .
          <source>No. 23 dated 11.12</source>
          .2017
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Pat. WO</surname>
          </string-name>
          2009143651 (
          <issue>А1</issue>
          ),
          <article-title>IPC7 G06T 5/00. Fast image segmentation using region merging 30 with a k-nearest neighbor graph /Mantao X</article-title>
          . [CN],
          <string-name>
            <surname>Qiyong</surname>
            <given-names>G.</given-names>
          </string-name>
          [CN],
          <string-name>
            <surname>Hongzhi</surname>
            <given-names>L.</given-names>
          </string-name>
          [CN], Jiwu
          <string-name>
            <surname>Z.</surname>
          </string-name>
          [CN].
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Redmon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Divvala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Girshick</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. Farhadi,</surname>
          </string-name>
          (
          <year>2016</year>
          )
          <article-title>You only look once: Unified, real-time object detection, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas</article-title>
          ,
          <string-name>
            <surname>NV</surname>
          </string-name>
          , USA,
          <year>2016</year>
          , pp.
          <fpage>779</fpage>
          -
          <lpage>788</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>W.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Anguelov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Erhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Szegedy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Reed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.-Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. C.</given-names>
            <surname>Berg</surname>
          </string-name>
          , Ssd: Single shot multibox detector,
          <source>2016 European Conference on Computer Vision</source>
          (ECCV), Amsterdam, Netherlands,
          <year>2016</year>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>T-Y Lin</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Girshick</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dollár</surname>
          </string-name>
          ,
          <article-title>Focal loss for dense object detection</article-title>
          ,
          <source>IEEE Transactions on Pattern Analysis &amp; Machine Intelligence</source>
          , Vol.
          <volume>42</volume>
          , No.
          <issue>2</issue>
          , pp.
          <fpage>318</fpage>
          -
          <lpage>327</lpage>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Lu, W. Zhang, CAD-Net:
          <article-title>A context-aware detection network for objects in remote sensing imagery</article-title>
          ,
          <source>IEEE Transactions on Geoscience and Remote Sensing</source>
          , Vol.
          <volume>57</volume>
          , No.
          <volume>12</volume>
          , pp.
          <fpage>10015</fpage>
          -
          <lpage>10024</lpage>
          , December,
          <year>2019</year>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y. Zhang,</surname>
          </string-name>
          <article-title>Arbitrary-Oriented Object detection in remote sensing images based on polar coordinates</article-title>
          ,
          <source>IEEE Access</source>
          , Vol.
          <volume>8</volume>
          , pp.
          <fpage>223373</fpage>
          -
          <lpage>223384</lpage>
          , November,
          <year>2020</year>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>X.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          , K. Fu,
          <article-title>SCRDet: Towards More Robust Detection for Small, Cluttered</article-title>
          and
          <string-name>
            <given-names>Rotated</given-names>
            <surname>Objects</surname>
          </string-name>
          ,
          <year>2019</year>
          IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea,
          <year>2019</year>
          , pp.
          <fpage>8231</fpage>
          -
          <lpage>8240</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Motornyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bilan</surname>
          </string-name>
          .
          <source>The Moving Object Detection and Research Effects of Noise on Images Based on Cellular Automata With a Hexagonal Coating Form and Radon Transform. Handbook of Research on Intelligent Data Processing and Information Security Systems. Handbook of Research on Intelligent Data Processing and Information Security Systems</source>
          . Edited by Bilan,
          <string-name>
            <given-names>S. M.</given-names>
            , &amp;
            <surname>Al-Zoubi</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. I. Hershey</surname>
          </string-name>
          , USA: IGI Global(
          <year>2019</year>
          ):
          <fpage>330</fpage>
          -
          <lpage>359</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Sobel</surname>
            ,
            <given-names>I.E. Camera Models</given-names>
          </string-name>
          <article-title>and Machine Perception</article-title>
          ,
          <source>PhD Thesis</source>
          , Stanford Univ, (
          <year>1970</year>
          ):
          <fpage>60</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Prewitt</surname>
            ,
            <given-names>J.M.S.</given-names>
          </string-name>
          <article-title>"Object Enhancement and Extraction"</article-title>
          .
          <source>Picture processing and Psychopictorics</source>
          . Academic Press. (
          <year>1970</year>
          ):
          <fpage>75</fpage>
          -
          <lpage>149</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Stepan</surname>
            <given-names>Bilan.</given-names>
          </string-name>
          (
          <year>2022</year>
          )
          <article-title>Operators for Edge Detection in an Image Based on Technologies of Cellular Automata</article-title>
          ,
          <source>International Conference "Information Technology and Interactions" (IT&amp;I2022)</source>
          .
          <source>Workshops Proceedings Kyiv, Ukraine</source>
          ,
          <year>2022</year>
          . Vol.
          <volume>3384</volume>
          :
          <fpage>142</fpage>
          -
          <lpage>150</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Motornyuk</surname>
            <given-names>R.L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>S. Bilan.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Methods for Extracting the Skeleton of an Image Based on Cellular Automata With a Hexagonal Coating Form and Radon Transform</article-title>
          .
          <source>In Handbook of Research on Intelligent Data Processing and Information Security Systems</source>
          , ed. by
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Bilan</surname>
          </string-name>
          , &amp;
          <string-name>
            <given-names>S. I.</given-names>
            <surname>Al-Zoubi</surname>
          </string-name>
          ,
          <fpage>289</fpage>
          -
          <lpage>329</lpage>
          , USA: IGI Global
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>