<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Method for Clusters and Classifying the Objects in Images</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Natalya Shakhovska</string-name>
          <email>nataliya.b.shakhovska@lpnu.ua</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksii Shamuratov</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>12 Bandera str., Lviv, 79013</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1913</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>This paper presents an algorithm by which it is possible to determine the main features of objects such as area and perimeter, radii of inscribed and circumscribed circles, the sides of the circumscribed rectangle, the number and relative position of angles, histogram gradient of object. On the basis of these features to carry out clustering and classification of the image. The success of the recognition based on the convolutional neural network was evaluated. According to the results, it is possible to conclude that the smaller the invariance of the class, the greater the accuracy of recognition. The amount of data in the training sample has little effect on the accuracy of the algorithm.</p>
      </abstract>
      <kwd-group>
        <kwd>Keywords1</kwd>
        <kwd>machine translation</kwd>
        <kwd>clustering</kwd>
        <kwd>object classification</kwd>
        <kwd>convolution artificial neural networks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        2021 Copyright for this paper by its authors.
selecting the essential features or properties that characterize this data from the total mass of
insignificant details [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        After analyzing the general information about the classification, we can describe the formal task of
classification. Many objects are given. They need to be classified. The set is represented by subsets
called classes. Specified: information about classes, a description of the whole set, and a description
of information about an object whose membership in a certain class is unknown. Based on the
available information about the classes and the description of the object, you need to determine to
which class this object belongs [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>The clustering problem is similar to the classification problem, but the difference between the
method is that the classes are not defined. Clustering does not make any statistical conclusions, but
makes it possible to analyze the data structure, i.e. the purpose of clustering. there is a search for
similar structures.</p>
      <p>
        Artificial neural networks are one of the most effective and common ways to represent and solve
image recognition problems. Neural networks perform very well in pattern recognition problems
because it combines mathematical and logical calculations. The neural network allows you to process
a large number of factors, regardless of their origin, this is a stable universal algorithm. Neural
networks allow to build dependencies of parameters in the form of a polynomial on the basis of
educational sample that very simplifies realization of recognition of objects [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Color image segmentation is the process of selecting one or more areas in an image that meet some
of the homogeneity criteria.</p>
      <p>Image segmentation divides the image into areas with similar characteristics. The main feature for
segmentation is the brightness for the monochrome image and the color component for color image.
Also, image boundaries and textures are used for segmentation. Consider the main groups of
segmentation methods, used in the analysis of cytological images.</p>
      <p>
        One of the most well-known methods is threshold segmentation. This method is used for
contrasting images. So, threshold segmentation used in conjunction with morphological operations of
dilatation and erosion, and the choice of threshold is usually based on a priori information about
micro-objects, stand out. The advantages of these methods are the high speed of image processing and
the presence of theoretical justification. However, in more complex images with weakly contrasting
boundaries or with an uneven background, these methods cannot be used [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>Morphological segmentation is another approach to solving the segmentation problem. It comes
down to finding the contour of the cell that would best fit its boundaries. This is achieved using a
morphological gradient. However, the presence of a large number of false branches and internal
contours does not allow to solve the problem. The high complexity of such methods is obvious, and,
as a consequence, low speed and high complexity.</p>
      <p>
        The third group is the methods of segmentation of building areas. Area scaling is widely used in
image analysis, because often the structure is the background and does not allow you to select the
object, because its elements have brightness and brightness levels, which match the brightness of the
background. Typically, the selection of segmentation starting points occurs either randomly or with
the help of a human operator on the ground certain a priori information [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Another group of methods is based on the principles of the cluster approach. During segmentation
in the image, the centers of the clusters are selected, and then sequentially all points are checked for
the distance to the centers in some metric.</p>
      <p>
        The disadvantage of this group of algorithms is the need to pre-specify the number of clusters. It
can be overcome using a neural network approach. The disadvantage of using neural networks is the
learning stage, which significantly increases the analysis time [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ].
      </p>
      <p>The paper is organized as follows: in the next section the procedure of object detection and
classification is developed. The Results section presents the described dataset and the accuracy of the
procedure. The Conclusion section summarizes the work.</p>
    </sec>
    <sec id="sec-2">
      <title>Materials and Methods</title>
      <p>The procedure of image analysis is considered in accordance with the so-called Marr paradigm.
This paradigm states that image analysis is based on several successive levels of the ascending
information line: from point representation (raster image, unstructured information) to symbolic
(vector and attributive data in a structured form, relational structures) and should be carried out on a
modular basis using the following stages of analysis:
• image pre-processing;
• primary image segmentation;
• selection of the geometric structure of the visible field;
• definition of the relative structure and semantics of the visible scene.</p>
      <p>The levels of processing associated with these stages are called analysis of the lower (noise
filtering, analysis of histograms, etc.), medium (segmentation, contour analysis, etc.) and high
(geometric transformations, classification, recognition, etc.) levels. While there is a lot of scientific
work to solve low-level image processing problems, medium- and high-level algorithms continue to
be a central field of research.</p>
      <p>To find an object in an image database of objects must exist. It is planned to fill the database at the
expense of system users, who will select objects for further animation. For such objects, the system
must select the features, classify the object, and save it in the database. The database of classified
objects will be used in the future to search for similar objects in the image and allow the user to
automatically select objects for subsequent animation.</p>
      <p>Therefore, the definition of the object in the image must be stable to the rotation of the object, the
scale of the image and the variety of perspectives of the image. The following features of the object
are selected:
• Determination of the area and perimeter;
• Determination of radii of inscribed and described circles;
• Definition of the sides of the described rectangle;
• Determining the number and relative position of the angles;
• Construction of a histogram of object gradients.</p>
      <p>Clustering (segmentation) is procedure for grouping objects based on the proximity of their
properties so that each cluster consists of similar objects, and objects of different clusters differ
significantly. Clustering helps to understand the natural grouping or structure of a data set. The
purpose of clustering is to define internal grouping in a set of unlabeled data segments. With the help
of clustering solve the following problems of data analysis: search and discovery of knowledge,
grouping and recognition of objects: search for homogeneous groups (reduction of data dimension),
search for natural clusters and description of their unknown properties, search for useful and
appropriate grouping, search for unusual objects. projects (emission detection).</p>
      <p>The main methods of clustering include hierarchical, partitioning, density-based or grid-based, and
artificial neural networks (ANN).</p>
      <p>Before clustering the object should be detected on the image. The different images can be analyses
with different objects. That is why the obkect segmentation should be provided. For object
segmentation, the following steps are used: determination of the perimeter, the histogram filter usage,
feature selection.</p>
      <p>The task of classification in machine learning is the task of assigning an object to one of the
predefined classes on the basis of its formalized features. Each of the objects in this problem is
represented as a vector in N-dimensional space, each dimension of which has a description of one of
the features of the object. Suppose we need to classify monitors: measurements in our parameter
space will be the size of the diagonal in inches, the aspect ratio, maximum resolution, the presence of
HDMI-interface, cost, etc. The case of classification of texts is a bit more complicated, for them the
term-document matrix is usually used.</p>
      <p>To teach a classifier, you must have a set of objects for which classes are predefined. This set is
called a training sample, its marking is done manually, with the involvement of specialists in the study
area.</p>
      <sec id="sec-2-1">
        <title>3.1. Determination of area and perimeter</title>
        <p>
          The standard formula for calculating the area of an object is calculated by counting the number of
elements related to the object. In the case of images entering the system, the area can be calculated by
the number of pixels in the image area.
where O is the set of an array of pixels A(x, y) belonging to the object [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
        </p>
        <p>To find the perimeter of the object P, it is enough to count the number of pixels belonging to the
contour of the object.</p>
        <p>In order to ensure invariance to the scale of the object, a normalized sign U is added.</p>
      </sec>
      <sec id="sec-2-2">
        <title>3.2. Determining the radii of inscribed and circumscribed circles</title>
        <p>First the determination of the geometric center of the object is provided,
where x and y are the row and column numbers of all pixels A(x,y) belonging to the object.</p>
        <p>At this stage, the radius of the inscribed and circumscribed circles can be determined. This is done
by choosing the minimum and maximum length of the vector, which starts at the center of the image
and ends at points inside the perimeter.</p>
        <p>
          ,
,
,
;
,
[
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>are the coordinates of the center, N is the number of pixels in the perimeter, and</p>
      </sec>
      <sec id="sec-2-3">
        <title>3.3. Determining the sides of the circumscribed rectangle</title>
        <p>Determine the maximum and minimum values of the abscissa and ordinate of the image of the
object xmax і xmin, ymax і ymin, and then determine the height H and length L of the rectangle:</p>
        <p>H
where N is the number of pixels in the perimeter, and</p>
      </sec>
      <sec id="sec-2-4">
        <title>3.4. Determining the number and relative position of angles</title>
        <p>To do this, estimate the distance l between the start and end points of the contour fragment.</p>
        <p>
          Then the condition , is checked, where H is the conditional threshold value. You will need
to implement its definition based on the object's past properties. If the condition is satisfied, then the
point belongs to the set of angular points [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-5">
        <title>3.5. Construction histogram gradient of object</title>
        <p>The basic idea is the assumption that the appearance and shape of the objects in the area of digital
image can be described by the distribution of intensity gradients. The direction of the gradient of each
pixel is calculated. For a two-dimensional function f(x,y) the gradient vector has the following form:</p>
        <p>
          The partial derivatives of the intensity functions at the x and y coordinates are the estimates of the
contrast in the direction of the corresponding coordinate axes [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. As an accurate assessment of
contrast in a direction corresponding coordinate axis can average the three different estimates of
contrast around the pixel of the formulas:
        </p>
        <p>
          For data evaluation, follow minimize operator mask image of the object contrast. To reduce
computational complexity is best to use a differential-dimensional mask [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ].
        </p>
        <p>Once counted intensity gradient should split the image at the cell and construct a histogram of
gradients object for each cell pocket G histogram module corresponds to the intensity gradient at
point.
,</p>
        <p>Then carry out smoothing histograms in cells to ensure invariance relative change of anger and
contrast.</p>
      </sec>
      <sec id="sec-2-6">
        <title>3.6. Object features selection</title>
        <p>To find the similarity between objects the semantic distance between signs should be calculated.</p>
        <p>,
where - is the Euclidean distance between points i and j, which characterize one feature of
different objects.. N is the smallest number of values taken into account for a feature between two
objects, k is the ordinal number of the feature value.</p>
        <p>The procedure for determining the number of clusters consists of the following steps:
1. The k-means algorithm for K clusters is started and the corresponding internal variance
.
x is the vector that characterizes the object included in the cluster ci, the total distance
between the features and the center of the cluster. The mi - parameter is the center of the
current cluster. A set ðk is constructed for different values of K.
2. Choosing the degree of transformation</p>
        <p>.
3. Calculate the jumps according to the formula</p>
        <p>
          .
4. For the final number of clusters choose a value that is equal to the maximum [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ].
        </p>
      </sec>
      <sec id="sec-2-7">
        <title>3.7. Object recognition</title>
        <p>
          The convolutional neural network architecture was chosen to search for objects in the image for
the following reasons:
• images have a large dimension;
• a large number of parameters and classes of objects;
invariance to image zoom changes, camera shooting angles and other geometric distortions of the
input signal [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]:
• Input layer. Will contain the input value of the pixels of the image, the dimension of the layer
[64x64x3].
• Convolutional layer. The size of the filter is 7x7x3. The analysis will be performed on 5
features, so the layer will consist of 5 planes with dimensions [62x62x1].
• Sublayer layer. The dimension of the mask is 2x2. It consists of 5 planes measuring
[31x31x1].
• Source layer. Dimension [1x1xN], where N is the number of clusters in the database.
The hyperbolic tangent was chosen as the activation function:
,
where - is the value of the searched element, a is the weighted sum of the signals of the previous
layer, A is the amplitude of this function, S is its position relative to the reference point [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ].
        </p>
        <p>This feature has a number of advantages to solve the issue:
• symmetric activation functions, such as hyperbolic tangent, provide faster convergence than
the standard logistics function;
• function has a continuous first derivative;
• function has a simple derivative that can be calculated because of its significance, giving a
lower computational complexity.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Results</title>
      <p>
        5 different object classes were used to test the classification. Samples were taken from the open
access of the TensorFlow project [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. This dataset consists of the ImageNet dataset resized to fixed
size.
      </p>
      <p>An artificial neural network was trained on image samples, each class contained from 2528 to
16185 images with a size of 64x64 pixels. 1000 images of objects of each class were then selected for
testing. The results are shown in the Table 1.</p>
      <p>So the lowest result was for the Flowers class, due to the fact that the training sample was not large
and the images of objects were flowers of more than one species. In the sample of the Cars class there
were also cars of different brands and cab types, but due to the large number of images in the sample
the final test result was higher. The highest result was obtained for the Apple class. This is because
the sample contained an object whose structure and shape differed little.</p>
    </sec>
    <sec id="sec-4">
      <title>Conclusion</title>
      <p>This paper presents an algorithm by which it is possible to determine the main features of objects
such as area and perimeter, radii of inscribed and circumscribed circles, the sides of the circumscribed
rectangle, the number and relative position of angles, histogram gradient of object. On the basis of
these features to carry out clustering and classification of the image. The success of the recognition
based on the convolutional neural network was evaluated. According to the results, it is possible to
conclude that the smaller the invariance of the class, the greater the accuracy of recognition. The
amount of data in the training sample has little effect on the accuracy of the algorithm.</p>
      <p>The task of classification in machine learning is the task of assigning an object to one of the
predefined classes on the basis of its formalized features. Each of the objects in this problem is
represented as a vector in N-dimensional space, each dimension of which has a description of one of
the features of the object. Suppose we need to classify monitors: measurements in our parameter
space will be the size of the diagonal in inches, the aspect ratio, maximum resolution, the presence of
HDMI-interface, cost, etc. The case of classification of texts is a bit more complicated, for them the
term-document matrix is usually used.</p>
      <p>Using pixel relationships in a neighborhood has a number of advantages over using individual
point characteristics:
• the ability to use for images of any type;
• increased resistance to segmentation of images in which objects are in close proximity to
each other. This advantage allows you to use this method for segmentation of cytological
images;
• reducing the impact of noise and distortion of the input image on the overall result of
segmentation by analyzing the image by several algorithms of previous markup;
• reducing the number of "undefined" points, ie points that are on the boundaries of areas
and with equal probability may belong to two areas.</p>
      <p>The disadvantages are:
• complexity of the segmentation process;
• influence of previous marking on work results.</p>
      <p>But this effect can be reduced by increasing the number of previous markups</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Tien</surname>
            <given-names>D. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ching</surname>
            <given-names>Y. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zi-Cai</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            <given-names>Y. T.</given-names>
          </string-name>
          , “Computer Transformation of Digital Images and Patterns”, p.
          <fpage>276</fpage>
          ,
          <year>1989</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y.-Q.</given-names>
            <surname>Wang</surname>
          </string-name>
          , “
          <article-title>An Analysis of Viola-Jones Face Detection Algorithm”</article-title>
          ,
          <source>IPOL Journal</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Khan</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Abdullah и M. Shamian Bin</surname>
            <given-names>Zainal</given-names>
          </string-name>
          , «
          <article-title>Efficient eyes and mouth detection algorithm using combination of viola jones and skin color pixel detection</article-title>
          »
          <source>International Journal of Engineering and Applied Sciences, №</source>
          Vol.
          <volume>3</volume>
          №
          <issue>4</issue>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Gaede</surname>
          </string-name>
          и
          <string-name>
            <given-names>O.</given-names>
            <surname>Gunther</surname>
          </string-name>
          , “
          <article-title>Multidimensional Access Methods”</article-title>
          , ACM Computing Surveys, pp.
          <fpage>170</fpage>
          -
          <lpage>231</lpage>
          ,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Rahmani</surname>
          </string-name>
          , Syed Afaq Ali Shah,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bennamoun</surname>
          </string-name>
          , G. Medioni, S. Medioni, “
          <article-title>A Guide to Convolutional Neural Networks for Computer Vision”</article-title>
          , Morgan &amp; Claypool, p.
          <fpage>207</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>[6] Sibt ul Hussain, “Machine Learning Methods for Visual Object Detection”</article-title>
          . p.
          <volume>160</volume>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>Y. J.</given-names>
          </string-name>
          (
          <year>2006</year>
          ).
          <article-title>An overview of image and video segmentation in the last 40 years</article-title>
          .
          <source>Advances in Image and Video Segmentation</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Ngugi</surname>
            ,
            <given-names>L. C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abelwahab</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Abo-Zahhad</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Recent advances in image processing techniques for automated leaf pest and disease recognition-A review</article-title>
          .
          <source>Information processing in agriculture, 8</source>
          (
          <issue>1</issue>
          ),
          <fpage>27</fpage>
          -
          <lpage>51</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shen</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2020</year>
          ,
          <article-title>August)</article-title>
          .
          <article-title>Efficient semantic video segmentation with per-frame inference</article-title>
          .
          <source>In European Conference on Computer Vision</source>
          (pp.
          <fpage>352</fpage>
          -
          <lpage>368</lpage>
          ). Springer, Cham.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>P.</given-names>
            <surname>Arabie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. J.</given-names>
            <surname>Hubert</surname>
          </string-name>
          , G. De Soete, “Clustering and Classification”, p.
          <fpage>500</fpage>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>D.</given-names>
            <surname>Parks</surname>
          </string-name>
          , “
          <article-title>Object Detection and Analysis: A Coherency Filtering Approach”</article-title>
          , p.
          <fpage>172</fpage>
          ,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Yongqiang</surname>
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seong</surname>
            <given-names>G. K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Quan</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yongmei</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <article-title>“Multi-band Polarization Imaging</article-title>
          and Applications” 1st ed., p.
          <fpage>204</fpage>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Manikandan</surname>
            <given-names>S.</given-names>
          </string-name>
          , “
          <article-title>Vision Based Assistive System for Label and Object Detection”</article-title>
          , p.
          <fpage>64</fpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Salma</surname>
            <given-names>H.</given-names>
          </string-name>
          ,
          <article-title>“Object Detection Using Histogram Of Gradients”</article-title>
          , p.
          <fpage>52</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Wu</surname>
            <given-names>J.</given-names>
          </string-name>
          , “
          <article-title>Advances in K-means Clustering: A Data Mining Thinking”</article-title>
          , Springer Science &amp; Business Media, p.
          <fpage>180</fpage>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Loy</surname>
          </string-name>
          , “
          <article-title>Neural Network Projects with Python: The ultimate guide to using Python to explore the true power of neural networks through six projects”</article-title>
          , Packt Publishing, p.
          <fpage>308</fpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Brannon</surname>
            <given-names>W. C.</given-names>
          </string-name>
          , “
          <article-title>Object Detection in Low-spatial-resolution Aerial Imagery Using Convolutional Neural Networks”</article-title>
          , p.
          <fpage>79</fpage>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>[18] Dataset https://knowyourdata-tfds.withgoogle.com</mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>