<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Speech Vocal Clustering Methods</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Cherkasy State Technological University</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cherkasy</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shevchenko blvd.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ukraine</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>t.utkina</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>k.rudakov</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>i.zubko</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>m.chychuzhko}@chdtu.edu.ua</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>fedorovee</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>@ukr.net</string-name>
          <email>ineks-kiev@ukr.net</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>E. O. Paton Electric Welding Institute</institution>
          ,
          <addr-line>Kyiv, Bozhenko str., 11, 03680</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1899</year>
      </pub-date>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>The problem of increasing the clustering efficiency of vocal speech sounds is considered. Centroid and medoid clustering methods that use normalized distances which increases the efficiency of clustering, are proposed. Characteristics and quality criteria based on them are suggested for these methods. These methods have been investigated on TIMIT database and are intended for intelligent biometric identification systems.</p>
      </abstract>
      <kwd-group>
        <kwd>vocal speech sound</kwd>
        <kwd>centroid</kwd>
        <kwd>medoid</kwd>
        <kwd>clustering</kwd>
        <kwd>inter-cluster distance</kwd>
        <kwd>intra-cluster distance</kwd>
        <kwd>compression ratio</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Automated biometric identification of a person means decision-making based on
acoustic and visual information, which improves the quality of recognition of the
person under investigation. Unlike the traditional approach [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], computer biometric
identification speeds up and increases the likelihood of recognition, which is
especially critical in conditions of limited time.
      </p>
      <p>
        A special class of biometric identification of a person is formed by methods based
on acoustic information [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>To increase the efficiency of analysis, storage and transmission of a speech signal,
knowledge of the structure of vocal speech sound is required, for which the following
clustering methods are used.</p>
      <p>
        Traditional methods (without the use of artificial neural networks and
metaheuristics) [
        <xref ref-type="bibr" rid="ref3 ref4 ref5 ref6">3-6</xref>
        ] search for a solution faster than metaheuristic methods, some methods do
not require setting of the number of clusters, but perform only a directed search for a
solution and some methods require setting of the number of clusters or additional
parameters.
      </p>
      <p>
        Neural network methods [
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7-10</xref>
        ] perform a search for a solution faster than
metaheuristic methods, but perform only a directed search for a solution and require exact
setting of the number of clusters.
      </p>
      <p>
        Metaheuristic methods [
        <xref ref-type="bibr" rid="ref11 ref12 ref13 ref14">11-14</xref>
        ] perform a random search for a solution, but the
search for a solution is longer than that of other methods, and require exact setting of
the number of clusters.
      </p>
      <p>Thus, only some traditional methods do not require an exact setting of the number
of clusters, but instead require setting of additional parameters.</p>
      <p>The aim of the work is to increase the clustering efficiency of vocal speech sounds
by providing an adaptive number of clusters, a minimum number of set parameters
and parallel information processing.</p>
      <p>To achieve this goal, it is necessary to solve the following tasks:</p>
    </sec>
    <sec id="sec-2">
      <title>1. to develop a centroid clustering method; 2. to create medoid clustering methods; 3. to determine the characteristics and quality criterion of the clustering method; 4. to conduct a numerical study of the proposed clustering methods.</title>
      <p>2</p>
      <p>Problem statement
The problem of increasing the clustering efficiency of vocal speech sound comes
down to the problem of finding such vector of parameters * that satisfies the
crite1
rion F  SSWB   min , where SSWB is the ratio of the sums of average
intra</p>
      <p>C 
cluster and inter-cluster distances, C is the compression ratio of speech sound.
3</p>
      <p>Literature review</p>
    </sec>
    <sec id="sec-3">
      <title>Traditional clustering methods include:</title>
      <p>1. Partition-based (partitioning-based) or center-based methods</p>
      <p>
        In this case, a cluster is a set of objects, each of which is closer to the center of
this cluster than to the center of any other cluster. The center of the cluster is
usually a centroid (average value of the coordinates of all objects in the cluster) or a
medoid (an object of the cluster, the average difference of which from other objects
in the cluster is minimal). These methods consider intersecting areas with a high
density of objects as different clusters, for example, K-means [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], PAM
(kmedoids) [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], FCM [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], ISODATA [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] methods. The advantage of these
methods consists in a quick search for a solution; some of them do not require setting of
the number of clusters and have poor sensitivity to noise or random emissions. The
disadvantage of these methods consists in the presence of only a directed search for
a solution; a problem with clusters of different shapes, sizes and densities; some of
them require setting of the number of clusters and are sensitive to noise or random
emissions or require setting of additional parameters.
2. Model mixture or distribution-based or model-based methods
      </p>
      <p>
        In this case, a cluster is described by the probability density function. These
methods should be used when clusters are of different sizes, and the set of objects
of all clusters can be described by a mixture of distribution densities, for example,
EM method [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. The advantage of these methods consists in a quick search for a
solution; determination of the shape and size of each cluster by the corresponding
distribution density. The disadvantage of these methods consists in the presence of
only a directed search for a solution; requirement for precise setting of the number
of clusters; sensitivity to noise or random emissions; the difficulty of choosing the
appropriate distribution density.
3. Density-based methods
      </p>
      <p>
        In this case, a cluster is an area with a high density of objects, which is separated
by areas with a low density of objects from areas with a high density of objects.
These methods consider intersecting areas with a high density of objects as a single
cluster. These methods should be used when noise and random emissions are
present, clusters have different shapes and sizes, and the number of clusters is
unknown. Methods that define clusters (for example, DBSCAN method [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]) and
those that provide visualization of clusters (for example, OPTICS method [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ]) are
distinguished. The advantage of these methods consists in the absence of a
requirement to set the number of clusters; the presence of various shapes and sizes of a
cluster; poor sensitivity to noise or random emissions. The disadvantage of these
methods consists in the presence of only a directed search for a solution; a problem
with clusters of heterogeneous density; slow search for a solution.
4. Hierarchal methods
      </p>
      <p>
        These methods provide visualization of a cluster tree called a dendrogram. On a
dendrogram, pairs of clusters that are joined (in the case of agglomerative methods)
or obtained as a result of separation (in the case of divisive methods) are connected
by a U-shaped arc, the height of which corresponds to the distance between clusters.
By the method of constructing a dendrogram, these methods are divided into
agglomerative or bottom up ones – each object is considered as a singleton cluster,
after which a stepwise combination of pairs of the closest clusters is performed (for
example, centroid communication, Ward, single communication, full
communication, group average methods [
        <xref ref-type="bibr" rid="ref22 ref23">22, 23</xref>
        ]) and divisive or top down ones – all objects
are considered as one cluster, and at each step one of the constructed clusters is
divided into a couple of clusters (for example, DIANA, DISMEA methods [
        <xref ref-type="bibr" rid="ref22 ref23">22, 23</xref>
        ]).
The advantage of these methods consists in the absence of a requirement to set the
number of clusters; visibility; some of them don’t strongly tend to clusters of a
certain shape and size and have poor sensitivity to noise or random emissions. The
disadvantage of these methods consists in the presence of only a directed search for a
solution; slow search for a solution; some of them strongly tend to clusters of a
certain shape and size, are sensitive to noise or random emissions, tend to discard
clusters of high power.
      </p>
      <p>Usually, the methods listed above either require setting of the number of clusters or
have a slow search for a solution or require setting of additional parameters, which
leads to a decrease in the clustering efficiency.</p>
      <p>Therefore, the urgent task is to increase the clustering efficiency of vocal speech
sounds by providing an adaptive number of clusters, a minimum number of set
parameters and parallel information processing.</p>
      <p>Method clustering of vocal speech sound
Centroid clustering of vocal speech sound based on minimum distance
method
Centroid clustering of vocal speech sound based on the author’s minimum distance
method includes the following steps:
1. Set a lot of samples of vocal speech sound S  {si (n)} , i 1, I , n 1, N , which are
in a single amplitude-time window, where I is the number of samples, N is the
length of the sample. Set the number of quantization levels of a speech signal L
(for an 8-bit sound sample L  256 ). Set normalized threshold  , 0    1 . Set
the number of clusters K  0 .
2. Calculate normalized squared distance between each pair of sound samples
3. Calculate the distance between each sound sample and a lot of sound samples
Dij 
4. Determine the number of sound sample with a minimum distance
5. Set a sound sample with a minimum distance as the new cluster center, i.e.
mK 1  si* , set the number of sound samples in the new cluster per unit, i.e.
aK 1  1 , increase the number of clusters, i.e. K  K 1 .
6. Set the sound sample number i  1 .
7. If i*  i , then go to step 13.
8. Calculate normalized squared distance between the ith sound sample and cluster
centers</p>
      <p>Dk 
si  mk
9. Calculate the smallest normalized squared distance between the ith sound sample
and cluster centers
10. Determine the number of the cluster with a minimum distance
d *  min Dk , k 1, K .</p>
      <p>k
k*  arg min Dk , k 1, K .</p>
      <p>k
a * mk*  si
11. If d *   , then calculate the new center of the k* th cluster, i.e. m *  k
k
increase the number of sound samples in the k* th cluster, i.e. ak*  ak* 1 .
12. If d *   , then set the ith sample as the new cluster center, i.e. mK 1  si , set the
number of sound samples in the new cluster per unit aK 1  1 , increase the number
of clusters, i.e. K  K 1 .
13. If i  I , then go to the new sample, i.e. i  i 1 , go to step 7.</p>
      <p>The method results in an adaptive set of centroids.
4.2</p>
      <p>Medoid clustering of vocal speech sound based on minimax distance
method
Medoid clustering of vocal speech sound based on minimax distance method, unlike
the traditional version, preliminarily determines the number of a sound sample with a
minimum distance for a non-random choice of the first cluster center and includes the
following steps:
1. Set a lot of samples of vocal speech sound S  {si (n)} , i 1, I , n 1, N , which are
in a single amplitude-time window, where I is the number of samples, N is the
length of the sample. Set the number of clusters c  0 .
2. Calculate the squared distance between each pair of sound samples
3. Calculate the distance between each sound sample and a lot of sound samples
2
Dij  si  s j , i 1, I , j 1, I .
4. Determine the number of sound sample with a minimum distance
5. Set a sound sample with a minimum distance as the new cluster center, i.e.</p>
      <p>c1 (n)  si* (n) , n 1, N , increase the number of clusters, i.e. c  c  1.
6. Calculate the squared distance between each sound sample and each cluster center
7. Calculate a minimax squared distance between sound samples and cluster centers
Dik  si k</p>
      <p>2 , i 1, I , k 1, c .
d *  max min Dik , i 1, I , k 1, c .</p>
      <p>i k
i*  arg max min Dik , i 1, I , k 1, c .</p>
      <p>i k
8. Determine the number of sound sample with a minimax squared distance
9. If c  1 , then set a sound sample with a minimax squared distance as the new
cluster center, i.e. c1 (n)  si* (n) , n 1, N , increase the number of clusters, i.e.
c  c  1, go to step 6.
10. Calculate the average squared distance between cluster centers
 </p>
      <p>2 c c
c2  c i1 ji i  j
2
.
11. Verification of the termination condition.</p>
      <p>If d * 

2</p>
      <p>, then set a sound sample with a minimax squared distance as the new
cluster center, i.e. c1 (n)  si* (n) , n 1, N , increase the number of clusters, i.e.
c  c  1, go to step 6.
12. Calculate the distance between each sound sample and each cluster center
13. Determine for each sound sample the cluster center closest to it</p>
      <p>Dik  si k , i 1, I , k 1, c .
Medoid clustering of vocal speech sound based on subtractive clustering method,
unlike the traditional version, uses the normalization of squared distances for
calculating the potentials of sound samples to specify normalized standard deviations and
includes the following steps:
1. Set a lot of samples of vocal speech sound S  {si (n)} , i 1, I , n 1, N , which are
in a single amplitude-time window, where I is the number of samples, N is the
length of the sample. Set the number of quantization levels of a speech signal L
(for an 8-bit sound sample L  256 ). Set the threshold to stop the method  ,
0    1 . Set standard deviations  a and  b ,  b   a , 0   a  1 , 0   b  1 .</p>
      <p>Set the number of clusters c  0 .
2. Calculate the potential of each sound sample</p>
      <p>j1
P(i)  I e 4sa2iNLs2j 2 , i 1, I .
i*  arg max P(i) , i 1, I .</p>
      <p>i
3. Determine the number of sound sample with a maximum potential
4. Set the potential P(i* ) as the new cluster potential, i.e. P(c  1)  P(i* ) .
5. Calculate the new potential of each sound sample
 4 sic1 2
P(i)  P(i)  P(c  1)e b2NL2
, i 1, I .</p>
    </sec>
    <sec id="sec-4">
      <title>6. Verification of the termination condition.</title>
      <p>P(c  1)
If</p>
      <p>P(1)</p>
      <p>  , then set a sound sample with the highest potential as the new
cluster center, i.e. c1(n)  si* (n) , n 1, N , increase the number of clusters, i.e. c  c  1,
go to step 3.
7. Calculate the distance between each sound sample and each cluster center</p>
      <p>Dik  si k , i 1, I , k 1, c .
8. Determine for each normalized sound sample the cluster center closest to it</p>
      <p>The method results in an adaptive set of medoids.</p>
      <p>Medoid clustering of vocal speech sound based on minimum average
distance method
Medoid clustering of vocal speech sound based on the author’s method of minimum
average distance includes the following steps:
1. Set a lot of samples of vocal speech sound S  {si (n)} , i 1, I , n 1, N , which are
in a single amplitude-time window, where I is the number of samples, N is the
length of the sample. Set the number of quantization levels of a speech signal L
(for an 8-bit sound sample L  256 ). Set the radius of the neighborhood of sound
samples  , 0    1 . Set the number of clusters c  0 . Set a lot of samples of
speech sound that haven’t fallen into existing clusters, S  {si (n)} .
2. Calculate normalized squared distance between each pair of sound samples
Dij 
3. Calculate the distance between each sound sample and a lot of sound samples
4. Determine the neighborhood of each sound sample</p>
      <p>Ui,  { j | Dij   , j 1, I} , i 1, I .
5. Determine the neighborhood of the sound sample with a minimum average
distance</p>
      <p>d
i*  arg min i ,</p>
      <p>iS S</p>
      <p>U *  Ui* , .
6. Calculate the new distance between each sound sample and a lot of sound samples
that haven’t fallen into existing clusters</p>
      <p>0, i U *
di  d   Dik , i U * , k Ui, U * , i  S .</p>
      <p> i
 k
7. Determine the new neighborhood for each sound sample that hasn’t fallen into
existing clusters.
8. Set a sound sample with a minimum average distance as the new cluster center, i.e.</p>
      <p>c1 (n)  si* (n) , n 1, N , increase the number of clusters, i.e. c  c  1.
9. Verification of the termination condition.</p>
      <p>If S \ U *   , then S  S \ U * , go to step 5.
10. Calculate the distance between each sound sample and each cluster center
(codebook vector)
11. Determine for each sound sample the cluster center closest to it
ui  arg min Dik , i 1, I , k 1, c .</p>
      <p>k
The method results in an adaptive set of medoids.
5</p>
      <p>Determination of characteristics and quality criterion for
clustering methods of vocal speech sound
To evaluate clustering methods, the following characteristics are used in the work:</p>
    </sec>
    <sec id="sec-5">
      <title>1. The sum of average intra-cluster distances: ─ in the case of centroid clustering methods ─ in the case of medoid clustering methods</title>
      <p> I
SSW  kK1  i1  AikI1(si )Ak s(isi)mk
 I
c    Ak (si ) si k
SSW  k1  i1 I1 i  Ak (si )
 Ak (si )  10,, ssii  AAkk ,
2 

 ,



2 

 ,


</p>
    </sec>
    <sec id="sec-6">
      <title>2. The sum of inter-cluster distances: ─ in the case of centroid clustering methods ─ in the case of medoid clustering methods</title>
      <p>3. The ratio of the sums of average intra-cluster and inter-cluster distances</p>
      <p>In addition, to assess the quality of clustering methods, the following
characteristics – compression ratio for speech sound – is proposed in the work:
where I is the number of samples of vocal speech sound, c – the number of clusters.</p>
      <p>The following criterion for the quality of clustering, which means choosing such a
value  that minimizes the sum of the inverse of compression ratio and the ratio of
the sums of average intra-cluster and inter-cluster distances, is formulated in the
paper:</p>
      <p>F  SSWB 
1
C
 min .</p>
      <p>
(1)</p>
      <p>For centroid clustering of vocal speech sound based on minimum distance method
   . For medoid clustering of vocal speech sound based on subtractive clustering
method   ( a , b , ) . For medoid clustering of vocal speech sound based on
minimum average distance method    .</p>
      <p>6</p>
      <p>Experiments and results
Numerical experiments were carried out using notebook Intel Core i5 8th Gen,
MATLAB package, CUDA technology of parallel information processing on the
GeForce 920M graphics card with the number of threads in the block N s = 1024. In this
case, the most time-consuming (computational complexity O(I 2 ) , where I is the
number of samples of vocal speech sound) step 2 of all four proposed clustering
methods was parallelized. This made it possible to speed up the search for a solution.</p>
      <p>For speech signals containing vocal sounds, the sampling frequency fd = 8 kHz
and the number of quantization levels L = 256 were set. The length of the sample of
vocal speech sound N  256 .</p>
      <p>For modified method of subtractive clustering, the following fixed parameter
values  =0.01,  b  1.25 a were set.</p>
      <p>The results of a numerical study of the proposed clustering methods for vocal
sounds of people from TIMIT database are presented in Table 1.</p>
      <p>The result presented in Table 1 shows that the author's method of minimum
average distance provides the smallest F value, calculated according to (1). The
compression ratio is approximately 7.5, i.e. the number of stored samples is reduced by about
7.5 times.</p>
      <p>Based on the experiments, the following conclusions can be drawn.</p>
      <p>The author’s method of minimum distance should be used only when centroids are
required, and not medoids, since it has the largest F and the lowest compression ratio.</p>
      <p>The modified method of minimax distance performs a coarser adjustment of the
number of clusters than the modified method of subtractive clustering and the author's
method of minimum average distance, because it does not use parameters. On the
other hand, an operator does not need to set any parameters, the values of which are
established empirically.</p>
      <p>The modified method of subtractive clustering performs finer adjustment of the
number of clusters than the modified method of minimax distance, because it uses
parameters. It requires more complex setting than the author's method of minimum
average distance, because it uses three parameters, the values of which are established
empirically.</p>
      <p>The author's method of minimum average distance performs finer adjustment of
the number of clusters than the modified method of minimax distance, because it uses
parameters. It requires simpler adjustment than the modified method of subtractive
clustering, because it uses only one parameter, the value of which is established
empirically.</p>
      <p>7</p>
      <p>Conclusions
The article considers the problem of increasing the clustering efficiency of vocal
speech sounds. The following clustering methods are proposed – the author’s method
of minimum distance, the author’s method of minimum average distance, the
modified method of minimax distance (unlike the traditional version, it preliminarily
determines the number of a sound sample with a minimum distance for non-random
choice of the first cluster center), the modified method of subtractive clustering
(unlike the traditional version, it uses squared distances normalization to set normal
standard deviations in calculating the potentials of sound samples). Characteristics and
quality criteria based on them are proposed for these methods. The proposed methods
allow to increase the clustering efficiency of vocal speech sounds by providing an
adaptive number of clusters, a minimum number of specified parameters and parallel
information processing. The proposed methods are intended for software
implementation on GPU using CUDA technology, which speeds up the process of finding a
solution. Software that implements the proposed methods has been developed and
researched on TIMIT database. The conducted experiments have confirmed the
operability of the developed software and allow to recommend it for use in practice when
solving problems of biometric identification of a person. Prospects for further
research are to test the proposed methods on a wider set of test databases.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khan</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shree</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <source>Applications of Speaker Recognition. Procedia Engineering</source>
          .
          <volume>38</volume>
          ,
          <fpage>3122</fpage>
          -
          <lpage>3126</lpage>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Weychan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marciniak</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dabrowski</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Implementation aspects of speaker recognition using Python language and Raspberry Pi platform</article-title>
          .
          <source>2015 Signal Processing: Algorithms</source>
          , Architectures, Arrangements, and
          <string-name>
            <surname>Applications</surname>
          </string-name>
          (SPA).
          <article-title>(</article-title>
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .1109/SPA.
          <year>2015</year>
          .7365153
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Brusco</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shireman</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Steinley</surname>
            ,
            <given-names>D.:</given-names>
          </string-name>
          <article-title>A comparison of latent class, K-means, and Kmedian methods for clustering dichotomous data</article-title>
          .
          <source>Psychological Methods</source>
          .
          <volume>22</volume>
          ,
          <fpage>563</fpage>
          -
          <lpage>580</lpage>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .1037/met0000095
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Lenarczyk</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Piotrowski</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Speaker recognition system based on GMM multivariate probability distributions built-in a digital watermarking token</article-title>
          .
          <source>Przeglad Elektrotechniczny</source>
          .
          <volume>89</volume>
          (
          <issue>2</issue>
          ),
          <fpage>59</fpage>
          -
          <lpage>63</lpage>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Ram</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jalal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jalal</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A Density Based Algorithm for Discovering Density Varied Clusters in Large Spatial Databases</article-title>
          .
          <source>International Journal of Computer Applications</source>
          .
          <volume>3</volume>
          ,
          <issue>1</issue>
          -
          <fpage>4</fpage>
          (
          <year>2010</year>
          ). doi 10.5120/
          <fpage>739</fpage>
          -
          <lpage>1038</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Krishnamurthy</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Balakrishnan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Efficient Active Algorithms for Hierarchical Clustering</article-title>
          .
          <source>CoRR. abs/1206</source>
          .4672, (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Haykin</surname>
            ,
            <given-names>S.S.:</given-names>
          </string-name>
          <article-title>Neural networks and learning machines</article-title>
          .
          <source>Pearson</source>
          , Delhi (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Du</surname>
          </string-name>
          , K.-L.,
          <string-name>
            <surname>Swamy</surname>
            ,
            <given-names>M.N.S.</given-names>
          </string-name>
          :
          <source>Neural Networks and Statistical Learning</source>
          . (
          <year>2014</year>
          ). doi:
          <volume>10</volume>
          .1007/978-1-
          <fpage>4471</fpage>
          -5571-3
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Sivanandam</surname>
            ,
            <given-names>S.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sumathi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deepa</surname>
            ,
            <given-names>S.N.</given-names>
          </string-name>
          :
          <article-title>Introduction to neural networks using MATLAB 6.0</article-title>
          .
          <string-name>
            <surname>Tata McGraw-Hill</surname>
            <given-names>Education</given-names>
          </string-name>
          , New Delhi (
          <year>2010</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Larin</surname>
            ,
            <given-names>V.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fedorov</surname>
            ,
            <given-names>E.E.</given-names>
          </string-name>
          :
          <article-title>Combination of PNN network and DTW method for identification of reserved words, used in aviation during radio negotiation</article-title>
          .
          <source>Radioelectronics and Communications Systems</source>
          .
          <volume>57</volume>
          ,
          <fpage>362</fpage>
          -
          <lpage>368</lpage>
          (
          <year>2014</year>
          ). doi:
          <volume>10</volume>
          .3103/s0735272714080044
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Das</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abraham</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Metaheuristic clustering</article-title>
          . Springer Berlin, Berlin (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Saemi</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hosseinabadi</surname>
            ,
            <given-names>A.A.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kardgar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Balas</surname>
            ,
            <given-names>V.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ebadi</surname>
          </string-name>
          , H.:
          <article-title>Nature Inspired Partitioning Clustering Algorithms: A Review and Analysis</article-title>
          .
          <source>Soft Computing Applications Advances in Intelligent Systems and Computing</source>
          .
          <volume>96</volume>
          -
          <fpage>116</fpage>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          - 62524-
          <issue>9</issue>
          _
          <fpage>9</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Brownlee</surname>
          </string-name>
          , J.:
          <article-title>Clever algorithms: nature-inspired programming recipes</article-title>
          .
          <source>Creative Commons</source>
          ,
          <string-name>
            <surname>Melbourne</surname>
          </string-name>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Subbotin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Oliinyk</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levashenko</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zaitseva</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <article-title>Diagnostic rule mining based on artificial immune system for a case of uneven distribution of classes in sample</article-title>
          .
          <source>Communications. 3</source>
          .
          <fpage>3</fpage>
          -
          <lpage>11</lpage>
          . (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Steinley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brusco</surname>
            ,
            <given-names>M.J.</given-names>
          </string-name>
          :
          <article-title>Initializing K-means Batch Clustering: A Critical Evaluation of Several Techniques</article-title>
          .
          <source>Journal of Classification</source>
          .
          <volume>24</volume>
          ,
          <fpage>99</fpage>
          -
          <lpage>121</lpage>
          (
          <year>2007</year>
          ).
          <source>doi: 10.1007/s00357- 007-0003-0</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Kaufman</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Rousseeuw</surname>
            ,
            <given-names>P.J.:</given-names>
          </string-name>
          <article-title>Finding groups in data: an introduction to cluster analysis</article-title>
          . Wiley-Interscience, Hoboken, NJ (
          <year>2005</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>K.-L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>M.-S.:</given-names>
          </string-name>
          <article-title>A novel fuzzy clustering algorithm based on a fuzzy scatter matrix with optimality tests</article-title>
          .
          <source>Pattern Recognition Letters</source>
          .
          <volume>26</volume>
          ,
          <fpage>639</fpage>
          -
          <lpage>652</lpage>
          (
          <year>2005</year>
          ). doi:
          <volume>10</volume>
          .1016/j.patrec.
          <year>2004</year>
          .
          <volume>09</volume>
          .016
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Memarsadeghi</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mount</surname>
            ,
            <given-names>D.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Netanyahu</surname>
            ,
            <given-names>N.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moigne</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          :
          <article-title>A Fast Implementation Of The Isodata Clustering Algorithm</article-title>
          .
          <source>International Journal of Computational Geometry &amp; Applications</source>
          .
          <volume>17</volume>
          ,
          <fpage>71</fpage>
          -
          <lpage>103</lpage>
          (
          <year>2007</year>
          ). doi:
          <volume>10</volume>
          .1142/S0218195907002252.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Hastie</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Friedman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tisbshirani</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          :
          <article-title>The Elements of statistical learning: data mining, inference, and prediction</article-title>
          . Springer, New York (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Borah</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhattacharyya</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>DDSC : A Density Differentiated Spatial Clustering Technique</article-title>
          .
          <source>Journal of Computers. 3</source>
          , (
          <year>2008</year>
          ).
          <source>doi: 10.4304/jcp.3.2</source>
          .
          <fpage>72</fpage>
          -
          <lpage>79</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhou</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
          </string-name>
          , N.:
          <source>VDBSCAN: Varied Density Based Spatial Clustering of Applications with Noise</source>
          .
          <source>2007 International Conference on Service Systems and Service Management</source>
          . (
          <year>2007</year>
          ). doi:
          <volume>10</volume>
          .1109/ICSSSM.
          <year>2007</year>
          .4280175
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Aggarwal</surname>
            ,
            <given-names>C.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reddy</surname>
            ,
            <given-names>C.K.</given-names>
          </string-name>
          :
          <article-title>Data Clustering: Algorithms and Applications</article-title>
          . Chapman and Hall/CRC, Boca Raton, FL (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Gan</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , Ma,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.</surname>
          </string-name>
          :
          <article-title>Data clustering: theory, algorithms, and applications</article-title>
          . Siam, Philadelphia (
          <year>2007</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>