<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Application of global optimization methods to increase the accuracy of classification in the data mining tasks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dorosh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>nko A. V.[</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>Lviv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>2</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>The article describes the solving of data mining task using neural-like structures of Successive Geometric Transformations Model (NLS SGTM). The main problems of this task are imbalanced dataset and different weigh of errors. Therefore, to take into account these features, the method of penalties and rewards was used, as well as a piecewise linear approach to classification. The supplement of the methods used by the final optimization procedure is proposed. The procedure of final optimization using simulated annealing.</p>
      </abstract>
      <kwd-group>
        <kwd>data mining</kwd>
        <kwd>classification</kwd>
        <kwd>imbalance problem</kwd>
        <kwd>cost-sensitive learning</kwd>
        <kwd>imbalanced data</kwd>
        <kwd>principal components</kwd>
        <kwd>neural-like structure of successive geometric transformations model</kwd>
        <kwd>NLS SGTM</kwd>
        <kwd>simulated annealing</kwd>
        <kwd>analysis of the principal components</kwd>
        <kwd>optimization methods</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>In the previous articles [1-2] the methods based on the combination of a Successive
Geometric Transformations Model with the method of penalties and rewards was
described. In addition, was developed the piecewise-linear approach to constructing
separating surfaces in classification tasks.</p>
      <p>The purpose of these methods is to solve the data mining tasks of the classification.
The main features of these tasks are large-size datasets, imbalanced dataset and
different weight of errors. The main goal of this research is increasing the accuracy of
classification and minimize the number of penalty points.</p>
      <p>In order to increase the accuracy of classification, we propose to supplement the
developed methods with final optimization procedures, in particular by the method of
random correction of decomposition elements and the method of simulated annealing.</p>
    </sec>
    <sec id="sec-2">
      <title>Problem statement</title>
      <p>Today the data mining tasks are widespread because most companies today have a
huge amount of accumulated data with information about sales, customers, orders,
and more. This information is a source of hidden knowledge. In turn, the possession
of this knowledge allows this company to take a leading position in the market, to win
the competitive struggle. Among such tasks, one of the most popular is the task of
classification. These tasks are formulated daily; in such spheres of life how as
commerce, telecommunication, and chemical industry, target marketing, insurance,
medicine, bioinformatics, and others. Researchers use different methods to solve
classification problems [1, 2, 3, 4].</p>
      <p>The main features of classification tasks in data mining are imbalanced data,
different weight of error, huge amounts of data. These features require using some
special additional methods to well-known methods of classification to provide high
accuracy of classification.</p>
      <p>Hereby as basic methods of classification, we used the neural-like structure of
successive geometric transformations model (NLS SGTM) [5, 6]. As an additional
method was used piecewise-linear approach [1] and cost-sensitive learning method [7,
8]. This allowed us to improve the classification accuracy and take into account the
specifics of a specific task. This article proposes to apply global optimization methods
to the neuro-like structure already trained as a result of previous experiments. This
will allow us to find such parameters of the neural-like structure, in which the sum of
points reaches the global max.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Increasing the accuracy of classification using random correction of decomposition elements.</title>
      <p>3.1.</p>
      <sec id="sec-3-1">
        <title>Analysis of the principal components</title>
        <p>The analysis of the principal components is the standard method used to reduce the
dimensionality of data in statistical pattern recognition system and signal processing
systems. However, it is also advisable to use the analysis of the principal components
to solve the problems of data mining because of their high dimensionality [9, 10].</p>
        <p>The main task of statistical recognition is the allocation of attributes - the process
in which the data space is transformed into space of attributes, which theoretically has
the same dimension as the input space. Conversions, however, are usually performed
so that a reduced number of the most effective features can represent the data space.
Consequently, only a substantial part of the information contained in the data remains,
the dimension of the data is reduced. If this approach is applied to data mining task,
we will reduce the size of the input data by extracting non-informative features
without losing significant data. Consider a more detailed analysis of the principal
components (in the theory of information is known as Karhunen-Loeve Transform) [11, 12].</p>
        <p>Assume that there exists a vector X of dimension m, which we want to convey with
the help of l numbers, where l &lt;m. If we simply cut the vector x, this will cause the
mean square error to be equal to the sum of the dispersions of the elements carved out
of the vector x. It is necessary to find some linear transformation T, for which the
value of the mean square error for the reduction of the vector X will be optimal. In
this case, the transformation T must have the property of a small dispersion for its
individual components. The analysis of the principal components maximizes the rate
of dispersion reduction and, accordingly, the probability of the correct choice [13,
14].</p>
        <p>Let X be an m-dimensional random vector from the initial set of data. The mean
value of this vector equal zero.
where E is the operator of statistical expectation. If X have a nonzero average, then
you can calculate this value before the analysis begins. Let q be a unit vector with
dimension m, the vector X is projected on vector q. This projection is defined as the
product of the vectors X and q:</p>
        <p>E(X)=0,</p>
        <p>A=XTq=qTX
with restriction</p>
        <p>||q||=(qTq)0.5 = 1. (3)</p>
        <p>Projection A is a random variable that has an average value and variance that is
related to random vector statistics X. Dispersion A equals</p>
        <p>σ2=E[A2]=E[(qTX) (XTq)]= qTE[(XXT]q = qTRq (4)</p>
        <p>The matrix R of dimension mm is the matrix of the correlation of the random
vector X defined as the expectation of the product of the random vector X on itself:</p>
        <p>R=E[XXT] (5)
The matrix R is symmetric, then:</p>
        <p>RT=R. (6)
It follows from (6) that if a and b are arbitrary vectors with dimension m1 then:
aTRb=bTRa (7)</p>
        <p>From equation (4) it follows that the projection A is a function of an odd vector q.
So, we have:
(q) =σ2= qTRq,
(1)
(2)
(8)
(q) – dispersion probe.</p>
        <p>Directly the principal components are defined as follows. Let the data vector x be a
realization of the random vector X. Since there are m possible values of the single
vector q, then we must consider m possible projections of the data vector x. By
formula (2)</p>
        <p>aj = qjTx=xTqj, j = 1,2,…,m (9)
where aj - the projections of the vector x on the main guides, represented by single
vectors qj. These projections are called principal components, their number
corresponds to the dimension of the data vector x. In this case formula (9) can be
considered as the analysis procedure.</p>
        <sec id="sec-3-1-1">
          <title>Vector of</title>
          <p>input data</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>Encoding device</title>
        </sec>
        <sec id="sec-3-1-3">
          <title>Vector of the main components</title>
          <p>
</p>
          <p>Allocation of the principal components (PC) is also proceeding on the outputs of
the hidden layer of the neural-like structure of GTM. After selecting the PC, we create
and teach an additional neural-like structure (Fig. 2), where the inputs are the vectors
of the PC, and the outputs are the significance of the corresponding outputs of the
initial training data [15, 16, 17].
The feature of this neural network is the representation of the output of the synapse
as y  im1 zi well as weak correlation zi. In the mode of application on the inputs
of the trained neural network, input vectors are:
0 0 ...0  y0</p>
          <p>Get the elements of decomposition z0  y0 , zi  PCii , i  1,2,..., m.
Initially we have
y  im0 zi or y  im0 zi ki ,
(10)
where zi  the elements of decomposition, ki =1, i = 1,2,…, m .</p>
          <p>For classification tasks, appropriate indicators are optimized (percentage of
properly classified specimens, number of penalty points, average arithmetic or mean
square error) by random correction of elements of decomposition of coefficients ki .
Random correction of coefficients ki can be performed simultaneously for all
components or for each component independently. It should be noted that
componentbased random correction is appropriate for prediction since it has been experimentally
confirmed that the components are practically independent [18, 19].</p>
          <p>Consider a more detailed algorithm for random correction.
3.2.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Increasing the accuracy of the classification tasks based on correction of decomposition elements by random correction</title>
        <p>Let us consider the case of optimizing the accuracy of the classification problem
solving for a case where the recognized data belong to one of the two classes.</p>
        <p>Re-coding the input file: outputs of the elements belonging to class 1, assign 1, the
outputs of the elements belonging to class 2, assign -1.</p>
        <p>The problem of recognition is solved as a prediction problem by the formula (3),
but after obtaining a value, we analyze it by the formula (11):</p>
        <p>class1, y  0,
y   (11)</p>
        <p>class2, y  0.</p>
        <p>For the task of recognizing the criteria by which optimization can be carried out,
there may be the number of penalty points, the percentage of incorrectly classified
representatives of class 1, the percentage of incorrectly classified representatives of
class 2.</p>
        <p>Then the percentage of incorrectly classified representatives of classes is calculated
as follows:</p>
        <p>ErrC1 =100% NWC1/NC1, ErrC2 =100% NWC2/NC2 (12)
where NC1 is the number of representatives in class 1 in the training sample, NC2 –
number of representatives of class 2 in the training sample, NWC1 – number of
incorrectly classified representatives of class 1, NWC2 – number of incorrectly classified
representatives of class 2.</p>
        <p>In order to calculate the number of penalty points, it is necessary to determine the
fines that count for erroneous recognition. So if P1 is a fine charged if the element of
class 1 is recognized as an element of class 2 and P2 is a fine charged if a class 2
element is recognized as an element of class 1.</p>
        <p>Then the value of the penalty function - the total number of received penalty points
(PP) is:</p>
        <p>PP=ErrC1P1+ErrC2P2. (13)</p>
        <p>Optimization method for classification tasks based on correction of decomposition
elements by random correction:
1. Initial values ki=1.</p>
        <p>2. We calculate the value of the error, which is optimization (PP, ErrC1 or
ErrC2).</p>
        <p>3. Using a random number generator with a uniform distribution, we choose the
value ΔD from the range (–D, D).</p>
        <p>4. Calculate new values ki = ki + ΔD.
5. Calculate the value of outputs with new ki .</p>
        <p>6. Convert the resulting value to the designation of the class to which this element
belongs.</p>
        <p>7. Calculate the new value of the error, which is optimization (PP, ErrC1 or
ErrC2).</p>
        <p>8. Compare the value of the calculated error with the pre-calculated value. If the
new value is less than the previous one, we will remember it as well as the current
coefficients ki . Go to step 3. Otherwise, go to step 3 without remembering.</p>
        <p>9. Continue optimization until the predefined desired optimization value is
reached, or until the time t has expired.</p>
        <p>After performing the optimization method by randomly correcting the
decomposition elements, the resulting coefficients are used for further classification.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Application of global optimization methods to increase the</title>
      <p>accuracy of classification in the tasks of the data mining
4.1.</p>
      <sec id="sec-4-1">
        <title>Method of simulated annealing</title>
        <p>Annealing method is an algorithmic analogue of the controlled cooling process. It
was proposed in 1953 by N. Metropolis and refined by numerous followers. Today it
is considered one of the few methods by which one can practically find the global
minimum of functions of several variables. Consider a more detailed method of
simulated annealing [20, 21].</p>
        <p>Algorithm of simulated annealing method:
1. Start the process from the starting point at a given initial temperature T=Tmax.
2. As long as T&gt;0, repeat L times the following actions:
• choose a new solution w' from the vicinity w;
• calculate the change of target function; Δ = E(w')- E(w);
• if Δ≤0, take w = w'; otherwise, if Δ&gt;0, take w = w' with probability exp(-Δ/T)
by generating a random number R from the interval (0,1), then comparing it
with the value exp(-Δ/T); if exp(-Δ/T) &gt; R, take a new solution w = w'; in other
case - ignore it.
3. Reduce the temperature ( T  rT ) using the reduction coefficient r, selected from
the interval (0,1) and return to step 2.
4. After lowering the temperature to zero, apply one of the deterministic methods
(Levenberg-Marquardt algorithm, error-return algorithm, fastest-speed algorithm,
etc.) to achieve the minimum of the target function.</p>
        <p>The concept of "temperature" in this algorithm is quite formal, since the presented
optimization model is only a mathematical analogy of the annealing process.</p>
        <p>The efficiency of the annealing algorithm has an extremely high impact with the
choice of parameters such as the initial temperature Tmax, the coefficient of reduction
of temperature r and the number of cycles L, performed at each temperature level.</p>
        <p>The main problem is to determine the threshold level optimal for each annealing
simulation process. For some practical tasks, this level may have different meanings,
but the overall range remains unchanged. As a rule, the initial temperature is selected
so as to ensure the implementation of about 50% of the subsequent random changes in
the solution. Therefore, knowledge of the pre-distribution of such changes makes it
possible to estimate the initial temperature approximately.</p>
        <p>Numerous computer experiments [22] prove that in the case where the time limit is
small, the best results give a single implementation. If simulation can be long lasting,
then statistically better results can be achieved thanks to the multiple implementation
of the annealing simulation if the value of the coefficient r is close to 1.</p>
        <p>If we compare genetic algorithms with an annealing algorithm, then, in spite of the
significant external difference between the algorithms, they are essentially similar in
nature. An annealing algorithm according to [23] can be considered a genetic
algorithm with a population consisting of one instance. Consequently, an algorithm for
simulating annealing of a metal can be regarded as an algorithm that has only a
mutation operation, but not cross-linking.</p>
        <p>In addition, if we compare these two algorithms from the applied point of view,
then it should be noted that, according to Kohonen's study [7], in the case when the
initial solution is sufficiently close to optimal, the annealing algorithm of the metal
has significant advantages over the genetic algorithms from a computational point of
view.</p>
        <p>Since in our study the initial data are pre-processed by methods of fines and
incentives with sampling alignment and piecewise linear classification on the basis of the
model of geometric transformations, then the initial solution of the problem is
sufficiently close to the optimal. Accordingly, in this case, it is more appropriate to choose
for optimization of the solution.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Improvement of accuracy of the decision of tasks of intellectual data analysis on the basis of correction of elements of decomposition by algorithm of simulated annealing</title>
        <p>Let's consider a more detailed algorithm of simulated annealing of metal in
combination with methods of fines and incentives and piecewise linear classification on the
basis of a model of geometric transformations.</p>
        <p>Fig. 3 depicts the structural scheme of the developed neural network based on the
model of geometric transformations, where x1, x2 ,..., xn primary features of
classification objects – input data, PC1, PC2, ..., PCn  the principal components derived from input
data, w1, w2 ,..., wn  weight coefficients, y  an output that indicate on belonging to
certain classes.
Functioning of such a neural network can be described by the formula 14.</p>
        <p>y  in1 PCwi (14)</p>
        <p>The method of simulating annealing of a metal is proposed to be used to
optimize weight coefficients so that the resulting amount of penalty points is minimal,
that is, the optimization parameter is the amount of penalty points [15, 17].</p>
        <p>As can be seen from the flowchart describing the solution of the data mining
problem by combining the method of fines and incentives, simulated annealing and
piecewise linear approach, a modified annealing method for which the target function is the
number of penalty points, will be applied separately for each cluster [22, 24].</p>
        <p>Accordingly, if we have a two-step division into clusters for a sample of n classes,
then we will have division into clusters, which is depicted in Fig. 4, and for each of
the clusters a modified annealing algorithm will be implemented.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Experimental results</title>
      <p>This article describes the solving of classification task, which was formulated in [1].
The training sample describes the transactions carried out by credit card holders
within two days and consists of 284,807 lines and 31 columns Also, the dataset
contains one target feature 'Class', which shows the client's affiliation to one of two
classes - frauds or ordinary clients. The main feature of the dataset is that the data set
is highly unbalanced - only 492 transactions out of 284807 (0.172% of all
transactions) have the value of the target field 1, that is, customers are fraudulent. The
dataset has been collected and analyzed during a research collaboration of Worldline and
the Machine Learning Group (http://mlg.ulb.ac.be) of ULB (Université Libre de
Bruxelles) on big data mining and fraud detection [7].</p>
      <p>According to the subject area, a matrix was formed. By analyzing this matrix, it
can be seen that a properly classified vector that belongs to the "fraud" class has a
much greater weight than a properly classified "ordinary client" vector. At the same
time, the case where an ordinary customer is classified as fraud has the highest
number of penalty points (Table 1).</p>
      <p>Then we used the modified method of imitation of annealing of the metal with
parameters: initial temperature T=Tmax=20895, L=100, R is a random number from the
interval (0,1), r=0,9.</p>
      <p>After lowering the temperature to zero, apply one of the deterministic methods
(Levenberg-Marquardt algorithm, error-return algorithm, fastest-speed algorithm,
etc.) to achieve the minimum of the target function. The results of classification are
described (Fig 5, Fig.6).
The application of global optimization methods to classification in the data mining
tasks allowed to increase the accuracy of classification, especially in combination
with other methods of classification, such as neural-like structure of successive
geometric transformations model. Also, the method of simulated annealing successfully
combined with such methods as piecewise-linear approach and cost-sensitive learning
method. The application of the method of simulated annealing made it possible to
reach the point of the global maximum and minimize the amount of penalty points for
this task.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Doroshenko</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Piecewise-Linear Approach to Classification Based on Geometrical Transformation Model for Imbalanced Dataset</article-title>
          .
          <source>In: Proceedings of the 2018 IEEE Second International Conference on Data Stream Mining &amp; Processing (DSMP)</source>
          , Lviv, pp.
          <fpage>231</fpage>
          -
          <lpage>235</lpage>
          (
          <year>2018</year>
          )
          <article-title>Tkachenko</article-title>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Doroshenko</surname>
          </string-name>
          ,
          <string-name>
            <surname>A.</surname>
          </string-name>
          :
          <article-title>Classification of Imbalanced Classes using the Committee of Neural Networks</article-title>
          .
          <source>In: Proceedings of the XIIIth International Scientific and Technical Conference Computer Sciences and Information Technologies (CSIT)</source>
          ,
          <year>Lviv</year>
          ,
          <fpage>11</fpage>
          -
          <issue>14</issue>
          <year>September 2018</year>
          .
          <article-title>(2018) Tepla</article-title>
          ,
          <string-name>
            <given-names>T.L.</given-names>
            ,
            <surname>Izonin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.V.</given-names>
            ,
            <surname>Duriagina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.А.</given-names>
            ,
            <surname>Tkachenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.О.</given-names>
            ,
            <surname>Trostianchyn</surname>
          </string-name>
          , А.М.,
          <string-name>
            <surname>Lemishka</surname>
          </string-name>
          , І.А.,
          <string-name>
            <surname>Kulyk</surname>
            ,
            <given-names>V.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kovbasyuk</surname>
          </string-name>
          , Т.М.:
          <article-title>Alloys selection based on the supervised learning technique for design of biocompatible medical materials Archives of Materials Science</article-title>
          and Engineering,
          <volume>93</volume>
          /1, pp.
          <fpage>32</fpage>
          -
          <lpage>40</lpage>
          (
          <year>2018</year>
          ) Izonin,
          <string-name>
            <given-names>I.</given-names>
            ,
            <surname>Trostianchyn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Duriagina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            ,
            <surname>Tkachenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Tepla</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Lotoshynska</surname>
          </string-name>
          <string-name>
            <surname>N.</surname>
          </string-name>
          :
          <article-title>The Combined Use of the Wiener Polynomial and SVM for Material Classification Task in Medical Implants Production</article-title>
          .
          <source>International Journal of Intelligent Systems and Applications (IJISA)</source>
          , Vol.
          <volume>10</volume>
          , No.
          <issue>9</issue>
          , pp.
          <fpage>40</fpage>
          -
          <lpage>47</lpage>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Tkachenko</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yurchak</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Polishchuk</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Neurolike networks on the basis of Geometrical Transformation Machine</article-title>
          .
          <source>In: Proceedings of the 2008 International Conference on Perspective Technologies and Methods in MEMS Design, Polyana</source>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>80</lpage>
          (
          <year>2008</year>
          ). doi:
          <volume>10</volume>
          .1109/MEMSTECH.
          <year>2008</year>
          .4558743 Polishchuk,
          <string-name>
            <given-names>U.</given-names>
            ,
            <surname>Tkachenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            ,
            <surname>Tkachenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Yurchak</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.</surname>
          </string-name>
          :
          <article-title>Features of the autoassociative neuro like structures of the geometrical transformation machine (GTM).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <source>In: Proceedings of the 2009 5th International Conference on Perspective Technologies and Methods in MEMS Design, Zakarpattya</source>
          , pp.
          <fpage>66</fpage>
          -
          <lpage>67</lpage>
          (
          <year>2009</year>
          )
          <string-name>
            <surname>Pozzolo</surname>
            ,
            <given-names>A.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caelen</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>R.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bontempi</surname>
          </string-name>
          , G.:
          <article-title>Calibrating probability with underdamping for unbalanced classification</article-title>
          .
          <source>In: Proceedings of the 2015 IEEE Symposium Series on Computational Intelligence</source>
          , Cape Town, pp.
          <fpage>159</fpage>
          -
          <lpage>166</lpage>
          (
          <year>2015</year>
          ) doi: 10.1109/SSCI.
          <year>2015</year>
          .33 Liu,
          <string-name>
            <surname>X.</surname>
          </string-name>
          :
          <article-title>A Benefit-Cost Based Method for Cost-Sensitive Decision Trees</article-title>
          .
          <source>In: Proceedings of the WRI Global Congress on Intelligent Systems, Xiamen</source>
          , pp.
          <fpage>463</fpage>
          -
          <lpage>467</lpage>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Ebied</surname>
            ,
            <given-names>H. M.</given-names>
          </string-name>
          :
          <article-title>Feature extraction using PCA and Kernel-PCA for face recognition</article-title>
          .
          <source>In: Proceedings of the 2012 8th International Conference on Informatics and Systems (INFOS)</source>
          , Cairo, pp.
          <fpage>MM</fpage>
          -72
          <string-name>
            <surname>-</surname>
          </string-name>
          MM-
          <volume>77</volume>
          (
          <year>2012</year>
          )
          <article-title>Subbotin, S.: Quasi-Relief Method of Informative Features Selection for Classification</article-title>
          .
          <source>In: Proceedings of the 2018 IEEE 13th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT)</source>
          , Lviv, pp.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Amar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leshem</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gastpar</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A greedy approach to the distributed KarhunenLoève transform</article-title>
          .
          <source>In: Proceedings of the 2010 IEEE International Conference on Acoustics, Speech and Signal Processing</source>
          , Dallas, TX, pp.
          <fpage>2970</fpage>
          -
          <lpage>2973</lpage>
          (
          <year>2010</year>
          ) Mattera,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Palmieri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Haykin</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.:</surname>
          </string-name>
          <article-title>An explicit algorithm for training support vector machines</article-title>
          .
          <source>In: Proceedings of the IEEE Signal Processing Letters</source>
          , vol.
          <volume>6</volume>
          , no.
          <issue>9</issue>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Riznyk</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yurchak</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Povshuk</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Synthesis of optimal recovery systems in distributed computing using ideal ring bundles</article-title>
          .
          <source>In: Proceedings of the 2016 XII International Conference on Perspective Technologies and Methods in MEMS Design (MEMSTECH)</source>
          , Lviv, pp.
          <fpage>220</fpage>
          -
          <lpage>222</lpage>
          (
          <year>2016</year>
          )
          <article-title>Li</article-title>
          , Xin, Qian, Xu, Wang, Ziqiang:
          <article-title>Classification rule mining using feature selection and genetic algorithm</article-title>
          .
          <source>In: Proceedings of the 2009 Asia-Pacific Conference on Computational Intelligence and Industrial Applications</source>
          , Wuhan, pp.
          <fpage>107</fpage>
          -
          <lpage>110</lpage>
          (
          <year>2009</year>
          ) Yue,
          <string-name>
            <surname>S.</surname>
          </string-name>
          , Kaizhang, Liu,
          <string-name>
            <given-names>W.</given-names>
            ,
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          :
          <article-title>A Fuzzy Clustering Approach Using Reward and Penalty Functions</article-title>
          .
          <source>In: Proceedings of the 2009 Sixth International Conference on Fuzzy Systems and Knowledge Discovery, Tianjin</source>
          , pp.
          <fpage>18</fpage>
          -
          <lpage>21</lpage>
          (
          <year>2009</year>
          )
          <article-title>Zhiwei</article-title>
          , Ye, Juan, Yang, Xu, Zhang, Zhengbing, Hu:
          <article-title>Remote Sensing Textual Image Classification based on Ensemble Learning</article-title>
          .
          <source>International Journal of Image, Graphics and Signal Processing(IJIGSP)</source>
          , Vol.
          <volume>8</volume>
          , No.
          <volume>12</volume>
          , pp.
          <fpage>21</fpage>
          -
          <lpage>29</lpage>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Tkachenko</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cutucu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Izonin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Doroshenko</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tsymbal</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          .
          <article-title>(2019) Noniterative Neural-like Predictor for Solar Energy in Libya</article-title>
          . In: Ermolayev V.,
          <string-name>
            <surname>SuárezFigueroa</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yakovyna</surname>
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mayr</surname>
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nikitchenko</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spivakovsky</surname>
            <given-names>A</given-names>
          </string-name>
          . (eds) In
          <source>: Proceedings of the 14-th International Conference ICTERI 2018</source>
          .
          <article-title>Volume I: Main Conference</article-title>
          . Kyiv, Ukraine, May
          <volume>14</volume>
          -
          <issue>17</issue>
          , pp.
          <fpage>35</fpage>
          -
          <lpage>45</lpage>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Oleynik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Subbotin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Parametrical synthesis of neural network models based on the evolutionary optimization</article-title>
          .
          <source>In: Proceedings of the 2009 10th International Conference The Experience of Designing and Application of CAD Systems in Microelectronics, Lviv-Polyana</source>
          , pp.
          <fpage>335</fpage>
          -
          <lpage>338</lpage>
          (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Feiden</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tetzlaff</surname>
          </string-name>
          , R.:
          <article-title>Iterative annealing: a new efficient optimization method for cellular neural networks</article-title>
          .
          <source>In: Proceedings of the 2001 International Conference on Image Processing (Cat. No.01CH37205)</source>
          , Thessaloniki, Greece,
          <year>2001</year>
          , pp.
          <fpage>549</fpage>
          -
          <lpage>552</lpage>
          vol.
          <volume>1</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Granville</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Krivanek</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rasson</surname>
          </string-name>
          , J.-P:
          <article-title>Simulated annealing: A proof of convergence</article-title>
          .
          <source>IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          .
          <volume>16</volume>
          (
          <issue>6</issue>
          ):
          <fpage>652</fpage>
          -
          <lpage>656</lpage>
          (
          <year>1994</year>
          ) doi:10.1109/34.295910.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hsieh</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Very fast simulated annealing for pattern detection and seismic applications</article-title>
          .
          <source>In: Proceedings of the IEEE International Geoscience and Remote Sensing Symposium</source>
          , Vancouver, BC, pp.
          <fpage>499</fpage>
          -
          <lpage>502</lpage>
          (
          <year>2011</year>
          ) Shafti,
          <string-name>
            <given-names>L. S.</given-names>
            ,
            <surname>Pérez</surname>
          </string-name>
          ,
          <string-name>
            <surname>E.</surname>
          </string-name>
          :
          <article-title>Data Reduction by Genetic Algorithms and Non-Algebraic Feature Construction: A Case Study</article-title>
          .
          <source>In: Proceedings of the 2008 Eighth International Conference on Hybrid Intelligent Systems, Barcelona</source>
          , pp.
          <fpage>573</fpage>
          -
          <lpage>578</lpage>
          (
          <year>2008</year>
          )
          <article-title>Zhengbing</article-title>
          , Hu.,
          <string-name>
            <surname>Bodyanskiy</surname>
            ,
            <given-names>Y.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tyshchenko</surname>
            ,
            <given-names>O. K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boiko</surname>
            ,
            <given-names>O. O.</given-names>
          </string-name>
          :
          <article-title>An Ensemble of Adaptive Neuro-Fuzzy Kohonen Networks for Online Data Stream Fuzzy Clustering</article-title>
          .
          <source>International Journal of Modern Education and Computer Science(IJMECS)</source>
          , Vol.
          <volume>8</volume>
          , No.
          <issue>5</issue>
          , pp.
          <fpage>12</fpage>
          -
          <lpage>18</lpage>
          (
          <year>2016</year>
          ). doi:
          <volume>10</volume>
          .5815/ijmecs.
          <year>2016</year>
          .
          <volume>05</volume>
          .02
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>