<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Deep Learning of Neural Networks Using Genetic Algorithms</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Serhii Lienkov</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sergey Shvorov</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksandr Sieliukov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Igor Tolok</string-name>
          <email>igortolok72@gmail.com</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nataliia Lytvynenko</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Taras Davydenko</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Kyiv National University of Construction and Architecture</institution>
          ,
          <addr-line>31 Povitroflotskyi Av., Kyiv, 03037</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The technologies of artificial intelligence (AI) are aimed at creating a "thinking machine", that is, a computer system with human-like intelligence. One of the current directions of intellectualization is the use of neural networks with the implementation of their deep learning. The paper analyzes modern approaches to learning neural networks and investigates the possibility of using genetic algorithms to solve the problems of deep learning of neural networks. The purpose of the paper is to develop the scientific and methodological foundations of learning neural networks using genetic algorithms. To achieve the goal, the following tasks were solved: the justification of the approach to learning neural networks using genetic algorithms was carried out and the task of optimizing the learning of neural networks using a genetic algorithm was solved using the example of forecasting the time series of the environmental temperature by the method of shortest descent. A biotechnical complex exposed to external disturbances (external temperature) was chosen as the object on that relevant research was conducted.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Genetic algorithm</kwd>
        <kwd>genes</kwd>
        <kwd>chromosomes</kwd>
        <kwd>optimal solution</kwd>
        <kwd>deep learning</kwd>
        <kwd>search domain</kwd>
        <kwd>evolution time</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        At first glance, this problem seems unsolvable, but it isn’t necessary to invent a neural network
"from scratch" - there are dozens of different neural network architectures, and the effectiveness of
many of them has been proven mathematically [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>At the second stage, the most difficult task is solved – it’s necessary to "teach" the selected network,
i.e. choose such values of its weights, so that the network works as required. In those neural networks
used in practice, the number of weights can be several tens of thousands, so learning is a complex
process. For many architectures, special learning algorithms have been developed that allow to adjust
the weights of the network in a certain way. However, for complex ANN architectures, solving this
problem requires further research. Solving the problem of deep learning of neural networks is especially
relevant.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Problem Formulation</title>
      <p>
        As shown in [
        <xref ref-type="bibr" rid="ref1 ref3">1,3</xref>
        ], depending on the functions performed by neurons in the network, three types can
be distinguished:
      </p>
      <p>1) input neurons are neurons that receive an input vector encoding an input action or an image of
the external environment; they usually don’t carry out computational procedures, information is
transferred from the input to the output of the neuron by changing its activation;
2) output neurons are neurons whose output values represent the output of the network;
3) intermediate neurons are neurons that form the basis of artificial neural networks.</p>
      <p>In most neural models, the type of neuron is related to its placement in the network. If a neuron has
only output connections, then it’s an input neuron, and on the contrary, it’s an output neuron.</p>
      <p>However, there may be a case where the output of a topologically internal neuron is considered as
part of the network output. In the process of functioning (evolution of the state) of the network, the
input vector is transformed into the output vector, i.e., some processing of information is carried out.</p>
      <p>
        The process of the neural network (NN) functioning and the set of actions it can perform mainly
depends on the values of synaptic connections. That’s why, having determined the structure of the
network corresponding to the selected problem area, the optimization of the weighting coefficients in
real time is an urgent task [
        <xref ref-type="bibr" rid="ref2 ref4">2,4</xref>
        ]. Its ability to function qualitatively and adequately depends on how
effectively the NN training will be performed.
      </p>
      <p>The result of adding  is:
where:  - result of addition;
  – synapse weight ( = 1, … ,  );  – displacement value;
  – component of the input vector (input signal) ( = 1, … ,  );
 - number of neuron inputs.</p>
      <p>The experiments on learning neural networks have shown that known methods of local and global
optimization (gradient, stochastic, Newton, Hessian, etc.) require a significant number of learning steps,
are sensitive to the accuracy of calculations, require a significant number of additional variables,
therefore, the search and development of new methods is an urgent task learning of neural networks
[57].</p>
      <p>Therefore, there is a need to use such approaches, that wouldn’t have the mentioned disadvantages.
The genetic algorithm (GA) stands out among optimization mathematical devices in the context of the
given task.</p>
      <p>The idea of GA was expressed by J. Holland in the late 1960s and early 1970s.</p>
      <p>The basic (classical, elementary or simple) genetic algorithm consists of the following steps:
1) selection of the initial population of chromosomes;
2) assessment of the fitness of chromosomes in the population;
3) checking the condition of stopping the algorithm;
4) selection of chromosomes;
5) application of genetic operators;
6) formation of a new population;
n
i=1
r = ∑ Vi ∙ xi + f
(1)
7) selection of the "best" chromosome.</p>
      <p>The formation of the initial population consists in the random selection of a given number of
chromosomes (individuals), that are represented by binary sequences of a fixed length (that is, the
alleles of all genes in the chromosome are equal to 0 or 1).</p>
      <p>The assessment of chromosome fitness in a population consists in calculating the fitness function
for each chromosome of this population. The greater the value of this function, the higher the "quality"
of the chromosome. The form of the fitness function depends on the nature of the problem tobe solved.
It’s assumed that the fitness function always takes non-negative values and, in addition, that this
function must be maximized to solve the optimization problem. If the original form of the fitness
function doesn’t satisfy these conditions, then an appropriate transformation is performed (for example,
the function minimization problem can be easily reduced to a maximization problem).</p>
      <p>
        The determination of the stopping condition of the genetic algorithm depends on its specific
application. In optimization tasks, if the maximum (or minimum) value of the fitness function is known,
then the algorithm can stop after reaching the expected optimal value, possibly with a given accuracy
[
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6-8</xref>
        ]. Stopping of the algorithm can also happen if its execution doesn’t lead to an improvement of the
already achieved value. The algorithm can be stopped after a certain execution time or after performing
a given number of iterations. If the stopping condition is met, the transitionto the final stage of selecting
the "best" chromosome is made. Otherwise, the selection is performed inthe next step.
      </p>
      <p>The chromosome selection consists in choosing (based on the values of the fitness function
calculated at the second stage) those chromosomes that will participate in the creation of offspring for
the next population, that is, for the next generation. Such a selection is made according to the principle
of natural selection, according to that chromosomes with the highest values of the fitness function have
the greatest chance of participating in the creation of new individuals. There are different methods of
selection. The most popular is the so-called roulette method, that got its name by analogy with the
wellknown gambling game.</p>
      <p>As a result of the selection process, a parent population (parental pool) is created with a number 
equal to the number of the current population. The application of genetic operators to chromosomes
selected by selection leads to the formation of a new population of descendants from the parent
population created in the previous step.</p>
      <p>Application of genetic operators. In the classic genetic algorithm, two basic genetic operators are
used: the crossover operator and the mutation operator. However, it should be noted that the mutation
operator plays a secondary role compared to the crossover operator. Since crossover in the classic
genetic algorithm occurs almost always, while mutation is quite rare.</p>
      <p>Formation of a new population. Chromosomes obtained as a result of the application of genetic
operators to the chromosomes of the temporary parent population are included in the composition of
the new population. It becomes the current population for this iteration of the genetic algorithm. At each
subsequent iteration, the values of the fitness function are calculated for all chromosomes of this
population, after that the condition for stopping the algorithm is checked and either the result is fixed
in the form of the “best” chromosome (that has the largest value of the fitness function), or the transition
is made to the next step of the genetic algorithm, i.e. to selection. In the classical genetic algorithm, the
entire previous population of chromosomes is replaced by a new population of descendants having the
same number.</p>
      <p>In the classical genetic algorithm, only the binary coding method is used: selection by the "roulette
wheel" method and point crossing (with one crossing point). To increase the efficiency of its work,
many modifications of the basic classical genetic algorithm have been created, that are associated with
the use of other selection methods, with the modification of genetic operators (first of all, the crossover
operator), with the transformation of the fitness function, as well as with other ways of encoding the
parameters of the problem in the form chromosomes.</p>
      <p>These algorithms simulated evolutionary processes in the generations of such chromosomes. They
implemented mechanisms of selection and reproduction similar to those used in natural evolution. Just
as in nature, genetic algorithms searched for "good" chromosomes without using any information about
the nature of the problem being solved. It was needed some estimate of each chromosome reflecting its
fitness. The mechanism of selection consists in choosing chromosomes with the highest score (that is,
the most adapted), that reproduce more often than individuals with a lower score (worse
adapted). The reproduction means the creation of new chromosomes as a result of the recombination of
the genes of the parental chromosomes. The recombination is a process that results in new combinations
of genes. Two operations are used for this: crossing over, that allows creating two completely new
offspring chromosomes by combining the genetic material of a pair of parents, and mutation, that can
cause changes in individual chromosomes.</p>
      <p>Genetic algorithms use a number of terms borrowed from genetics, primarily genes and
chromosomes, as well as population, individual, allele, genotype, phenotype.</p>
      <p>Genetic algorithms are used in software development, in artificial intelligence systems, optimization,
artificial neural networks and in other fields of knowledge. It should be noted that with their help
problems are solved for that only neural networks were previously used. In this case, the genetic
algorithms act simply as an alternative method independent of neural networks, designed to solve the
same problem. The genetic algorithms are often used in conjunction with neural networks. They can
support neural or jointly interact within the framework of a hybrid system designed to solve a specific
task. The genetic algorithms are also used in conjunction with fuzzy systems.</p>
      <p>The genetic algorithm is a method that reflects the natural evolution of problem solving methods,
and primarily optimization problems. The genetic algorithms are search procedures based on
mechanisms of natural selection and heredity. They use the evolutionary principle of survival of the
fittest. Genetic algorithms differ from traditional optimization methods in the following main
properties:</p>
      <p>In terms of the speed of determining the optimal value of the objective function, the genetic
algorithms are several orders of magnitude ahead of random search. However, the genetic algorithms
aren’t the only way to solve optimization problems. In addition to it, there are two main approaches to
solving such problems - exhaustive and local-gradient, each of that has its advantages and
disadvantages.</p>
      <p>The iterative method is the easiest to program. To search for the optimal solution, it’s necessary to
calculate consistently the value of the objective function at all possible points, remembering the
maximum (or minimum) of them. The disadvantage of the method is the high computational
complexity, however, if it’s possible to go through all options in a reasonable time, then the solution
found is optimal.</p>
      <p>The second approach is based on the gradient descent method. First, some random parameter values
are chosen, and then these values are gradually changed, achieving the highest growth rate of the
objective function. When a local maximum (minimum) is reached, this method stops, so additional
measures are required to find the global optimum.</p>
      <p>The gradient methods work quickly, but don’t guarantee the optimality of the solution found. They
are ideal for solving unimodal problems, where the objective function has a single local optimum
(global). The practical problems, as a rule, are multimodal and multidimensional, for them there are
no universal methods that allow to find quickly absolutely accurate solutions. By combining the
screening and gradient methods, it can to get approximate solutions, the accuracy of that will increase
as the calculation time increases.</p>
      <p>
        The paper [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] shows the main differences between GA and standard optimization algorithms:
 the search for a suboptimal solution, based on the optimization of randomly given set of solutions,
rather than one solution, that allows simultaneous analysis of several ways of approaching the
extremum; evaluation of such solutions at each step allows synthesizing new solutions on the basis of
old ones, i.e. evolutionary development of optimal solutions takes place;
      </p>
      <p> the solutions are considered as some coded structures, and not as a set of parameters, that allows
in some cases to significantly reduce the time of data transformation, i.e. to increase the speed of finding
optimal solutions;</p>
      <p> to assess the "suitability" of a decision for further evolutionary development, along with the use
of the objective function, the "Rules of Survival" are additionally modeled, that increase the diversity
of the set of decisions and determine the evolutionary development;</p>
      <p> the initialization, transformation and other types of decision operations use probabilistic rather
than deterministic rules that introduce elements of randomness into the genetic search; therebysolving
the problem of leaving local optima;</p>
      <p> there is no need to calculate the derivatives of the target function (as in gradient methods) or the
matrix of derivatives of the second order (as in quasi-Newtonian methods);
 non-critical to the number of components of the admissible solution vector.</p>
      <p>
        The authors of [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4-6</xref>
        ] consider the issue of operator adaptation in evolutionary computations and its
application to optimize the structure of neural networks, effective multi-objective search for neural
architecture using Lamarck evolution, such works indicate the need for further research, and especially, in
out;
solved.
the direction of the application of genetic algorithms for learning neural networks.
      </p>
      <p>The purpose of the paper is to develop the scientific and methodological foundations of learning
neural networks using genetic algorithms.</p>
      <p>To achieve the goal, the following tasks were solved:
1. The justification of the approach to learning neural networks using genetic algorithms was carried
2. The optimization problems of learning using the genetic algorithm on the example of forecasting
the time series of the temperature of the natural environment by the method of shortest descent are</p>
    </sec>
    <sec id="sec-3">
      <title>3. Research materials and methodology</title>
      <p>
        In the classical formulation, the task of learning neural network is considered as the task of finding
the minimum of the learning error, that depends on the parameters of the network [
        <xref ref-type="bibr" rid="ref7 ref8">7,8</xref>
        ]. The quality of
learning directly affects the prediction capabilities of the neural network, and therefore the accuracyof
the problems being solved. Taking into account the above, it’s possible to determine the network
structure that corresponds to the chosen problem. The optimization of weighting factors in real time is
relevant. The adequacy of its functioning depends on this.
      </p>
      <p>The main functional purpose of an artificial neural network is the transformation of input signals
(some scattered information about the external environment) into output signals (concepts about the
external environment). Based on (1), the neural network in this case is represented as some
multidimensional function  :</p>
      <p>→  .</p>
      <p>If the set of weight coefficients   of the input signals of the neurons of the network isn’t ordered,
then the function  implemented by the network is arbitrary. The set of all weights of all neurons
corresponds to the vector  . The set of vectors 
forms the state space of the neural network. Let us
correspond to the initial state of the network with some arbitrary vector  0. Then the trained neural
network corresponds to  ∗, i.e. such a state in that the one-valued mapping is realized  : 
→  . In this
case, the task of learning NN is formally reduced to the task of transition from some initial stateof the
network corresponding to  0, to the final state corresponding to  ∗.</p>
      <p>When learning a neural network, the problem of error minimization is set

 =1
 ( ) = ∑(  −   )
(2)
where   - the value of the  th output of the neural network, is the known value of the  th output;
 - the number of neurons in the output layer.</p>
      <p>If the network doesn’t make errors, then  = 0, that is, the goal of learning neural network is the
task of finding the minimum of the error function (2) in the state space  .</p>
      <p>
        In order to increase the reliability of decisions made on the basis of a neural network, it’s necessary
to investigate alternative optimization algorithms that allow finding the global extremum. Therefore,
there is a need to use such approaches, that wouldn’t have the indicated disadvantages. The considered
genetic optimization algorithms are the most promising in this regard [
        <xref ref-type="bibr" rid="ref7 ref8">7-8</xref>
        ].
      </p>
      <p>To perform the optimization procedure using the genetic algorithm, it’s necessary:
1) to choose a representation of optimization parameters in the form of a certain data format: line,
vector, table, array, etc.;
features of the search space;
2) to develop or choose from a set of genetic operators those that best take into account the</p>
      <sec id="sec-3-1">
        <title>3) to determine the size of the initial population;</title>
        <p>4) to develop a technique for using genetic operators;
5)to set the fitness function (the objective function by that variants are selected in the population);
6) to develop a method of selection’s variants to a new population;
7) to set the criterion for stopping the evolutionary process.</p>
        <p>To minimize the learning error of neural networks based on the genetic algorithm, each variant of
the vector of weighting coefficients  is matched with some chromosome, presented in the form of a
bit string. In the selection process, a directed search is made for chromosomes that provide the
extremum of a given objective function, as the neural network learning error function  is used in the
neural network training procedure.</p>
        <p>The concept of learning neural network when using a genetic algorithm, in contrast to traditional
learning methods, has a different meaning: learning here consists in applying genetic operators to the
genotype of the vector   , i.e. to the chromosome, and the training sample serves to calculate the
learning error  of the neural network with specific values of the weighting coefficients   .</p>
        <p>Thus, the procedure for optimizing the learning process of neural network using genetic algorithm
is also iterative and includes the stages of synthesis of new chromosomes and their selection into the
new population.</p>
        <p>The scheme of the genetic algorithm of the learning neural network procedure is shown in Fig. 1.</p>
        <p>START</p>
      </sec>
      <sec id="sec-3-2">
        <title>The selection of the initial population of chromosomes</title>
      </sec>
      <sec id="sec-3-3">
        <title>The calculation of the value of the weighting coefficients</title>
      </sec>
      <sec id="sec-3-4">
        <title>The creation of a new population by applying genetic operators</title>
      </sec>
      <sec id="sec-3-5">
        <title>The assessment of the membership function (fitness of chromosomes in the population)</title>
        <p>No</p>
      </sec>
      <sec id="sec-3-6">
        <title>Is the condition ended?</title>
        <p>Yes</p>
      </sec>
      <sec id="sec-3-7">
        <title>The «best» combination of weighting factors</title>
        <p>END</p>
        <p>The process continues until an optimal solution or a given number of generations is obtained. At the
same time, each subsequent population must be better than the previous one. The solution to this
problem is a chromosome with the minimum value of the membership function, that determines the
optimal vector of weighting coefficients   , while the learning error  becomes less than the given
value  . If the optimal solution isn’t found, and the given number of generations is reached, then the
learning procedure ends with the selection of an elite chromosome in one or more generations.</p>
        <p>Depending on the type of genetic operators and selection schemes used, different genetic algorithms
can be constructed, each of them will be effective in terms of convergence speed and the best approach
to the extremum when solving real problems.</p>
        <p>When solving the problem of forecasting the time series of the temperature of the natural
environment, an appropriate neural networks were synthesized, where the input and output values are
the temperature values.</p>
        <p>As a result of solving the optimization problem by the gradient method of the shortest descent, the
best NNs were selected: the radial basis function (errors: training - 2.617 0С, control - 2.617 0С, test
2.06 0С), the linear with two neurons in the input layer (errors: training - 0.103 0С, control - 0.086 0С,
test - 0.097 0С), the linear with three neurons in the input layer (errors: training - 0.103 0С, control
0.086 0С, test - 0.096 0С), the multilayer perceptron with five neurons in the hidden layer (errors:
training - 0.077 0С, control - 0.068 0С, test - 0.074 0С), the multilayer perceptron with two neurons in
the hidden layer (errors: training - 0.073 0С, control - 0.065 0С, test - 0.07 0С).</p>
        <p>To implement the learning algorithm of the neural network with the help of genetic algorithms, it
will use the network with the smallest errors - a multilayer perceptron with two neurons in the hidden
layer (Fig. 2).
neurons in the hidden layer (MLP 2)</p>
        <p>The parameters of the solved problem are the weighting coefficients  , that is, the problem will
have 17 parameters, and the set of these parameters determines the point of the search space, and
accordingly, the possible solution.</p>
        <p>We assume that the solution search area  is a finite set of solutions, in that each admissible solution</p>
        <p>∈  is an  -dimensional vector  ̅ =  1,  2, … ,   ), where  = 17. Suppose that each component
  ,  = 1,  of the vector 
∈</p>
        <p>
          can take values from 0 to 1with a step of 0.004, then the
value of the weighting coefficient   is encoded using a non-negative integer   ∈ [0,   ], where  =
0,250   is the number of possible discrete values of the  -th variable in the search area  [
          <xref ref-type="bibr" rid="ref5 ref6 ref7">5-7</xref>
          ]. To
minimize the learning error of the neural network based on the genetic algorithm, we will match each
variant of the vector of weighting coefficients with a chromosome, presented in the form of a bit string.
        </p>
        <p>That is, let's match each vector   with a vector   , for the representation of that in the binary code,
it’s necessary to determine the maximum number of binary symbols  , that is sufficient to represent
any value   from the range of its permissible values   ∈ [0,   ]. The value 
must meet the
requirement  ≤ 2 , where  = 251 is the number of possible discrete values   of the variable, then 
= 8 (Fig.3).</p>
      </sec>
      <sec id="sec-3-8">
        <title>Where   - the binary symbol 0 or 1 (</title>
        <p>=1,  ),; ℎ(  ) - the symbolic record of a variable   . To
present an admissible solution  ∈  , it’s necessary to combine the symbol records of code sequences
describing all  components of the vector ̅</p>
        <p>= ( 1,  2, . . ,   ), where  = 17, in the form of a linear
sequence of bit strings from binary symbols (Fig. 4).</p>
        <p>Thus, the symbolic model of the solution vector of the given problem  ∈  , can be represented in
the form of a bit string, that is used to specify a set of admissible solutions   , belonging to the search
domain  .</p>
        <p>A binary combination ℎ(  ) is taken as a gene, that determines the fixed value of the parameter  
in the binary code. And the smallest indivisible unit amenable to evolution is a person  1 ( - th e number
of the person in the population,  - the moment in time of the evolutionary process), characterized by
 -genes, each of that is responsible for the corresponding variable (Fig. 5).</p>
        <p>The chromosome that has specific values of alleles in its loci is called a genotype, that contains all
the hereditary genetic information of a person  1. The finite set of all admissible genotypes is the</p>
        <p>The assessment of the fitness of chromosomes in the population will be determined by calculating
the membership function (fitness function) for each chromosome of this population. In our case, this
assessment is performed using the membership function, that determines the difference between the
calculated and real output value with the same input action and represents the numerical value of the
function calculated for an admissible solution to the problem  ∈  −  ( 1) = (
−  )2, and the
smaller the value of the membership function, the better the quality of the chromosome. The fitness
function always takes a non-negative value, in addition, to solve the optimization problem, this function
needs to be minimized.</p>
        <p>A collection of individuals (  , … ,   )</p>
        <p>forms a population   , where  - the size of the
population, and  = 0,1, … ,  , where  determines the period of its evolution. The goal of population
evolution is to increase the average value of the membership function of the population as a whole:</p>
        <p>The effectiveness of the genetic algorithm, the quality of the obtained solution and the evolution as
a whole largely depends on the structure and quality of the initial population.</p>
        <p>Thus, the optimization procedure using the genetic algorithm is iterative and includes two stages:
- the synthesis of new chromosomes (crossing and mutation);
- the selection of chromosomes in a new population.</p>
        <p>The process continues until the optimal solution and the given number of generations are obtained.
At the same time, the goal of population evolution (2) is taken into account, that is, each subsequent
population should be better than the previous one. The solution of the problem corresponds to the
chromosome with the minimum value of the membership function, that determines the optimal vector
of weighting coefficients   , while the learning error (1) is less than the specified value  
. If the stop
of the algorithm cannot be fulfilled according to the condition, then the procedure is completed
according to the option and with the selection of the best chromosome in one or more generations.</p>
        <p>The optimal weighting coefficients of a multilayer perceptron type neural network with two neurons
in the hidden layer (Fig. 1) were determined using a genetic algorithm with the following parameters:
the number of chromosomes in the population is 10; the number of populations - from 20 to 100;
crossover operator - one-point crossover; percentage of gene mutation - 0.001; the selection is elite.
Figure 6 shows the probability of finding extrema of the function from a given number of populations.</p>
        <p>
          The obtained results show that the speed of approaching the extremum is high and increases with
the increase in the number of populations [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref9">9-13</xref>
          ]. It’s also possible to draw a conclusion about the low
percentage of experiments in that the exact value of the minimum was obtained, depending on the total
number of runs of the genetic algorithm, that is, on the number of populations [
          <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref14 ref15 ref9">9-15</xref>
          ].
(3)
        </p>
        <p>
          Analyzing the projections of time series of temperature (Fig. 7) [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], it should be noted that NN
relatively correctly predicted its decrease, increase and stabilization.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>1. Experiments on learning neural networks showed that known methods of local and global
optimization (gradient, stochastic, Newton, Hessian, etc.) require a significant number of learning steps,
that are sensitive to the accuracy of calculations, require a significant number of additional variables.
In most cases, they make it possible to find a local rather than a global extremum. That is why the search
and development of new methods of learning neural networks is an urgent task.</p>
      <p>2. In order to increase the reliability of decisions made on the basis of a neural network, it’s
necessary to investigate alternative optimization algorithms that allow finding the global extremum.
Therefore, there is a need to use such approaches, that wouldn’t have the mentioned disadvantages. The
genetic optimization algorithms are the most promising in this regard.</p>
      <p>3. The scientific and methodical foundations of learning neural networks are developed in the paper.
In contrast to traditional learning methods, the concept of learning a neural network using a genetic
algorithm is defined, that is the most effective method of optimizing weighting coefficients that
minimize the value of the error of the network.</p>
      <p>4. The study of the effectiveness of the application of the genetic algorithm to optimize the
functioning and deep learning of neural networks in operation established the perspective of such an
approach for biotechnical complexes that are exposed to external disturbances (external temperature).</p>
    </sec>
    <sec id="sec-5">
      <title>5. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Korchemny</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Lysenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. V.</given-names>
            <surname>Chapny</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Shtepa</surname>
          </string-name>
          , Neural Networks, Kyiv, Agrar Media Group,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M. O.</given-names>
            <surname>Korchemny</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Lysenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. V.</given-names>
            <surname>Chapny</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Shtepa</surname>
          </string-name>
          ,
          <source>Neural Networks: Theory and Practice</source>
          , 4th ed.,
          <source>Kyiv</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>I. Bratko</surname>
          </string-name>
          ,
          <article-title>Algorithms of artificial intelligence in the PROLOG language</article-title>
          , Moscow, Williams,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Igel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kreutz</surname>
          </string-name>
          ,
          <article-title>Operator adaptation in evolutionary computation and its application to structure optimization of neural networks</article-title>
          , Neurocomputing, No.
          <volume>55</volume>
          (
          <issue>1-2</issue>
          ),
          <year>2003</year>
          , pp.
          <fpage>347</fpage>
          -
          <lpage>361</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Metzen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hutter</surname>
          </string-name>
          ,
          <article-title>Efficient multi-objective neural architecture search via lamarckian evolution</article-title>
          ,
          <source>ArXiv</source>
          preprint arXiv:
          <year>1804</year>
          .09081,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Elsken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Metzen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hutter</surname>
          </string-name>
          ,
          <article-title>The search for neural architecture: an overview</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          , No.
          <volume>20</volume>
          (
          <issue>1</issue>
          ),
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Ridkokasha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Golder</surname>
          </string-name>
          ,
          <source>Basics of Artificial Intelligence Systems</source>
          , Cherkasy,
          <string-name>
            <surname>ECHOPLUS</surname>
          </string-name>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V. V.</given-names>
            <surname>Kruglov</surname>
          </string-name>
          ,
          <article-title>Artificial neural networks</article-title>
          .
          <source>Theory and practice</source>
          , Moscow, Hotline - Telecom,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Lysenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Zayets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Shtepa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. O.</given-names>
            <surname>Dudnyk</surname>
          </string-name>
          ,
          <source>Neural Network Forecasting of Time Series of Ambient Temperature, Bioresources and nature conservation 3</source>
          .
          <fpage>3</fpage>
          -
          <lpage>4</lpage>
          (
          <year>2011</year>
          )
          <fpage>102</fpage>
          -
          <lpage>108</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Zayets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Shtepa</surname>
          </string-name>
          ,
          <article-title>The Use of A Genetic Algorithm For Solving Optimization Problems In Electrical Engineering</article-title>
          ,
          <source>Scientific Bulletin of the National University of Bioresources and Nature Management of Ukraine 166. 4</source>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Zayets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Shvorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M.</given-names>
            <surname>Shtepa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. O.</given-names>
            <surname>Osypa</surname>
          </string-name>
          ,
          <article-title>The Use of Genetic Algorithms to Calculate the Optimal Settings for the Operation of a Robotic Complex, Collection of Scientific Works of the Military Institute of Taras Shevchenko Kyiv National University 38 (</article-title>
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Pasichnyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Shvorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. A.</given-names>
            <surname>Gunchenko</surname>
          </string-name>
          , I. Sharipova,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Tereshchenko</surname>
          </string-name>
          ,
          <article-title>Methodological Bases of Construction of Dispatchers Intensive Training Simulators of AirTraffic Control</article-title>
          ,
          <source>in: Proceedings of IEEE 6th International Conferenceon Methods and Systems of Navigation and Motion Control</source>
          ,
          <string-name>
            <surname>MSNMC</surname>
          </string-name>
          ,
          <year>2020</year>
          , рр.
          <fpage>122</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Shvorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Lysenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pasichnyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Lukin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Martsyfei</surname>
          </string-name>
          ,
          <article-title>The method of determining the amount of yield based on the results of remote sensing obtained using UAV on the example of wheat</article-title>
          ,
          <source>in: Proceedings of 15th International Conference on Advanced Trends in Radioelectronics</source>
          , Telecommunications and Computer Engineering, TCSET
          <year>2020</year>
          ,
          <year>2020</year>
          , рр.
          <fpage>245</fpage>
          -
          <lpage>248</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y. A.</given-names>
            <surname>Gunchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Shvorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Rudnichenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. D.</given-names>
            <surname>Boyko</surname>
          </string-name>
          ,
          <article-title>Methodical complex of accelerated training for operators of unmanned aerial vehicles</article-title>
          ,
          <source>in: Proceedings of IEEE 4th International Conference Methods and Systems of Navigation and Motion Control</source>
          ,
          <string-name>
            <surname>MSNMC</surname>
          </string-name>
          <year>2016</year>
          ,
          <year>2016</year>
          , рр.
          <fpage>130</fpage>
          -
          <lpage>133</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Voronin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Yasinsk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Shvorov</surname>
          </string-name>
          .
          <article-title>Synthesis of compromise-optimal trajectories of mobile objects in conflict environment</article-title>
          ,
          <source>Journal of Automation and Information Sciences</source>
          <volume>34</volume>
          (
          <issue>2</issue>
          ) (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>