<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Forecast Method Based on the Time-Delay Mean Field Boltzmann Machine</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oleg Grygor</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eugene Fedorov</string-name>
          <email>fedorovee75@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olga Nechyporenko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cherkasy State Technological University</institution>
          ,
          <addr-line>Shevchenko blvd., 460, Cherkasy, 18006</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>114</fpage>
      <lpage>124</lpage>
      <abstract>
        <p>The problem of insufficient forecast efficiency for supply chain management is solved. A neural network forecast model based on the Time-Delay Mean Field Boltzmann Machine with time delays in the visible layer has been created. In the process of adjusting the structure of the developed model, the length of the hidden layer was determined, and the calculation of the model parameters was carried out on the basis of the parallel computing platform CUDA. Improving forecast accuracy and speed of calculations makes it possible to improve the quality of the forecast, resulting in increased supply flexibility and reduced logistics costs. A software toolkit based on the Matlab package has been developed, which makes it possible to implement the proposed method. The developed software tools are used to solve the problem of supply chains forecasting. Forecast efficiency, supply chain management problem, neural network forecast model, Time-Delay Mean Field Boltzmann Machine, positive and negative learning phase Supply chains are complex adaptive systems characterized by structural and dynamic complexity, operating under a large number of random factors. Supply chain management is based on forecasting the demand for the final product. This requires efficient and intelligent supply chain planning. Planning challenges include, but are not limited to, fragmented data across the organization and difficulty in forecasting deliveries. This leads to low accuracy of sales plans, a large volume of illiquid products and, as a result, to losses for the company.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>




logical forecasting methods based on classification and regression trees [5];
forecasting methods based on exponential smoothing [6];
regression and autoregressive forecasting methods [7];
neural network forecasting methods [8, 9];
structural forecasting methods based on Markov chains [10].</p>
      <p>Using artificial neural networks for forecasting provides the following advantages:
assumptions about the distribution of input features are not required;
analysis of systems with a high degree of nonlinearity is possible;</p>
      <p>2022 Copyright for this paper by its authors.
 high adaptability;
 rapid model development;
 the relationships between the input features are investigated on ready-made models;
 a priori information about the input features may be missing;
 the original data may be incomplete or contain noise, as well as highly correlated;
 analysis of systems with a large number of input features is possible;
 analysis of systems with heterogeneous characteristics is possible;
 a complete enumeration of all possible models is not required.</p>
      <p>Therefore, a neural network forecasting method will be used in the article.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Formal problem statement</title>
      <p>Let the training set S  {(x , d )},  1, P be given for the forecast.</p>
      <p>Then the problem of improving the forecast accuracy for the Time-Delay Mean Field Boltzmann
Machine (TDMFBM) model is g(x,W ) , where x – is the input vector, W – is the vector of
parameters, represented as the problem of finding such a vector of parameters W * for this model, that
1 P
satisfies criterion F  (g(x ,W * )  d )2  min .</p>
      <p>P  1</p>
      <p>The aim of the work is to create an effective forecasting method for supply chain management. To
achieve this goal, the following tasks were set and solved:
 analyze existing neural network forecasting methods;
 create a neural network forecast model based on the mean field Boltzmann machine;
 choose a criterion for evaluating the effectiveness of a neural network forecast model based
on the mean field Boltzmann machine;
 develop a method for identifying the parameters values of the neural network forecast model
based on the mean field Boltzmann machine;
 perform numerical studies.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Literature review</title>
      <p>The number of publications demonstrates significant attention to the application of advanced
analytics, methods and modern computer tools of artificial intelligence in the field of supply chain
management, but also leaves unresolved and insufficiently studied a number of problems regarding
the development and synthesis of methods and models of artificial intelligence.</p>
      <p>The most commonly used forecast neural networks are:
1. Gateway neural networks:
 long short-term memory (LSTM) [11, 12];
 bidirectional long short-term memory (BLSTM) [13, 14];
 gateway recurrent unit (GRU) [15-17];
 bidirectional gateway recurrent unit (BGRU) [18, 19].
2. Reservoir neural networks:
 echo state network (ESN) [20, 21];
 liquid state machine (LSM) [22-24].</p>
      <p>Table 1 shows the comparative characteristics of forecasting neural networks.</p>
      <p>
        The learning rate is directly proportional to the computational complexity. For LSTM
computational complexity ~PN(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )(5M(0)+ 3M(0)S+24S+S2), for BLSTM computational complexity
~2PN(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )(5M(0)+ 3M(0)S+24S+S2)), for GRU computational complexity ~PN(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )6(M(0)+N(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )), for BGRU
computational complexity ~PN(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )6(M(0)+N(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )), for ESN computational complexity
~PN(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )(M(0)+N(
        <xref ref-type="bibr" rid="ref2">1</xref>
        ))+(max{P,M(0)+N(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )})2, for LSM computational complexity ~PN(r)(N(r)M(0)+ N(
        <xref ref-type="bibr" rid="ref2">1</xref>
        )),
      </p>
      <sec id="sec-3-1">
        <title>Network</title>
      </sec>
      <sec id="sec-3-2">
        <title>Criterion</title>
        <p>The presence of feedback
Low probability of getting
into a local extremum</p>
        <p>High learning speed</p>
        <p>
          Possibility of batch training
4. Materials and methods
+
where M(0) – the number of unit delays for the input layer, S – the number of cell, N(
          <xref ref-type="bibr" rid="ref2">1</xref>
          ) – the number of
neurons in the first layer, N(r) – the number of neurons in the reservoir layer, P – training set
cardinality, N(
          <xref ref-type="bibr" rid="ref2">1</xref>
          )&lt;&lt;P, N(r)&lt;&lt;P. According to Table 1, none of the networks meets all the criteria.
        </p>
        <p>Thereby, the creation of a neural network that will eliminate the specified drawback is relevant.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Block diagram of a neural network forecast model</title>
      <p>In contrast to the traditional mean field Boltzmann machine (MFBM) [25, 26], time delays are
used for the neurons of the visible layer, and the neurons of the visible layer are not connected with
each other. TDMFBM type 1 has time delays in the input layer. TDMFBM type 2 has time delays in
the input and output layers.
4.2.</p>
    </sec>
    <sec id="sec-5">
      <title>Forecasting model based on TDMFBM type 1</title>
      <sec id="sec-5-1">
        <title>Positive phase (steps 1-3)</title>
        <p>2. Initialization of the state of visible input, hidden and output neurons</p>
        <p>xin(  t)  0 , t 1, M in .</p>
        <p>xin( )  xin , xh ( )  0 , xout ( )  0 .
3. Computation of the state of hidden neurons ( j 1, N h ) at time </p>
        <p>M in Nin Nout Nh
s hj ( )  bhj    wtiinjh xiin (  t)   wiojuth xiout ( )   wihjh xih ( ) ,
t0 i1 i1
x hj ( ) 
i1
1
1  exp s hj ( )
,
where wtiinjh – synaptic weights between the visible input layer (taking into account unit delays) and
wouth – synaptic weights between the visible output and the hidden layer,</p>
        <p>ij
whh – synaptic weights inside the hidden layers,</p>
        <p>ij
bhj – bias of neurons of the hidden layer,
M in – the number of unit delays for the visible input layer,
N in – the number of neurons in the input layer,
N h – the number of neurons in the hidden layer,
N out – the number of neurons in the output layer.
where bojut – bias of neurons of the visible output layer.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Forecasting model based on TDMFBM type 2</title>
      <p>2. Initialization of the state of visible input and output neurons</p>
      <p>xin ( )  xin , xh ( )  0 , xout ( )  0 .
3. Computation of the state of hidden neurons ( j 1, N h ) at time 
sh ( )  bhj  
j
 wtiinjh xiin (  t)  
 wtoiujth xout (  t)   whh xh ( ) ,</p>
      <p>i ij i
1  exp sh ( )
j
where wtiinjh – synaptic weights between the visible input layer (taking into account unit delays) and
wtoiujth – synaptic weights between the visible output layer (taking into account unit delays) and the
whh – synaptic weights inside the hidden layers,</p>
      <p>ij
bh – bias of neurons of the hidden layer,</p>
      <p>j
M in – the number of unit delays for the visible input layer,
M out – the number of unit delays for the visible output layer,
N in – the number of neurons in the input layer,
N h – the number of neurons in the hidden layer,
N out – the number of neurons in the output layer.</p>
      <p>4. Computation of the state of visible output neurons ( j 1, N out ) at time 
sout ( )  bout  
j j
 wtiinjout xiin (  t)   w0oiujth xh ( ) ,</p>
      <p>i</p>
      <sec id="sec-6-1">
        <title>Negative phase (step 4)</title>
        <p>,
where bout – bias of neurons of the visible output layer.</p>
        <p>j</p>
        <p>The result is vector (x1out ( ),..., xNouotut ( )) .
4.4.
model</p>
        <p>Criterion for evaluating the effectiveness of a neural network forecast
In this work, for training the TDMFBM model, a model adequacy criterion was chosen, which
means the choice of such values of parameters W  {winh , wtoiujth , wtiinjin , wtoiujtout} , which deliver a
tij
minimum of the mean squared error (the difference between the model output and the desired output):</p>
        <p>
          P N out
(
          <xref ref-type="bibr" rid="ref2">1</xref>
          )
        </p>
        <p>
          The training of the TDMFBM model is subject to criterion (
          <xref ref-type="bibr" rid="ref2">1</xref>
          ).
4.5.
        </p>
        <p>Method for determining the parameter values of the forecasting model
based on TDMFBM type 1</p>
        <p>
          1. Number of training iteration n  1 , initialization by means of uniform distribution on the
interval (
          <xref ref-type="bibr" rid="ref2">0,1</xref>
          ) or [-0.5, 0.5] bias bout (n) , i 1, N out , bh (n) , j 1, N h , and weights
i j
wtiinjh (n) ,
t  0, M in , i 1, N in , j 1, N h , wtiinjout (n) , t  0, M in , i 1, N in , j 1, N out , wouth (n) , i 1, N out ,
ij
j 1, N h , whh (n) , i 1, N h , j 1, N h , wtiinih (n)  0 , wtiinjout (n)  0 , wouth (n)  0 ,
ij ii
wtiinjh (n)  wtijnih (n) , wtiinjout (n)  wtijniout (n) , wouth (n)  wojiuth (n) , whh (n)  whjih (n) , where M in is
ij ij
the number of unit delays for visible input neurons.
        </p>
        <p>
          2. A training set {(xin, xout ) | xin  (
          <xref ref-type="bibr" rid="ref2">0,1</xref>
          )N in , xout  (
          <xref ref-type="bibr" rid="ref2">0,1</xref>
          )N out } ,
 1, P
is set, where xin

–  th
training vector of states of visible input neurons, xout –  th training vector of states of visible output

neurons, P – is the power of the training set.
w
ii
hh
(n)  0
,
Initialization of the state of the visible input neurons of the time delay
        </p>
        <p>Positive phase (steps 3-6)
4. Initialization of the state of visible input and output neurons
5. Computation of the state of hidden neurons ( j 1, N h ) at time 
6. Preservation of the state of neurons in the positive phase at time  , i.e. x1in ( )  xin ( ) ,
x1out ( )  xout ( ) , x1h ( )  xh ( ) . If   P , then    1, go to 4.</p>
        <p>N out
i1</p>
        <p>1
1  exp sh ( )
j
Initialization of the state of the visible input neurons of the time delay
8. Initialization of the state of visible input and output, hidden neurons</p>
        <p>xin(  t)  0 , x2in (  t)  0 , t 1, M in .</p>
        <p>xin( )  x1in ( ) , xout ( )  x1out ( ) , xh ( )  x1h ( ) .
9. Computation of the state of visible output neurons ( j 1, N out ) at time 
sojut() bojut(n)wtiinjout(n)xiin( t)wouth(n)xih(),</p>
        <p>ij
10. Computation of the state of hidden neurons ( j1,Nh ) at time 
shj() bhj(n)wtiinjh(n)xiin( t) wouth(n)xiout()wihjh(n)xih(),
ij</p>
        <p>Nh
i1
Min Nin
biout(n) biout(n) 1 Px1iout() x2iout(), i1,Nout ,</p>
        <p>1 P 
 P 1 P 1 </p>
        <p>
bhj(n)  bhj(n) 1 Px1hj() x2hj() , j1,Nh ,</p>
        <p>1 P 
 P 1 P 1 </p>
        <p>
wtiinjh(n)  wtiinjh(n)(tij tij), t0,M in , i1,Nin , j1,Nh ,</p>
        <p>tij  x1iin( t)x1hj(), tij  x2iin( t)x2hj(),
wtiinjout(n)  wtiinjout(n)(tij tij) , t0,M in , i1,Nin , j1,Nout ,
tij  1 Px1iin( t)x1ojut(), tij  1 Px2iin( t)x2ojut() ,
wouth(n)  wiojuth(n)(ij ij) , i1,Nout , j1,Nh ,</p>
        <p>ij
ij  1 Px1iout()x1hj(), ij  1 Px2iout()x2hj(),</p>
        <p>wihjh(n)  wihjh(n)(ij ij), i, j1,Nh ,
ij  1 P</p>
        <p>x1ih( t)x1hj(), ij  1 Px2ih( t)x2hj().</p>
        <p>1 P Nout 
13. If | x1iout() x2iout()|  , then n  n 1, go to 2.</p>
        <p> 
P 1 i1 
1 P
P 1
P 1
P 1
P 1
11. Saving the state of neurons in the negative phase at time  , i.e. x2in()  xin(),
x2out()  xout() , x2h()  xh() . If   P , then   1, go to 8.</p>
        <p>12. Adjustment of synaptic weights and bias based on Boltzmann's rule
4.6. Method for determining the parameter values of the forecasting model
based on TDMFBM type 2</p>
        <p>
          1. Number of training iteration n 1, initialization by means of uniform distribution on the
interval (
          <xref ref-type="bibr" rid="ref2">0,1</xref>
          ) or [-0.5, 0.5] bias biout(n), i1,Nout , bhj(n) , j1,Nh , and weights wtiinjh(n),
t0,M in , i1,Nin , j1,Nh , wtiinjout(n) , t0,M in , i1,Nin , j1,Nout , wtoiujth(n), t0,Mout ,
i1,Nout , j1,Nh , wihjh(n), i1,Nh , j1,Nh , wtiinih(n)  0, wtiinjout(n) 0, wtoiuith(n)  0,
wihih (n)  0 , wtiinjh (n)  wtijnih (n) , wtiinjout (n)  wtijniout (n) , wtoiujth (n)  wtji
outh (n) , wihjh (n)  whjih (n) ,
where M in is the number of unit delays for visible input neurons, M out is the number of single
delays for visible output neurons.
vector of states of visible input neurons, xout –  th training vector of states of visible output neurons,

P is the power of the training set.
        </p>
      </sec>
      <sec id="sec-6-2">
        <title>Positive phase (steps 3-6)</title>
        <p>4. Initialization of the state of visible input and output neurons</p>
        <p>xin ( )  xin , xh ( )  0 , xout ( )  xout .
5. Computation of the state of hidden neurons ( j 1, N h ) at time </p>
        <p>M in N in M outN out N h
s hj ( )  bhj (n)   wtiinjh (n)xiin (  t)    wtoiujth (n)xiout (  t)   wihjh (n)xih ( ) ,
t 0 i1 t 0 i1 i1</p>
        <p>1
x hj ( ) </p>
        <p>1  exp s hj ( )
6. Saving the state of neurons in a positive phase at time
 , i.e. x1in ( )  xin ( ) ,
x1out ( )  xout ( ) , x1h ( )  xh ( ) . If   P , then    1, go to 4.</p>
      </sec>
      <sec id="sec-6-3">
        <title>Negative phase (steps 7-11)</title>
        <p>9. Computation of the state of visible output neurons ( j 1, N out ) at time </p>
        <p>M in N in N h
s ojut ( )  bojut (n)   wtiinjout (n)xiin (  t)   w0oiujth (n)xih ( ) ,
t 0 i1 i1
xoj ut ( ) 
.
11. Saving the state of neurons in the negative phase at time  , i.e. x2in ( )  xin ( ) ,
x2out ( )  xout ( ) , x2h ( )  xh ( ) . If   P , then    1, go to 8.
12. Adjustment of synaptic weights and bias based on Boltzmann's rule
biout (n)  biout (n)   1 P x1iout ( )   x2iout ( ) , i 1, N out ,</p>
        <p>1 P 
 P  1 P  1 
tij 
 tij 
ij 
13. If
bhj (n)  bhj (n)   1 P x1hj ( ) 
 P  1
1 P </p>
        <p> x2hj ( ) , j 1, N h ,</p>
        <p>P  1 
wtiinjh (n)  wtiinjh (n)  (tij  tij) , t  0, M in , i 1, N in , j 1, N h ,
 tij 
1 P</p>
        <p> x1iin (  t)x1hj ( ) ,  tij 
P  1
1 P</p>
        <p> x2iin (  t)x2hj ( ) ,</p>
        <p>P  1
wtiinjout (n)  wtiinjout (n)  (tij  tij) , t  0, M in , i 1, N in , j 1, N out ,
1 P</p>
        <p> x1iin(  t)x1ojut ( ) , tij 
P  1
1 P</p>
        <p> x2iin(  t)x2ojut ( ) ,</p>
        <p>P  1
wtoijuth (n)  wtoijuth (n)  (tij  tij) , i 1, N out , j 1, N h ,
1 P</p>
        <p> x1iout (  t)x1hj ( ) ,
P  1
tij 
1 P</p>
        <p> x2iout (  t)x2hj ( ) ,</p>
        <p>P  1
wihjh (n)  wihjh (n)  (ij  ij ) , i, j 1, N h ,
1 P</p>
        <p> x1ih (  t)x1hj ( ) ,
P  1
 ij 
1 P</p>
        <p> x2ih (  t)x2hj ( ) .</p>
        <p>P  1
1 P  N out </p>
        <p> | x1iout ( )  x2iout ( ) |   , then n  n  1 , go to 2.</p>
        <p>P  1 i1 </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Experiments and results</title>
      <p>To determine the structure of the forecasting model based on TDMFBM with 16 input neurons, i.e.
determining the amount of hidden neurons, a number of experiments were carried out, the results of
which are presented in Figure 3. As input data to determine the values of the parameters of the neural
network forecasting model, a sample of values of the economic activities of the logistics company
«Ekol Ukraine» was used. The criterion for choosing the structure of the neural network model was
the minimum mean squared error (MSE) of forecasting. The dataset capacity for the "cost of
transportation" indicator was 1000. The dataset was divided into three parts - training data (60%), test
data (20%), test data (20%). The training took place over 100 epochs. The change in the MSE value
chosen as the loss function depended on the training epoch number and occurred exponentially. The
common parameters for all neural networks were the number of neurons in the hidden layer. As can
be seen from Figure 3, with an increase in the amount of hidden neurons the error value decreases.
For the forecast, it is sufficient to use 32 hidden neurons, since with a further increase in their amount
the change in the error value is insignificant. The work investigated forecasting neural networks
according to the criterion of the minimum mean squared error (MSE) of the forecast (Table 2).</p>
      <p>According to Table 2, TDMFBM type 2 has the highest forecast accuracy. TDMFBM type 1 can
train in burst mode unlike other networks. Thus, type 1 TDMFBM has the fastest learning rate.
1,8
1,6
1,4
1,2
SE 1
M
0,8
0,6
0,4
0,2
0
4
8
12
16
20
24
28
32
36
40
44
48</p>
      <p>52
6. Conclusions</p>
      <p>1. To solve the problem of improving forecast quality for effective supply chain management,
forecast methods were analyzed. According to the studies carried out, neural networks are currently
the most effective forecasting tool.</p>
      <p>2. In order to improve the forecast efficiency, the MFBM neural network was selected, modified
(by introducing time delays in the visible layer), and the structure of its model was identified in the
process of numerical study. The conducted study showed that with 32 neurons in the hidden layer, the
value of the root mean square error changes little, and the proposed network performs the forecast
with a minimum error.</p>
      <p>3. A method for calculating the values of the parameters of the created neural network forecast
model was proposed. This ensures high accuracy and speed of the forecast.</p>
      <p>4. The developed approach can be used for forecasting in various intelligent computer systems of
general and special purpose.</p>
    </sec>
    <sec id="sec-8">
      <title>7. References</title>
      <p>Computing, volume 754, Springer, Cham, 2019, pp. 735–745. doi:
10.1007/978-3-319-910086_72.
[7] R. T. Baillie, G. Kapetanios, F. Papailias, Modified information criteria and selection of long
memory time series models, in: Computational Statistics and Data Analysis, volume 76, 2014,
pp. 116–131. doi: 10.1016/j.csda.2013.04.012.
[8] L. Lyubchyk, E. Bodyansky, A. Rivtis, Adaptive harmonic components detection and forecasting
in wave non-periodic time series using neural networks, in: Proceedings of the ISCDMCI'2002,
Evpatoria, 2002, pp. 433-435.
[9] S. N. Sivanandam, S. Sumathi, S. N. Deepa, Introduction to Neural Networks using Matlab 6.0,</p>
      <p>The McGraw-Hill Comp., Inc., New Delhi, 2006.
[10] A. Wilinski, Time series modelling and forecasting based on a Markov chain with changing
transition matrices, in: Expert Systems with Applications, volume 133, 2019, pp. 163–172. doi:
10.1016/j.eswa.2019.04.067.
[11] M. Sundermeyer, R. Schluter, H. Ney, LSTM neural networks for language modeling, in:
Thirteenth Annual Conference of the International Speech Communication Association, 2012,
pp. 194-197. doi: 10.1.1.248.4448.
[12] P. Potash, A. Romanov, A. Rumshisky, Ghostwriter: using an LSTM for automatic rap lyric
generation, in: Proceedings of the 2015 Conference on Empirical Methods in Natural Language
Processing, 2015, pp. 1919– 1924. doi:10.18653/v1/D15-1221.
[13] E. Kiperwasser, Y. Goldberg, Simple and Accurate Dependency Parsing Using Bidirectional
LSTM Feature Representations, Transactions of the Association for Computational Linguistics 4
(2016) 313–327. doi: 10.1162/tacl_a_00101.
[14] A. Graves, J. Schmidhuber, Framewise phoneme classification with bidirectional LSTM and
other neural network architectures, Neural Networks 18 (2005) 602–610,
doi:10.1016/j.neunet.2005.06.042.
[15] J. Chung, C. Gulcehre, K. Cho, Y. Bengio, Empirical evaluation of gated recurrent neural
networks on sequence modeling, arXiv preprint arXiv:1412.3555, 2014.
[16] R. Dey, F. M. Salem, Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks,
arXiv:1701.05923, 2017. – URL: https://arxiv.org/ftp/arxiv/papers/1701/1701.05923.pdf.
[17] E. Fedorov, T. Utkina, О. Nechyporenko, Forecast method for natural language constructions
based on a modified gated recursive block, in: CEUR Workshop Proceedings, vol. 2604, 2020,
pp. 199-214.
[18] Q. Lu,; Z. Zhu, F. Xu, D. Zhang, W. Wu, Q. Guo, Bi-GRU Sentiment Classification for Chinese
Based on Grammar Rules and BERT, International Journal of Computational Intelligence
Systems 13 (2020) 538-548. doi: 10.2991/ijcis.d.200423.001.
[19] T. Fan, J. Zhu, Y. Cheng, Q. Li, D. Xue, R. Munnoch, A New Direct Heart Sound Segmentation
Approach using Bi-directional GRU, in: Proceedings of the 2018 24th International Conference
on Automation and Computing, 2018, pp. 1–5. doi: 10.23919/IConAC.2018.8749010.
[20] H. Jaeger, Tutorial on Training Recurrent Neural Networks, Covering BPPT, RTRL, EKF and
the Echo State Network Approach, GMD Report 159, German National Research Center for
Information Technology, 2002.
[21] H. Jaeger, M. Lukosevicius, D. Popovici, U. Siewert, Optimization and applications of echo state
networks with leakyintegrator neurons, in: Neural Networks volume 20, 2007, pp. 335–352.
doi:10.1016/j.neunet.2007.04.016.
[22] T. Natshlager, W. Maas, H. Markram, The liquid computer: A novel strategy for real-time
computing on time series, in: Special Issue on Foundations of Information Processing of
Telematik, 2002, pp. 39–43.
[23] W. Maass, Liquid state machines: motivation, theory, and applications, in: Computability in
context: computation and logic in the real world, 2011, pp. 275–296.
doi: 10.1142/9781848162778_0008.
[24] T. Neskorodieva, E. Fedorov, I. Izonin, Forecast method for audit data analysis by modified
liquid state machine, in: CEUR Workshop Proceedings, 2020, volume 2631, pp. 145-158.
[25] G. E. Hinton, A Practical Guide to Training Restricted Boltzmann Machines, Technical Report</p>
      <p>UTML TR 2010–003, University of Toronto, 2010.
[26] A. Fischer, C. Igel, Training Restricted Boltzmann Machines: An Introduction, Pattern
Recognition 47 (2014) 25-39. doi: 10.1016/j.patcog.2013.05.025.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <article-title>2. A training set {(xin, xout ) | xin  (0,1)N in , xout  (0,1)N out } is set</article-title>
          ,  1,
          <string-name>
            <surname>P</surname>
          </string-name>
          , where xin - 
          <source>th training</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Cox</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.G.</given-names>
            <surname>Schleher</surname>
          </string-name>
          , Theory of Constraints Handbook, New York, NY,
          <string-name>
            <surname>McGraw-Hill</surname>
          </string-name>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Smerichevska</surname>
          </string-name>
          et al,
          <article-title>Cluster Policy of Innovative Development of the National Economy: Integration and Infrastructure Aspects: monograph</article-title>
          , S. Smerichevska (Eds.),
          <source>Wydawnictwo naukowe WSPIA</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Goldratt</surname>
          </string-name>
          ,
          <article-title>My saga to improve production, Selected Readings in Constraints Management, Falls Church, VA: APICS (</article-title>
          <year>1996</year>
          )
          <fpage>43</fpage>
          -
          <lpage>48</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Goldratt</surname>
          </string-name>
          ,
          <article-title>Production: The TOC Way (Revised Edition) including CD-ROM Simulator and Workbook, Revised edition</article-title>
          , Great Barrington, MA: North River Press,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B.</given-names>
            <surname>Choubin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Zehtabian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Azareh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Rafiei-Sardooi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sajedi-Hosseini</surname>
          </string-name>
          , Ö. Kişi,
          <article-title>Precipitation forecasting using classification and regression trees (CART) model: a comparative study of different approaches</article-title>
          ,
          <source>Environ Earth Sci</source>
          <volume>77</volume>
          ,
          <issue>314</issue>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .1007/s12665-018-7498-z.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>P.</given-names>
            <surname>Bidyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Prosyankina-Zharova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Terentiev</surname>
          </string-name>
          ,
          <article-title>Modelling nonlinear nonsta-tionary processes in macroeconomy and finances</article-title>
          , in: Z.
          <string-name>
            <surname>Hu</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Petoukhov</surname>
          </string-name>
          , I. Dychka, M. He (Eds.), Advances in Computer Science for Engineering and Education.
          <source>Advances in Intelligent Systems and</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>