<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>COLINS-</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Method for Recognizing Linguistic Constructions Based on Stochastic Neural Networks</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Eugene Fedorov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olga Nechyporenko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cherkasy State Technological University</institution>
          ,
          <addr-line>Shevchenko blvd., 460, Cherkasy, 18006</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2022</year>
      </pub-date>
      <volume>6</volume>
      <fpage>12</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>The paper proposes a method for recognizing linguistic constructions based on stochastic neural networks. The novelty of the study lies in the fact that in order to ensure the interaction of software agents representing subjects that operate within supply chains, two models of an artificial neural network were created to recognize natural language structures based on the restricted Boltzmann machine (in contrast to it, the neurons of the hidden layer were interconnected), a criterion for evaluating the effectiveness of training the proposed models was chosen, the parameters of the proposed models were identified based on the contrastive divergence. The proposed models and methods for their parametric identification make it possible to improve the recognition accuracy of natural language constructions. The proposed method for recognizing linguistic structures based on stochastic neural networks can be used in various intelligent systems that use the recognition of natural language structures.</p>
      </abstract>
      <kwd-group>
        <kwd>1 supply chain</kwd>
        <kwd>multi-agent interaction</kwd>
        <kwd>artificial neural network</kwd>
        <kwd>restricted Boltzmann machine</kwd>
        <kwd>contrastive divergence</kwd>
        <kwd>linguistic constructions</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>• difficulty in forming a representative sample;
• high probability for the training and adaptation method hitting a local extremum;
• inaccessibility to human understanding of the knowledge accumulated by the network (it is
impossible to represent the relationship between input and output in the form of rules), since they
are distributed among all elements of the neural network and are presented in the form of its
weight coefficients.</p>
      <p>The following recurrent networks are most often used as neural networks for recognition:
• Elman neural network (ENN) or simple recurrent network (SRN) [15, 16], which is a
recurrent two-layer network and is based on a multilayer perceptron (MLP). The advantages of this
network are a simpler architecture and higher learning rate than in gated and bidirectional
networks. The disadvantage is the insufficient recognition accuracy compared to bidirectional
networks;
• bidirectional recurrent neural network (BRNN) [17, 18], which is a recurrent two-layer
network and is built on the basis of two Elman neural networks. The advantage of this network is a
higher recognition accuracy than in a conventional Elman neural network. The disadvantages are a
higher complexity of determining the architecture, a lower learning rate than in a conventional
Elman neural network;
• long short-term memory (LSTM) [19, 20], which is a recurrent network and is built on the
basis of memory blocks (containing one or more cells) and input, output forgetting gates (FIR
filters). The advantage of this network is a higher recognition accuracy than in a conventional
Elman neural network. The disadvantages are a higher complexity of determining the architecture,
a lower learning rate than in a conventional Elman neural network;
• bidirectional recurrent neural network (BLSTM) [21, 22], which is a recurrent network and is
built on the basis of two LSTM neural networks. The advantage of this network is a higher
recognition accuracy than in a conventional LSTM. The disadvantages are a higher architecture
definition complexity, a lower learning rate than conventional LSTM;
• gated recurrent unit (GRU) [23, 24], which is a recurrent two-layer network and is built on the
basis of hidden blocks and reset and update gates (FIR filters). The advantage of this network is a
higher recognition accuracy than in a conventional Elman neural network. The disadvantages are a
higher complexity of determining the architecture, a lower learning rate than in a conventional
Elman neural network;
• idirectional recurrent neural network (BGRU) [25], which is a recurrent network and is built
on the basis of two GRU neural networks. The advantage of this network is a higher recognition
accuracy than in a conventional GRU. The disadvantages are a higher complexity of architecture
definition, a lower learning rate than in conventional GRU.</p>
      <p>Thus, none of the networks satisfies all the criteria.</p>
      <p>The aim of the work is to develop a method for recognizing natural language constructions. To
achieve the goal, the following tasks were set and solved:
• analyze existing recognition methods;
• propose neural network recognition models;
• choose a criterion for evaluating the effectiveness of neural network recognition models;
• propose methods for determining the values of the parameters of neural network recognition
models;
• conduct numerical study.
2. Block diagram of neural network recognition models
3. Neural network recognition models
3.1. Recognition model based on a unidirectional RBMRHL</p>
      <p>Positive phase (steps 1-2)
1.</p>
      <p>Computation of the state of hidden neurons</p>
      <p>1
 N in N out N h 
1 + exp − bhj − ∑ wiijn−h xiin − ∑ wiojut −h xiout − ∑ wihj −h xih 
 i=1 i=1 i=1 </p>
      <p>1,
x hj = 
0,
 N h 
1 + exp − bout − ∑ wout −h xh 
 j ij i 
 i=1 
1,
xout = 
j
0,</p>
      <sec id="sec-1-1">
        <title>The result is vector (x1out ,..., xNouotut ) .</title>
        <p>3.2.</p>
        <p>Recognition model based on a bidirectional RBMRHL
1.
2.</p>
        <p>xin = x1in , xout = 0
Computation of the state of hidden neurons
1
N out
1
N out
1
P =</p>
        <p>j
P =
j

</p>
        <p>N in</p>
        <p>N in
1 + exp − bhj − ∑ wLiinj−h xiin − ∑ wLoijut −h xiout − ∑ wLhij−h xLhi 
 i=1 i=1 i=1 
1 + exp − bhj − ∑ wRiijn−h xiin − ∑ wRiojut −h xiout − ∑ wRihj −h
 i=1 i=1 i=1
N h</p>
        <p>N h
1,
xRh = 
j
0,</p>
        <p>Computation of the state of visible output neurons</p>
      </sec>
      <sec id="sec-1-2">
        <title>The result is vector (x1out ,..., xNouotut ) .</title>
        <p>models
4. Criteria for evaluating the effectiveness of neural network recognition</p>
        <p>
          In this work, for training a unidirectional and bidirectional RBMRHL model, the model adequacy
criterion was chosen, which means choosing such parameter values W = {wiijn−h , wiojut −h , w
(matching the model output and the desired output):
} respectively, that deliver maximum accuracy
F =
1
The training of the RBMRHL model is subject to the criterion (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ).
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
5. Methods for determining the values of parameters of neural network
recognition models
5.1. Principles for determining the parameters of neural network
recognition models
        </p>
        <p>RBMRHL parameter values are determined based on the CD-1 contrastive divergence method,
which speeds up supervised learning, since instead of stabilizing the state of neurons, it performs only
one step of tuning their state. RBMRHL classification operates in two phases - positive and negative.</p>
        <p>For RBMRHL recognition in the positive phase, visible input and output neurons are fixed, and
RBMRHL functions until hidden neurons are established. In the negative phase, firstly hidden
neurons trained in the positive phase are fixed, and RBMRHL functions until visible input and output
neurons are established, after which visible input and output neurons trained in the negative phase are
fixed, and RBMRHL functions until hidden neurons are established.
5.2. Method for determining the parameter values of a unidirectional
RBMRHL model for recognition based on contrastive divergence
1.</p>
        <p>
          Number of training iteration n = 1 , initialization by means of uniform distribution on the
interval (
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ) or [-0.5, 0.5] offsets (thresholds) biin (n) , i ∈1, N in , biout (n) , i ∈1, N out ,
b hj (n) , j ∈1, N h , and weights win−h (n) , i ∈1, N in , j ∈1, N h , wout−h (n) , i ∈1, N out , j ∈1, N h ,
ij ij
wihj−h (n) , i ∈1, N h , j ∈1, N h , win−h (n) = 0 , wout−h (n) = 0 , wihj−h (n) = 0 , win−h (n) = wijni −h (n) ,
ii ii ij
wout−h (n) = wojiut−h (n) , wihj−h (n) = w hji−h (n) .
        </p>
        <p>ij</p>
        <p>Training set{(xµin , xµout ) | xµin ∈{0,1}Nin , xµout ∈{0,1}N out } , µ ∈1, P is specified, where xµin – µ th
training vector of states of visible input neurons, xout – µ th training vector of states of visible
µ
output neurons, P is the power of the training set.</p>
        <p>Positive phase (steps 3-7)
2.
3.</p>
        <p>x1µin = xµin , x1µout = xµout , µ ∈1, P .
µ =1, x1µh −1 = 0 ,
xin = x1µin , xout = x1µout , xh = x1µh −1 .</p>
        <p>Computation of the state of hidden neurons</p>
        <p>1
Pj =  N in
 − bhj (n) − ∑ win−h (n)xiin −
 i=1 ij
1 + exp N out



</p>
        <p>N h 
 ∑ wout −h (n)xiout − ∑ wihj −h (n)xih </p>
        <p>ij
 i=1 i=1 
x1µh = xh . If µ &lt; P , then µ = µ +1 , go to 5.</p>
        <p>
          1, Pj ≥ U (
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
x hj = 
0, Pj &lt; U (
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
10. Computation of the state of visible output neurons
12. x2µin = xin , x2µout = xout . If µ &lt; P , then µ = µ +1, go to 9.
13. µ =1, x2µh−1 = 0 ,
14. xin = x2µin , xout = x2µout , xh = x2µh−1.
15. Computation of the state of hidden neurons
11. Computation of the state of visible input neurons
16. x2µh = xh . If µ &lt; P , then µ = µ +1, go to 14.
17. Adjustment of synaptic weights based on Boltzmann's rule

 i=1
−bhj(n) − ∑win−h(n)xiin −
 ij
1+ exp i=1
Nout
        </p>
        <p>Nh</p>
        <p>
∑wout−h(n)xiout − ∑wihj−h(n)xih 
ij</p>
        <p>
 Nh 
1+ exp−bijn(n) − ∑win−h(n)xih 
 ij
 i=1 
1
1
Nin
wout−h(n +1) = wiojut−h(n) +η(ρi+j −ρi−j) , i∈1,Nout , j∈1,Nh ,
ij
ρi+j = 1 P 1 P</p>
        <p>∑x1µouit x1µhj , ρi−j = ∑x2µouit x2µhj ,
P µ=1
P µ=1</p>
        <p>h
wihj−h(n +1) = wihj−h(n) +η(ρi+j −ρi−j) , i ∈1, N , j∈1,Nh ,</p>
        <p>
          1, Pj ≥U(
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
xhj = 
0, Pj &lt;U(
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
i=1
, j∈1,Nh .
biout(n +1) = biout(n) +η 1 P ∑x2µouit , i∈1,Nout ,
        </p>
        <p>
∑x1µouit − 1 P
bhj(n +1) = bhj(n) +η 1 ∑P x1µhj − ∑x2µhj  , j∈1,Nh ,</p>
        <p>1 P 
win−h(n +1) = wiijn−h(n) +η(ρi+j −ρi−j), i∈1,Nin , j∈1,Nh ,
ij
ρi+j = 1 P 1 P</p>
        <p>∑x1µini x1µhj , ρi−j = ∑x2µini x2µhj ,
∑ x1µh −1,i x1µhj , ρ i−j =
∑ ∑| x1µouit − x2µouit | &gt; ε , then n = n + 1, go to 2.</p>
        <p>Method for determining parameter values of a bidirectional RBMRHL
model for recognition based on contrastive divergence</p>
        <p>
          Number of training iteration n = 1 , initialization by means of uniform distribution on the
interval (
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ) or [-0.5, 0.5] offsets (thresholds)
biin (n) , i ∈1, N in , b out (n) , i ∈1, N out ,
i
bLhj (n) , j ∈1, N h , bR hj (n) , j ∈1, N h , and weights wLiinj−h (n) , i ∈1, N in , j ∈1, N h , wRiijn−h (n) ,
i ∈1, N in , j ∈1, N h , wLoijut−h (n) , i ∈1, N out , j ∈1, N h , wRiojut−h (n) , i ∈1, N out , j ∈1, N h ,
wL
wLoijut−h (n) = wLojuit−h (n) , wRiojut−h (n) = wR ojiut−h (n) .
        </p>
        <p>Training set {(xµin , xµout ) | xµin ∈{0,1}N in , xµout ∈{0,1}N out } , µ ∈1, P , is specified, where xµin –
µ th training vector of states of visible input neurons, xout – µ th training vector of states of visible
µ
output neurons, P is the power of the training set.</p>
        <p>Positive phase (steps 3-11)</p>
        <p>x1µin = xµin , x1µout = xµout , µ ∈1, P .</p>
        <p>h
µ = 1, xL1µ −1 = 0 ,
xin = x1µin , xout = x1µout , xLh = xL1µh −1
Computation of the state of hidden neurons</p>
        <p>P =
j
1 + exp

 − bL j (n) − ∑ wL</p>
        <p>h
</p>
        <p>N in
i=1
 N h
 i=1
 ∑ wLhij−h (n)xLhi
xL1µh = xL . If µ &lt; P , then µ = µ + 1 , go to 5.</p>
        <p>h
µ = P , xR1µh +1 = 0 .</p>
        <p>xin = x1µin , xout = x1µout , xRh = xR1µh +1 .
10. Computation of the state of hidden neurons
1</p>
        <p>N out
i=1
in−h (n)xiin − ∑ wLout−h (n)xiout − 
ij ij






14. Computation of the state of visible output neurons
1
1
1
, j∈1,Nh ,
 Nin Nout 
−bRhj(n) − ∑wRiijn−h(n)xiin − ∑wRiojut−h(n)xiout −
1+ exp i=1 i=1 </p>
        <p>Pj =
Pj =
15. Computation of the state of visible output neurons
16. x2µin = xin , x2µout = xout . If µ &lt; P , then µ = µ +1, go to 13.
17. µ =1, xL2µh−1 = 0 ,
18. xin = x2µin , xout = x2µout , xLh = xL2µh−1
19. Computation of the state of hidden neurons
 Nh
 i=1
∑wLhij−h(n)xLhi
 Nin Nout 
−bLhj(n) − ∑wLiinj−h(n)xiin − ∑wLoijut−h(n)xiout −
1+ exp i=1 i=1 </p>
        <p>
          1, Pj ≥U(
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
xLhj = 
0, Pj &lt;U(
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          )
23. Computation of the state of hidden neurons






bRhj(n +1) = bRhj(n)+η 1 P
        </p>
        <p>1 P 
∑xR1µhj − ∑xR2µhj , j∈1,Nh ,
biin(n +1) = biin(n) +η 1 P</p>
        <p>1 P 
∑x1µini − ∑x2µini  , i∈1,Nin ,
biout(n +1) = biout(n) +η 1 P</p>
        <p>
∑x1µouit − 1 P</p>
        <p>∑x2µouit , i∈1,Nout ,
bLhj(n +1) = bLhj(n)+η 1 P</p>
        <p>1 P 
∑xL1µhj − ∑xL2µhj  , j∈1,Nh ,
24. xR2µh = xRh . If µ &gt;1, then µ = µ −1, go to 22.
25. Adjustment of synaptic weights based on Boltzmann's rule
26. If
P⋅ Nout</p>
        <p>P Nout
µ=1 i=1
∑∑| x1µouit − x2µouit | &gt;ε , then n = n +1, go to 2.</p>
        <p>P µ=1
P µ=1
P µ=1



P µ=1
P µ=1
P µ=1
P µ=1
P µ=1</p>
        <p>P µ=1
P µ=1
P µ=1
P µ=1</p>
        <p>P µ=1
P µ=1</p>
        <p>P µ=1
wLhij−h(n +1) = wLhij−h(n)+η(ρi+j −ρi−j), i∈1,Nh , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑xL1µh−1,i xL1µhj , ρi−j = ∑xL2µh−1,i xL2µhj
wRihj−h(n +1) = wRihj−h(n) +η(ρi+j −ρi−j) , i∈1,Nh , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑xR1µh+1,i xR1µh,j , ρi−j = ∑xR2µh+1,i xR2µhj
wLiinj−h(n +1) = wLiinj−h(n) +η(ρi+j −ρi−j), i∈1,Nin , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑x1µini xL1µhj , ρi−j = ∑x2µini xL2µhj ,
wRiijn−h(n +1) = wiijn−h(n) +η(ρi+j −ρi−j), i∈1,Nin , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑x1µini xR1µhj , ρi−j = ∑x2µini xR2µhj ,
wLoijut−h(n +1) = wLoijut−h(n) +η(ρi+j −ρi−j), i∈1,Nout , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑x1µouit xL1µhj , ρi−j = ∑x2µouit xL2µhj
wRiojut−h(n +1) = wiojut−h(n) +η(ρi+j −ρi−j), i∈1,Nout , j∈1,Nh ,
ρi+j = 1 P 1 P</p>
        <p>∑x1µouit xR1µhj , ρi−j = ∑x2µouit xR2µhj</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>6. Numerical study</title>
      <p>The numerical study of the proposed methods for determining the parameter values was carried out
in the Google Colaboratory environment using the Tensorflow package.</p>
      <p>To determine the structure of a classification model based on RBMRHL with 200 input neurons
(corresponding to the number of analyzed words in each text), i.e. determining the number of hidden
neurons, a number of experiments were carried out, the results of which are presented in Figure 3.</p>
      <p>1
0,9
0,8
0,7
0,6
y
c
rau0,5
c
c
A0,4
0,3
0,2
0,1
0
20
40
60
80
100
120
160
180
200
220
240</p>
      <p>260</p>
      <p>The standard IMDB data set was used as the input data to determine the values of the parameters
of the neural network classification model. The criterion for choosing the structure of the neural
network model was the classification accuracy. As can be seen from the Figure 3, with an increase in
the number of hidden neurons, the accuracy value increases. For prediction, it is sufficient to use 200
hidden neurons, since with a further increase in the number of hidden neurons, the change in the
accuracy value is insignificant.</p>
      <p>Table 1 presents a comparative description of neural networks for recognition, where BRBMRHL
means bidirectional RBMRHL.
According to Table 1, BRBMRHL has the highest recognition accuracy.</p>
      <sec id="sec-2-1">
        <title>Network</title>
      </sec>
      <sec id="sec-2-2">
        <title>Criterion Accuracy</title>
        <p>7. Conclusions
1. To solve the problem of increasing the accuracy of recognition of natural language structures,
the existing methods of neural network classification were investigated. These studies have shown
that today the most effective is the use of recurrent neural networks.
2. To improve the quality of recognition of natural language constructions, mathematical models
of unidirectional and bidirectional stochastic neural networks RBMRHL (Restricted Boltzmann
machine with recurrent hidden layer) were created, in which, unlike the traditional RBM
(Restricted Boltzmann machine), hidden layer neurons are interconnected.
3. For models of unidirectional and bidirectional stochastic neural networks RBMRHL, methods
for identifying their parameters based on contrastive divergence were proposed.
4. In the course of a numerical study of models of unidirectional and bidirectional stochastic
neural networks RBMRHL, their structure was determined. The experiments showed that with 200
hidden neurons (corresponding to the number of input neurons), the accuracy value does not
change significantly, and the selected network gives recognition results with maximum accuracy.
5. The proposed approach can be used in various intelligent systems that use the recognition of
natural language constructs. For example, in supply chain management systems, where natural
language interaction between subjects, which are represented by software agents, plays an
important role.
8. References
[16] T. Mikolov, M. Karafi´at, L. Burget, J. Cernock`y, S. Khudanpur, Recurrent neural network
based language model, in: 11th Annual Conference of the International Speech Communication
Association, 2010, pp. 1045-1048.
[17] M. Sundermeyer, T. Alkhouli, J. Wuebker, H. Ney, Translation modeling with bidirectional
recurrent neural networks, in: Proceedings of the Conference on Empirical Methods on Natural
Language Processing, 2014, pp. 14-25.
[18] M. Berglund, T. Raiko, M. Honkala, L. Kärkkäinen, A. Vetek, J. Karhunen, Bidirectional
recurrent neural networks as generative models, in: Proceedings of the 28th International
Conference on Neural Information Processing Systems, 2015, pp. 856−864.
[19] M. Sundermeyer, R. Schluter, H. Ney, LSTM neural networks for language modeling, in:
Thirteenth Annual Conference of the International Speech Communication Association, 2012,
pp. 194-197. doi: 10.1.1.248.4448.
[20] P. Potash, A. Romanov, A. Rumshisky, Ghostwriter: using an LSTM for automatic rap lyric
generation, in: Proceedings of the 2015 Conference on Empirical Methods in Natural Language
Processing, 2015, pp. 1919– 1924. doi:10.18653/v1/D15-1221.
[21] E. Kiperwasser, Y. Goldberg, Simple and Accurate Dependency Parsing Using Bidirectional
LSTM Feature Representations, Transactions of the Association for Computational Linguistics 4
(2016) 313–327. doi: 10.1162/tacl_a_00101.
[22] A. Graves, J. Schmidhuber, Framewise phoneme classification with bidirectional LSTM and
other neural network architectures, Neural Networks 18 (2005) 602–610,
doi:10.1016/j.neunet.2005.06.042.
[23] J. Chung, C. Gulcehre, K. Cho, Y. Bengio, Empirical evaluation of gated recurrent neural
networks on sequence modeling, arXiv preprint arXiv:1412.3555, 2014.
[24] R. Dey, F. M. Salem, Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks,
arXiv:1701.05923, 2017. URL: https://arxiv.org/ftp/arxiv/papers/1701/1701.05923.pdf.
[25] S. A. Khan, S. M. D. Khalid, M. A. Shahzad, F. Shafait, Table structure extraction with
bidirectional gated recurrent unit networks, in: Proceedings of the 15th International Conference on
Document Analysis and Recognition, vol. 4, no. 2, 2019, pp. 78–88.
doi:10.1109/ICDAR.2019.00220
[26] G. E. Hinton, A Practical Guide to Training Restricted Boltzmann Machines, Technical Report</p>
        <p>UTML TR 2010–003, University of Toronto, 2010.
[27] A. Fischer, C. Igel, Training Restricted Boltzmann Machines: An Introduction, Pattern
Recognition 47 (2014) 25-39. doi: 10.1016/j.patcog.2013.05.025.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>P. F.</given-names>
            <surname>Dominey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hoen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Inui</surname>
          </string-name>
          ,
          <article-title>A neurolinguistic model of grammatical construction processing</article-title>
          ,
          <source>in: Journal of Cognitive Neuroscience</source>
          , vol.
          <volume>18</volume>
          , issue 12,
          <year>2006</year>
          , pp.
          <fpage>2088</fpage>
          -
          <lpage>2107</lpage>
          . doi:
          <volume>10</volume>
          .1162/jocn.
          <year>2006</year>
          .
          <volume>18</volume>
          .12.
          <year>2088</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>N.</given-names>
            <surname>Khairova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sharonova</surname>
          </string-name>
          ,
          <article-title>Modeling a logical network of relations of semantic items in super phrasal unities</article-title>
          ,
          <source>in: Proceedings of the 2011 9th East-West Design &amp; Test Symposium (EWDTS)</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>360</fpage>
          -
          <lpage>365</lpage>
          . doi:
          <volume>10</volume>
          .1109/EWDTS.
          <year>2011</year>
          .
          <volume>6116585</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J. F.</given-names>
            <surname>Cox</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.G.</given-names>
            <surname>Schleher</surname>
          </string-name>
          , Theory of Constraints Handbook, New York, NY,
          <string-name>
            <surname>McGraw-Hill</surname>
          </string-name>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Goldratt</surname>
          </string-name>
          ,
          <article-title>My saga to improve production, Selected Readings in Constraints Management, Falls Church, VA: APICS (</article-title>
          <year>1996</year>
          )
          <fpage>43</fpage>
          -
          <lpage>48</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Goldratt</surname>
          </string-name>
          ,
          <article-title>Production: The TOC Way (Revised Edition) including CD-ROM Simulator and Workbook, Revised edition</article-title>
          , Great Barrington, MA: North River Press,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S. N.</given-names>
            <surname>Sivanandam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sumathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N.</given-names>
            <surname>Deepa</surname>
          </string-name>
          , Introduction to Neural
          <source>Networks using Matlab 6</source>
          .0,
          <string-name>
            <surname>The</surname>
            <given-names>McGraw-Hill</given-names>
          </string-name>
          <string-name>
            <surname>Comp</surname>
          </string-name>
          ., Inc., New Delhi,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Haykin</surname>
          </string-name>
          ,
          <article-title>Neural networks and Learning Machines</article-title>
          , Upper Saddle River, New Jersey: Pearson Education, Inc.,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>K.-L. Du</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. S. Swamy</surname>
          </string-name>
          ,
          <source>Neural Networks and Statistical Learning</source>
          , Springer-Verlag, London,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>E.</given-names>
            <surname>Fedorov</surname>
          </string-name>
          , T. Utkina, О. Nechyporenko,
          <article-title>Forecast method for natural language constructions based on a modified gated recursive block</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , vol.
          <volume>2604</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>199</fpage>
          -
          <lpage>214</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Vairappan</surname>
          </string-name>
          ,
          <article-title>A novel learning method for Elman neural network using local search</article-title>
          ,
          <source>in: Neural Information Processing - Letters and Reviews</source>
          , vol.
          <volume>11</volume>
          ,
          <year>2007</year>
          , pp.
          <fpage>181</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Dey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Salem</surname>
          </string-name>
          ,
          <article-title>Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks</article-title>
          ,
          <source>arXiv:1701.05923</source>
          ,
          <year>2017</year>
          . - URL: https://arxiv.org/ftp/arxiv/papers/1701/1701.05923.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. van Merrienboer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gulcehre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Bougares</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Schwenk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>Learning phrase representations using RNN encoder-decoder for statistical machine translation</article-title>
          ,
          <source>in: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</source>
          , Doha, Qatar,
          <year>2014</year>
          , pp.
          <fpage>1724</fpage>
          -
          <lpage>1734</lpage>
          . doi:
          <volume>10</volume>
          .3115/v1/
          <fpage>D14</fpage>
          -1179.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Jaeger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Maass</surname>
          </string-name>
          , J. Prıncipe,
          <article-title>Special issue on echo state networks and liquid state machines</article-title>
          ,
          <source>Neural Networks</source>
          <volume>20</volume>
          (
          <year>2007</year>
          )
          <fpage>287</fpage>
          -
          <lpage>289</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.neunet.
          <year>2007</year>
          .
          <volume>04</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A. H. S.</given-names>
            <surname>Hamdany</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R. O.</given-names>
            <surname>Al-Nima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. H.</given-names>
            <surname>Albak</surname>
          </string-name>
          ,
          <article-title>Translating cuneiform symbols using artificial neural network</article-title>
          ,
          <source>in: TELKOMNIKA Telecommunication, Computing, Electronics and Control</source>
          , volume
          <volume>19</volume>
          ,
          <source>No. 2</source>
          ,
          <issue>2021</issue>
          , pp.
          <fpage>438</fpage>
          -
          <lpage>443</lpage>
          . doi:
          <volume>10</volume>
          .12928/telkomnika.v19i2.
          <fpage>16134</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Wysocki</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Ławryńczuk</surname>
          </string-name>
          ,
          <article-title>Predictive control of a multivariable neutralisation process using Elman neural networks</article-title>
          ,
          <source>in: Advances in Intelligent Systems and Computing</source>
          , Springer:Heidelberg,
          <year>2015</year>
          , pp.
          <fpage>335</fpage>
          -
          <lpage>344</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>319</fpage>
          -15796-234.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>