<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>ORCID:</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Hybrid GMDH Deep Learning Networks - State-of Art and New Prospective Trends</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yuriy Zaychenko</string-name>
          <email>zaychenkoyuri@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Galib Hamidov</string-name>
          <email>galib.hamidov@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>03056</institution>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Information Technologies Department</institution>
          ,
          <addr-line>Azershig,str. K. Kazim-zade 20, Baku, , AZ1008</addr-line>
          ,
          <country country="AZ">Azerbaijan</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Institute for Applied System Analysis, Igor Sikorsky Kyiv Polytechnic Institute</institution>
          ,
          <addr-line>Peremogy avenue 37, Kyiv</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1950</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>In this paper new class of deep learning (DL) neural networks is considered and investigatedso-called hybrid DL networks based on self-organization method Group Method of Data Handling (GDMDH). The application of GMDH enables not only to train neural weights but to construct the network structure as well. As nodes of this structure different elementary neurons with two inputs may be used. So, the advantage of such structure is a small number of tuning parameters. In the paper the following types of neurons are considered: WangMendel network with two inputs and neo-fuzzy neurons. The advantage of the neo-fuzzy neurons is unlike general fuzzy neurons is absence of fuzzy membership functions training and less computational time for training. The application of GMDH enables to train neuron weights sequentially layer after layer in the process of construction network structure until the stop criterion holds. Such approach allows to exclude drawbacks of DL training algorithms -decay or explosion of gradient. The process of structure construction and optimization using GMDH algorithm is presented. The numerous applications of suggested hybrid DL networks for solution of AI problems like forecasting of share prices and market indicators at various stock exchanges are considered and analyzed. The comparison with conventional DL networks is performed which enables to estimate their efficiency and advantages. Hybrid deep learning networks, self-organization, structure optimization, forecasting</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Last years deep learning (DL) networks are widely used in different problems of artificial
intelligence: forecasting, pattern recognition, medical diagnostics, etc.[
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1-4</xref>
        ]. For its training various
algorithms were developed usually based on Back propagation method. Presence of many layers when
using gradient algorithm usually lead to occurrence drawbacks as vanishing or explosion of gradient.
Therefore, the approach was suggested how to exclude this drawback to perform layer after layer
training using stacked encoder-decoder or stacked restricted Boltsmann machines [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. However, the
problem is left how to choose the number of layers in DL network. The existing methods of DL don’t
enable to generate structure of Dl networks. But the training process will be more efficient if to adapt
not only neuron weights but the structure of network as well. For this goal the application of GMDH
method seems very promising. GMDH is based on principle of self- organization and enables to
construct
network
structure
automatically
in
the
process
of
algorithm
run
[
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5-7</xref>
        ].
      </p>
      <p>
        In the previous years GMDH-neural networks having active neurons [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5-7</xref>
        ], R-neurons [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], Q-neurons
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] as nodes were developed; in the area integrating fuzzy GMDH and neural networks the
GMDHneuro-fuzzy systems and GMDH-neo-fuzzy systems [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] were developed;
GMDH-wavelet-neurofuzzy systems [
        <xref ref-type="bibr" rid="ref14 ref15">14,15</xref>
        ] were also elaborated. The very important property of GMDH is that as building
blocks for construction of a structure of DL networks elementary models with only two inputs, so-called
      </p>
      <p>2022 Copyright for this paper by its authors.
partial descriptions, are used. This allows to cut substantially training time for hybrid Dl network as
compared with conventional DL networks.</p>
      <p>
        Therefore, GMDH-hybrid neuro-fuzzy system was developed in [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] that combines advantages of
the traditional GMDH and DL fuzzy networks and may be trained with simple learning procedures. The
nodes of this network are Wang-Mendel elementary neural networks with only two inputs. The
experimental investigations of this class of hybrid DL networks have shown their efficiency and
preference over conventional DL networks. But the drawbacks of application of Wang-Mendel
networks as nodes of hybrid Dl networks lies herein that it’s necessary to train not only neural weights
but membership functions as well.
      </p>
      <p>
        Therefore, later another class of hybrid networks – GMDH – neo-fuzzy networks were developed
wherein as nodes of network -neo-fuzzy neurons with two inputs are used [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. For their training its
necessary to adapt only neuron weights that demands less computational resources and cuts training
time. That’s very important for DL networks with a large number of hidden layers. The experimental
investigations of hybrid -neo-fuzzy networks and comparison with conventional DL network have
shown their efficiency and less computational calculations for training.
      </p>
      <p>The goal of this paper is to investigate different hybrid GMDH-neo-fuzzy networks with small
number of adjusted parameters and estimate their efficiency for structure optimization and forecasting.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Hybrid network structure optimization based on GMDH method</title>
      <p>The GMDH method was used to synthesize the structure of the hybrid network based on the principle
of self-organization. The successive increase in the number of layers is carried out until the value of the
external criterion of optimality MSE begins to increase for the best model of the current layer. In this
case it is necessary to return to the previous layer, to find there the best model with the minimum value
of criterion. Then we move backward, go through its connections, find the corresponding neurons of
the previous layer. This process continues until we reach the first layer and the corresponding structure
is automatically determined.</p>
      <p>The process of synthesis of the network structure in the forward direction is shown in Fig. 1 where
in green color the outputs which passed through selection block (SB)are shown while in red color
outputs which were dropped (excluded) by SB.</p>
      <p>The process of restoring the desired structure in the backward direction is shown in Fig. 2. In the
yellow color nodes and their connections selected by this process are indicated</p>
      <p>The corresponding optimal constructed structure of the hybrid network for this forecasting problem
is shown in Fig. 3. It consists of 3 layers: first layer has 3 neo-fuzzy neurons, second layer- two neurons
and the last- one neuron.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Experimental investigations of hybrid GMDH -fuzzy networks in forecasting problems</title>
      <p>For efficiency estimation of hybrid GMDH DL networks the problems of the forecasting share price
and market indices at the stock exchanges were considered. The experimental investigations for stock
prices forecasting were carried out. In the first experiment as a forecasted variable the RTS index
in 2013 with time step one week was chosen. As external regressors (inputs) stock prices of the leading
companies were used. Total sample had 55 points that was used while searching the optimal partial
description in the GMDH.</p>
      <p>As the accuracy criteria of the obtained models MAPE and RMSE were used. In the first experiment
the dependence of MAPE on inputs number was explored. The forecasting results for hybrid
neurofuzzy network are presented in the table 1. For comparison the corresponding results for full cascade
neo-fuzzy network (NFN) network are presented.</p>
      <p>As it follows, hybrid GMDH-neuro-fuzzy network has higher accuracy than the cascade neuro-fuzzy
network due to properties of hybrid networks.</p>
      <p>In the next experiment the problem of forecasting share prices of Microsoft corp was considered. As
input sample the stock prices of Microsoft corp. since 01.11.14 to 29.12.14 were used. The sample size
was 64 points. The training sample included 62 points the test sample 4 points. The forecasting interval
was for 4 steps ahead, the first two steps are checked with available data. The constructed
GMDHneuro-fuzzy network had 6 fuzzy inputs. The experimental results are presented in Table 2 and Table 3.
As it follows from the experimental results the GMDH-neuro-fuzzy network showed better forecasting
accuracy than the cascade neuro-fuzzy network. Its MAPE value doesn’t exceed 0.4%.
number of inputs</p>
      <p>In the next experiment training time for GMDH-neuro-fuzzy network, and cascade fuzzy network
were compared. In the Table 4 the training time in seconds for hybrid GMDH- neuro-fuzzy network
and full cascade neuro-fuzzy network is presented. As an initial sample Microsoft stock prices in the
same period since 01.11.14 to 29.12.14 was used.</p>
      <p>In the next experiments efficiency of hybrid neo-fuzzy network in forecasting index NASDAQ was
explored. The data was taken in the period from 13.11. 17 till 29.11.19. The sample size was 510 points.
As an output variable the closing price of the index NASDAQ next day was taken. In the first
experiment the accuracy dependence on number of inputs for hybrid neo-fuzzy network was
investigated. In the table 5 the forecasting results are presented under different inputs number with 8
membership functions per variable (parameter h) and ratio training/test =70/30. In the next experiment
the investigation of error dependence on number of MF per variable (parameter h) was performed.
Number of inputs was n=5, training/test ratio was 70/30. The results are presented in Fig.4.</p>
      <p>
        Analyzing these results, one may conclude that with MF number rise MAPE first falls, then attains
minimum and after then begins to rise. That fully matches to self- organization principle of GMDH
method [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. The best value was obtained with the following parameters values: number of inputs n=5,
h=8, number of layers 4 and MAPE value is 3,91.
      </p>
      <p>
        For forecasting efficiency estimation of the hybrid network, it was compared with a cascade
neofuzzy network [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and GMDH at the same data. In the cascade neo-fuzzy network, the following
parameters values were used: number of inputs n=9, number of rules 9, cascades number is 3. The
comparative forecasting results are presented in the table 6, training sample – 70%.
      </p>
      <p>Analyzing these results one can easily conclude the suggested hybrid neo-fuzzy network and
neurofuzzy network have the best accuracy, the second one is GMDH method and the worst is the cascade
neo-fuzzy network. The forecasting accuracy of both hybrid networks differs insignificantly.</p>
      <p>In the next experiments the training time of different hybrid networks and alternative NN was
investigated and compared. In the table 7 the training time in seconds for GMDH-neuro-fuzzy and
neo-fuzzy network and full cascade neuro-fuzzy network are presented. As an initial sample we used
Microsoft stock prices in the period since 01.11.14 to 29.12.14., a sample size is 64 points.</p>
      <p>As it follows from the presented results the least training time has hybrid neo-fuzzy network, the
second place takes hybrid neuro fuzzy network and the last is full cascade network</p>
    </sec>
    <sec id="sec-4">
      <title>4. Optimization of hybrid GMDH -neo-fuzzy network in the problem of forecasting</title>
      <p>In the next experiments investigations of hybrid GMDH -neo-fuzzy network in the problem of Dow
Jones Index forecasting were performed and compared with FNN ANFIS.</p>
      <p>The Dow Jones is the stock index of the 30 largest American companies, which was founded in
1896. The initial data was taken from Yahoo, a leading financial information provider owned by
Yahoo! To prepare the initial data, data were uploaded at various intervals, namely the value of the
stock index by days, weeks and months. Each of the sets contains the following data:
 Date - data period;
 Open - opening price;
 High - the highest price for the period;
 Low - the lowest price for the period;
 Close - the price at the end of the period;

</p>
      <p>Adj Close - average closing price;</p>
      <p>Volume - sales for the period.</p>
      <p>The data set for the interval of one day contains 4867 records, of which non-zero records are 4788
ones. The data set for the interval one month contains 1001 records, of which 1000 records are
nonzero. The data set for the interval of one month contains 195 records, of which 195 are non-zero.</p>
      <p>Data normalizing. Reduction to a single scale is provided by normalization of each variable to the
range of its values. In the simplest case, it is a linear transformation
m ̅ =</p>
      <p>−  
   −</p>
      <p>
        in the interval   ∈ [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ].
      </p>
      <p>To find the most informative features as an input vector, we have alternately trained the network on
data sets that transmit only the following features subsets:
('Open', 'High', 'Low', 'Volume', 'Close'); ('Open', 'High', 'Low', 'Volume');
('Open', 'High', 'Low', 'Close'); ('Open', 'High', 'Low'); ('Open', 'High', 'Close');
('Open', 'High', 'Volume'); ('Open', 'Close', 'Low'); ('Open', 'Volume', 'Low');
('High', 'Low', 'Close'); ('Open', 'High'); ('High', 'Close'); ('Low', 'Close'); ('Open', 'Volume').</p>
      <p>The main network parameters that can be configured include the size of the input vector, the number
of rules, and the function that sets them, the number of parameters that are transferred to the next layer.</p>
      <p>The size of the input vector is determined by the number of informative features that are transmitted
for training, and the number of days on the basis of which the network gives the predicted value. Also,
the number of network functions that can be set includes the number of membership functions and their
appearance, as well as the degree of freedom of choice of the system.</p>
      <p>To select these parameters, it is necessary to conduct an experiment, training the system, setting
these parameters in the interval, and keeping those that give the best results in the test sample.</p>
      <p>The following parameters were investigated:
 n – number of preceding days, based on which the forecasting is performed (sliding window
size). N ∈ [1; 6];
 h - number of membership functions in each node, h ∈ [2; 9];
 s – membership function parameter  [− ( −  )2], where  = ( –  ) ( ∗ (ℎ − 1));
2 ℎ
1 month</p>
      <p>МАЕ
0.0147
'Open', 'High', 'Low',
'Open', 'High', 'Volume' 5 2 5 0.1
 – an interval beginning;
 ∈ [0.01; 1.5];
ℎ – membership functions number, which cover the interval;
 − number of parameters which are transferred to the network next layer (freedom of choice).</p>
      <p>To set of initial data was divided into a training sample and test sample in the ratio of 70% and 30%,
respectively. Having launched GMDH-neo-fuzzy system for training, values of MAE and MAPE
criteria were obtained with different combinations of these parameters. For the Dow Jones stock index
with different forecast intervals, the best parameters for the different set of informative features were
obtained as a result of training and testing, which are shown in Table 8.
with different prediction intervals</p>
      <p>Sets of informative</p>
      <p>features
'Open', 'High', 'Low',</p>
      <p>'Volume', 'Close'
'Open', 'High', 'Low',
'Volume'</p>
      <p>'Close'
'Open', 'High', 'Low'
'Open', 'High', 'Close'
'Open', 'Low', 'Close'
'Open', 'Volume', 'Low'
'High', 'Low', 'Close'
'Open', 'High'
'Open', 'Close'
'High', 'Close'
'Low', 'Close'
'Open', 'Volume'</p>
      <p>Thus, analyzing presented results one may conclude that the most informative for GMDH-neo-fuzzy
system are the following sets of features: ['Open', 'High', 'Close'], ['Open', 'Low', 'Close'], ['High', 'Low',</p>
      <p>For the Dow Jones stock index for one month forecast period, the following optimal configurations
'Close'], ['High', 'Close'], ['Low', 'Close'].
of GMDH-neo-fuzzy network were obtained:
the number of informative features - 3;
the number of periods on the basis of which the forecast is made - 1;
the number of membership functions in each of the nodes - 2;
the number of layers - 2;
the number of nodes in the first layer – 3;
number of nodes on the second layer – 1.</p>
      <p>For the Dow Jones stock index for the one week forecast period, the following optimal
configurations of the GMDH-neo-fuzzy system were obtained:
the number of informative features- 3;
the number of periods on the basis of which the forecast is made - 2;
the number of membership functions in each of the nodes - 4;
 the number of layers - 2;
 the number of nodes on the first layer - 15;
 the number of nodes on the second layer – 1.</p>
      <p>The form of the membership function for forecasting interval of one week is shown in the Figure 5.</p>
      <p>For the Dow Jones stock index for one week forecast period, the following optimal configurations
of GMDH-neo-fuzzy network were obtained:
 number of informative features - 3;
 the number of periods on the basis of which the forecast is made - 5;
 the number of membership functions in each of the nodes - 2;
 the number of layers - 2;
 the number of nodes in the first layer - 105;
 the number of nodes in the second layer - 1</p>
      <p>Next, experiments were performed to find the optimal values of the parameters of FNN ANFIS. The
size of the input vector is determined by the number of informative features that are transmitted for
training, and the number of days of prehistory, on the basis of which the forecasting is performed.</p>
      <p>To select these parameters, an experiment was performed, including training of the network, setting
these parameters in the interval, and choosing those that give the best results at the test sample.</p>
      <p>The following intervals for parameters were set:
 n is the number of previous days on the basis of which the forecast is made (the size of the
sliding window), n∈ [1; 6];
 h - the number of membership functions in each of the nodes, h ∈ [2; 9].</p>
      <p>The set of initial data was divided into a training sample and test data in the proportion of 70% and
30%, respectively. By launching the ANFIS network with different combinations of these parameters,
data on MAE and MAPE criteria were obtained.</p>
      <p>For the Dow Jones stock index one month forecast period, the following optimal ANFIS network
configurations were obtained:
 number of informative features - 3;
 number of nodes – 6;
 the number of periods on the basis of which the forecast is made - 2;
 the number of membership functions in each of the nodes - 6.</p>
      <p>After finding all the optimal parameters of GMDH-neo-fuzzy system and training parameters, the
system was trained, and then the data for prediction was provided. Training and testing of the system
took place on data for the period up to 01.01.2021 for monthly periods, and until 01.06.2021 for weekly
and day periods. Forecasting was based on data for the period after 01.01.2021. for monthly periods
and after 01.06.2021 for day and week periods. For Dow Jones index with a forecast period of one
month, the following forecasting data were obtained: MAE - 0.02952; MAPE - 0.0335, forecasting time
- 0.00025s. Learning and forecasting results are shown in Figure 6.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Comparison of forecasting results of GMDH-neo-fuzzy system and ANFIS network</title>
      <p>Experimental investigations of the accuracy of Dow Jones index forecasting with forecasting
intervals of one month, one week and one day were performed. using a hybrid GMDH-neo-fuzzy
network. For each prediction interval the optimal parameters found in previous experiments were
selected. A comparative analysis with the forecasting results obtained by FNN ANFIS was performed.
According to the results of forecasting, values of MAE, MAPE and training time for each type of neural
network were obtained. All comparison results are summarized in Tables 10 – 12.</p>
      <p>As one can see, for all forecasting intervals the best forecasting results were obtained for hybrid
GMDH-neo-fuzzy system. The worst forecasting result for ANFIS network was obtained for one month
forecasting period. The largest difference in the accuracy of forecasting by both criteria was obtained
for the forecasting period of one month (over 200%). As the forecasting period decreases, the gap
between the networks accuracy also decreases. In addition, training and direct prediction times were
also significantly less for hybrid GMDH-neo-fuzzy system.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In the paper hybrid GMDH -neuro-fuzzy and neo-fuzzy networks are considered and investigated.</p>
      <p>The algorithm of hybrid network structure synthesis is presented and demonstrated at the problem
of forecasting. The experimental investigations of the hybrid networks were carried out and compared
with conventional DL networks. The experiments have shown that forecasting accuracy of hybrid
neuro-fuzzy and neo- fuzzy networks at the considered problems are approximately equal and it’s better
than for alternative DL cascade neo- fuzzy networks and GMDH. The problem of forecasting Dow
Jones Index with application of hybrid neo- fuzzy networks was considered, investigated and compared
with FNN ANFIS at the different forecasting intervals. The optimal parameters of hybrid neo- fuzzy
networks were found. The experimental results have shown the forecasting accuracy of hybrid
neofuzzy networks is much better than for FNN ANFIS. The training time is the least for hybrid neo- fuzzy
networks as compared with all considered alternative DL networks.</p>
      <p>In a whole the hybrid DL networks based on GMDH are free from drawbacks of conventional DL
networks- decay or explosion of gradient. Besides, they enable to construct optimal network structure
automatically in the process of algorithm GMDH run and additionally they demand less computational
costs for training due to small number of tunable parameters (only two) in every hidden node as
compared with DL networks of general structure. That’s is especially significant for DL networks with
large number of layers.</p>
    </sec>
    <sec id="sec-7">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>I.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <article-title>Deep Learning</article-title>
          . MIT Press,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Hinton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Osindero</surname>
          </string-name>
          , and Y.-W. Teh, “
          <article-title>A fast learning algorithm for deep belief nets</article-title>
          ,
          <source>” Neural Computation</source>
          , vol.
          <volume>18</volume>
          , no.
          <issue>7</issue>
          , pp.
          <fpage>1527</fpage>
          -
          <lpage>1554</lpage>
          , May
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          , Y. LeCun, and G. Hinton, “Deep learning,” Nature, no.
          <issue>521</issue>
          , pp.
          <fpage>436</fpage>
          -
          <lpage>444</lpage>
          , May
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Schmidhuber</surname>
          </string-name>
          , “
          <article-title>Deep learning in neural networks: an overview,” Neural Networks</article-title>
          , no.
          <issue>61</issue>
          , pp.
          <fpage>85</fpage>
          -
          <lpage>117</lpage>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.G.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.A.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            <surname>Mueller</surname>
          </string-name>
          ,
          <article-title>Self-organization of the neural networks with active neurons</article-title>
          .
          <source>Pattern Recognition and Image Analysis</source>
          <volume>4</volume>
          ,
          <issue>2</issue>
          (
          <year>1994</year>
          ):
          <fpage>177</fpage>
          -
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.G.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wuensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.A.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <article-title>Inductive sorting-out GMDH algorithms with polynomial complexity for active neurons of neural networks</article-title>
          .
          <source>Neural Networks</source>
          <volume>2</volume>
          (
          <year>1999</year>
          ):
          <fpage>1169</fpage>
          -
          <lpage>1173</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>G.A.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <article-title>Self-organization of neuronet with active neurons for effects of nuclear test explosions forecasting</article-title>
          .
          <source>System Analysis Modeling Simulation</source>
          <volume>20</volume>
          (
          <year>1995</year>
          ):
          <fpage>107</fpage>
          -
          <lpage>116</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zgurovsky</surname>
          </string-name>
          , Yu. Zaychenko,
          <article-title>Fundamentals of computational intelligence- System approach</article-title>
          . Springer,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.-X.</given-names>
            <surname>Wang</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Mendel</surname>
          </string-name>
          , “
          <article-title>Fuzzy basis functions, universal approximation, and orthogonal least-squares learning”</article-title>
          .
          <source>IEEE Trans. on Neural Networks</source>
          <volume>3</volume>
          , №
          <volume>5</volume>
          (
          <year>1992</year>
          ):
          <fpage>807</fpage>
          -
          <lpage>814</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>J.-S</surname>
          </string-name>
          . Jang, “ANFIS:
          <article-title>Adaptive-network-based fuzzy inference systems”</article-title>
          .
          <source>IEEE Trans. on Systems, Man, and Cybernetics</source>
          .
          <volume>23</volume>
          (
          <year>1993</year>
          ,):
          <fpage>665</fpage>
          -
          <lpage>685</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yamakawa</surname>
          </string-name>
          , E. Uchino,
          <string-name>
            <given-names>T.</given-names>
            <surname>Miki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kusanagi</surname>
          </string-name>
          ,
          <article-title>A neo-fuzzy neuron and its applications to system identification and prediction of the system behavior</article-title>
          ,
          <source>in: Proc. 2nd Intеrn. Conf. Fuzzy Logic and Neural Networks «LIZUKA-92». Lizuka</source>
          ,
          <year>1992</year>
          , pp.
          <fpage>477</fpage>
          -
          <lpage>483</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Teslenko</surname>
            and
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Grimm</surname>
          </string-name>
          ,
          <article-title>Hybrid evolving neural network using kernel activation functions</article-title>
          ,
          <source>in: Proc. 17th</source>
          Zittau
          <string-name>
            <surname>East-West Fuzzy</surname>
            <given-names>Colloquium</given-names>
          </string-name>
          , Zittau/Goerlitz, HS,
          <year>2010</year>
          , pp.
          <fpage>39</fpage>
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Ye</surname>
            . Bodyanskiy, Yu. Zaychenko, E. Pavlikovskaya,
            <given-names>M.</given-names>
          </string-name>
          <article-title>Samarina and Ye. Viktorov, The neofuzzy neural network structure optimization using the GMDH for the solving forecasting and classification problems</article-title>
          ,
          <source>Proc. Int. Workshop on Inductive Modeling</source>
          , Krynica, Poland,
          <year>2009</year>
          , pp.
          <fpage>77</fpage>
          -
          <lpage>89</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Vynokurova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Dolotov</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Kharchenko</surname>
          </string-name>
          ,
          <article-title>Wavelet-neuro-fuzzy network structure optimization using GMDH for the solving forecasting tasks</article-title>
          ,
          <source>in: Proc. 4th Int. Conf. on Inductive Modelling ICIM</source>
          <year>2013</year>
          , Kyiv,
          <year>2013</year>
          , pp.
          <fpage>61</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Vynokurova</surname>
            and
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Teslenko</surname>
          </string-name>
          ,
          <article-title>Cascade GMDH-wavelet-neuro-fuzzy network</article-title>
          ,
          <source>in: Proc. 4th Int. Workshop on Inductive Modeling «IWIM</source>
          <year>2011</year>
          », Kyiv, Ukraine,
          <year>2011</year>
          , pp.
          <fpage>22</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>O. Boiko</given-names>
          </string-name>
          <string-name>
            <surname>Yu</surname>
          </string-name>
          . Zaychenko, G. Hamidov,
          <article-title>Evolving Hybrid GMDH-Neuro-Fuzzy Network and Its Applications</article-title>
          ,
          <source>in: Proceedings of the Intern conference SAIC</source>
          <year>2018</year>
          , Kiev, Ukraine,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Evgeniy</surname>
            <given-names>Bodyanskiy</given-names>
          </string-name>
          , Yuriy Zaychenko, Olena Boiko, Galib Hamidov,
          <string-name>
            <given-names>Anna</given-names>
            <surname>Zelikman</surname>
          </string-name>
          .
          <article-title>The hybrid GMDH-neo-fuzzy neural network in forecasting problems in financial sphere</article-title>
          ,
          <source>in: Proceedings of the International conference IEEE SAIC</source>
          <year>2020</year>
          , Kiev, Ukraine,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>D. T.</given-names>
            <surname>Pham</surname>
          </string-name>
          and
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Neural Networks for Identification, Prediction and Control</article-title>
          . London, Springer-Verlag,
          <year>1995</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ohtani</surname>
          </string-name>
          ,
          <article-title>Automatic variable selection in RBF network and its application to neuro-fuzzy GMDH</article-title>
          ,
          <source>Proc. Fourth Int. Conf. on Knowledge-Based Intelligent Engineering Systems and Allied Technologies</source>
          ,
          <year>2000</year>
          , V.2, pp.
          <fpage>840</fpage>
          -
          <lpage>843</lpage>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>