<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Investigation of Hybrid Deep Learning Networks in Forecasting Energy Supply</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yuriy Zaychenko</string-name>
          <email>zaychenkoyuri@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Helen Zaichenko</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksii Kuzmenko</string-name>
          <email>oleksii.kuzmenko@ukr.net</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>The Group</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>GMDH-neo-fuzzy network</institution>
          ,
          <addr-line>GMDH</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Technical University of Ukraine “Igor Sikorsky Kyiv Polytechnic Institute”, Institute for Applied</institution>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>System Analysis</institution>
          ,
          <addr-line>Prospect Beresteiskyi, 37, Kyiv, 03056</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper the intelligent methods for solving the problem of short- and middle-term forecasting electricity sales to network for short- and middle-term forecasting was performed. Experimental studies of hybrid Problems of forecasting non-stationary time series and market indexes at stock exchanges pay great attention of managers of enterprises and various scientific researchers. For its solution were developed and applied for a long time powerful statistical methods, first of all ARIMA [1, 2]. In recent years, various intelligent methods, and technologies, such as fuzzy logic systems and neural networks, have also been proposed and widely used for forecasting in economics and technology.</p>
      </abstract>
      <kwd-group>
        <kwd>Keywords1</kwd>
        <kwd>Hybrid DL network</kwd>
        <kwd>GMDH</kwd>
        <kwd>ARIMA</kwd>
        <kwd>short-term</kwd>
        <kwd>middle-term forecasting</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>hybrid</p>
      <p>ARIMA
for short- and
middle-term
forecasting have been conducted. The accuracy of the obtained forecasts was compared. The
expediency of applying the researched methods of artificial intelligence for the considered
intervals is substantiated.</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>2023 Copyright for this paper by its authors.
overall training time. Based on the studied two classes of hybrid DL networks, their effectiveness for
forecasting in the financial sphere was compared.</p>
      <p>In the previous papers these methods were applied and investigated in the problem of forecasting
in financial sphere for market indices Dow Jones industrial average and NASDAQ. That is why it is
interesting to investigate the effectiveness of ARIMA, GMDH and hybrid DL networks in forecasting
in other areas, such as technology and production, specifically in short- and middle-term forecasting
tasks. The goal of this paper is to investigate the accuracy of intelligent methods – hybrid DL
networks, GMDH and ARIMA at the problem of forecasting Electricity Sales to Ultimate Customers,
Residential (USA), June 7, 2023 – at the different forecasting intervals (short-term and middle-term),
compare their efficiency and to determine which computational intelligence methods are the most
perspective for forecasting in the economy and technology.</p>
    </sec>
    <sec id="sec-3">
      <title>2. A review of the evolving hybrid GMDH-neo-fuzzy network</title>
      <p>The architecture of the evolving hybrid DL-network is shown in Fig. 1. The input of the system
accepts an ( × 1)-dimensional vector of signals that are considered input. Then the first hidden layer
receives this signal. At this level there are  1 =  2 nodes, each of which has strictly two inputs.</p>
      <p>
        Outputs  [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] of the first hidden layer form the output signals to be further transmitted to the
selection block located after the first hidden layer.
precise signals by some chosen criterion (mostly by the mean squared error  2
      </p>
      <p>
        It selects among the output signals  ̂[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]  1 ∗ (where  1 ∗=  is so-called freedom of choice) most




best outputs of the first hidden layer  ̂[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] ∗  2 pairwise combinations  ̂[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] ∗,  ̂ [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] ∗ are formed. These


signals are fed to the second hidden layer, that is formed by neurons  [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. After training these neurons
output signals of this layer  ̂[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] are transferred to the selection block   [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] which choses  best
neurons by accuracy (e.g. by the value of  2
      </p>
      <p>
        [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]) if the best signal of the second layer is better than the
best signal of the first hidden layer  ̂1[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] ∗. Other hidden layers work in a similar way. The evolution of
the system continues until the best signal of the selection block   [ +1] is worse than the best signal
received on the previous s-h layer. After that, you need to return to the previous layer to select the
best node neuron  [ ], which will have some output signal  ̂[ ]. The sequential movement from this
neuron (node) back takes place along its connections and passes through all previous layers, which
makes it possible to build the resulting structure of the GMDH-neo- fuzzy network.
      </p>
      <p>
        As a result, due to the GMDH algorithm, it is possible to obtain a well-trained network with an
optimal structure that was synthesized automatically. High-dimensionality problems, as well as
vanishing or exploding gradients, are avoided because the learning is sequential layer-by-layer.


[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]). Among these  1 ∗
2.1.
      </p>
    </sec>
    <sec id="sec-4">
      <title>The role of the Neo-fuzzy neuron in the hybrid GMDH system</title>
      <p>
        In Fig. 2 shows the architecture of the node selected as the quality for the proposed GMDH
system. This is a neo-fuzzy neuron (NFN) proposed by Takeshi Yamakawa et al. in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. It is a
Using a conventional stochastic gradient descent algorithm, it can be minimized.
      </p>
      <p>
        In the case of a predefined dataset, the training process can be performed in a single epoch in batch
mode. For this purpose, the conventional least squares method is used [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]
and fuzzy inference is realized in the form: if   is   then the output is   ,where   is the synaptic
weight in the consequent,   is a fuzzy set whose membership function is   [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
2.2.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Neo-fuzzy neuron training algorithm</title>
      <p>
        The standard local quadratic error function is used as the goal function (i.e., the learning criterion):
 ( )=
( ( )−  ̂( )) =  ( )2 =
( ( )− ∑
∑     (  ( )))
non-linear system that has one output and several inputs. In the proposed GMDH system,
neo-fuzzy-neurons with only two inputs are used, which implements the following mapping:
of
where  ̂ is the output of the system,   is the input signal  ( = 1,2, … ,  ). The nonlinear synapses
  are the building blocks of a neo-fuzzy neuron. Their task is to convert the input signal in the form
 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]( )= (∑  [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]( ) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] ( ))
      </p>
      <p>
        ∑  [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]( ) ( )=  [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]( )∑  [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]( ) ( )
where (•)+ denotes the pseudo-inverse of the Moore-Penrose (y(k) is assumed to be the real value of
the external reference signal).
      </p>
      <p>1
2

 =1
2
1
2
1
2
+ 
 =1
2 ℎ
 =1  =1</p>
      <p>2

 =1
(1)
(2)
(3)
(4)
3. Data set
,
(5)
  ( − 1)( ( )− (   ( − 1))   ( ( )))  ( ( ))</p>
      <p>1 + (  ( ( )))   ( − 1)  ( ( ))
  ( − 1)  ( ( ))(  ( ( )))   ( − 1)</p>
      <p>1 + (  ( ( )))   ( − 1)  ( ( ))</p>
      <p>.</p>
      <p>With the sequential receipt of training observations, i.e., in the online mode, the recurrent form of
the ANN can be represented as</p>
      <p>As the data set for forecasting were taken monthly Electricity Sales to Ultimate Customers,
Residential (USA) since 01-2002 till 01-2023 taken. The whole sample consisted of 251 instances.
The sample was divided into training and test subsamples. The dynamics of monthly energy power
supply to Ultimate Customers, Residential (USA) is shown in the Fig. 3.
preceding and conceding values and the process is periodical with period 6 months. Autocorrelation
function (ACF) was determined for this process of power supply which is shown in the Fig. 5.
The check for stationarity of this process was performed using Dickey-Fuller test.</p>
      <p>P-value: 0.5117527467140699 &gt; 0.05. As it follows from this test the initial time series is not
stationary. Using differencing this time series was transformed to the stationary one that’s confirmed
by Dickey-Fuller test with P-value: 1.3594288749888985e-14 &lt; 0.05.</p>
    </sec>
    <sec id="sec-6">
      <title>4. Experimental investigations</title>
      <p>In the investigations was explored the forecasting accuracy of hybrid DL neo-fuzzy networks at
various forecasting intervals: short-term forecasting with intervals 1, 3 5 and 7 days and middle term
forecasting with intervals 10 and 20 days. At the first step the variable experimental parameters of
hybrid network were chosen which are presented in the Table 1.</p>
      <p>The optimization of these parameters was performed, in result the following optimal values were
determined inputs: 3; linguistic variables: 3; ratio: 0,7.</p>
      <p>After that the structure optimization of hybrid DL neo-fuzzy network was constructed using
GMDH method. The process of structure generation is presented in the Table 2.</p>
      <p>In result the optimal structure of three layers was determined: at the first layer 3 inputs, second
layer – two neurons, third layer – one output neuron.</p>
      <p>Further the training of the best hybrid network was carried out using method SGD (stochastic
gradient descent with variable step. Flow chart of forecasting results for interval 1day in presented in
the Fig. 6. The values of MSE and MAPE for this experiment are shown in the Table 3.</p>
      <p>In the Fig. 6. flow chart of MAPE values for the best model of hybrid network is shown.</p>
      <p>Further the similar experiments of hybrid network were performed with forecasting interval 3, 5,
10 and 20 days. After optimization the parameters and structure of hybrid network it was trained
using training subsample. The forecasting accuracy at the test sample for interval 3 days is presented</p>
      <sec id="sec-6-1">
        <title>Criterion Min: Avg: Max:</title>
        <p>SB 1
0.138
0.23
0.116
at the Table 4. In the succeeding experiments forecasting accuracy of Hybrid neo-fuzzy network was
investigated with forecasting intervals 5, 10 and 20 days.</p>
        <p>In the Table 5 accuracy of forecasting of the hybrid NFN optimal structure is presented with
forecasting interval 10 days and in the table 6 with forecasting interval 20 days.</p>
      </sec>
      <sec id="sec-6-2">
        <title>Criterion Min: Avg: Max:</title>
      </sec>
      <sec id="sec-6-3">
        <title>Criterion Min: Avg: Max:</title>
      </sec>
      <sec id="sec-6-4">
        <title>Criterion Min: Avg: Max:</title>
        <p>For estimating forecasting accuracy of hybrid DL network, it was compared with alternative
methods: ARIMA and GMDH. The forecasting accuracy of GMDH for interval 1 day is shown in the
Table 7 and for 5 days in the Table 8. The flowchart of forecasting results for the interval 5 days is
shown in the Fig. 7 and for 20 days in the Fig. 8.</p>
        <p>In the next experiments forecasting efficiency of method ARIMA was investigated and analyzed.
After the preliminary investigations the optimal parameters for ARIMA were found which were used
in the following experiments. The forecasting accuracy of ARIMA for interval 1 day is presented in
the Table 9 and for interval 5 days in the Table 10. The flowchart of real and forecasting results for
ARIMA with interval 20 days is shown in the Fig. 9.</p>
        <p>The comparative experiments were performed in which the accuracy of forecasting by hybrid DL
network, GMDH and ARIMA at the different forecasting intervals was estimated and compared. The
corresponding results are presented in the Tables 11, 12 and Fig. 10, 11.</p>
        <p>Analyzing the presented results in the Fig. 10 and 11 one may conclude that GMDH method
appears to be the best at short term forecasting 1, 3 days which complies the theory.</p>
        <p>Hybrid deep learning neo-fuzzy networks are the best at middle-term forecasting 5, 7, 10, 20 days.
ARIMA appeared to be the worst by accuracy as compared with intelligent methods – hybrid DL
networks and GMDH.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>5. Conclusion</title>
      <p>In this paper the investigations of artificial intelligence methods: hybrid Deep learning networks
and GMDH and ARIMA were carried out in the problem of forecasting Electricity Sales to Ultimate
Customers, Residential (USA) since 01-2002 till 01-2023.</p>
      <p>During the experiments the optimal structure and optimal parameters: number of inputs, number of
linguistic values, ratio training/test samples of hybrid neo-fuzzy networks were determined.</p>
      <p>After optimization of hybrid neo-fuzzy networks and parameters of GMDH method the
experiments on forecasting Electricity Sales to Ultimate Customers, were performed at different
intervals: 1, 3, 5, 7 (short term forecast) and 10, 20 days (middle term forecast).</p>
      <p>The accuracy of forecasting by Hybrid DL networks was compared with alternative methods –
GMDH and ARIMA.</p>
      <p>The analysis of obtained results have shown that GMDH method is the best at short term
forecasting 1, 3 days while hybrid deep learning neo-fuzzy networks are the best at middle-term
forecasting 7, 10, 20 days. Method ARIMA appeared to be the worst by accuracy as compared with
intelligent method – hybrid DL networks and GMDH.</p>
    </sec>
    <sec id="sec-8">
      <title>6. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Brockwell</surname>
            ,
            <given-names>Peter J</given-names>
          </string-name>
          . Introduction to time series and forecasting / Peter J. Brockwell and
          <string-name>
            <given-names>Richard A.</given-names>
            <surname>Davis</surname>
          </string-name>
          . - 2nd ed. Springer,
          <year>2002</year>
          . - 429 p.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Robert</surname>
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Shumway</surname>
            ,
            <given-names>David S.</given-names>
          </string-name>
          <string-name>
            <surname>Stoffer</surname>
          </string-name>
          .
          <source>Time Series Analysis and Its Applications With R Examples. 4 th edition</source>
          . Springer,
          <year>2017</year>
          . - 562 p.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Mueller</surname>
          </string-name>
          ,
          <article-title>Self-organization of the neural networks with active neurons, Pattern Recognition</article-title>
          and
          <string-name>
            <given-names>Image</given-names>
            <surname>Analysis</surname>
          </string-name>
          ,
          <year>1994</year>
          ,
          <volume>4</volume>
          , №2, pp.
          <fpage>177</fpage>
          <lpage>188</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Wuensch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. A.</given-names>
            <surname>Ivakhnenko</surname>
          </string-name>
          ,
          <article-title>Inductive sorting-out GMDH algorithms with polynomial complexity for active neurons of neural networks</article-title>
          ,
          <source>Neural Networks</source>
          ,
          <year>1999</year>
          , 2, pp.
          <fpage>1169</fpage>
          <lpage>1173</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Haykin</surname>
          </string-name>
          ,
          <article-title>Neural networks: a comprehensive foundation</article-title>
          , 2nd ed. Upper Saddle River,
          <string-name>
            <surname>N.J</surname>
          </string-name>
          : Prentice Hall,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Osowski</surname>
            <given-names>S.</given-names>
          </string-name>
          <article-title>Neural networks for processing information</article-title>
          . - Publishing House of the University of Technology Warszawska, Warszawa, (in Polish) -
          <year>2000</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Wang</surname>
            <given-names>F</given-names>
          </string-name>
          .
          <source>Neural Networks Genetic Algorithms and Fuzzy Logic for Forecasting // Proc. Intern. Conf. Advanced Trading Technologies</source>
          . - New York,
          <year>1992</year>
          . - pp.
          <fpage>504</fpage>
          <lpage>532</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Yamakawa</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uchino</surname>
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miki</surname>
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kusanagi H</surname>
          </string-name>
          .
          <article-title>A neo-fuzzy neuron and its applications to system identification and prediction of the system behavior //</article-title>
          <source>Proc. 2nd Intеrn. Conf. Fuzzy Logic and Neural Networks «LIZUKA-92». - Lizuka</source>
          ,
          <year>1992</year>
          . - pp.
          <fpage>477</fpage>
          <lpage>483</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>I.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          , Deep Learning, MIT PRESS,
          <year>2016</year>
          . URL: http://www.deeplearningbook.org.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Yuriy</surname>
            <given-names>Zaychenko</given-names>
          </string-name>
          , Yevgeniy Bodyanskiy, Oleksii Tyshchenko, Olena Boiko,
          <string-name>
            <given-names>Galib</given-names>
            <surname>Hamidov</surname>
          </string-name>
          .
          <article-title>Hybrid GMDH-neuro-fuzzy system and its training scheme</article-title>
          .
          <source>Int. Journal Information theories and Applications</source>
          ,
          <year>2018</year>
          . vol.
          <volume>24</volume>
          , Number 2. - pp.
          <fpage>156</fpage>
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Evgeniy</surname>
            <given-names>Bodyanskiy</given-names>
          </string-name>
          , Yuriy Zaychenko, Olena Boiko, Galib Hamidov,
          <string-name>
            <given-names>Anna</given-names>
            <surname>Zelikman</surname>
          </string-name>
          .
          <article-title>Structure Optimization and Investigations of Hybrid GMDH-Neo-fuzzy Neural Networks in Forecasting Problems</article-title>
          .
          <source>System Analysis &amp; Intelligent Computing. Ed. Michael Zgurovsky</source>
          ,
          <string-name>
            <given-names>Natalia</given-names>
            <surname>Pankratova</surname>
          </string-name>
          .
          <source>Book Studies in Computational Intelligence, SCI</source>
          , vol.
          <volume>1022</volume>
          . Springer,
          <year>2022</year>
          . - pp.
          <fpage>209</fpage>
          <lpage>228</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>