<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>International Journal of Forecasting 37.2
(2021) 547</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1016/j.ijforecast.2020.07.005</article-id>
      <title-group>
        <article-title>Improvement  of  time  series  forecasting  quality  by  means  of  multiple models prediction averaging </article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Denis Petrusevich</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>MIREA - Russian Technological University</institution>
          ,
          <addr-line>Moscow, 119454, Prospekt Vernadskogo, 78</addr-line>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>19</volume>
      <fpage>308</fpage>
      <lpage>319</lpage>
      <abstract>
        <p>   Construction of time series models is usually based on Akaike and Bayes information criteria. Requirement of model simplicity is built inside the information criteria structure and the best fitted models aren't always best in terms of criteria values. Often there are a few best models that fit investigated time series well and there's problem of choice between them. Usually information criteria values allow to choose among them. But if one needs the best model by forecasts quality, best fitted models, or there are other thoughts a research has to choose and test models manually. At the same time when a few models are chosen it's possible to construct their combination. The simplest way is to count mean value of their forecasts and to use it as a combined prediction. Practical researches confirm that forecast error gets lower in this approach. Also, more complex construction than averaging of forecasts can be used (for example, weighted voting that is widely used in bagging technique in solution of classification problems). But this approach hasn't got enough theoretical base. From theoretical point of view confidence intervals of time series forecasts (usually here they're considered as prediction intervals) is also very complex task. Intervals tend to be very wide even for models with good prediction quality in terms of mean forecast errors. Thus, prediction intervals are rare in use. In this paper a few time series models are constructed for wage and income indices of Russian macroeconomic time series. Their predictions are combined into one forecast and it's quality is compared to individual ones. Transformation of forecast variance and prediction intervals in case of simple (moving averages MA(q) and autoregressions AR(p) of low order) models is also considered but is a part of further work.</p>
      </abstract>
      <kwd-group>
        <kwd> 1  Time series forecasting</kwd>
        <kwd>prediction averaging</kwd>
        <kwd>ARIMA</kwd>
        <kwd>information criteria</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction </title>
      <p>Time series modeling is usually based on information criteria evaluation (Bayes or Akaike criteria)
in order to choose the best model for investigated time series [1-3]. In practice often a few models have
got good criteria values and it’s difficult to choose the best one. At the same time it’s possible to make
averaging of forecasts made with each model. Such technique is investigated in [4-7] and it has got
good results. But this approach isn’t confirmed well theoretically. Prediction quality is usually evaluated
with confidence intervals (for example, in case of linear regression model). But in case of time series
investigation usual 95% prediction intervals are too wide and are difficult for evaluation [1, 4, 8, 9].
Thus, there’s no theoretical connection between forecast averaging and narrowing of prediction
intervals. If used models are of the same type their prediction intervals remain the same and prediction
interval of combined model isn’t better. Also, it should be mentioned that prediction intervals behaviour
in case of averaging of different models (for example, ARIMA and GARCH models) still isn’t
investigated. In this research practical experiments on time series models averaging are presented.
Prediction intervals behaviour for simple models (autoregression AR(p) models of low order, p&lt;3, and
moving average MA(q) models) and for their combination are in the scope of current research program
and is going to be under further investigation.
2. Time series forecast errors and prediction intervals </p>
      <p>
        One of the most widely used time series models is ARIMA that’s constructed out of two parts: the
p-order autoregression AR(p) model and the q-order moving average MA(q) model [1, 9]. They are
used in combination with time series integration technique in order to get stationary time series. These
models have got wide prediction intervals. They usually have got complex form (for example, in case
of autoregression models AR) and are difficult to evaluate. In order to evaluate forecasts often mean
square errors are used (formula (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ):
      </p>
      <p>
         
or root mean square errors are implemented (formula (
        <xref ref-type="bibr" rid="ref2">2</xref>
        )):
      </p>
      <p>MSE  t
,
  </p>
      <p>
                                                             (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) 
 ( (t)  ts(t))2
      </p>
      <p>N
 ( (t)  ts(t))2</p>
      <p>t
 </p>
      <p>RMSE </p>
      <p>
        .                                                               (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) 
      </p>
      <p>N</p>
      <p>Here τ(t) denotes predicted values of the processed time series, ts(t) is real values of the investigated
time series, t enumerates all predicted time points and N is their quantity.</p>
      <p>
        In practice also mean absolute error of forecast is used (formula (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ):
      </p>
      <p> | (t)  ts(t) |
   MAE  t .   </p>
      <p>
                                                             (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) 
      </p>
      <p>N</p>
      <p>Thus, quality of forecasts at certain period is used instead of prediction intervals evaluation. At the
same time there’s idea of averaging forecasts made by means of a few models. If all the models make
good forecasts it’s obvious that combination of their forecasts should behave well also [4-7, 10]. But
this result needs more thorough check and strict confirmation. Also, idea of connection between forecast
quality improvement and prediction intervals narrowing should be tested. If forecasts become better, it
should have some influence at prediction interval. Evaluation of prediction intervals of time series
forecasters has been in scope of scientific research for a long time [8].</p>
      <p>The idea to select the best models in some set their by predictions evaluating is considered in [11].
But there the best model is chosen. In this research predictions of a set of models are combined into one
forecast. Prediction intervals should also be evaluated as well as prediction quality. Implementation of
non-linear combination forecasters [12] is an idea of next stage of research in this area comparing to
linear combination technique used in [4-7]. Competitive selection of best models by their prediction
intervals quality is under investigation of [13]. So, this problem is in scope of modern research.</p>
      <p>The tested method is close to bagging technique. In the classification problem at which bagging is
used very often there can be constructed a lot of weak classifiers. Their accuracy must be higher than
50% but isn’t supposed to be very high. This technique constructs strong classifier (with high accuracy)
out of a lot of weak ones. Each weak classifier is constructed using its own training set and test set, its
own set of variables (features). One of main examples of bagging principle implementation is random
forest classifier. This approach also takes place in combining linear regression models into a new one
with higher accuracy [14].</p>
      <p>
        In this research quality of combination constructed out of ARIMA models [1, 9] is under
investigation: their RMSE and MAE values (formulae (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) and (
        <xref ref-type="bibr" rid="ref3">3</xref>
        )). First of all, mean value of all
predictors’ forecasts is counted:
      </p>
      <p> f (t)
 (t  1)  N</p>
      <p>.   </p>
      <p>N</p>
      <p>Here τ(t+1) denotes predicted value of the processed time series, f(t) are forecasters’ predictions and
N is their quantity.</p>
      <p>
         
                                                                 (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) 
      </p>
      <p>
        At the same time voting technique used in random forests classifiers [14] is tested. At each
experiment time series forecaster makes some error (difference between forecast at time point and its
real value). Level of confidence to each model is constructed into the model (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ). Weight w of model’s
vote is increased if its prediction is close to real value and it’s decreased if other models make better
predictions:
 (t 1)   wi fi (t),
      </p>
      <p>
        N
                                                 
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) 
 wi  1.
      </p>
      <p>K</p>
      <p>Here w is weight of each vote. Weights change at training stage. Model’s weight is increased when
its forecast is close to reality and is decreased otherwise.</p>
      <p>
        Classifiers used in [14] should be trained at different sets of data. In time series models this
requirement can be reinterpreted in use of different training periods for each time series model in
combination and also in implementation of different by structure time series predictors. If ARIMA (p,
d, q) models are considered, models with varying p and q parameters should be used. Thus, models with
close orders p and q are going to make close predictions and they can be considered as dependent. If
models have got different structure (varying p and q in large intervals) these models’ predictions tend
to independence. Also, predictions of different models by type (for example, GARCH and ARIMA)
should have good results and are supposed to be implemented. But it’s task for further investigation
because prediction intervals in such combinations can’t be constructed with usual statistics. There’s
some research that can be used as basis for GARCH models testing in this approach [15]. Is one takes
into account that information criteria values depend not only on model quality (as difference between
model’s prediction and real value) but requirement of model simplicity is also used [1], other metrics
of time series quality can be used: for example, only likelihood function value of each time series. At
this stage even game-based techniques [16, 17] can be used.
                                                       
(
        <xref ref-type="bibr" rid="ref7">7</xref>
        ) 
 
 
      </p>
    </sec>
    <sec id="sec-2">
      <title>3. Experiments </title>
      <p>
        The Dynamic series of macroeconomic statistics of the Russian Federation (monthly wage index
and income index) [18] have been handled in the experiment section. Last 12 values were used as test
period predicted by models. Good ARIMA (p, d, q) models by RMSE and MSE of their forecasts were
chosen to construct combined models. These models aren’t trained with regard to information criteria
values or likelihood function level. Their coefficients are mean values of appropriate coefficients of the
chosen set of models (in accordance with expression (
        <xref ref-type="bibr" rid="ref6">6</xref>
        )). Also mean value of all chosen models’
predictions is obtained and compared to real values.
3.1. ARIMA  (p,  d,  q)  models  of  wage  and  income  indices  of  Russian 
macroeconomic statistics and combination of models  
      </p>
      <p>
        The ARIMA (p, d, q) models of orders p&lt;4 and q&lt;4 have been tested for the monthly income index
[18] of Russian macroeconomic statistics. According to automatic fitting function that’s based on time
differentiation and on choice of the best model by means of information criteria values [1, 9], the
ARIMA (
        <xref ref-type="bibr" rid="ref1 ref2 ref2">2, 1, 2</xref>
        ) is best fitted model. In this experiment other “good models” are defined as models
that have got good forecasts. At the same time it’s necessary to select models with varying structure.
So, if there are two models with orders differing only by 1 and they’ve got close values of coefficients,
only one of these models is selected for further experiment. These models are presented in the Table 1.
Columns contain Akaike information criterion values, RMSE and MAE of forecasts at period of 12
months.
      </p>
      <sec id="sec-2-1">
        <title>Table 1  </title>
        <p>The ARIMA (p, d, q) models of the income index </p>
        <p>ARIMA(p, d, q) models </p>
        <p>
          It’s clearly seen that result of automatic fitting procedure (ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2 ref2">2, 1, 2</xref>
          )) isn’t the best model
according to errors of forecasts. The best model according to automatic fitting procedure is marked with
bold font. Mean forecast of all these models’ predictions is presented in the end of the table. It’s better
than forecasts of some other models and it can be used in practice. Quality of the forecast made by the
combined model is shown in the last row. It’s much closer to the real values than all other models shown
above.
        </p>
        <p>Appropriate models of wage index [18] are shown in the Table 2.</p>
        <p>ARIMA(0, 1, 1) 
ARIMA(0, 1, 3) 
ARIMA(1, 1, 2) 
ARIMA(2, 1, 2) 
ARIMA(3, 1, 1) </p>
        <p>Mean forecast </p>
        <p>Combined model </p>
      </sec>
      <sec id="sec-2-2">
        <title>Table 2  </title>
        <p>The ARIMA (p, d, q) models of the wage index </p>
        <p>ARIMA (p, d, q) models </p>
        <p>ARIMA(0, 1, 3) 
ARIMA(1, 1, 1) 
ARIMA(1, 1, 2) 
ARIMA(2, 1, 1) 
ARIMA(2, 1, 2) 
ARIMA(3, 1, 2) </p>
        <p>Mean forecast 
Combined model </p>
        <p>RMSE of 
forecast 
30.12 
30.59 
30.21 
30.29 
30.47 
30.22 
26.36 
RMSE of 
forecast 
21.60 
22.63 
22.90 
22.25 
21.43 
21.54 
20.93 
20.64 </p>
        <p>MAE of forecast </p>
        <p>Here the combined model and mean forecast are closer to real values than other ARIMA models.
Mean forecast has got the best forecast according to MSE value and the combined model’s forecast is
the best according to RMSE value.</p>
        <p>
          Plots of predictions made by the best fitted model ARIMA(
          <xref ref-type="bibr" rid="ref1 ref2 ref3">3, 1, 2</xref>
          ), the combined model, mean
forecast of five models shown in the Table 2 and real values are presented as Figure 1:, Figure 2:, Figure
3:, Figure 4:.
3.2.
        </p>
        <p>
          Combination of ARIMA (p, d, q) models trained with voting technique  
The voting technique based on expression (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ) is also used analyzing the monthly wage index [18]
of Russian macroeconomic statistics. 2/3 of training set was used to construct ARIMA models. Five
“good” models were chosen to use them at the next stage. Then equal weights for their votes were added
into the model. Last 1/3 part of the training set was used to implement voting technique. All chosen
models made prediction for one month ahead. Their predictions were compared to the real data. Weight
of the model with the best prediction was increased. Other weights were decreased. Their sum must
equal to 1. At the same time it’s necessary to control that weights shouldn’t become less than zero.
        </p>
        <p>
          In case of the income index the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref1">1, 1, 0</xref>
          ), ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2">2, 1, 0</xref>
          ), ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">3, 1, 0</xref>
          ) models were
chosen. Weights of the combination: 0.15 for the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref1">1, 1, 0</xref>
          ) model, 0.03 for the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2">2, 1,
0</xref>
          ) model and 0.82 for the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">3, 1, 0</xref>
          ) one. This result itself can be used for evaluation of models
and their predictions.
        </p>
        <p>
          Unfortunately, variance of long-term prediction is very high and it’s impossible to compare
constructed model with models of previous section trained at 100% of data. But one can compare it
with the models used in this combination. Though in the «voting» model data of the predicted period
have been implemented as training set. So, evaluation method for such models is still to be discussed
and constructed. Forecasts have been made for 12 months ahead. Their comparison is shown in the
Table 3.
In the voting model weight of the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">3, 1, 0</xref>
          ) model is maximal and they give close predictions.
Their forecasts are even better than the automatically fitted function prediction.
        </p>
        <p>
          In case of the wage index the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">0, 1, 3</xref>
          ), ARIMA (
          <xref ref-type="bibr" rid="ref1 ref1 ref3">1, 1, 3</xref>
          ), ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2">2, 1, 0</xref>
          ), ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2 ref2">2, 1,
2</xref>
          ) and ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">3, 1, 0</xref>
          ) models were chosen. Because of “winner takes it all” principle usage at the end
of the voting training stage there are only two models with non-zero weights: weight of the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref3">0,
1, 3</xref>
          ) is about 0.2 and weight of the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2">2, 1, 0</xref>
          ) model is about 0.8. Comparison of the voting
model for the wage index with the chosen models are shown in the Table 4 (the best fitted model hasn’t
been shown because that’s ARIMA (
          <xref ref-type="bibr" rid="ref1 ref1">0, 1, 1</xref>
          ) and it can’t make stable predictions).
        </p>
        <p>One can conclude that this voting model has shown the best result. Usually conclusions made by
researcher are based on consideration of a lot of models and of their forecasts. So, such approach can
give one more model in such set and here this model has become the best one by prediction quality.</p>
        <p>
          Predictions of the ARIMA (
          <xref ref-type="bibr" rid="ref1 ref2">2, 1, 0</xref>
          ) model, voting model at period of 12 months and real values of
the wage index at the predicted period are shown at Figure 5:Figure 6:,Figure 7:.
        </p>
        <p>Figure 5: Prediction of the wage index by the model ARIMA (2, 1, 0) (12 months) 
Figure 6: Prediction of the voting model (12 months) 
Figure 7: Real values of the wage index at the predicted period </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Conclusion </title>
      <p>In this research three ways of combining ARIMA models into a new one is investigated. The main
goal is to construct a new model that’s going to make better forecasts or predictions of the same quality.
Russian macroeconomical statistics was used in the experimental part of the research.</p>
      <p>The first method of averaging is just constructing mean forecast out of predictions of a few good
models. Their quality is confirmed with prediction at test period. This approach has been tested in the
first experiment and it has got good results that are close to the best models.</p>
      <p>The second method is averaging of the best models. The model with maximal orders p and q inside
of set of the best ones (ARIMA (p, d, q) models) is constructed. Its coefficients are mean values of
appropriate coefficients of this set. Combined model makes the best prediction in case of income index
and good one in case of the wage index.</p>
      <p>The third type of the models is based on implementation of voting technique. First of all, good
models are chosen while comparing quality of their predictions. At the first stage of votes evaluation
all of them are equal and this model is the same like in the second case. But then if a model has got
forecast closest to reality its weight is increased, weights of other models are decreased. Here, only
winners’ weight is increased. But in further research “soft” methods implementation (like softmax
method) are going to be investigated. Combined models are the best ones by prediction quality in the
both experiments.</p>
      <p>Also, prediction intervals of combined models are in scope of further research. One can conclude
that combination of models has got forecast of the same quality or even better. This conclusion is
confirmed in [4-7] and in this paper. But narrowing of prediction intervals for combination of models
is still investigated [12, 13] and is in the scope of further research.</p>
      <p>Forecasting with combination of time series models is close to bagging technique [14] used in
classification and regression tasks. But there are requirements to weak classifiers. Analogical
requirements should be formulated and tested for time series models combined into a set. This problem
is going to be under investigation in further work.</p>
    </sec>
    <sec id="sec-4">
      <title>5. References </title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Hyndman</surname>
          </string-name>
          , G. Athanasopoulos,
          <article-title>Forecasting: principles and practice, 2nd edition</article-title>
          , OTexts: Melbourne, Australia,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E. G</given-names>
            <surname>Andrianova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Golovin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. V.</given-names>
            <surname>Zykov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Lesko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. R.</given-names>
            <surname>Chukalina</surname>
          </string-name>
          ,
          <article-title>Review of modern models and methods of analysis of time series of dynamics of processes in social, economic and socio-technical systems</article-title>
          ,
          <source>Russian Technological Journal (In Russ) 8</source>
          .4 (
          <issue>2020</issue>
          )
          <fpage>7</fpage>
          -
          <lpage>45</lpage>
          . doi:
          <volume>10</volume>
          .32362/
          <fpage>2500</fpage>
          -316X-2020-8
          <issue>-4</issue>
          -7-45.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Petrusevich</surname>
          </string-name>
          ,
          <article-title>Time series forecasting using high order ARIMA functions</article-title>
          ,
          <source>in: Proceedings of the International Multidisciplinary Scientific GeoConference: SGEM</source>
          <year>2019</year>
          , Volume
          <volume>19</volume>
          (
          <issue>2</issue>
          .1), Sofia, Bulgaria,
          <year>2019</year>
          , pp.
          <fpage>673</fpage>
          -
          <lpage>679</lpage>
          . doi:
          <volume>10</volume>
          .5593/sgem2019/2.1/S07.088.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>F.</given-names>
            <surname>Petropoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Hyndman</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Bergmeir, Exploring the sources of uncertainty: Why does bagging for time series forecasting work?</article-title>
          <source>European Journal of Operational Research 268.2</source>
          (
          <year>2018</year>
          )
          <fpage>545</fpage>
          -
          <lpage>554</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.ejor.
          <year>2018</year>
          .
          <volume>01</volume>
          .045.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Dal Molin Ribeiro</surname>
          </string-name>
          , L. dos
          <string-name>
            <surname>Santos</surname>
            <given-names>Coelho</given-names>
          </string-name>
          ,
          <article-title>Ensemble approach based on bagging, boosting and stacking for short-term prediction in agribusiness time series</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>86</volume>
          (
          <year>2020</year>
          )
          <article-title>105837</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.asoc.
          <year>2019</year>
          .
          <volume>105837</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Bagging based ensemble learning approaches for modeling the emission of PCDD/Fs from municipal solid waste incinerators</article-title>
          ,
          <source>Chemosphere</source>
          <volume>274</volume>
          (
          <year>2021</year>
          )
          <article-title>129802</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.chemosphere.
          <year>2021</year>
          .
          <volume>129802</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <article-title>Regularized ELM bagging model for Tropical Cyclone Tracks prediction in South China Sea</article-title>
          ,
          <source>Cognitive Systems Research</source>
          <volume>65</volume>
          (
          <year>2021</year>
          )
          <fpage>50</fpage>
          -
          <lpage>59</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.cogsys.
          <year>2020</year>
          .
          <volume>09</volume>
          .005.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chatfield</surname>
          </string-name>
          , Prediction Intervals for Time Series Forecasting, Principles of Forecasting (
          <year>2001</year>
          )
          <fpage>475</fpage>
          -
          <lpage>494</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-0-
          <fpage>306</fpage>
          -47630-3_
          <fpage>21</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Stock</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. W.</given-names>
            <surname>Watson</surname>
          </string-name>
          , Introduction to Econometrics, Pearson,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Sarkisov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. V.</given-names>
            <surname>Lomonosova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. V.</given-names>
            <surname>Zolkina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Sarkisov</surname>
          </string-name>
          ,
          <article-title>Integration of digital technology in mining and metallurgy industries, Tsvetnye Metally (in Russ) 3 (</article-title>
          <year>2020</year>
          )
          <fpage>7</fpage>
          -
          <lpage>14</lpage>
          . doi:
          <volume>10</volume>
          .17580/tsm.
          <year>2020</year>
          .
          <volume>03</volume>
          .01.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lunde</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Nason</surname>
          </string-name>
          ,
          <article-title>Model confidence sets for forecasting models</article-title>
          ,
          <source>Econometrica</source>
          <volume>79</volume>
          .2 (
          <year>2005</year>
          ):
          <fpage>453</fpage>
          -
          <lpage>497</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>W.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <article-title>A novel method for time series prediction based on error decomposition and nonlinear combination of forecasters</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>426</volume>
          (
          <year>2021</year>
          )
          <fpage>85</fpage>
          -
          <lpage>103</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.neucom.
          <year>2020</year>
          .
          <volume>10</volume>
          .048.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>