<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yevgeniy Bodyanskiy</string-name>
          <email>yevgeniy.bodyanskiy@nure.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ivan Izonin</string-name>
          <email>ivan.v.izonin@lpnu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iryna Pliss</string-name>
          <email>iryna.pliss@nure.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olha Chala</string-name>
          <email>olha.chala@nure.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sergiy Popov</string-name>
          <email>serhii.popov@nure.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Kharkiv National University of Radio Electronics</institution>
          ,
          <addr-line>14 Nauky av., Kharkiv, 61166</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>12 Stepana Bandery St, Lviv, 79000</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Adaptive nonlinear neo-fuzzy bagging metamodel and its fast online learning procedure are proposed intended for optimal combining of the results of different computational intelligence systems that simultaneously solve the same problem. It is assumed that the data is processed by the members of the ensemble and by the metamodel online in real time. The proposed metamodel is intended for solving a wide class of Data Stream Mining problems under the conditions of data non-stationarity, and when the processing speed is of the utter importance. Simulation based on the short-term electric load forecasting problem confirmed theoretical results. The metamodel demonstrated significant improvement over the results of the member models.</p>
      </abstract>
      <kwd-group>
        <kwd>adaptive nonlinear bagging</kwd>
        <kwd>neo-fuzzy metamodel</kwd>
        <kwd>optimal combining</kwd>
        <kwd>fast online learning 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Recently, computational intelligence has emerged as a powerful tool across various fields,
particularly in Data Mining. It encompasses a variety of systems such as artificial neural networks,
fuzzy systems, neuro-fuzzy systems, and neo-fuzzy systems. These systems are employed to tackle
an extensive array of tasks. Just to mention a few:</p>
      <p>On the other hand, every task can be addressed using various computational intelligence systems,
each with distinct architectures, training principles, data requirements, and operational speeds,
which leads to different data processing results. For instance:


</p>
      <p>Artificial Neural Networks (ANNs): Known for their high accuracy, especially deep neural
networks, they require vast volumes of training samples (which may not always be feasible
due to data scarcity) and computational resources, i.e. a lot of time and computing power for
their training in multi-epoch mode.</p>
      <p>Fuzzy Systems: These systems are known for their ability to handle imprecise or incomplete
data, making them suitable for real-world problems where data may be uncertain.
Neuro-Fuzzy Systems: Combining the strengths of neural networks and fuzzy logic, these
systems offer both learning capabilities and robustness to noisy data.</p>
      <p>
        Selecting the appropriate computational intelligence system for a given problem is an intricate
task that lacks a formal solution, primarily relying on the empirical knowledge of researchers and
occasional intuition. This process is not trivial due to the vast array of systems available, each with
its own strengths and weaknesses. To address this challenge, one promising approach is bagging [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
where an ensemble of different systems [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ] is employed. These systems operate concurrently but
independently to solve the same problem. Their outputs are then combined through a metamodel –
a higher-level model that integrates these signals into an optimal solution. Typically, this integration
is achieved through a linear combination of individual system outputs.
      </p>
      <p>
        However, nonlinear approaches to bagging have been largely overlooked, with only limited
research available in this domain [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Most existing solutions operate in batch mode, processing data
offline, which can be restrictive in dynamic environments where real-time decision-making is
crucial. Adaptive systems such as [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ] that function online offer a potential solution to this
limitation, capable of responding and adapting as new data arrives.
      </p>
      <p>
        The aggregation of system outputs poses another challenge. While linear combinations have been
extensively studied [
        <xref ref-type="bibr" rid="ref7 ref8 ref9">7-9</xref>
        ], achieving optimal adaptive combinations remains an open area for
research. Introduction of adaptive fast-acting nonlinear metamodels could significantly enhance the
quality of results, yet such innovations are still in their infancy and not widely adopted.
      </p>
      <p>
        In light of these considerations, we propose a novel approach using the neo-fuzzy methodology
[
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10-12</xref>
        ], tailored to meet the demands of adaptive nonlinear bagging. This method is designed for
real-time environments where both adaptability and nonlinearity are essential. By modifying the
neo-fuzzy framework specifically for this purpose, we aim to bridge existing gaps in system selection
and integration, offering a more robust solution to complex problems.
      </p>
      <p>In summary, while choosing the right computational intelligence system remains challenging due
reliance on experience and intuition, innovative approaches like bagging ensembles and adaptive
metamodels offer promising pathways. The exploration of nonlinear solutions, particularly through
neo-fuzzy approaches, holds the potential to advance our ability to effectively combine diverse
systems for improved problem-solving in dynamic environments.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Architecture of the adaptive neo-fuzzy bagging metamodel</title>
      <p>
        The architecture of the adaptive neo-fuzzy bagging metamodel is illustrated in Figure 1. This system
represents an enhanced version of the traditional neo-fuzzy neuron [
        <xref ref-type="bibr" rid="ref10 ref11 ref12">10-12</xref>
        ], which is distinguished
by its strong approximation capabilities and the ability to adjust its parameters dynamically, or
“online.” At the core of this architecture lies a layer composed of nonlinear synapses, which serve as
the primary element within the neo-fuzzy neuron. These synapses essentially perform an operation
known as the F-transform [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], i.e. it is a universal approximator.
      </p>
      <p>The metamodel receives input signals from  ensemble members that work simultaneously to
solve the same problem. Each member provides its output in the form of a time sequence:
 ( ), … ,  ( ), … ,  ( ), where  represents the discrete time index  = 1,2, …
These sequences
are organized into a vector format denoted as  ( ) =  ( ), … ,  ( ), … ,  ( ) , which is then
passed to the layer of nonlinear synapses.
 
 
 
(1)
(2)</p>
      <p>, which are determined through a learning process tailored
to optimize performance.</p>
      <p>At the output of each of the nonlinear synapses 
a signal is formed
 ( ) =
 
 ( ) =  
,

 ( ) = 
 ( ) , … , 
Triangular functions in the following form are most often used as membership functions in
 ( ) =</p>
      <p>( ) − 
⎧
⎪  , − 
⎨
⎪ 
⎩0 otherwise,
, −  ( )
, −  ,
,</p>
      <p>, ,  , ,
, if  ( ) ∈  , , 
where 
, ,  , ,</p>
      <p>, are coordinates of the membership functions centers, which are usually
distributed evenly along the coordinate axes.</p>
      <p>One significant benefit of utilizing triangular functions within fuzzy logic systems lies in their
efficiency during the learning process. When processing incoming observations – denoted as  ( )
– these triangular functions ensure that only two neighboring membership functions are activated
at any given time. This means that, rather than recalculating or adjusting all relevant weights across
the system, which would be computationally intensive, only a specific subset of synaptic weights
needs to be updated.</p>
      <p>By having only two neighboring functions triggered at each step of the learning process, we
minimize the number of adjustments required – specifically, adjusting 2 synaptic weights per
moment. This targeted approach not only reduces computational load but also enhances efficiency,
making real-time processing more feasible. Moreover, this property contributes to scalability since
it limits the complexity that can arise as data volume increases.</p>
      <p>In essence, the use of triangular functions streamlines the learning process by confining
adjustments to a localized set of weights and membership functions, thereby optimizing performance
in dynamic or large-scale applications. This efficiency is crucial for maintaining responsiveness and
accuracy in systems where resources are constrained or rapid data processing is necessary.
 
can increase computational demands, which may impact real-time performance if not carefully
managed.
constraints:</p>
      <p>The output signals of nonlinear synapses 
–  ( ), … ,  ( ), … ,  ( ) are fed to an
additional layer of tuned weights  ∗, … ,  ∗, … ,  ∗ (which is absent in the standard neo-fuzzy
neuron), that are present in most known bagging systems and are subject to unbiasedness
 ∗ =   ∗ = 1,
where  – ( × 1 ) vector formed by ones.</p>
      <p>Signals  ( ) after passing through synaptic weights  ∗ are combined in the output adder,
forming the optimal output signal
 ∗( ) =</p>
      <p>∗ ( ) =  ( ) ∗,
where  ∗ =  ∗, … ,  ∗, … ,  ∗ ,  ( ) =  ( ), … ,  ( ), … ,  ( ) .
(3)
(4)
 ( ) = 
( − 1)
 ∗( ) =  ( ) ∗( − 1),
which, after passing through the set of weights  ∗, are combined in the output adder in the form
where  ( − 1),  ∗( − 1) – estimates obtained based on  − 1 previous observations.</p>
      <p>Adaptive metamodel learning can be implemented based on errors backpropagation. First, the
parameter vector  ∗ is tuned, and then all the parameter vectors of nonlinear synapses  ,  =
1,2, … ,  are tuned.</p>
      <p>A standard quadratic learning criterion can be used to tune the vector  ∗</p>
      <p>Thus, in the process of its training, two sets of parameters are configured in the proposed
metamodel: ( × 1) vector  ∗ and  ( + 1) vectors of parameters of nonlinear synapses  ,  =
1,2, … ,  .</p>
    </sec>
    <sec id="sec-3">
      <title>3. Adaptive neo-fuzzy bagging metamodel learning</title>
      <p>When the signal  ( ) is fed to the metamodel, the following signals are formed at the outputs of
nonlinear synapses
(5)
(6)
(7)
(8)
(10)
(11)
(12)
1
 ∗( ) =</p>
      <p>2
subject to the unbiasedness constraints
 ( ) −  ∗( )
  ∗ = 1,</p>
      <p>( )
 ( ) = ,</p>
      <p>( )‖ ( )‖ −  ( − 1)  ( )
we get the optimal learning procedure in terms of speed
where  ( ) – external reference signal, also used for all ensemble members training.</p>
      <p>To solve the problem, quadratic programming can be used to find the saddle point of the Lagrange
function
 ( ∗,  ) = 1 ( ( ) −  ( ) ∗) +    ∗ − 1 , (9)</p>
      <p>2
where  is an undefined Lagrangian multiple, which in this case is also a tuned parameter.</p>
      <p>To find a saddle point, it is convenient to use the Arrow-Hurwitz algorithm, which in this case
takes the form
 ∗( ) =  ∗( − 1) −  ( )∇ ∗ ( ∗,  ),
 ( ) =  ( − 1) +  ( )  ( ∗,  )⁄  ,
or
 ∗( ) =  ∗( − 1) +  ( )  ( ) ( ) −  ( − 1) ,
 ( ) =  ( − 1) +  ( )   ∗( ) − 1 ,
where  ( ) =  ( ) −  ( ) ∗( − 1) – learning error,  ( ),  ( ) – search step parameters
that determine the convergence rate of the search process.</p>
      <p>
        The process of tuning the parameter vector  ∗ can be optimized for speed by appropriately
choosing the step  ( ). By specifying this step in the form [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]
=
1
2
1
2
 ( ) −
 ∗( )
      </p>
      <p>,
 ( ) =
 ( ) −  ( ) ∗( )
=
 ( ) −
 ∗( ) ( )</p>
      <p>=
1
2
 ∗( ) =</p>
      <p>( − 1) ∗  ( ) .
 ( ) =  ( − 1) +
 ( ) −</p>
      <p>
        ( − 1) ∗  ( )
 ∗  ( )
 ∗  ( )
To adjust the parameters vector  ( ), the same optimal Kaczmarz-Widrow-Hoff algorithm can
or its modification with additional smoothing properties, which has proven effective in training
be used in the form
neo-fuzzy neurons [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]
(13)
(14)
(15)
(16)
(17)
in this case, during the learning process, 
parameters of nonlinear synapses 
,  = 1,2, … , 
must be determined.
      </p>
      <p>Introducing (</p>
      <p>× 1) vector of tuned parameters of all nonlinear synapses  ( ) =

( ), … , 
( ), … , 
( ), 
( ), … , 
( ), … , 
( ), 
( )
and a modified vector
formed
 ∗( )
 ∗( )
by
all
membership
functions
and
adjusted
output
weights  ∗  ( ) =
 ( ) , … ,  ∗( )
 ( ) ,  ∗( )
 ( ) , … ,  ∗( )
 ( ) , … ,  ∗( )</p>
      <p>, it is easy to write the transform implemented by the
metamodel with the adjusted parameters vector  ∗( ) in the form
 ∗( ) =  ∗( − 1) +
 ( ) =  ( − 1) +  ( )   ∗( ) − 1 .</p>
      <p>( )  ( ) ( ) −  ( − 1)
 ( )‖ ( )‖ −  ( − 1)  ( )
,</p>
      <p>
        It is easy to see that in the absence of unbiasedness constraints, this procedure completely
coincides with the optimal one-step Kaczmarz-Widrow-Hoff algorithm [
        <xref ref-type="bibr" rid="ref17 ref18">17, 18</xref>
        ].
      </p>
      <p>After adjusting the output parameters vector  ∗( ), we can proceed to adjusting the nonlinear
synapses parameters based again on the quadratic criterion
 ( ) =  ( − 1) + 
( )  ( ) −</p>
      <p>( − 1) ∗  ( )  ∗  ( ) ,
 ( ) =  ( − 1) +  ∗  ( )</p>
      <p>, 0 ≤  ≤ 1,
where  is a smoothing parameter defining a compromise between the filtering and tracking
properties of the learning algorithm.</p>
      <p>The proposed training procedure for the adaptive neo-fuzzy nonlinear bagging metamodel is
designed for online information processing when data is received by the system in real time.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Simulation results</title>
      <p>We applied the proposed bagging approach to the short-term electric load forecasting problem
(STLF), specifically focusing on 1-step ahead forecasting of daily electric load in one of Ukraine’s
regional power systems. To see, why bagging is the approach of choice to solve STLF problems, let’s
start with a short overview of STLF itself.</p>
      <sec id="sec-4-1">
        <title>4.1. Overview of short-term electric load forecasting</title>
        <p>Electric load forecasting is a critical task for utility companies, enabling them to manage electricity
generation and distribution efficiently. Among various forecasting horizons, one-day ahead
(onestep) forecasting stands out as particularly challenging due to its dynamic nature and the need for
real-time accuracy.</p>
        <p>Here are the most common challenges inherent to short-term electric load forecasting.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.1.1. Data complexity</title>
        <p>The first challenge lies in the sheer volume and diversity of data that must be processed to generate
accurate forecasts. Historical consumption patterns provide a foundation for predictions, but this
dataset is augmented by numerous other variables like weather conditions, economic indicators,
calendar events, grid conditions, etc. The integration of these diverse data points is essential for
accurate predictions but presents a significant challenge due to their different measurement scales,
varying nature, and potential inconsistencies.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.1.2. Dynamic nature of demand</title>
        <p>Electricity demand fluctuates continuously, shaped by human activities such as turning on
appliances, working schedules, and leisure time consumption patterns. This dynamic behavior makes
it difficult to predict with high precision even just one day ahead. Difficulties arise from time-of-day
variations, weekday vs. weekend pattern differences, and seasonal changes.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.1.3. Nonlinear relationships</title>
        <p>
          The relationship between various factors influencing electricity demand is often nonlinear and
complex. Traditional statistical models, which rely on linear relationships, may struggle to capture
these nuances effectively. The impact of temperature on load isn’t always directly proportional; there
can be saturation points beyond which further changes in temperature don’t significantly affect
demand [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. The combined effect of weather and economic conditions might not be merely additive
but could interact in more complex ways. There are other sources of nonlinearity as well.
        </p>
        <p>This complexity necessitates the use of advanced modeling techniques capable of handling
nonlinear relationships, such as machine learning algorithms like Artificial Neural Networks and
Support Vector Machines.</p>
      </sec>
      <sec id="sec-4-5">
        <title>4.1.4. External uncertainties</title>
        <p>External factors beyond immediate control can disrupt even the most sophisticated forecasting
models. These include, but are not limited to sudden weather changes, unforeseen events (e.g.
sporting events, strikes, or other unexpected occurrences), technical problems in the grid, etc.</p>
        <p>These uncertainties require forecasting models to be robust and adaptable, capable of
incorporating real-time adjustments as new information becomes available.</p>
      </sec>
      <sec id="sec-4-6">
        <title>4.1.5. Computational demands</title>
        <p>The need for real-time processing and high-frequency updates imposes significant computational
demands. To keep forecasts accurate, data must be processed quickly enough to reflect the latest
conditions. Advanced models require substantial computational resources for training and updating
as new data comes in.</p>
        <p>This challenge is compounded by the need for scalability, ensuring that forecasting systems can
handle increased data loads without compromising performance or accuracy.</p>
      </sec>
      <sec id="sec-4-7">
        <title>4.1.6. Model selection and validation</title>
        <p>Selecting the right forecasting model is a significant challenge due to the dynamic nature of load
data. Different models perform better under various conditions, requiring careful selection based on
historical performance and expected future scenarios. Ensuring that chosen models remain effective
over time requires ongoing validation and adjustment as patterns evolve.</p>
      </sec>
      <sec id="sec-4-8">
        <title>4.1.7. Strategies to address the challenges</title>
        <p>To overcome these challenges, experts employ a variety of strategies, including but not limited to:


</p>
        <p>The use of various classes of models – from linear regression to deep neural networks to find
a better match to a particular forecasting task.</p>
        <p>Online data processing – using real-time data feeds to update forecast as new information is
received.</p>
        <p>Ensemble forecasting – combining predictions from multiple models leverages diverse
strengths and reduces reliance on any single model’s potential biases. This is the focus of this
paper.</p>
        <p>Hence, short term electric load forecasting is a multifaceted challenge requiring advanced
techniques, robust computational infrastructure, careful model selection and tuning, dynamic data
processing capabilities, and ongoing validation. The complexity stems from the intricate interplay of
numerous variables, the nonlinear relationships between factors influencing demand, and the need
for real-time accuracy. Applying bagging approaches can help cope with at least some of the
mentioned challenges.</p>
      </sec>
      <sec id="sec-4-9">
        <title>4.2. Test problem details</title>
        <p>The original time series consisted of  = 337 samples. We generated six forecast series ( = 6),
each with 337 samples, using six different independent computational intelligence models. In this
setup, we treated the time series as a data stream, meaning that both forecasting and metamodel
operations were performed in an online mode. This approach ensured that the entire dataset was
processed sequentially – sample by sample ( = 1,2, . . . ,  ) – without requiring multiple passes or
divisions into training, validation, and test sets.</p>
        <p>
          The original time series (see Figure 3) exhibited several distinct trends corresponding to different
seasons, periodic patterns (primarily weekly), sudden changes, and outliers. The presence of these
features made the forecasting task particularly challenging. Additionally, there was a significant
random component in the data because electric load in large power systems depends on numerous
external factors, many of which are inherently unpredictable or chaotic – for example, weather
conditions [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. This randomness contributed to the nonstationary nature of the time series, meaning
its statistical properties changed over time. Furthermore, the series contained noise and was highly
variable, making it a difficult target for forecasting models.
        </p>
        <p>Given these characteristics – non-stationarity, noise, and complexity – the performance of any
single forecasting model or method is unlikely to be consistently superior across the entire dataset.
In other words, different models tend to perform better on specific parts of the series but
underperform on others. This variability highlights a common challenge in time series forecasting:
no single model or method dominates in all situations. It is precisely this kind of scenario where
ensemble methods like bagging come into play.</p>
        <p>By leveraging the strengths of multiple models through bagging, we aimed to improve overall
forecasting accuracy by combining their predictions. The idea was to minimize errors that might
arise from relying on a single model and instead capture more robust insights by aggregating results
from diverse perspectives within the ensemble. This approach has shown promise in addressing the
inherent limitations of individual forecasting methods while providing a more balanced and accurate
prediction across the entire time series.</p>
        <p>In our simulation, we utilize six specialized STLF models within an ensemble framework. Each of
these models has unique inputs and distinct structural differences, which collectively contribute to
a diverse predictive capability. Some models focus on historical weather patterns as inputs, while
others prioritize calendar events like holidays that affect electricity usage. The diverse structures
ensure that each model interprets and processes these inputs differently – some may use linear
regression techniques, while others employ neural networks capable of identifying complex patterns.</p>
        <p>Deploying six specialized STLF models with varied inputs and structures allows us to
comprehensively capture the multifaceted nature of electric load data. This strategic diversity
ensures that our ensemble can account for a wide array of factors influencing demand across
different parts of the time series.</p>
        <p>Figure 4 illustrates the time series data and forecasts for the past 30 days. While long-term trends
are reasonably captured by all models, the variability in the data within shorter timeframes is difficult
for every forecasting method used. As a result, none of the models demonstrate a significant
advantage over the others.</p>
        <p>We utilized the proposed metamodel to derive an optimal combination  ∗( ) of the six forecasts
from the ensemble member models. By visually inspecting the plot, it becomes evident that the
combined forecast  ∗( ) closely aligns with the actual data series  ( ). This proximity is further
corroborated by an in-depth error comparison presented in Table 1.</p>
        <p>To evaluate the accuracy of our predictions, we employed the Mean Absolute Percentage Error
(MAPE) criterion. Widely recognized in short-term electric load forecasting research, MAPE is a
robust metric that quantifies forecast errors as a percentage of actual values, which makes a clear
physical sense. The best performing ensemble member model achieved a MAPE of 4.858%. Through
the application of the nonlinear bagging procedure, the MAPE was reduced to 3.3751%. This
represents an approximate reduction by a factor of 1.44 times compared to the original best model’s
performance alone.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this study, we introduced an innovative bagging method based on an adaptive nonlinear
neofuzzy metamodel. This system is designed to integrate the results from multiple computational
intelligence systems working towards solving similar tasks, such as approximating values or making
forecasts.</p>
      <p>Our approach processes data in real-time and is particularly suited for environments where data
conditions change rapidly over time (non-stationarity). This adaptability is crucial because many
real-world applications, such as electric load forecasting, involve dynamic factors influencing
demand. The speed of processing is also vital; ensuring timely decisions are made based on current
data.</p>
      <p>We validated our metamodel through a simulation involving short-term electric load forecasting.
Utilizing an ensemble of six independent forecasting models to predict electricity consumption, we
compared their results with those generated by our proposed metamodel. The outcomes
demonstrated that our method outperformed the individual models, decreasing the MAPE error by
a factor of 1.44 times relative to the best model in the ensemble.</p>
      <p>Looking ahead, we aim to enhance this metamodel’s structure by increasing its flexibility. This
will enable it to more effectively adapt to various types of relationships in data, particularly nonlinear
ones, which are common in complex real-world scenarios like electricity demand forecasting. By
doing so, we expect the metamodel to become even more robust and versatile in handling diverse
and dynamic data environments.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This research was partially funded by the European Union (through the EURIZON H2020 project,
grant agreement 871072).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Breiman</surname>
          </string-name>
          , Bagging predictors,
          <source>Machine Learning</source>
          <volume>24</volume>
          (
          <year>1996</year>
          )
          <fpage>126</fpage>
          -
          <lpage>140</lpage>
          . https://doi.org/10.1007/BF00058655.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <article-title>Ensembling neural networks: Many could be better than all</article-title>
          ,
          <source>Artificial Intelligence 137</source>
          .
          <fpage>1</fpage>
          -
          <lpage>2</lpage>
          (
          <year>2002</year>
          )
          <fpage>239</fpage>
          -
          <lpage>263</lpage>
          . https://doi.org/10.1016/S0004-
          <volume>3702</volume>
          (
          <issue>02</issue>
          )
          <fpage>00190</fpage>
          -
          <lpage>X</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , Y. Ma (Eds)
          <article-title>Ensemble Machine Learning</article-title>
          .
          <source>Methods and Applications</source>
          , Springer, New York, NY,
          <year>2012</year>
          . https://doi.org/10.1007/978-1-
          <fpage>4419</fpage>
          -9326-7.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Friedman</surname>
          </string-name>
          , P. Hall,
          <article-title>On bagging and nonlinear estimation</article-title>
          ,
          <source>Journal of Statistical Planning and Inference 137.3</source>
          (
          <year>2007</year>
          )
          <fpage>669</fpage>
          -
          <lpage>683</lpage>
          . https://doi.org/10.1016/j.jspi.
          <year>2006</year>
          .
          <volume>06</volume>
          .002.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bifet</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Holmes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Pfahringer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gavaldà</surname>
          </string-name>
          ,
          <article-title>Improving Adaptive Bagging Methods for Evolving Data Streams</article-title>
          , in: ZH.
          <string-name>
            <surname>Zhou</surname>
          </string-name>
          , T. Washio (Eds),
          <source>Advances in Machine Learning</source>
          , Vol.
          <volume>5828</volume>
          of LNCS, Springer, Berlin,
          <year>2009</year>
          , pp.
          <fpage>23</fpage>
          -
          <lpage>37</lpage>
          . https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -05224-
          <issue>8</issue>
          _
          <fpage>4</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E.</given-names>
            <surname>Lughofer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pratama</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Skrjanc</surname>
          </string-name>
          ,
          <article-title>Online bagging of evolving fuzzy systems</article-title>
          ,
          <source>Information Sciences 570</source>
          (
          <year>2021</year>
          )
          <fpage>16</fpage>
          -
          <lpage>33</lpage>
          . https://doi.org/10.1016/j.ins.
          <year>2021</year>
          .
          <volume>04</volume>
          .041.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Levinson</surname>
          </string-name>
          ,
          <article-title>The ensemble approach to forecasting: A review and synthesis</article-title>
          , Transportation Research Part C: Emerging Technologies,
          <volume>132</volume>
          (
          <year>2021</year>
          )
          <article-title>103357</article-title>
          . https://doi.org/10.1016/j.trc.
          <year>2021</year>
          .103357
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          , Ensemble Learning, in: Y.
          <string-name>
            <surname>Yang</surname>
          </string-name>
          (Ed)
          <article-title>Temporal Data Mining Via Unsupervised Ensemble Learning</article-title>
          , Elsevier,
          <year>2017</year>
          , pp.
          <fpage>35</fpage>
          -
          <lpage>56</lpage>
          . https://doi.org/10.1016/B978-0
          <source>-12-811654-8</source>
          .
          <fpage>00004</fpage>
          -X
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Ye</surname>
            . Bodyanskiy,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Otto</surname>
            , I. Pliss,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Popov</surname>
          </string-name>
          ,
          <article-title>An Optimal Algorithm for Combining Multivariate Forecasts in Hybrid Systems</article-title>
          , in: V.
          <string-name>
            <surname>Palade</surname>
            ,
            <given-names>R.J.</given-names>
          </string-name>
          <string-name>
            <surname>Howlett</surname>
          </string-name>
          , L. Jain (Eds)
          <article-title>Knowledge-Based Intelligent Information and Engineering Systems</article-title>
          .
          <source>KES</source>
          <year>2003</year>
          , volume
          <volume>2774</volume>
          <source>of LNCS</source>
          , Springer, Berlin, Heidelberg,
          <year>2003</year>
          . https://doi.org/10.1007/978-3-
          <fpage>540</fpage>
          -45226-3_
          <fpage>132</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>T.</given-names>
            <surname>Yamakawa</surname>
          </string-name>
          , E. Uchino,
          <string-name>
            <given-names>T.</given-names>
            <surname>Miki</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kusanagi</surname>
          </string-name>
          ,
          <article-title>A neo fuzzy neuron and its applications to system identification and prediction of the system behavior</article-title>
          ,
          <source>in: Proc. 2nd Int. Conf. on Fuzzy Logic and Neural Networks</source>
          ,
          <year>1992</year>
          , pp.
          <fpage>477</fpage>
          -
          <lpage>483</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>E.</given-names>
            <surname>Uchino</surname>
          </string-name>
          , T. Yamakava,
          <article-title>Neo-fuzzy neuron based new approach to system modeling with application to actual system</article-title>
          ,
          <source>in: Proc. Sixth International Conference on Tools with Artificial Intelligence</source>
          , New Orlean, LA, USA,
          <year>1994</year>
          , pp.
          <fpage>564</fpage>
          -
          <lpage>570</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Miki</surname>
          </string-name>
          , T. Yamakawa,
          <article-title>Analog implementation of neo-fuzzy neuron and its on-board learning</article-title>
          ,
          <source>in: Computational Intelligence and Applications</source>
          , WSES Press, Piraeus,
          <year>1999</year>
          , pp.
          <fpage>144</fpage>
          -
          <lpage>149</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>I. Perfilieva</surname>
          </string-name>
          ,
          <article-title>Fuzzy transforms: Theory and applications</article-title>
          ,
          <source>Fuzzy Sets and Systems 157.8</source>
          (
          <year>2006</year>
          )
          <fpage>993</fpage>
          -
          <lpage>1023</lpage>
          . https://doi.org/10.1016/j.fss.
          <year>2005</year>
          .
          <volume>11</volume>
          .012.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , A. Knoll,
          <string-name>
            <surname>Constructing Fuzzy Controllers with B-Spline Models</surname>
          </string-name>
          .
          <source>Principles and Applications</source>
          ,
          <source>International Journal of Intelligent Systems 13, N. 2/3</source>
          (
          <year>1998</year>
          )
          <fpage>257</fpage>
          -
          <lpage>286</lpage>
          . https://doi.org/10.1002/(SICI)
          <fpage>1098</fpage>
          -
          <lpage>111X</lpage>
          (
          <issue>199802</issue>
          /03)13:
          <issue>2</issue>
          /3%3C257:
          <article-title>:AID-INT9%3E3.0</article-title>
          .CO;2-
          <string-name>
            <surname>Z.</surname>
          </string-name>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>V.</given-names>
            <surname>Kolodyazhniy</surname>
          </string-name>
          , Ye. Bodyanskiy,
          <article-title>Cascaded multiresolution spline-based fuzzy neural network</article-title>
          ,
          <source>in: Proc. International Symposium on Evolving Intelligent Systems</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>29</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Mykhalyov</surname>
            ,
            <given-names>I. Pliss</given-names>
          </string-name>
          ,
          <article-title>Adaptyvne vyyavlennya rozladnan' v ob'ektakh keruvannya za dopomogoyu shtuchnykh neyronnykn merezh, Systemni tekhnilogii</article-title>
          , Dnipro,
          <year>2000</year>
          . In Ukrainian.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Kaczmarz</surname>
          </string-name>
          ,
          <article-title>Approximate solution of systems of linear equations</article-title>
          ,
          <source>International Journal of Control 57.6</source>
          (
          <year>1993</year>
          )
          <fpage>1269</fpage>
          -
          <lpage>1271</lpage>
          . https://doi.org/10.1080/00207179308934446.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>B.</given-names>
            <surname>Widrow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.E.</given-names>
            <surname>Hoff</surname>
          </string-name>
          ,
          <article-title>Adaptive switching circuits, in: 1960 IRE WESCON Convention Record, Part 4</article-title>
          ,
          <string-name>
            <surname>IRE</surname>
          </string-name>
          , New York,
          <year>1960</year>
          , pp.
          <fpage>96</fpage>
          -
          <lpage>104</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Kokshenev</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Kolodyazhniy</surname>
          </string-name>
          ,
          <article-title>An adaptive learning algorithm for a neo fuzzy neuron</article-title>
          ,
          <source>in: Proc. 3rd Conference of the European Society for Fuzzy Logic and Technology</source>
          , Zittau, Germany,
          <year>2003</year>
          , pp.
          <fpage>375</fpage>
          -
          <lpage>379</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Ye. Bodyanskiy</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Popov</surname>
          </string-name>
          , T. Rybalchenko,
          <article-title>Feedforward neural network with a specialized architecture for estimation of the temperature influence on the electric load</article-title>
          ,
          <source>in: Proc. 2008 4th International IEEE Conference Intelligent Systems, Varna, Bulgaria</source>
          ,
          <year>2008</year>
          , pp.
          <fpage>7</fpage>
          -
          <issue>14</issue>
          -7-18. https://doi.org//10.1109/IS.
          <year>2008</year>
          .
          <volume>4670444</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>