<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A modified neuro-fuzzy counterpropagation network and its fast adaptive learning</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sergiy Popov</string-name>
          <email>serhii.popov@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iryna Pliss</string-name>
          <email>iryna.pliss@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olha Chala</string-name>
          <email>olha.chala@nure.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Olexii Holovin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Central Scientific Research Institute of Armament and Military Equipment of the Armed Forces of Ukraine</institution>
          ,
          <addr-line>Kyiv, 03049</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Kharkiv National University of Radio Electronics</institution>
          ,
          <addr-line>14 Nauky av., Kharkiv, 61166</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A neuro-fuzzy counterpropagation network is introduced that employs a modified fuzzy C-means clustering procedure in an online mode, enhancing both learning rate and accuracy while maintaining the same simple architecture as traditional CPN networks. This modification allows handling of overlapping classes, when an observation can belong to multiple classes simultaneously. Consequently, several output layer neurons can be activated at once. An optimized algorithm is introduced for the output layer tuning with a better control over its filtering and followingcharacteristics through the use of a special adjustable parameter. Experiments demonstrate that this innovative approach outperforms traditional counterpropagation networks in various performance metrics.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Counterpropagation network</kwd>
        <kwd>neuro-fuzzy network</kwd>
        <kwd>overlapping classes</kwd>
        <kwd>increased learning rate</kwd>
        <kwd>short training set 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In recent years, artificial neural networks (ANN) have become a popular solution to various
information processing challenges. These include tasks such as pattern recognition (classification),
clustering, and forecasting (extrapolation). The success of ANNs can be attributed to their ability to
approximate complex functions (universal approximation properties) and learn by adjusting their
parameters based on optimization procedures.</p>
      <p>Deep neural networks (DNNs), a subset of ANNs, have demonstrated remarkable results in solving
numerous data analysis problems. However, DNNs also have significant drawbacks. One major
limitation is the requirement for large amounts of training data, which may not always be available.
Additionally, DNNs can be slow during parameter adjustment in multi-epoch learning mode. DNNs
also face challenges when tackling real-time data stream mining tasks under conditions of
nonstationarity and limited input information. Similar limitations apply to classic multilayer perceptron
(MLP) models trained using the error backpropagation procedure.</p>
      <p>It is worth noting that classic radial basis function networks (RBFN) [1, 2] exhibit a higher learning
rate but may encounter issues related to the “curse of dimensionality” as the number of input signals
increases.</p>
      <p>In today’s data-driven world, there is a growing need for neural networks that can efficiently
handle data stream mining tasks in online mode with limited training data. Among various neural
network models, the counterpropagation neural network (CPN), introduced by R. Hecht-Nielsen [3-5],
stands out as a viable solution despite its architectural simplicity.</p>
      <p>Advantages of CPNs:

</p>
      <p>High learning rate: CPNs are known for their ability to learn quickly.</p>
      <p>Simple architecture: only two layers formed by simple nodes, CPNs offer computational
efficiency.</p>
      <p>However, there is an inherent trade-off. While CPNs excel in learning rate, their approximation
properties – the ability to model complex functions – are inferior compared to traditional MLPs and
RBFNs, and of course modern DNNs, which are generally more powerful in function approximation.</p>
      <p>Despite these limitations, ongoing research focuses on enhancing the approximation capabilities
of CPNs while maintaining their high learning and processing rates. These efforts aim to bridge the
gap between performance and efficiency without compromising on speed. Recent applications of
CPNs include but are not limited to classification [6-9], prediction [10], parameter identification [11],
structural optimization [12], extreme learning machine optimization [13], digital image watermarking
[14], navigation systems development [15] and others. One promising direction for improvement
involves integrating hybrid systems of computational intelligence [16]. Specifically, neuro-fuzzy
approaches, which combine neural networks with fuzzy logic, offer a potential solution. By leveraging
these methods [2, 17, 18] it may be possible to enhance the characteristics of CPNs. In conclusion,
while CPNs present unique challenges compared to more sophisticated neural architectures like
DNNs and MLPs, ongoing research explores innovative solutions that could unlock their full
potential.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Counterpropagation network basics</title>
      <p>From a theoretical point of view, the counterpropagation network is intended for restoring the
nonlinear mapping y = F ( x ) (forward-only CPN architecture shown in Fig. 1 is sufficient), as well as
the inverse mapping x= F−1 ( y ) (full CPN architecture is required, see Fig. 2), i.e. identifying a
nonlinear transform</p>
      <p>
        F : X → Y ( Rn → Rm)
from the training samples x (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) , y (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) , … , x ( k ) , y ( k ) , … , x ( N ) , y ( N ), where
x ( k )=( x1 ( k ) , … , xi ( k ) , … , xn ( k ))T ∈ Rn, y ( k )=( y1 ( k ) , … , yi ( k ) , … , ym ( k ))T ∈ Rm,
k =1,2 , … , N is the observation index in the dataset, or the index of the current discrete time, if the
data is being processed in online mode.
      </p>
      <p>CPN contains two layers of neurons: the first hidden layer, called the T. Kohonen layer, and the
output layer, called the S. Grossberg layer. In this paper, we will focus on the forward-only
architecture, but the proposed methods are equally applicable to the full CPN architecture as well.</p>
      <p>
        The input signals x ( k ) arrive sequentially from the receptive layer to the first hidden layer, which
is usually a Kohonen’s self-organizing map (SOM) [19, 20] designed to solve the crisp clustering
problem, i.e. dividing the data set into h non-overlapping classes/clusters in the self-learning mode.
SOM implements the following “Winner Takes All” mapping
ul ( k )={¿ 1 , if wlK ( k ) is a winner , i . e .‖x ( k )−wlK ( k )‖≤‖x ( k )−wiK ( k )‖∀ i=1,2 , … , h
¿ 0 otherwise
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
where u ( k )=(u1 ( k ) , … , ul ( k ) , … , uh ( k ))T, W K ={wlKi } – ( h × n) tuned matrix of synaptic weights
that define centroids of the clusters.
      </p>
      <p>
        The Kohonen layer learning is based on the same “Winner Takes All” (WTA) principle, when only
one winning neuron is tuned at each iteration k =1,2 , … When observation x ( k ) is received, the
closest to x ( k ) neuron is determined (in the Euclidean metrics sense), which is called a “winner” (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ).
Then only this “winner” neuron’s vector of weights wlK ( k ) is being tuned according to the rule:
wlK ( k +1)={¿ wlK ( k )+ηK ( k )( x ( k )−wlK ( k )) , if wlK ( k ) is a winner ,
¿ wlK ( k ) otherwise
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
G ^xn ( k )
      </p>
      <p>G
wn+1</p>
      <p>
        G
wn+m
^y1 ( k )
^ym ( k )
x1 ( k )
xn ( k )
y1 ( k )
ym ( k )
x1 ( k )
xn ( k )
y1 ( k )
ym ( k )
w1K
w2K
whK
w1K
w2K
w3K
whK
u1 ( k )
u2 ( k )
uh ( k )
u1 ( k )
u2 ( k )
u3 ( k )
uh ( k )
(here 0&lt;ηK ( k )&lt;1 is the Kohonen layer learning rate parameter, which is usually chosen
empirically). Note also that when ηK ( k )=k−1 procedure (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) calculates the arithmetic mean (centroid)
of the lth cluster, i.e. it actually implements the popular crisp K-means clustering algorithm.
      </p>
      <p>
        The output layer is formed by the so-called Grossberg outstars, which are essentially modifications
of the standard linear element (Adaline) and implement the mapping
^y ( k )=W G u ( k ) ,
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
where W G={w Gjl } – ( m × h) matrix of synaptic weights tuned in the controlled learning mode.
Outstar neurons of the output (Grossberg) layer are usually trained using a fairly simple algorithm
w Gjl ( k +1)=w Gjl ( k )+ηG ( k ) ul ( k ) ( yl ( k )−w Gjl ( k ))
or using vector notation,
w Gj ( k +1)=w Gj ( k )+ηG ( k ) uT ( k )⊙ ( yl ( k ) Eh−wlG ( k )) ,
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
where 0&lt;ηG ( k )&lt;1 is the Grossberg layer learning rate parameter, Eh is a (1 × h) vector of ones, ⊙
is the element-wise product symbol
      </p>
      <p>So, this neural network acts like a simple lookup table. It gives outputs in steps rather than
smoothly, which limits its ability to model complex relationships. Furthermore, using a “Winner
Takes All” approach in the Kohonen layer means that during training, only one outstar is adjusted at a
time. This makes the overall training process slower.</p>
      <p>Given these limitations, modifying the Counterpropagation Neural Network along with its
learning methods could enhance its ability to model functions more accurately while speeding up the
training process.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Neuro-fuzzy counterpropagation network (NFCPN)</title>
      <p>The proposed NFCPN maintains the same architecture as traditional CPN networks, but introduces
key improvements. Instead of employing a traditional Self-Organizing Map within the Kohonen layer,
which traditionally uses a recurrent version of the crisp K-means clustering algorithm, our approach
utilizes a modified fuzzy C-means clustering procedure (FCM) [21, 22] in the recurrent form [23],
enhancing both learning efficiency and approximation accuracy.</p>
      <p>This modification allows for effective handling of situations where classes overlap in feature space,
enabling an observation to belong to multiple classes simultaneously. Furthermore, by applying a
nonlinear strategy, the network can activate several output layer neurons at once. In contrast, classic
CPN networks only trigger one Grossberg outstar during learning, which inherently slows down the
process.</p>
      <p>Overall, these changes significantly improve the network’s performance and adaptability in
complex scenarios.</p>
      <sec id="sec-3-1">
        <title>3.1. Kohonen layer learning</title>
        <p>To improve the quality and speed of the SOM clustering, we use the so-called “Winner Takes More”
(WTM) rule, instead of WTA. This approach utilizes a neighborhood function ψ (l , g , k ) that
determines the proximity of all other neurons w gK ( k ) , g=1,2 , … , l−1 , l +1 , … , h to the “winner”
wlK ( k ). For g=l, ψ (l , l , k )=1, and the value of ψ (l , g , k ) decreases with the increase of the distance
between vectors wlK ( k ) and w gK ( k ).</p>
        <p>
          All centroids – vectors of synaptic weights are tuned according to the modified learning rule
wlK ( k +1)=wlK ( k )+ηK ( k ) ψ (l , g , k ) ( x ( k )−wlK ( k ))∀ l=1,2 , … , h
(
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
        </p>
        <p>
          It is readily seen that (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ) is a generalization of the WTA algorithm (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ), for which the neighborhood
function is a singleton. Unfortunately, there are no formal rules for determining neighborhood
functions ψ (l , g , k ), hence their selection is based on empirical considerations.
        </p>
        <p>Considering a more practical situation, when each observation can belong to several or all clusters
simultaneously, it is beneficial to use a recurrent modification of J.C. Bezdek’s FCM algorithm [21]
related to optimization of the following objective function</p>
        <p>N h</p>
        <p>K 2
J ( μl ( k ) , wlK )=∑ ∑ μlβ ( k )‖x ( k )−wl ‖</p>
        <p>k=1 l=1
subject to constraints</p>
        <p>Here μl ( k ) – degree of fuzzy membership of observation x ( k ) to lth cluster, β &gt;0 – fuzzifier
(usually β =2), wlK – centroid of lth cluster.</p>
        <p>
          Solving the optimization problem based on finding the saddle point of the Lagrange function
L ( μl ( k ) , wlK , λ ( k ))= ∑N ∑h μlβ ( k )‖x ( k )−wlK‖2+ ∑N λ ( k )(∑h μl ( k )−1)
k=1 l=1 k=1 l=1
(
          <xref ref-type="bibr" rid="ref6">6</xref>
          )
(here λ ( k ) – Lagrange multipliers) for β =2 leads to the standard FCM algorithm
        </p>
        <p>h
∑ μl ( k )=1∀ k =1,2 , … , N ,
l=1</p>
        <p>N
0&lt;∑ μl ( k )&lt; N ∀ l=1,2 , … , h .</p>
        <p>k=1
¿ μl ( k )= h</p>
        <p>,
N
∑ μl2 ( k )
k=1
¿ wlK = k=1
.</p>
        <p>
          To solve the fuzzy clustering problem in online mode, i.e. training the fuzzy Kohonen map,
consider a local modification of the Lagrange function (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) in the form [24, 25]
        </p>
        <p>h 2 h
L ( μl ( k ) , wlK ( k ) , λ ( k ))=∑ μlβ ( k )‖x ( k )−wlK ( k )‖ + λ ( k )(∑ μl ( k )−1).</p>
        <p>l=1 l=1</p>
        <p>Optimizing it with the K.J. Arrow, L. Hurwitz, H. Uzawa procedure [26], we obtain the following
result
which coincides with the D.C. Park, I. Dagher algorithm [27] when β =2:
{
{
¿ wlK ( k +1)=wlK ( k )+ηK ( k ) μlβ ( k )( x ( k )−wlK ( k )) ,
¿ μl ( k )= h</p>
        <p>‖x ( k )−wlK ( k )‖
∑ ‖x ( k )−w gK ( k )‖
g=1
¿ μl ( k )= h</p>
        <p>‖x ( k )−wlK ( k )‖
∑ ‖x ( k )−w gK ( k )‖
g=1
−β</p>
        <p>−β
−2
−2
,</p>
        <p>
          It is easy to see that (
          <xref ref-type="bibr" rid="ref7">7</xref>
          ), (
          <xref ref-type="bibr" rid="ref8">8</xref>
          ) structurally coincide with the WTM algorithm (
          <xref ref-type="bibr" rid="ref5">5</xref>
          ), but here the
neighborhood function is being chosen automatically.
        </p>
        <p>Next, the calculated membership degrees μl ( k )∀ l=1,2 , … , h are fed to the output layer of the
network, i.e. vector u ( k ) is formed not by a single one and a set of zeros, but by membership degrees
μl ( k ), which activate all neurons of the Grossberg output layer.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Grossberg layer learning</title>
        <p>
          As described above, the Grossberg layer receives the vector μ ( k )=( μ1 ( k ) , … , μl ( k ) , … , μh ( k ))T as
input, instead of u ( k ) in classic CPN. This leads to acceleration of the Grossberg layer learning,
because all weights are being updated at each iteration, not only the ones connected to the “winner” of
the Kohonen layer. Hence, instead of (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ), the learning process of this layer can be rewritten as
w Gj ( k +1)=w Gj ( k )+ηG ( k ) μT ( k )⊙ ( yl ( k ) Eh−wlG ( k )) .
        </p>
        <p>The output layer learning rate parameter ηG ( k ) can be optimized, considering the objective
function
and its gradient optimization procedure</p>
        <p>w Gj ( k +1)=w Gj ( k )+ηG ( k )( yl ( k )−wlG ( k ) μ ( k )) μT ( k ) .</p>
        <p>
          Optimizing (
          <xref ref-type="bibr" rid="ref10">10</xref>
          ) for speed leads to the Kaczmarz-Widrow-Hoff algorithm [28-30] in a form
w Gj ( k +1)=w Gj ( k )+ yl ( k )−wlG ( k2) μ ( k ) μT ( k ) .
        </p>
        <p>‖μ ( k )‖</p>
        <p>
          The balance between filtering and following properties of (
          <xref ref-type="bibr" rid="ref11">11</xref>
          ) can be chosen by the following
modification[31, 32]:
{
¿ w Gj ( k +1)=w Gj ( k )+ α−1 ( k )( yl ( k )−wlG ( k ) μ ( k )) μT ( k ) ,
        </p>
        <p>
          2
¿ α ( k )=γα ( k −1)+‖μ ( k )‖ , 0 ≤ γ ≤ 1 ,
(
          <xref ref-type="bibr" rid="ref9">9</xref>
          )
(
          <xref ref-type="bibr" rid="ref10">10</xref>
          )
(
          <xref ref-type="bibr" rid="ref11">11</xref>
          )
(
          <xref ref-type="bibr" rid="ref12">12</xref>
          )
which coincides with (
          <xref ref-type="bibr" rid="ref11">11</xref>
          ) when γ =0, and becomes a stochastic approximation procedure when
γ =1.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental results</title>
      <sec id="sec-4-1">
        <title>4.1. Experimental setup</title>
        <p>For the sake of comparison between classic counterpropagation network (CPN) and the proposed
neuro-fuzzy counterpropagation network (NFCPN), we use a simple test case with n=2 inputs,
h=9 neurons in the Kohonen layer, and m=1 output. Inputs x1 ( k ) , x2 ( k ) are sampled from the
uniform distribution over the interval [ 0,1], the corresponding output is calculated as
1
y ( k )=( x1 ( k )2+ x2 ( k )2)2 .</p>
        <p>The first N =1000 observations form the training set, another T =1000 observations form the
test set. Both networks operate in online mode, processing all N training observations sequentially
and only once, updating their parameters after each step k. Also, after each step, the mean absolute
error (MAE) is calculated over the entire test set, i.e. we monitor how the out-of-sample error changes
during the online training process.</p>
        <p>
          First, both networks are trained under the same conditions ηK ( k )=ηG ( k )=0.1∀ k =1,2 , … , N ,
hence we compare WTA principle in CPN versus WTM in NFCPN (Fig. 3). Then, an optimized
learning algorithm (
          <xref ref-type="bibr" rid="ref12">12</xref>
          ) with various values of parameter γ is used for NFCPN in order to further
improve its performance (Fig. 4). Numerical results are presented in Table 1.
        </p>
        <p>The analysis conducted on the results reveals significant differences between traditional
counterpropagation networks (CPN) and neuro-fuzzy counterpropagation networks (NFCPN). These
findings highlight the advantages of using NFCPN in achieving faster learning rates and improved
accuracy.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Learning efficiency: MAE comparison across networks</title>
        <p>Mean Absolute Error (MAE), a key metric for evaluating model performance, was calculated at
various stages of training. The results demonstrate that:

</p>
        <p>Classic CPN: At k =399 iterations, the classic CPN achieved MAE level of 0.1.</p>
        <p>NFCPN: In comparison, the neuro-fuzzy counterpart reached a similar MAE level of 0.1 at
k =264 iterations.</p>
        <p>This indicates that NFCPN requires fewer training cycles to achieve comparable accuracy,
suggesting superior learning efficiency compared to classic CPN. Further results reinforce this
conclusion.</p>
        <p>At k =1000:

</p>
        <p>Classic CPN: The MAE stabilized at 0.085.</p>
        <p>NFCPN: Achieved an improved MAE of 0.065, showcasing greater accuracy even as training
progressed.</p>
        <p>These results collectively demonstrate that, with the same training parameters, NFCPN learns
approximately 1.5 times faster than CPN and is by 24% more accurate in performing the given task.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Adjusting learning dynamics: the role of gamma parameter (γ )</title>
        <p>
          The study also explored different configurations for enhancing learning performance using algorithm
(
          <xref ref-type="bibr" rid="ref12">12</xref>
          ) and adjusting the gamma parameter (γ ).
        </p>
        <p>Initially, γ was set to 0. This setting significantly increased the learning rate by over 10 times in
comparison to the classical learning algorithm with ηK ( k )=ηG ( k )=0.1. However, this improvement
came at a cost – training became noisy, lacking effective filtering properties. To address this trade-off
and improve the filtering characteristics of the algorithm without compromising learning speed,
gamma was gradually increased. With γ =0.9, errors comparable to those achieved with a fixed
ηG ( k )=0.1 were observed. Additionally, this configuration maintained an impressive learning rate
that was about 2.2 times faster than with the classical learning algorithm.</p>
        <p>This experimentation underscores the importance of fine-tuning gamma to achieve a balance
between noise reduction and efficient learning rates. By carefully controlling gamma, it is possible to
optimize both filtering properties and following characteristics (i.e., adaptability to changes in
nonstationary data streams).</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Key findings summary</title>
        <p>

</p>
        <p>Learning rate: NFCPN consistently outperforms CPN by achieving comparable or better MAE
with fewer training iterations.</p>
        <p>Accuracy enhancement: The improved performance of NFCPN results in a 24% increase in
accuracy over classic CPN under the same conditions.</p>
        <p>Parameter optimization: Modifying gamma allows for precise control over learning dynamics,
balancing between noisy and stable training processes. Adjusting gamma to higher values
enhances filtering properties without significantly compromising on learning speed.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>We have introduced a fuzzy modification to a counterpropagation network, enhancing its ability to
handle situations where data categories overlap. This means an item can belong to multiple classes
simultaneously, which is common in real-world scenarios.</p>
      <p>Our modifications improve the network’s learning efficiency and enable it to address a wider range
of problems in real-time data processing. Additionally, this enhanced version is simpler
mathematically and requires less training data compared to traditional methods. It also adapts
smoothly as new, varied data arrives, which is crucial for handling dynamic information streams.</p>
      <p>Experiments demonstrate that our modified network performs effectively and outperforms the
standard CPN model. These results suggest that neuro-fuzzy counterpropagation networks hold
significant potential in real-time data processing tasks where both efficiency and accuracy are critical.
The ability to adjust gamma parameter offers flexibility, enabling the network to adapt to varying
levels of non-stationarity in input data streams.</p>
      <p>Further research could explore additional parameter configurations or investigate how the
proposed approach generalizes across different tasks. We also aim to explore different clustering
techniques within the hidden layer of NFCPN to further enhance its capabilities. Such advancements
would likely enhance the applicability of counterpropagation networks across a broader range of
realworld scenarios.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Grammarly in order to: Grammar and spelling
check. After using this tool, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.
[14] A. Taheri “A New Robust Digital Image Watermarking Technique Using Relations between
Wavelet Coefficients Transform and Full Counter Propagation Neural Network” International
Journal of Recent Research in Electrical and Electronics Engineering (IJRREEE) 7 (2020): 1–12.
[15] D-J. Jwo, A. Biswal, I. Mir “Artificial Neural Networks for Navigation Systems: A Review of</p>
      <p>Recent Research.” Applied Sciences 13 (2023):4475.
[16] J. Kacprzyk, W. Pedrycz (Eds.) Springer Handbook of Computational Intelligence, Springer,</p>
      <p>Berlin-Heidelberg, 2015.
[17] T. Yamakawa, E. Uchino, T. Miki, H. Kusanagi, A neo fuzzy neuron and its applications to system
identification and prediction of the system behavior, in: Proceedings of 2nd Int. Conf. on Fuzzy
Logic and Neural Networks, 1992, pp. 477-483.
[18] T. Miki, T. Yamakawa, Analog implementation of neo-fuzzy neuron and its on-board learning, in:</p>
      <p>Computational Intelligence and Applications, WSES Press, Piraeus, 1999, pp. 144-149.
[19] T. Kohonen, Self-Organizing Maps, Springer-Verlag, Berlin, 1995.
[20] T. Kohonen, Improved versions of learning vector quantization, in: Proceedings of 1990 IJCNN
International Joint Conference on Neural Networks, volume 1, San Diego, CA, USA, 1990, pp.
545-550.
[21] J. C. Bezdek. Pattern Recognition with Fuzzy Objective Function Algorithms, Plenum Press, New</p>
      <p>York, 1981.
[22] F. Hoeppner, F. Klawonn, R. Kruse, T. Runkler, Fuzzy cluster analysis and image recognition,</p>
      <p>Wiley &amp; Sons, Chichester, 1999.
[23] Ye. Bodyanskiy, Computational intelligence techniques for data analysis, in: Lecture Notes in</p>
      <p>Informatics, volume 72, GI, Bonn, 2005, pp. 15-36.
[24] Ye. Bodyanskiy, Ye. Gorshkov, V. Kolodyazhniy, New recursive learning algorithms for fuzzy
Kohonen clustering network, in: Proceedings of 17th Int. Workshop on Nonlinear-Dynamics of
Electronic Systems. Rapperswil, Switzerland, June 21-24, 2009, pp. 58-61.
[25] Ye. Bodyanskiy, A. Deineko, F. Eze “Kernel Fuzzy Kohonen’s Clustering Neural Network and It’s</p>
      <p>Recursive Learning.” Automatic Control and Computer Sciences 52 (2018): 166-174.
[26] K. J. Arrow, L. Hurwicz, H. Uzawa. Studies in linear and non-linear programming, Stanford</p>
      <p>University Press, 1958.
[27] D. C. Park, I. Dagher, Gradient based fuzzy c-means (GBFCM) algorithm, in: Proceedings of 1994
IEEE International Conference on Neural Networks (ICNN’94), volume 3, Orlando, FL, USA, 1994,
pp. 1626-1631
[28] S. Kaczmarz “Angenäherte Auflösung von Systemen linearer Gleichungen.” Bulletin
International de l’Académie Polonaise des Sciences et des Lettres. Classe des Sciences
Mathématiques et Naturelles. Série A, Sciences Mathématiques 35 (1937): 355–357.
[29] S. Kaczmarz “Approximate solution of systems of linear equations.” International Journal of</p>
      <p>Control 57 (1993): 1269-1271.
[30] B. Widrow, M. Hoff, Adaptive Switching Circuits, in: 1960 IRE WESCON Convention Record,
part 4, 1960, pp. 96-104.
[31] Ye. Bodyanskiy, O. Chala, I. Izonin, S. Popov Simple neuro-fuzzy system with combined learning
for pattern recognition under conditions of short training set in medical diagnostics tasks, in:
Proceedings of the 5th International Conference on Informatics &amp; Data-Driven Medicine (IDDM
2022), Lyon, France, November 18-20, 2022, pp. 1-8.
[32] Ye. Bodyanskiy, I. Kokshenev, V. Kolodyazhniy, An adaptive learning algorithm for a neo fuzzy
neuron, in: Proceedings of the 3rd Int. Conf. of European Union Society for Fuzzy Logic and
Technology (EUSFLAT 2003), Zittau, Germany, 10-12 September, 2003, pp. 375-379.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rojas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Neural</given-names>
            <surname>Networks</surname>
          </string-name>
          .
          <source>A Systematic Introduction</source>
          , Springer-Verlag Berlin,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L. H.</given-names>
            <surname>Tsoukalas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. E.</given-names>
            <surname>Uhrig</surname>
          </string-name>
          , Fuzzy and Neural Approaches in Engineering, John Wiley and Sons,
          <year>1997</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hecht-Nielsen</surname>
          </string-name>
          ,
          <article-title>Counter-propagation Network</article-title>
          ,
          <source>in: Proceedings of IEEE First International Conference on Neural Networks</source>
          , volume
          <volume>2</volume>
          ,
          <year>1987</year>
          , pp.
          <fpage>19</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hecht-Nielsen</surname>
          </string-name>
          .
          <article-title>“Counterpropagation networks</article-title>
          .
          <source>” Applied optics 26</source>
          (
          <year>1987</year>
          ):
          <fpage>4979</fpage>
          -
          <lpage>4984</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R.</given-names>
            <surname>Hecht-Nielsen</surname>
          </string-name>
          .
          <article-title>“Applications of counterpropagation networks</article-title>
          .
          <source>” Neural networks 1</source>
          (
          <year>1988</year>
          ):
          <fpage>131</fpage>
          -
          <lpage>139</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B.</given-names>
            <surname>Bajželj</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          <article-title>Drgan “Hepatotoxicity Modeling Using Counter-Propagation Artificial Neural Networks: Handling an Imbalanced Classification Problem”</article-title>
          .
          <source>Molecules</source>
          <volume>25</volume>
          (
          <year>2020</year>
          ):
          <fpage>481</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Rahmat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Harahap</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <article-title>Rachmawati “Counter-propagation Neural Network for Brain Tumor Classification”</article-title>
          <source>J. Phys.: Conf. Ser</source>
          .
          <volume>1566</volume>
          (
          <year>2020</year>
          )
          <fpage>012128</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Sutha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Gnanambigai</surname>
          </string-name>
          , P. Dinadayalan “
          <article-title>Extended Self-Organizing Map With Ubiquitous Counter Propagation Network In Classification For Diabetic Database” Solid State Technology 63 (</article-title>
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>N.</given-names>
            <surname>Vivekanandan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Rajeswari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Salve</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kanna</surname>
          </string-name>
          ,
          <article-title>Precision Smoke Detection System with Counter Propagation Neural Network and Electronic Olfactory</article-title>
          . In: De,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.P.</given-names>
            ,
            <surname>Pati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Biswas</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . (
          <article-title>eds) Recent Trends in Mechanical Engineering</article-title>
          .
          <source>ICRAME 2024. Lecture Notes in Mechanical Engineering</source>
          . Springer, Singapore, pp.
          <fpage>235</fpage>
          -
          <lpage>247</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>B.</given-names>
            <surname>Pratap</surname>
          </string-name>
          “
          <article-title>Analysis of mechanical properties of fly ash and bauxite residue based geopolymer concrete using ANN, Random Forest and Counter propagation neural network</article-title>
          .
          <source>” Asian J Civ Eng</source>
          <volume>25</volume>
          (
          <year>2024</year>
          ):
          <fpage>4303</fpage>
          -
          <lpage>4317</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mutra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Srinivas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rządkowski</surname>
          </string-name>
          “
          <article-title>An optimal parameter identification approach in foil bearing supported high-speed turbocharger rotor system</article-title>
          ”
          <source>Arch Appl Mech</source>
          <volume>91</volume>
          (
          <year>2021</year>
          ):
          <fpage>1557</fpage>
          -
          <lpage>1575</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kaveh</surname>
          </string-name>
          ,
          <article-title>Structural Optimization by Gradient-Based Neural Networks</article-title>
          ,
          <source>In: Applications of Artificial Neural Networks and Machine Learning in Civil Engineering. Studies in Computational Intelligence</source>
          , Vol
          <volume>1168</volume>
          ,
          <year>2024</year>
          , Springer, Cham, pp
          <fpage>147</fpage>
          -
          <lpage>163</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>G.</given-names>
            <surname>Kayhan</surname>
          </string-name>
          , İ.
          <source>İşeri “Counter Propagation Network Based Extreme Learning Machine.” Neural Process Lett</source>
          <volume>55</volume>
          (
          <year>2023</year>
          ):
          <fpage>857</fpage>
          -
          <lpage>872</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>