<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Hybrid Neuro-Fuzzy Model with Immune Training for Recognition of Objects in an Image</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Kharkiv National University of Radio Electronics</institution>
          ,
          <addr-line>Kharkiv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>0000</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Modern systems for image processing and analysis are characterized by the active use of artificial neural networks, for training of which, as a rule, gradient methods are used, but their main limitation of the implementation is high computational cost. The use of the principles of hybridization of neural networks, fuzzy logic and evolutionary algorithms allows you to create new types of models that have a higher recognition quality while reducing the computational cost of training. A hybrid neuro-fuzzy recognition model is proposed, which consists of two modules: a convolutional module (CNN) and a neuro-fuzzy classifier module (NFC) built on the basis of a modified ANFIS network. The CNN module, which is trained by the method of back propagation error, acts as a kind of expert system for the NFC module. It is proposed to perform NFC training based on the use of artificial immune systems by presenting all adjustable parameters in the form of a structured adaptive multiantibody, and consists in adjusting the NFC parameters and structure. Experimental studies have been carried out on test samples, confirming the effectiveness of the proposed model for recognizing objects in an image.</p>
      </abstract>
      <kwd-group>
        <kwd>object recognition</kwd>
        <kwd>hybrid model</kwd>
        <kwd>convolutional module</kwd>
        <kwd>neurofuzzy classifier</kwd>
        <kwd>immune learning</kwd>
        <kwd>multi-antibody</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Today, one of the most rapidly developing scientific and technological areas is image
processing and analysis. In recent years, a number of methods, models, and
algorithms have been introduced to address these challenges [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], among which
artificial neural networks (ANN) are the most effective [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. There are a large number
of standard ANN architectures, and the use of different types of ANN has proven to
be quite effective in solving a wide range of problems. The emergence of a new type
of ANN - the Convolutional Neural Network (CNN) [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ] gives a new impetus in this
area of research, it is widely used in various fields, including image recognition,
speech recognition, natural language processing, video analysis, etc. [
        <xref ref-type="bibr" rid="ref10 ref5 ref6 ref7 ref8 ref9">5-12</xref>
        ].
      </p>
      <p>
        When solving applied problems in order to improve accuracy and reduce complexity,
the task of finding the optimum neural network (NN) topology and according to it
Copyright © 2020 for this paper by its authors. Use permitted under Creative
License Attribution 4.0 International (CC BY 4.0). ICST-2020
Commons
structural (determining the number of hidden layers and neurons in them, interneuronal
connections of individual NN) and parametric (adjusting the weight coefficients of NN)
optimization. Typically, methods that require the calculation of the gradient of the selected
functional are used for NN training [
        <xref ref-type="bibr" rid="ref2">2, 13</xref>
        ]. Methods for solving this problem include
the organization of parallel and distributed computing on specialized hardware, which
can significantly imcrease productivity in solving many practical problems, as well as
the use of new evolutionary approaches to the learning of ANN, in particular artificial
immune systems (AIS).
      </p>
      <p>The construction of hybrid NN, consisting of different types of NN, each of which
is trained by a specific algorithm in layers, in most cases can significantly improve the
efficiency of their functioning. The study of the principles of NN hybridization, fuzzy
logic and evolutionary algorithms allows us to create new types of models that have a
higher quality of recognition in the case of simultaneous reduction of computational
costs [14]. One such approach that proposed in this work is hybrid neuro-fuzzy model
for recognizing objects in an image on the basis of the CNN and the neuro-fuzzy
classifier (NFC), which is taught by AIS.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Analysis of approaches to recognize object in the image</title>
      <p>The purpose of the recognition procedure is to answer the question: does the object
described by the given characteristics relate to the grouping of objects that we are
interested in, and if so, to which one. This is a fundamental problem in the computer
vision and forms the basis for solving the other tasks, such as the detection,
localization and segmentation of objects in the image [15]. Any recognition system
includes both the process of synthesizing images, that is, forming descriptions of
recognition objects and their classes, and the process of analyzing images, that is, the
decision-making process itself. The current level of computing allows to combine in
recognition systems both different approaches to describe images as well as methods
used in the recognition process.</p>
      <p>The use of CNN, especially deep learning models, has influenced the development
of methods for detecting and classifying objects in an image [16]. Deep learning
models that use several levels of nonlinear information processing to extract and
transform features, as well as analyze and classify patterns, have become the leading
architecture for most image recognition, classification, and detection tasks [17, 18].
Deep CNN (DCNN), which use GPUs and large datasets for deep learning, have
implemented technologies related to various aspects of their improvement [19-22]:
1) network architecture; 2) nonlinear activation functions; 3) mechanisms of
regularization; 4) optimization methods, etc. However, the resulting models were
often large and slow to calculate.</p>
      <p>
        The main limitations of the known methods and technologies currently in use arise
from inefficiency of solving the problem of ANN training, adjusting and adapting to
the problem area, processing incomplete and inaccurate source information,
interpreting data and accumulating expert knowledge, uniform presentation of
incoming information from different sources, etc. Therefore, one of the leading trends
is the development of integrated, hybrid systems based on deep learning [
        <xref ref-type="bibr" rid="ref6">6, 14</xref>
        ]. Such
systems consist of different elements (components), united in order to achieve the set
goals. The integration and hybridization of different methods and technologies allows
to solve complex problems that cannot be solved on the basis of individual methods
or technologies. Today, there is a tendency for hybridization of neural network
models and fuzzy logic systems, which combine the ability to represent and process
fuzzy knowledge as a base of fuzzy production rules, and the ability to learn on a
limited set of examples with a further generalization of the knowledge gained.
      </p>
      <p>This paper aims to develop and study a model that combines ANN, fuzzy logic and
AIS technologies for effective image recognition in the image.The result of the
research is the creation of a recognition system that, using the developed models as
the base architecture, will perform the task of recognizing the objects in an image.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Implementation of a hybrid neuro-fuzzy recognition model</title>
      <p>A combination of two modules was used to create the hybrid neuro-fuzzy model of
recognition: the CNN convolutional module and the ANFIS-based NFC module. The
general architecture of the neuro-fuzzy convolution network is presented in Fig. 1. In
this hybrid model, the CNN module acts as a kind of expert system for the module
with neuro-fuzzy output. CNN receives three channels of 174 x 174 pixel RGB image
on its inputs. Thereafter, it goes through four stages of alternating convolution of the
input matrix using the 3x3 coagulation kernel and the process of subsampling the
maximum values using the 2x2 working matrix. As a result, each of the 4 layers, after
completing the full lifecycle, will highlight certain features of the image objects,
which as a result of the identifications will be presented as feature maps on the source
layer CNN 9x9x128.</p>
      <p>e
g
a
m
I</p>
      <sec id="sec-3-1">
        <title>Convolutional module</title>
      </sec>
      <sec id="sec-3-2">
        <title>Neuro-fuzzy</title>
        <p>classificator module
s
s
a
l
C</p>
        <p>Each element of the resulting subsample is connected to the input layer of the
NFC, where the next step is the phasing process. Each fuzzy signal creates three fuzzy
membership functions that characterize the degree of signal saturation (low, middle,
high). Based on the activation of production rules from the knowledge base of the
NFC, the values of the signals with the highest value are selected, and then the final
result of the system operation is formed on the layers of aggregation and dephasing.
The CNN convolutional module consists of two types of layers: convolutional layers
and subsampling layers. As they alternate, they form the input feature vector for a
hybrid neuro-fuzzy classifier. CNN learning algorithm is based on backpropagation.
CNN can run quickly on a sequential machine and learn quickly by parallelizing the
convolution process on each card, as well as the reverse convolution when spreading
a network error. Compared to other algorithms for processing and synthesizing the
features of the objects in the image, CNN uses relatively little preprocessing. For the
hybrid pattern recognition model, the architecture of a 4-level convolutional CNN
module with one input 3-channel RGB layer was taken (Fig. 1). For convolution
feature kernels 3x3 size and subsample maps, the value of which is equal to 2x2
elements were used on all network layers.</p>
        <p>The convolutional layer is the main building block of CNN. The layer parameters
consists of a set of filters (kernels) for training that have a small receptive field. In the
direct pass, each filter performs convolution by rows and columns of the input matrix,
calculating the scalar product of the filter and input data, forming a 2-dimensional
excitation map of this filter. As a result, the network learns, with the filters activated
when the network detects a particular specific type of feature in a particular spatial
position at the input. The size of all maps of the convolutional layer is the same and is
determined by the formula:</p>
        <p>( r, c ) = f ( mC − kC + 1, nR − hR + 1) , (1)
where ( r, c ) is the calculated size of the convolution card; mC is number of
columns the previous map; nR is number of rows of the previous map; kC
number of core columns ; hR – number of core rows.</p>
        <p>A kernel is a filter or window that slides across the input map area and finds certain
features of objects. The kernel moves through the input map and performs the
convolution operation used to process the images:
( f * g )m, n = 
k ,l
f ( m − k, n −1) * g[k, l] ,
(2)
where f – the basic image matrix; g – convolutional kernel. A window of kernel
size g   passes through the specified step through all the image f , elementwise
multiply the contents of the window by kernel g , the result is summed up and
written to the original result matrix. In a simplified form, the convolutional layer can
be described by the expression:
(3)
(4)
xl = f ( xl−1 * k l + bl  ) ,
where xl – the output of layer l ; f () – the activation function;  bl – the shift factor
of layer l ; * – convolution operation of input x with kernel k . In this case, due to
the boundary effects, the size of the output matrices decreases:</p>
        <p> l 
xlj = f  (xil−1 * kil + bi )  ,</p>
        <p> i 
where xil – map of the signs of the i-th card (input of layer l ); f () – the activation
function; bil – the shift factor of the i-th map of layer l ; kil – the convolution kernel
of the i-th map of layer l ; * – convolution operation of input x with kernel k .</p>
        <p>
          In order for the neural network to have good adaptive properties, it must have
nonlinear components or perform nonlinear transformations. Given this, within CNN,
the weighted sum of inputs (the result of the convolution operation) passes through
the ReLU function [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ], i.e., each element of the output convolution matrix is an
artificial neuron with this type of activation function. Thus, the values in the final
object maps are not really the sum of the product of values of the kernels of features
and input matrices, but the result of the activation function applied to them.
        </p>
        <p>The subsample layer, like the downsample layer, has maps, but the number is the
same as the previous convolutional layer associated with it. The purpose of the layer
is to reduce the dimension of the original matrix of the previous layer. If some
features have already been detected in the previous convolution operation, then such a
detailed image is no longer required for further processing and is compacted to a less
detailed one. In the process of scanning with the core of the subsample layer (filter) of
the map of the previous layer, the selected areas of the scanning core do not intersect
unlike the convolutional layer. Each of them does work related only to their own data
area. Each card has a 2x2 size kernel, which allows you to reduce the previous
downsample matrices by 2 times. The entire feature map is divided into 2x2 element
area, from which the maximum value elements are selected. The work of the
subsample layer is described by the formula:
xl = f ( al * subsample ( xl−1 ) + bl ) ,
(5)
where xl – the output of the layer l ; f () – the activation function; al ,  bl – the shift
coefficients of layer l ; subsample() –operation of selecting local maximum values.</p>
        <p>
          Like all direct signal propagation neural networks, CNN are well trained using the
error propagation algorithm [
          <xref ref-type="bibr" rid="ref2 ref4">2, 4</xref>
          ]. However, the use of this approach in this type of
network has its own characteristics. One of the biggest problems with CNN training is
the calculation of the δ-error on the subsample layer. This process is represented in
several variations within the network. The first case is when the subsample layer is in
front of the output signal, then it has neurons and connections of the same type as in
the associated layer, so calculating the δ-error is no different from calculating the
δerror of the next layer. The second case is when the subsample layer is in front of the
convolutional level. Then the calculation of the δ-error occurs by backward
convolution. The purpose of a backward convolution (deconvolution) is to find the
equation of a direct convolution given in the form f * g = h , where h - is the
recorded signal and f is the signal that need to be restored, and it is known that the
first signal is obtained by convolution of the second signal with some known signal
g . However, if we do not know g in advance, then we need to estimate it. This is
usually done by using statistical estimation methods.
        </p>
        <p>Usually, the previous layer after the convolution layer is subsample layer, and
according to the task we need to calculate the δ-errors of the current layer
(convolutional) on the basis of knowledge about the subsample δ-errors. In fact, the
δerror is not calculated but copied. When the signal is propagated directly, the neurons
of the subsample layer are formed by nonoverlapping scan windows over a
downsampling layer, in which the neurons with the maximum value are selected. In
reverse propagation, the δ-errors are returned only to the neuron that was previously
selected with the maximum output signal in its subgroup. Others get a zero δ-error.
3.2</p>
      </sec>
      <sec id="sec-3-3">
        <title>Neuro-fuzzy classifier with immune training</title>
        <p>The NFC structure is based on a modified ANFIS network, which is a multilayer
network with direct signal propagation that implements the first-order Takagi-Sugeno
fuzzy output algorithm for n input variables described by  m fuzzy sets [23, 24]:
  R i : IF  xi  is Ai1  AND AND x j  is Aij  AND  AND xm  is Aim  ,  
          THEN  y = ki1x1 +  + kij x j +  +kim xm + ki0 , i = 1,, n ,
(6)
where x j , j =1 , m</p>
        <p>– input variables,  yi , i =1 , n – output variables Aij – linguistic
terms representing fuzzy sets;  kij – coefficients of arguments of function; ki0 – shifts.</p>
        <p>The network consists of 5 structural layers (Fig. 3). The first layer is the entry
level. It consists of n nodes, where n - the number of inputs to the system. The
elements of this layer are combined with the output data obtained from the
convolution module. This establishes a link between the two modules – from each
element of the feature matrix a separate entry channel into the NFC is allocated. The
second layer performs the function of fuzzification the input. The outputs of the neurons of
this layer represent the values of the membership functions (MF)  Aji
( x j ) for input
variables x j , j =1 , m . A Gaussian MF is used, which is described by the expression:
  x − c 2 
 ( x ) = exp −    ,
    
(7)
where parameter C denotes the center of the fuzzy set; parameter   – the coefficient of
stretching of the function. The choice of a Gaussian MF is stemmed from its sufficient
flexibility and simplicity, which reduces the dimension of the optimization problem when
performing fuzzy model training. In this layer are m  p elements, where m – the
number of input variables, and p – the number of membership functions. For
fuzzification of the input variables, three membership functions are used, which
describe the degree of signal saturation, which can vary from 0 to 255.</p>
        <sec id="sec-3-3-1">
          <title>Level 3</title>
        </sec>
        <sec id="sec-3-3-2">
          <title>Level 4</title>
        </sec>
        <sec id="sec-3-3-3">
          <title>Level 5</title>
          <p>The fifth layer implements the function of activation of neurons (consequence), each
neuron in it is described by the expression:</p>
          <p>wi f = wi ( ki0 + ki1x1 + ki2 x2 + ... + kim xm ) = yi = 01,,  iiff y yi i   00,,
where kij , i =1 , pn , j =1 , m – parameters of i -th rule. The outputs of this layer are
the outputs from the network, each result of which is the value of the ratio of the
result to a particular class, which are predetermined according to the conditions set by
the recognition task.</p>
        </sec>
        <sec id="sec-3-3-4">
          <title>Structure of neuro-fuzzy classificator</title>
          <p>In the third layer, the input signals are multiplied and the weights of the rule
(aggregation) are formed:
where pn – number of fuzzy rules.</p>
          <p>Each i -th neuron in the fourth layer serves to calculate the ratio of the weight of the
ith rule to the sum of the weights of all the rules (normalization):</p>
          <p>m
wi =  Aij ( x j ) ,  i =1 , pn ,</p>
          <p>j=1
wi = 
wi
mn
i=0wi</p>
          <p>,  i =1 , pn .</p>
          <p>Level 1
x1</p>
          <p>The second and fifth layers are adaptive, the settings of which are made during the
training of the network. In the second layer, the MF parameters (centers of fuzzy sets C
and stretching factors  are adjusted for the Gaussian MF), and in the fifth - the
parameters of fuzzy rules kij , i =1 , pn , j =1 , p . Some neurons in the NFC are
adaptive, and each neuron output depends on the parameters related to that neuron. The
learning rule determines how these parameters change to minimize the fuzzy output error.</p>
          <p>The NFC training represents the following optimization task:</p>
          <p>1
n * m
im=1nj=1  yi ( x j , P ) − yir 2   mPin,
(11)
where yir
– the required values of the NFC outputs; P
– the vector of NFC
parameters, cij ,   ij and kij , i =1 , pn , j =1 , m   are used as parameters, yi ( yi , P )
– i -th output NFC value with inputs values x j , dependent from parameter P .</p>
          <p>We will perform NFC training on the basis of AIS [25] in order to adjust the
parameters of the MF input variables and fuzzy rules coefficients, as well as to
synthesize NFC – to remove redundant rules and to corresponding change the
network structure. The main idea of teaching NFC with the use of AIS is to present a
solvable problem in the form of an antigen, and its possible solutions - in the form of
antibodies [26]. We form a population of antigens from the training sample examples
Ag =  Ag1, Ag2 ,, AgM , , where M – the size of the antigen population
corresponding to the number of examples in the training sample. Each antigen is
represented by a fixed length vector: Agi = x1i , x2i ,, xni , y1i , y2i , yki , , i =1 , M ,
where x1i , x2i ,, xni – input variables, and y1i , y2i , yki – output NFC variables for
i -th example of training sample. In one antibody, all configurable parameters of the
NFC are encoded:  cij ,  ij and kij , i =1 , pn ,  j =1 , m .</p>
          <p>To train the NFC, a model is used for representing all tunable parameters in the
form of a structured adaptive multi-antibody [27, 28], presented in fig. 2 where cij ,
 ij , i =1 , pn ,  j =1 , m – parameters of Gaussian MF for n input variables, each of
them has m terms; ksv , s =1 , n ,  v = 0, m – coefficients n of fuzzy rules (6).</p>
          <p>c11, , c1m ,, cn1, , cnm ;
 11, , 1m ,, n1, , nm</p>
        </sec>
        <sec id="sec-3-3-5">
          <title>Part 1</title>
          <p>Ab0
k10 , , k1n
Ab1
…
…</p>
        </sec>
        <sec id="sec-3-3-6">
          <title>Part 2</title>
          <p>kv0 , , kvn</p>
          <p>AbL−1</p>
          <p>Fig. 4. Multi-antibody structure
The
population
of
multi-antibodies
can
be
represented
as
mAb = mAb1, mAb2 ,, mAbn  ,
where
mAbi =  Ab0 , Ab1,, AbL−1 ,
i =1 , N – i-th multi-antibody, which is a structured vector whose length L changes
during the execution of the immune algorithm; N – multi-antibody population size.
Each multi-antibody mAbi  , i =1 , N</p>
          <p>of the population is characterized by a full set
of tunable parameters of NFC. In the structure of the multi-antibody shown in Fig. 4,
separation of configurable parameters is used in two independent parts: part 1 with
MF parameters cij ,   ij , i =1 , pm , j =1 , m , and part 2 with the parameters of fuzzy
rules ks,v , s =1 , n, v = 0, m.</p>
          <p>Part 1 of multi-antibody represents a single antibody Ab0 , whose length is fixed,
because the number of phase transitions for the input variables is constant. Part 2 of
multi-antibody consists of ( L − 1 )  independent antibodies Ab1,, AbL−1 , each of
them contains coefficients ks,v , s =1 , n, v = 0, m for one fuzzy input rule. Part 2 of
the multi-antibody is adaptive because in the learning process, in addition to
optimizing the coefficients of fuzzy rules contained in this part, their total number
also changes, i.e. structural synthesis is performed. Structural synthesis, in turn, leads
to a change in the number of neurons in the hidden layers of the NFС (Fig. 3).</p>
          <p>A structured method of forming multi-antibodies allows to increase the efficiency
of the immune algorithm due to the separate use of immune operators to each part of
the antibody - the immune operators are applied to the parameters of the MF and the
coefficients of the fuzzy rules separately. The affinity calculation is performed for the
multi-antibody as a whole, using both parts. The initial population of multi-antibodies
is generated randomly. The size of the population of multi-antibodies is chosen small,
because the use of a large number of multi-antibodies in a population leads to large
computational costs.</p>
          <p>The immune algorithm that implements the training of NFC is based on the
principle of clonal selection and the theory of the immune network [25, 26]. The
adjustment of the NFC parameters is performed according to the principle of clonal
selection. The clonal selection algorithm provides support for the diversity of
multiantibodies in a population by globally viewing the range of tunable parameters,
avoiding local minimum. The application of the provisions of the theory of the
immune network allows us to evaluate the interaction of antibodies with each other
and do a suppression, thus eliminating the redundancy of the fuzzy inference rules.
Formally, the immune learning algorithm of NFC can be represented as follows:
ImmAlg ( PL , L, mAb, N , Ag, M , Op, nc , Nc , d , net , Alph, A, Aff , gen, t ) 
(12)
where P L</p>
          <p>– the search range; L – the search range dimension; mAb
–multiantibody population where mAb = mAb1, mAb2 ,..., mAbN  ; mAbi – i -th
multiantibody of population
mAb : mAbi =  Ab0 , Ab1, Ab2 ,, AbL−1 ; N
antibody population size; Ag – antigen population: Ag =  Ag1, Ag2 ,, AgM  ;
M
– learning sample size;</p>
          <p>Op</p>
          <p>– set of used immune operators where
Op = Clone, Mutate, Edit, Suppress ;  nCl
–
number
of
multi-antibodies for
cloning; NCl – number of clones of one multi-antibody d – the number of
multiantibodies with the worst affinity to be replaced when editing the population of
antibodies;  net – network compression ratio; Alph – the alphabet with which
antibodies are encoded; A – the power of alphabet Alph ; Aff – affinity function;
gen – work generation of immune algorithm; t – algorithm termination criterion.</p>
          <p>NN training algorithm is an iterative procedure of the sequential identifying
observations from the training sample. Antigens are examples of a training sample.
Each antibody encodes one possible solution, and the amount of antibodies in the
population corresponds to the number of examples in the training set. The training
algorithm of NN is the following sequence of steps:</p>
          <p>1. Initialization of the initial population of multi-antibodies mAb - performed
randomly. Calculation of the affinity of each multi-antibody for antigen
2. Affinity calculation AffmAb− Ag of each multi-antibody for antigen:
−1
AffmAb− Ag = (1 + dmAb− Ag  ) ,
(13)
(14)
where dmAb− Ag – Hamming distance between the obtained output values of NN
yi , i = 1, n  and desired yir for all S population antigens Ag :
dmAb− Ag = in=0yi , where yi = 1,  if  yi   yir  .</p>
          <p>0, if  yi = yir</p>
          <p>To calculate the affinity of a multi-antibody, it is necessary to substitute the
parameters encoded in the multi-antibody into the NFC. The input features xm are
fed to the network input and the values of the output ys variables are calculated.
Thus, the affinity of each multi-antibody mAbi  is calculated in relation to the entire
population of antigens Ag .</p>
          <p>3. Cloning of multi-antibodies is proportional to their affinity and formation of a
clone population Cl . The parameters of the cloning operator are the number of
antibodies  nCl
clone and the cloning ratio of multi-antibodies
NCl   . A fixed
parameter value  nCl is used. The cloning ratio of multi-antibody NCl   is regulated
during the operation of the immune algorithm, depending on the affinity of the
multiantibody according to the ratio:</p>
          <p>NCl   =  * NCl _ min + (1 − ) * NCl _ max  ,
where   =</p>
          <p>Affbest − AffmAb− Ag ;  NCl _ min and NCl _ max – minimum and maximum</p>
          <p>Affbest
cloning ratio of multi-antibody; Affbest – the best affinity value in the current
generation.</p>
          <p>4. Clone mutation is inversely proportional to the affinity of multi-antibodies and
the formation a population of mutated clones MC . Mutation of selected
multiantibody parameters mAb performed by adding a gaussian noise:</p>
          <p>mAbi+1 = mAbi + N ( 0, i )
To
change
the
variance
of
a
random
variable</p>
          <p>Affbest − AffmAb− Ag
ratio   i =  i</p>
          <p>Affbest − Affworst
in the current generation.</p>
          <p>5. Calculation the affinity of a population of mutated clones MC in order to (13).
If, as a result of a mutation, affinity improves, then replace with clones the
corresponding multi-antibodies in the population  mAb .</p>
          <p>6. Calculation the affinity of antibodies within part 2 of multi-antibodies.
Suppression of antibodies with affinity greater than a predetermined threshold  net .</p>
          <p>is used, where Affworst – the worst affinity value
The affinity calculation is performed in accordance to the expression:
Aff Ab− Ab = (1+ | Ab1 − Ab2 |)
−1</p>
          <p>
= 1 +

Perform suppression by removing antibodies
with affinity greater than a
given threshold  net , allows to reduce the number of neurons and connections
between them in the hidden layer and eliminate their redundancy.</p>
          <p>7. Checking the stopping criterion. As a stopping criterion, either the achievement
of a given affinity threshold or the achievement of a given number of generations of
the algorithm's operation is used. The result of the algorithm will be a multi-antibody
with the best affinity by the population, which determines the structure of the NFC
and containing its configured parameters.</p>
          <p>Parts 1-5 of the algorithm correspond to the principle of clonal selection. At these
stages, the algorithm works with both parts of the multi-antibody. Step 6 correspond
to the principle of networking. If previously the multi-antibody was processed as a
normal antibody, then at this step the work is performed only with part 2 of the
multiantibody, which consists of individual antibodies representing the parameters of fuzzy
 i
(16)
the
output rules kij , i =1 , pn ,  j =1 , m . The structure of the NFC (the number of
neurons in the hidden layers) is set in accordance to the number of fuzzy rules
encoded in the antibody.</p>
          <p>The result of the algorithm is an antibody with the best affinity by the population
that contains the parameters of the fuzzy output rules and the MF of the input
variables.
4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experimental results</title>
      <p>Ready-made architectural solutions (libraries) were used to build the models for
each individual module. The generation of the CNN module was performed based on
keras-sharp library which is used to generate deep learning networks. In its turn, the
NFC module was based on the fuzzy-class-net software package. But its training was
done using AIS. System that was obtained as a result of compilation of two modules
was run on a computer that had the following characteristics:
– ОS Windows 10;
– the number of physical cores – 2;
– RAM – 4 GB;
– CPU – Intel core i5-3210M 2500МГц.</p>
      <p>For the objective analysis of the results, the obtained data of the investigated
hybrid neuro-fuzzy model were compared with similar output values obtained using
the standard CNN network. This approach makes it possible to evaluate the
advantages and disadvantages of the proposed model.</p>
      <p>A set of 174 x 174 pixel images was selected for the training. Each image has one
target object that needed to be identified during the test experiments. The entire
sample of images is divided into two classes: class 1 - cats images and class 2 dogs
images. Each set of test data images consists of 1000 instances. 750 copies of each set
of test data were randomly selected to train the networks. And for testing the model
250 copies remaining in each of the sets, plus 250 were randomly selected from those
that are already used for training in both classes.</p>
      <p>The results of the training are shown in Fig. 5, where the error values are on the Y
scale and a number of epochs on the X scale. As we can see from Fig. 5, during the
course of 1500 training epochs, the total error value produced by the hybrid NFN
stopped at 0.2. For more accurate metrics, you need to increase the number of initial
training sample and the number of training epochs. From the 350 epoch relatively
correct answers began to form at the output of the classifier with an accuracy of the
result in the range of 65 - 80%. In turn, for standard CNN, more accurate output
values began to form from the 300 epoch, and their values ranged from 64 to 79%. In
the last 1500 epoch, the value of network error was approximately the same.</p>
      <p>As a result of network training, one feature was noticed. The time to perform each
training iteration of the hybrid NFN epoch was approximately 5 percent greater than
time of a standard CNN network. This is due to the fact that during the training of the
hybrid NFN, at the time of neuro-fuzzy classifier training, additional operations are
performed to adjust the structure and parameters of the production rules.</p>
      <p>The next phase of the experiments was to test the trained systems on how they
would recognize the test data set, which consisted of new, not involved in training and
used images. Each model input received 500 images from each class. The following
results were obtained as a result of the test. The accuracy of object recognition on the
image for the hybrid NFN was 80%, and in the CNN network the ratio was fixed at
78%. Detailed information on this parameter for each of the test classes is given in
Table. 1.</p>
      <p>Table 1 – Comparison of the recognition accuracy of classes
Image class Class 1 Class 2</p>
      <p>Model of Standard Hybrid Standard
recognition CNN NFN CNN</p>
      <p>Recognition 77.27% 80.12% 78.49%
accuracy (in %)</p>
      <p>Number of 372 398 374
recognized images</p>
      <sec id="sec-4-1">
        <title>Hybrid NFN 79. 84% 388</title>
        <p>The other equally important indicator is the number of images on which object
classes were correctly recognized. In order for an object to be recognized, the system
must give an accuracy value greater than or equal to 65% in this test. As a result of
the experiment, it was found that for the hybrid NFN, it was 786 precisely found
objects, and for CNN this figure was 746 objects. The final indicator difference is 40
images.</p>
        <p>From the results of the experiments, it follows that the hybrid NFN algorithm
developed slightly slower than the CNN algorithm during the training. Because when
you perform the training iteration, additional calculations are performed to set up
production rules, which takes up to 5% of the time. But the proposed model did better
in recognizing the objects in the image. It was found that the result of image
recognition accuracy is greater than that of CNN by 1.5% and the number of correctly
recognized images is greater by 40 elements, which is 8% of the total sample of test
data.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>The construction of hybrid NN, consisting of NN of different types, each of which is
trained according to a specific algorithm, makes it possible to increase the accuracy
and reduce the complexity of solving practical problems, including the problem of
recognizing objects in an image. The use of NN hybridization principles, fuzzy logic
and evolutionary computations allows the creation of new types of models that have a
higher quality of recognition while reducing computational training costs.</p>
      <p>To solve the task of recognizing objects in an image, it is proposed to use a hybrid
neuro-fuzzy model, which consists of two modules: a convolutional module and a
neuro-fuzzy classifier (NFC) module. The convolutional module is implemented on
the basis of CNN, which is trained using the algorithm of back propagation error. The
NFC module is implemented on the basis of a modified ANFIS network.</p>
      <p>It is proposed to train the neuro-fuzzy classifier on the basis of the immune
approach by adjusting its parameters and structure. A model which is proposed for the
representation of all tunable network parameters in the form of a structured adaptive
multi-antibody, consisting of two parts. The structured method for the formation of a
multi-antibody allows to increase the efficiency of immune algorithms due to the
separate application of immune operators to each part of the multi-antibody and thus
solve the task of synthesizing a neuro-fuzzy classifier.</p>
      <p>Experimental studies have shown that the proposed hybrid neuro-fuzzy model
showed better performance in object recognition in an image than the standard CNN.
2018.
11. Jungmo Ahn, JaeYeon Park, Donghwan Park, Jeongyeup Paek, JeongGil Ko.</p>
      <p>Convolutional neural network-based classification system design with compressed
wireless sensor network images / Published: May 8, 2018.
12. Mingyuan Xin, Yong Wang. Research on image classification model based on deep
convolution neural network /EURASIP Journal on Image and Video Processing, Article
number: 40, 2019.
13. Y. Lecun ; L. Bottou ; Y. Bengio ; P. Haffner Gradient-based learning applied to
document recognition / Proceedings of the IEEE ( Volume: 86 , Issue: 11 , Nov. 1998 ), .
2278 – 2324.
14. Castillo O. Melin P. Hybrid Intelligent Systems in Control, Pattern Recognition and
Medicine / Studies in Computational Intelligence. Springer Nature Switzerland AG,
2020, 381 p.
15. Karpathy, A., &amp; Fei-Fei, L. Deep visual-semantic alignments for generating image
descriptions / In Proceedings of the IEEE Conference on Computer Vision and Pattern
Recognition. Red Hook, NY: Curran. 2016. pp. 3128–3137.
16. Boukaye BoubacarTraoreabBernardKamsu-FoguemaFanaTangara Deep convolution
neural network for image recognition / Ecological Informatics // Vol. 48, 2018, pp.
257268.
17. LeCun, Y. Handwritten digit recognition with a backpropagation neural network IY.</p>
      <p>LeCun, B. Boser, J. S. Denker, D. Henderson, R. Howard, W. Hubbard, L. Jackel
//Advances in Neural Information Processing Systems. 1990. № 2. pp. 396-404.
18. LeCun Y., Bengio Y., Hinton G, Deep Learning / in Nature 521 (7553): 436-44, 2015.
19. Krizhevsky, A., Sutskever, I., &amp; Hinton, G.E. ImageNet Classification with Deep
Convolutional Neural Networks / Part of: Advances in Neural Information Processing
Systems 25 (NIPS 2012), Red Hook, NY, 2012, pp. 1097–1105.
20. Li Deng, Dong Yu. Deep Learning: Methods and Applications / MSR-TR-2014-21, 2014.
21. Karen Simonyan, Andrew Zisserman. Very Deep Convolutional Networks for
Largescale Image Recognition / arXiv: 1409.1556, 2014.
22. Matthew D. Zeiler, Rob Fergus. Visualizing and Understanding Convolutional</p>
      <p>Networks / Computer Vision – ECCV 2014, pp 818-833.
23. Zadeh, Lotfi A. A Theory of Approximate Reasoning. In: Hayes, Jean E. Michie, Donald
and L. I. Mikulich (eds.): Machine Intelligence, vol. 9 (based on the International
Machine Intelligence Workshop), New York: Elsevier, 1979, pp. 149–194.
24. Kasabov N. The Evolution of the Evolving Neuro-Fuzzy Systems: From Expert Systems
to Spiking-, Neurogenetic-, and Quantum Inspired / A Bradford Book The MIT Press
Cambridge, Massachusetts London, England, 2013, pp. 165-175.
25. Dasgupta D. Immunological computation, Theory and applications / D. Dasgupta, L.F.</p>
      <p>Nino – Taylor &amp; Francis Group, 2009. – 278 p.
26. Dasgupta D. Recent Advanced in Artifical Immune Systems: Models and Applications /</p>
      <p>D. Dasgupta, S. Yu, F. Nino // Applied Soft Computing. Elsevier, 2011. – P. 1574-1587.
27. Korablev N. Immune Approach for Neuro-Fuzzy Systems Learning Using Multiantibody
Model / N. Korablev, I. Sorokina // ICARIS 2011, Springer Lecture Notes in Computer
Science. – 2011. – Vol. 6825. – P. 395–405.
28. Korablyov M., Axak N., Soloviov D. Hybrid evolutionary decision-making model based
on neural network and immune approaches // 2018 IEEE 13th International Scientific and
Technical Conference on Computer Sciences and Information Technologies (CSIT) 2018.
2018. V.1. P. 378–381.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Duda</surname>
            <given-names>R.O. Pattern classification</given-names>
          </string-name>
          / R.O.
          <string-name>
            <surname>Duda</surname>
            ,
            <given-names>P.E.</given-names>
          </string-name>
          <string-name>
            <surname>Hart</surname>
            ,
            <given-names>D.G.</given-names>
          </string-name>
          <string-name>
            <surname>Stork</surname>
          </string-name>
          . - Wiley &amp; Sons,
          <year>2010</year>
          . - 738 p.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Khaikin</surname>
            <given-names>S.</given-names>
          </string-name>
          <article-title>Neural networks. Full course: A Comprehensive Foundation, 2nd edn</article-title>
          , Williams, Moscow,
          <year>2006</year>
          . - 1104 p.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>LeCun</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <article-title>Convolutional networks for images, speech, and timeseries I Y</article-title>
          . LeCun,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <article-title>Bengio II The Handbook of Brain Theory</article-title>
          and
          <string-name>
            <given-names>Neural</given-names>
            <surname>Networks</surname>
          </string-name>
          .
          <year>1995</year>
          . pp.
          <fpage>255</fpage>
          -
          <lpage>258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4. GoodfellowI. Deep learning / I. Goodfellow,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          : -The MIT Press,
          <year>2016</year>
          . - 800 p.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Zheng</surname>
            <given-names>YI</given-names>
          </string-name>
          .
          <article-title>Evaluation and Implementation of Convolutional Neural Networks in Image Recognition / First International Conference on Advanced Algorithms and Control Engineering</article-title>
          .
          <source>IOP Conf. Series: Journal of Physics: Conf. Series</source>
          <volume>1087</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>Waseem</given-names>
            <surname>Rawat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Zenghui</given-names>
            <surname>Wang</surname>
          </string-name>
          .
          <source>Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review / Neural Computation 29</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>2352</fpage>
          -
          <lpage>2449</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Qingdong</given-names>
            <surname>Wei</surname>
          </string-name>
          , Fengjing Shao,
          <source>Ji Lin. Research Summary of Convolution Neural Network in Image Recognition / Proceedings of the International Conference on Data Processing and Applications</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>39</fpage>
          -
          <lpage>44</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Jason</given-names>
            <surname>Brownlee</surname>
          </string-name>
          .
          <article-title>Convolutional Neural Network Model Innovations for Image Classification / Deep Learning for Computer Vision</article-title>
          ,
          <year>2019</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Venkatesan</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <source>Li Baoxin. Convolutional Neural Networks in Visual Computing</source>
          , CRC Press,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10. Keunyoung Park,
          <string-name>
            <surname>Doo-Hyun Kim</surname>
          </string-name>
          .
          <source>Accelerating Image Classification using Feature Map Similarity in Convolutional Neural Networks / Journals Applied Sciences</source>
          , Vol.
          <volume>9</volume>
          ,
          <issue>Iss</issue>
          .
          <volume>1</volume>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>