<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Information Control Systems &amp; Technologies, September</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Multicriteria optimization of hybrid convolutional neural network structural synthesis using evolutionary algorithms</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Illia Boryndo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Siveglazov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Michael Z. Zgurovsky</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Aviation University</institution>
          ,
          <addr-line>Kyiv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Technical University of Ukraine</institution>
          ,
          <addr-line>Kyiv</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>2</volume>
      <fpage>3</fpage>
      <lpage>25</lpage>
      <abstract>
        <p>This paper defines and describes the promising architectural solutions for convolutional neural networks and considers their key parameters for further structural and parametric synthesis. It s been proven that for better qualitive results these networks should include functional blocks such as SRU (Spatial Reconstruction Unit), CRU (Channel Reconstruction Unit), dense residual attention unit, etc., to the traditional supportive layers (batch normalization layer, 1x1 convolutional layer, dropout layer, etc). Described the advantages and disadvantages of different blocks as well as reasoning of their usage during structural synthesis. It is proposed to use a genetic evolutionary algorithm for structural-parametric synthesis and reviewed modern approaches. It is shown and described the process of configuring evolutionary algorithm. Based on optimization criteria the fitness function, selection, mutation and crossover approaches were defined. The results of experimental evolutionary process were shown and analyzed. It is considered an example of model generated by evolutionary algorithms that is based on using the functional blocks aggregated from different CNN architectural approaches. The performance criteria for each model during synthesis process is calculated, including average training time shortening, their advantages and architectural integration details. Based on the experimental results it is proven that utilizing complex structural blocks instead of traditional layers with flexible configuration of fitness function for both qualitive and performant criteria shows significant improvement for resulting model. structural-parametric synthesis; convolutional neural networks; genetic algorithm.1</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Nowadays, even with significant progress in computer vision and the use of advanced
convolutional networks, especially visual transformers [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], many image processing challenges
remain unresolved and need tailored solutions. This is largely because of the unique characteristics
of training datasets and the high computational demands of complex neural network topologies,
like transformers, which face hardware constraints. Most CNN architectures suffer from low
performance, slow learning rates, and simultaneously demand high-quality, well-balanced training
datasets.
      </p>
      <p>
        The natural way out of this situation is to create hybrid convolutional neural networks [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
However, many problems arise in this process: optimal choice of the basic topology of the
convolutional neural network, optimal choice of various structural blocks to be used in the process
of synthesizing the structure of the hybrid convolutional neural network, optimal choice of their
locations. From the point of view of machine learning, there are problems of selecting learning
criteria when we have a single-criteria or multi-criteria optimization problem, hitting a local
extremum, overtraining, gradient drop, pre-training, pre-training, pre-training, creating a hybrid
learning algorithm.
      </p>
      <p>This paper will outline and discuss the research findings on the application and architectural
characteristics of hybrid convolutional neural networks (HCCN) and various building blocks
essential for their synthesis. It involves studying the performance of individual components,
examining contemporary CNN architectures, and applying various training data to address
performance challenges. The primary focus of this study is to identify the best kinds and
arrangements of current CNNs, extract operational components, and incorporate them in the
HCNN synthesis using evolutionary algorithms to attain favorable performance, accuracy
outcomes and other optimization criteria. The main goal of this paper is to develop the
evolutionary mechanism that will utilize structural components of different CNN architectures to
create model that will satisfy predefined optimization criteria.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related works and existing approaches</title>
      <p>Convolutional Neural Networks have revolutionized the field of computer vision since their
inception, offering unparalleled performance in image classification, object detection, and
segmentation. Despite their success, the quest for improved accuracy and versatility has led
researchers to explore hybrid models that combine CNNs with other neural network architectures.
These hybrid models leverage the unique strengths of different networks, addressing the
limitations of pure CNNs and enhancing their applicability across a wider range of tasks.</p>
      <p>For structural synthesis of HCNNs there two main components evolutionary algorithm (EA)
itself and building components to be utilized by this algorithm. Ther
of evolutionary algorithms in integration with CNNs, but most of them have their own limitations.</p>
      <p>Modern evolutionary algorithms are inspired by the process of natural selection and genetic
evolution. They operate on a population of potential solutions, iteratively selecting, recombining,
and mutating individuals to produce new generations of solutions. Key modern EAs include:
Genetic Algorithms (GAs):
•
•
•</p>
      <p>Operate with a population of candidate solutions.</p>
      <p>Use crossover and mutation to explore new architectures.</p>
      <p>Selection mechanisms favor better-performing individuals.</p>
      <p>
        Genetic Programming (GP) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]:
•
•
      </p>
      <p>Similar to GAs but operates on tree-like structures.</p>
      <p>Can evolve entire programs or architectures.</p>
      <p>
        NeuroEvolution of Augmenting Topologies (NEAT) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]:
•
•
•
•
Evolutionary Strategies (ES) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]:
      </p>
      <p>Evolves both the weights and the architecture of neural networks.</p>
      <p>Introduces innovations by tracking genes and structures through generations.</p>
      <sec id="sec-2-1">
        <title>Focus on the optimization of continuous parameters.</title>
        <p>Use strategies like covariance matrix adaptation to guide the search.</p>
        <p>
          The architecture of Hybrid Convolutional Neural Networks (HCNNs) [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] integrates CNNs with
various other neural network models, capitalizing on their respective strengths to create more
powerful and versatile models. Here, we delve into the specific architectures and mechanisms of
three prominent types of HCNNs: CNN-RNN hybrids, CNN-GNN hybrids, and CNN-Transformer
hybrids.
        </p>
        <p>The architecture of Hybrid Convolutional Neural Networks exemplifies the synergy achieved by
combining the complementary strengths of different neural network models. CNNs excel at spatial
feature extraction, making them a crucial component in these hybrid architectures. When paired
with RNNs, they can model temporal sequences effectively; when combined with GNNs, they can
handle relational data adeptly; and when integrated with transformers, they can capture both local
and global dependencies in data. These hybrid architectures have demonstrated superior
performance across various domains, from video classification and image captioning to social
network analysis and molecular graph prediction, showcasing their versatility and efficacy.</p>
        <p>The goal of this paper is to define main criteria for CNN synthesis such as accuracy,
computational cost, model robustness, etc., analyze and extract structural blocks of modern CNN
architectures, modify the existing solution of evolutionary algorithms and apply it to synthesize
optimal HCNN architecture.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Problem Statement</title>
      <p>Hybrid Convolutional Neural Networks have demonstrated significant improvements in
performance across various complex tasks by integrating the strengths of Convolutional Neural
Networks with other neural network architectures. Despite their potential, designing optimal
HCNN architectures remains a challenging task due to the vast search space of possible
configurations and the intricate balance required between different components.</p>
      <p>
        To address this, multicriteria optimization [
        <xref ref-type="bibr" rid="ref2 ref7">2, 7</xref>
        ] provides a framework for systematically
evaluating and optimizing multiple performance metrics simultaneously. Evolutionary algorithms
(EAs) are particularly well-suited for this task due to their ability to explore large, complex search
spaces and their flexibility in handling multiple objectives.
      </p>
      <p>The goal of this research is to develop and configure evolutionary algorithm for the
multicriteria optimization of CNN structural synthesis using evolutionary algorithms. The key
objectives include:</p>
      <p>
        Defining the Optimization Criteria: Establishing a set of relevant performance metrics that
reflect the quality and efficiency of HCNN architectures. These criteria typically include [
        <xref ref-type="bibr" rid="ref8">8, 9</xref>
        ]:
•
•
•
•
•
•
•
•
      </p>
      <p>Accuracy: The ability of the HCNN to correctly classify or predict data.</p>
      <p>Computational Efficiency: Metrics such as inference time and memory usage.</p>
      <p>more complex tasks.</p>
      <p>CNN Structural Design Space: Identifying and parameterizing the components and
configurations of HCNNs, including:
•
•
•
•</p>
      <p>Type and Depth of CNN Layers: The number of convolutional, pooling, and fully connected
layers.</p>
      <p>Type and Integration Method of Hybrid Components: The choice between structural blocks
and components used for structural synthesis, types of integration components (e.g., SRU,
CRU, LSTM, GRU, GCN, GAT, etc.).</p>
      <p>Connectivity Patterns: How different components are connected and the flow of data
between them.</p>
      <p>Hyperparameters: Parameters such as learning rates, batch sizes, and dropout rates.</p>
      <p>Develop Evolutionary Algorithm: Design and implement an evolutionary algorithm tailored for
CNN structural synthesis. This involves:</p>
      <p>Encoding CNNs: Developing an encoding scheme for representing HCNN architectures in a
manner suitable for evolutionary operations.</p>
      <p>Fitness Function: Formulating a multi-objective fitness function that balances accuracy,
efficiency, robustness, and scalability.</p>
      <p>Selection Mechanism: Implementing a selection process to maintain diversity and guide the
search towards optimal solutions.</p>
      <p>Crossover and Mutation Operators: Creating operators to generate new HCNN
architectures by recombining and modifying existing ones based on preselected structural
blocks.</p>
      <p>By addressing these objectives, this research aims to significantly advance the state-of-the-art in
hybrid neural network design and optimization, providing a powerful tool for the development of
more efficient, robust, and scalable deep learning models.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Decomposition of modern CNN architectures and evaluation of their structural blocks</title>
      <p>Currently, there is continuous advancement in the topology analysis of modern convolutional
neural networks, with new architectural solutions and applications being continuously proposed.
Due to increased applied tasks instead of iteratively increasing the complexity most modern
convolutional neural network architectures started to implement functional structural components.
These functional structural components incorporate the set of basic layers enhanced with unique
connectivity approach, combinations and functional postprocessing. Such components could
greatly increase qualitive parameters of networks and could be both used independently or as part
of more complex structure.</p>
      <p>For structural synthesis process of CNNs the main components are the structural blocks that
will be utilized to form the generated architecture. To define the list of suitable blocks and to label
them based on their qualitive parameters and functional means it is necessary to analyze and
define such blocks within modern CNN architectures. Most of them are well known so we will take
a look on modern corner cases instead.</p>
      <p>
        Each of these blocks possesses its own distinct conceptual framework and attributes,
necessitating their evaluation both as individual entities and as paired combinations within a single
neural network. Subsequently, it is essential to conduct performance evaluations to analyze their
intrinsic properties, effects on the overall system performance, and variations in accuracy [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ].
      </p>
      <p>In our experimental study, we will examine these individual blocks as well as pairs of blocks
while employing genetic algorithms for HCNN synthesis. The designated CNN should be
extremely uncomplicated and easy to understand. Simplicity reduces outside influences and
randomness, making it easier to focus on each block's internal impact. The testing and training will
be conducted using a sample known as "CIFAR-100".</p>
      <p>Following the implementation of genetic algorithms to produce a fundamental CNN lacking
distinct blocks, the system achieves an initial accuracy of 86.3% with a learning duration of 5.3
hours. These values will act as the baseline for future performance test comparisons. It's important
to recognize that the overall training time depends on the hardware utilized, so only the differences
in time should be taken into account.</p>
      <p>
        The fundamental test-driven CNN structure is created using a series of predefined blocks, as
per the algorithm. Next, the generated result model undergoes performance analysis [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ]. By
adding structural blocks to generation process the result values are listed in Table 1.
      </p>
      <p>
        When facing issues with low learning performance, several solutions can be implemented. One
effective approach is to enhance the current system's architecture by incorporating supportive
blocks. The main ones are: batch normalization layer [15]; 1x1 convolution layer; dropout layer
[
        <xref ref-type="bibr" rid="ref11">12</xref>
        ]; residual block.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Proposed structural synthesis of HCNN using evolutionary algorithm with functional modules and optimization-based construction blocks</title>
      <sec id="sec-5-1">
        <title>5.1. Optimal combination and placement of structural blocks</title>
        <p>
          The structural synthesis of Hybrid Convolutional Neural Networks involves the integration of
various advanced building blocks to enhance network performance, efficiency, and robustness.
This process requires a careful selection of building blocks such as Channel Boosting-Based CNNs
(CB-CNN), Squeeze-and-Excitation (SE) blocks [
          <xref ref-type="bibr" rid="ref12">13</xref>
          ], Split-Combine-Convolutions (SCConv), and
attention-based blocks. Each of these components brings unique strengths to the architecture, and
their optimal combination can significantly improve the overall capabilities of HCNNs.
        </p>
        <p>Comprehensive Integration: The optimal synthesis of HCNNs involves a strategic combination
and placement of the aforementioned building blocks to leverage their strengths synergistically.
For instance, an HCNN might start with CB-CNN layers to enhance initial feature extraction,
followed by SCConv layers to capture diverse patterns. SE blocks can be interspersed throughout
the network to recalibrate channel-wise features, while attention-based blocks can be integrated
into deeper layers to focus on important contextual information.</p>
        <p>
          Task-Specific Configuration: The choice and arrangement of these building blocks should be
tailored to the specific requirements of the task at hand. For instance, in image classification, a
configuration emphasizing CB-CNN and SE blocks might be optimal, whereas in object detection, a
combination of SCConv and attention-based blocks could provide the best performance [
          <xref ref-type="bibr" rid="ref1">1, 16</xref>
          ].
        </p>
        <p>Computational resources and efficiency should also guide the integration strategy. SCConv and
attention-based blocks can be computationally intensive, so their use should be balanced with the
overall resource budget.</p>
        <p>This involves:
•
•
•
features.</p>
        <p>Scaling Width: Increasing the number of channels in each layer to improve feature
representation.</p>
        <p>Scaling Resolution: Using higher resolution input images to capture more detail.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Evolutionary algorithm for structural synthesis of Hybrid Convolutional</title>
      </sec>
      <sec id="sec-5-3">
        <title>Neural Networks</title>
        <p>Genetic algorithms are part of evolutionary computing, a field of artificial intelligence. They are
inspired by evolution and natural selection, where the strongest traits are passed down from
generation to generation. The multicriteria genetic algorithm (MCGA) is an
extension of this process. It focuses on optimizing multiple objectives simultaneously. Each
solution provided by the algorithm is associated with a set of objective function values. The BCGA
optimizes these values and provides a set of Pareto-optimal solutions.</p>
        <p>
          In multi-objective optimization problems, there are several conflicting objectives that need to be
optimized. This results in a set of possible solutions, known as Pareto solutions, where no other
solution can improve all objectives simultaneously. Therefore, the goal is not to find a single
optimal solution, but to generate a set of Pareto-optimal solutions that provide a trade-off between
the conflicting objectives [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>Multi-criteria genetic algorithms, such as NSGA-II1, SPEA3, have demonstrated strong
performance in various engineering optimization problems. By utilizing selection, mutation, and
crossover operators in iteration, competitive individuals can be generated, drawing inspiration
from the evolutionary theory of "survival of the fittest." These individuals, who cannot surpass
each other in every aspect, form a group known as the non-dominance front. From the point of
view of physical optimization problems, in which evaluations are always computationally complex,
the population size in MOGA is usually small due to limited computing resources. On the Figure 1
presented the logical flow for application of evolutionary algorithm to structurally synthesize
optimal CNN structure. This is a high-level block-diagram and it does not reflect low level logic.</p>
        <p>For structural synthesis of target CNNs we consider to user modified SPEA-3[8] algorithm to
overcome the aforementioned problems.</p>
      </sec>
      <sec id="sec-5-4">
        <title>5.2.1. Defining an Individual</title>
        <p>In context of this paper, an individual is representing a specific CNN architecture. Based on the
criteria defined in problem statement section, the target individual considered to be encoded
directly into the string that explicitly describes the architecture. Genome structure is considered to
encapsulate the following:
•</p>
        <p>number of layers;
block, with specific parameters encoded within it.</p>
        <p>The qualitive criteria on which the final individual is selected are defined and described in
details in the next section.
format:</p>
        <p>Simplified example of the genome could be the following and represented in the unified JSON
[{"type": "Conv", "filters": 32, "kernel_size": 3, "stride": 1, "padding": "same", "activation": "relu"},
{"type": "DenseBlock", "num_layers": 4, "growth_rate": 12, "bottleneck_size": 4},
{"type": "SEBlock", "reduction_ratio": 16},
{"type": "SCConv", "filters": 64, "kernel_size": 3, "stride": 1, "padding": "same"},
{"type": "Pooling", "pool_size": 2, "stride": 2, "pool_type": "MaxPooling"},
{"type": "FC", "units": 10, "activation": "softmax"}].</p>
      </sec>
      <sec id="sec-5-5">
        <title>5.2.2. Formulating a Multi-Objective Fitness function and evaluation criteria</title>
        <p>To quantify the performance of Hybrid Convolutional Neural Networks based on multiple
objectives, we need to construct a</p>
        <p>multi-objective fitness function that integrates various
performance
metrics.</p>
        <sec id="sec-5-5-1">
          <title>These</title>
          <p>metrics typically
include
accuracy, computational efficiency,
robustness, and scalability. The proposed fitness function balances these aspects to provide a
comprehensive evaluation of HCNN architectures.</p>
          <p>Criteria that is considered for formulating fitness functions are following:</p>
          <p>Accuracy (A): The primary measure of how well the network performs on the task, such as
classification accuracy on a validation dataset.</p>
          <p>Computational Efficiency (E): Metrics like inference time and memory usage indicate the
efficiency of the network.</p>
          <p>Robustness (R): The network's resilience to adversarial attacks or noisy data, often
measured by accuracy under adversarial conditions or performance degradation.
Scalability (S): The ability to maintain performance when scaled to larger datasets or more
complex tasks, often evaluated by the change in accuracy and efficiency when scaling the
network.</p>
          <p>Let   ,   ,</p>
          <p>be the weights assigned to each performance metric to reflect their
relative importance. The fitness function FFF is formulated as a weighted sum of the normalized
performance metrics:
+  ⋅
 =   ⋅  
−  ,
−  ,
 −</p>
          <p>−  
+   ⋅ ( ⋅
+   ⋅ ( ⋅
Δ
Δ
−Δ
− 
−  
,
Δ
Δ
 
 
−Δ
−Δ
Accuracy Normalization:  −</p>
          <p>− 
maximum  
possible values.</p>
          <p>normalizes the accuracy  between its minimum  
and
Inference Time Normalization:
normalizes the inference time 
, where  
are the minimum and maximum inference times.</p>
        </sec>
        <sec id="sec-5-5-2">
          <title>Memory Usage Normalization:</title>
          <p>normalizes memory usage 
, where  
are the minimum and maximum memory usages.</p>
          <p>and are weights to balance the contributions of inference time and memory usage.
Robustness Normalization:
normalizes the robustness metric  
, which
 
 
 
 
measures accuracy under adversarial conditions.</p>
          <p>Accuracy Scaling Normalization:</p>
          <p>) when scaling the network.</p>
          <p>Efficiency Scaling Normalization: Δ
) when scaling the network.</p>
          <p>,
,
,
−Δ</p>
          <p>−Δ
−Δ
−Δ
and
,
,
,
normalizes the change in accuracy
normalizes the change in efficiency
are weights to balance the contributions of
weights could vary from 0 to 1.
accuracy and efficiency scaling. The weights   ,   ,  
  , as well as the sub-weights , , ,
and</p>
          <p>, are chosen based on the specific requirements and priorities of the task. For example, if
accuracy is paramount,   should be set higher relative to the other weights. The values of these</p>
        </sec>
      </sec>
      <sec id="sec-5-6">
        <title>5.2.3. Selection of Individual</title>
        <p>Selection is the process of choosing individuals from the current population to serve as parents for
the next generation. The goal is to favor individuals with higher fitness scores, ensuring that
better-performing CNN architectures are more likely to propagate their genetic material. Fitness
Evaluation: Evaluate each individual's fitness using a predefined fitness function described in
previous section. Regarding the selection method, the are several ones, such as:
straight forward evaluation-based approach.</p>
      </sec>
      <sec id="sec-5-7">
        <title>5.2.4. Defining crossover approach</title>
        <p>Roulette Wheel Selection: Assign a selection probability to each individual proportional to
its fitness. Randomly select individuals based on these probabilities.</p>
        <p>Tournament Selection: Randomly select a subset of individuals (tournament) and choose
the best-performing one as a parent. Repeat to select the required number of parents.
Rank-Based Selection: Rank individuals by fitness and assign selection probabilities based
on rank, giving higher-ranked individuals a better chance of being selected.</p>
        <p>In our approach we choose the rank-based selection due to its implementation simplicity and

Crossover combines the genetic material of two parent solutions to produce one or more offspring.
This
process
creates
new
architectures by
mixing
and
matching
structural blocks
and
hyperparameters from the parents. For our implementation we decided to utilize modified
multipoint crossover. This
method enhances genetic diversity and allows for a
more extensive
exploration of the solution space compared to single-point crossover. In the context of CNN
structural synthesis, multi-point crossover involves combining different structural blocks and
hyperparameters from two parent CNN architectures to create new offspring architectures.</p>
        <sec id="sec-5-7-1">
          <title>Proposed crossover process is the following [1, 17, 18]: 1. Selection of Crossover Points: Choose multiple crossover points randomly within the genomes of the parent CNN architectures.</title>
          <p>complexity of the problem.</p>
          <p>The number of crossover points is typically predefined and can vary depending on the
2. Splitting and Combining:</p>
        </sec>
        <sec id="sec-5-7-2">
          <title>3. Reconstruction of Offspring:</title>
          <p>Split the parent genomes at the chosen crossover points.</p>
          <p>Alternate the segments from each parent to form the offspring genomes.
•
•
•
•
•
•
•
•</p>
          <p>Reconstruct the offspring genomes by merging the alternating segments.</p>
          <p>Ensure that the resulting offspring are valid CNN architectures.</p>
          <p>Example:
Parent 1: [Conv, 32, 3x3, relu]-[DenseBlock, 4, 12, 4]-[SEBlock, 16]-[SCConv, 64, 3x3]
Parent 2: [Conv, 64, 5x5, relu]-[SCConv, 32, 3x3]-[SEBlock, 8]-[DenseBlock, 6, 16, 4]
Offspring 1: [Conv, 32, 3x3, relu]-[SCConv, 32, 3x3]-[SEBlock, 16]-[DenseBlock, 6, 16, 4]
Offspring 2: [Conv, 64, 5x5, relu]-[DenseBlock, 4, 12, 4]-[SEBlock, 8]-[SCConv, 64, 3x3]</p>
        </sec>
      </sec>
      <sec id="sec-5-8">
        <title>5.2.5. Mutation of Individuals</title>
        <p>Mutation introduces random changes to the genetic material of individuals to maintain diversity in
the population and explore new regions of the search space. For proposed implementations we
propose the following transformations of individuals:</p>
        <p>Parameter Mutation: Randomly alter the parameters of a structural block.</p>
        <p>Example: Change the number of filters in a Conv layer from 32 to 64.</p>
        <p>Block Mutation: Replace one structural block with another of the same type but with
different parameters.</p>
        <p>Example: Replace [DenseBlock, 32, 3x3, relu] with [SCConv, 64, 5x5].</p>
        <p>Layer Addition/Removal: Randomly add or remove a structural block in the genome.</p>
        <p>Example: Add [SEBlock, 16] after a Conv layer or remove an existing DenseBlock.
Swap Mutation: Swap the positions of two structural blocks.</p>
        <p>Example: Swap [DenseBlock, 4, 12, 4] with [SEBlock, 16].</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Experiment</title>
      <p>In the same way as other implementations of evolutionary algorithms, our implementation
requires a large computational capability to perform the task that makes it hard to evaluate it on
large and complex - -100 to synthesize
set of potentially good architectures and then migrate it to larger-scale environment.</p>
      <sec id="sec-6-1">
        <title>6.1. Settings and results</title>
        <p>- of 32x32 color images. The
task is to receive most suitable CNN model based on criteria described in fitness function definition
section. Population size is set to 10, generations 50, mutation rate 5.5%, initial population
randomly generated CNN architecture with a selective mix of convolutional layers, Dense blocks,
SE-blocks, SCConv blocks, attention-based modules, SRU, CRU, Batch Normalization layers, 1x1
Conv layers, Dropout layers, and Fully Connected layers. The results based on the described
evolutionary algorithm is shown in Table 2. and accuracy change dynamics is shown on Figure 2.</p>
        <p>The maximum error percentage decreased significantly from 37% to 16% over 50 generations.
This indicates that even the worst-performing models in the population improved significantly.
The midpoint error rate saw a substantial improvement, reflecting overall population
improvement.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>The implementation of a multi-objective fitness function enabled a comprehensive evaluation of
CNN architectures. This function incorporated normalized metrics for accuracy, inference time,
memory usage, robustness under adversarial conditions, and scalability. By weighting these
metrics appropriately, we could balance their contributions and achieve a holistic assessment of
network performance</p>
      <p>Our experimental results shown in Table 2 and Figure 2, obtained through training on the
CIFAR-100 dataset, illustrated the significant impact of different structural blocks on HCNN
performance. By testing blocks both individually and in combination, we identified optimal
configurations that enhanced feature representation and reduced computational load.</p>
      <p>Based on the experimental results, we received optimal CNN architecture for recognition based
on CIFAR-100 dataset. Due to computational and time restrictions, the experiment was executed in
low-
continue research within more complex conditions using more advanced hardware support in
future.</p>
      <p>In conclusion, this research advances the state-of-the-art in hybrid neural network design and
optimization. The proposed hybrid learning algorithm and multi-objective optimization framework
offer powerful tools for developing more efficient, robust, and scalable deep learning models.</p>
      <sec id="sec-7-1">
        <title>International Conference</title>
        <p>[16] S. Bell, C. Zitnick, K. Bala, R. Girshick, "Inside-Outside Net: Detecting Objects in Context with
Skip Pooling and Recurrent Neural Networks" in: 2016 IEEE Conference on Computer Vision
and Pattern Recognition, Las Vegas, USA, 2016 pp. 2874-2883, doi: 10.1109/CVPR.2016.314.
[17] A. Newell, K. Yang, an</p>
        <p>ECCV (2016). doi: 10.1007/978-3-319-46484-8_29.
[18] V. Sineglazov, A. Kot, "Design of hybrid neural networks of the ensemble structure"
EasternEuropean Journal of Enterprise Technologies 77 (2022) 31-45, doi:
10.15587/17294061.2021.225301.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wen</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <article-title>"SCConv: Spatial and Channel Reconstruction Convolution for Feature Redundancy,"</article-title>
          <source>in: 2023 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          , Vancouver, BC, Canada,
          <year>2023</year>
          , pp.
          <fpage>6153</fpage>
          -
          <lpage>6162</lpage>
          , doi: 10.1109/CVPR52729.
          <year>2023</year>
          .
          <volume>00596</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jiang</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>"A Strength Pareto Evolutionary Algorithm Based on Reference Direction for Multiobjective and Many-Objective Optimization,"</article-title>
          <source>in: IEEE Transactions on Evolutionary Computation</source>
          , vol.
          <volume>21</volume>
          , no.
          <issue>3</issue>
          ,
          <issue>2017</issue>
          , pp.
          <fpage>329</fpage>
          -
          <lpage>346</lpage>
          . doi:
          <volume>10</volume>
          .1109/TEVC.
          <year>2016</year>
          .
          <volume>2592479</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zgurovsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sineglazov</surname>
          </string-name>
          , E. Chumachenko,
          <source>Classification and Analysis of Multicriteria Optimization Methods n: Artificial Intelligence Systems Based on Hybrid Neural Networks</source>
          , vol
          <volume>904</volume>
          ,
          <year>2021</year>
          , pp.
          <fpage>59</fpage>
          -
          <lpage>174</lpage>
          , doi: 10.1007/978-3-
          <fpage>030</fpage>
          -48453-
          <issue>8</issue>
          _
          <fpage>2</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Meza-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Clemente</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.C.</given-names>
            <surname>Rodríguez-Liñán</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Olague</surname>
          </string-name>
          ,
          <article-title>"Synthetic-analytic behavior-based control framework: Constraining velocity in tracking for nonholonomic wheeled mobile robots"</article-title>
          <source>Information Sciences</source>
          <volume>501</volume>
          (
          <year>2019</year>
          )
          <fpage>436</fpage>
          -
          <lpage>459</lpage>
          , doi: 10.1016/j.ins.
          <year>2019</year>
          .
          <volume>06</volume>
          .025.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5] ,
          <source>in: Parallel Problem Solving from Nature</source>
          , Springer, Berlin, Heidelberg, vol.
          <volume>4193</volume>
          ,
          <year>2006</year>
          . doi:
          <volume>10</volume>
          .1007/11844297_
          <fpage>68</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N.</given-names>
            <surname>Hansen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. V.</given-names>
            <surname>Arnold</surname>
          </string-name>
          and A.
          <volume>8</volume>
          , doi: 10.1007/978-3-
          <fpage>319</fpage>
          - 07124-4_
          <fpage>13</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>V.</given-names>
            <surname>Sineglazov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Riazanovskiy</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          .
          <source>Klanove [8] Biol. Med</source>
          .
          <volume>147</volume>
          (
          <year>2022</year>
          ) 105800, doi: 10.1016/j.compbiomed.
          <year>2022</year>
          .
          <volume>105800</volume>
          .
          <string-name>
            <surname>Optimization</surname>
          </string-name>
          on Learning Representations,
          <year>Jan 2017</year>
          , pp.
          <fpage>1</fpage>
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Nagpal</surname>
          </string-name>
          and
          <string-name>
            <given-names>S. R.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <article-title>"A Performance Evaluation of Convolutional Neural Networks for Face Anti Spoofing,"</article-title>
          <source>in: International Joint Conference on Neural Networks</source>
          , Budapest, Hungary,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . doi:
          <volume>10</volume>
          .1109/IJCNN.
          <year>2019</year>
          .
          <volume>8852422</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>N.</given-names>
            <surname>Shone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Ngoc</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. D.</given-names>
            <surname>Phai</surname>
          </string-name>
          and
          <string-name>
            <given-names>Q.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <article-title>"A Deep Learning Approach to Network Intrusion Detection,"</article-title>
          <source>in: IEEE Transactions on Emerging Topics in Computational Intelligence</source>
          , vol.
          <volume>2</volume>
          , no.
          <issue>1</issue>
          ,
          <issue>2018</issue>
          , pp.
          <fpage>41</fpage>
          -
          <lpage>50</lpage>
          . doi:
          <volume>10</volume>
          .1109/TETCI.
          <year>2017</year>
          .
          <volume>2772792</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [11]
          <string-name>
            <surname>M. K. Yadav</surname>
            and
            <given-names>K. P.</given-names>
          </string-name>
          <string-name>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>"Intrusion Detection System using Machine Learning Algorithms: A Comparative Study,"</article-title>
          <source>in: 2nd International Conference on Secure Cyber Computing and Communications</source>
          , Jalandhar, India,
          <year>2021</year>
          , pp.
          <fpage>415</fpage>
          -
          <lpage>420</lpage>
          , doi: 10.1109/ICSCCC51823.
          <year>2021</year>
          .
          <volume>9478086</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>Tan</surname>
          </string-name>
          ,
          <article-title>"Feedback Convolutional Neural Network for Visual Localization and Segmentation"</article-title>
          ,
          <source>in: IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          , vol.
          <volume>41</volume>
          , no.
          <issue>7</issue>
          ,
          <issue>2019</issue>
          , pp.
          <fpage>1627</fpage>
          -
          <lpage>1640</lpage>
          . doi:
          <volume>10</volume>
          .1109/TPAMI.
          <year>2018</year>
          .
          <volume>2843329</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Albanie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sun</surname>
          </string-name>
          and
          <string-name>
            <given-names>E.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>"Squeeze-and-Excitation Networks"</article-title>
          <source>in: IEEE Transactions on Pattern Analysis and Machine Intelligence</source>
          , vol.
          <volume>42</volume>
          , no.
          <issue>8</issue>
          ,
          <issue>2020</issue>
          , pp.
          <fpage>2011</fpage>
          -
          <lpage>2023</lpage>
          . doi:
          <volume>10</volume>
          .1109/TPAMI.
          <year>2019</year>
          .
          <volume>2913372</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>F.</given-names>
            <surname>Chollet</surname>
          </string-name>
          ,
          <article-title>"Xception: Deep Learning with Depthwise Separable Convolutions," in: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu</article-title>
          ,
          <string-name>
            <surname>HI</surname>
          </string-name>
          , USA,
          <year>2017</year>
          , pp.
          <fpage>1800</fpage>
          -
          <lpage>1807</lpage>
          . doi:
          <volume>10</volume>
          .1109/CVPR.
          <year>2017</year>
          .
          <volume>195</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>