<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Kirill Smelyakov 1, Yaroslav Honchar 1, Oleksandr Bohomolov 1 and Anastasiya Chupryna 1</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Kharkiv National University of Radio Electronics</institution>
          ,
          <addr-line>14 Nauky Ave., Kharkiv, 61166</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The article is devoted to the analysis of the effectiveness of the application of modern machine learning models of convolutional neural networks, which are used for image classification. To conduct such an analysis, an actual dataset is selected and divided into training, validation, and test subsets in a standard proportion. The dataset which is selected consists of images of birds. Classification efficiency indicators are determined. ResNet and EfficientNet V2 neural networks are trained using a full training cycle and Transfer Learning technology on frozen and free weights. Pytorch framework is used to train ResNet model and Tensorflow framework is used to train EfficientNet V2 model. The effectiveness of the use of neural networks is evaluated. The evaluation is done by analyzing popular classification metrics, such as precision, recall, and f1 score. The results of experiments are given, along with conclusions and practical recommendations on the use of machine learning models.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Convolutional Neural Network</kwd>
        <kwd>Image Classification</kwd>
        <kwd>Machine Learning Model</kwd>
        <kwd>Transfer Learning</kwd>
        <kwd>Metrics</kwd>
        <kwd>Efficiency Estimation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The successful application of neural networks to solve actual computer vision problems has led to the
emergence of many new models of convolutional neural networks (CNN) and their various modifications
used for image detection and classification in recent years. Many CNN models are pre-trained. This
means that they are focused on the detection or classification of images of a certain list of classes. This is
enough for some applications. However, most often the user needs to expand the list of classes with which
the CNN works. To do this, it is required to train the CNN to work with images of new classes. For such
training, there are relatively large number of alternative techniques and training methods. And only on the
basis of an experimental analysis of effectiveness, it is possible to answer the question of which training
method will be better and by what criterion.</p>
      <p>The aim of the work is to ensure the efficiency of machine learning of modern convolutional neural
networks.</p>
      <p>The goals of the work are to develop a plan and set up a series of experiments on the application of
widely used machine learning methods in relation to modern CNNs based on actual data, evaluate the
effectiveness of machine learning on free and frozen weights, formulate recommendations on the practical
application of machine learning techniques and methods.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        In famous reviews [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ] it is shown that in recent years, convolutional neural networks (CNN) have
become the standard model and technology for a wide range of computer vision tasks in both image
detection and classification. Mainly thanks to recent advances in deep learning [
        <xref ref-type="bibr" rid="ref1 ref3">1, 3</xref>
        ] and highly efficient
post-image processing. In this regard, a detailed analysis of the effectiveness of applying the most
common CNN models for image classification is provided in this work [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The relationship between the
components of the CNN architecture and the effectiveness of their application is shown. The paper [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
proposed a solution for the classification of moving vehicles based on the application of CNN. The paper
[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] describes the use of new FPGA technology to implement training and improve performance with
testing on VGG-16 and ResNet-50 networks. The paper [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] describes the most important models and
technologies of deep learning and the effective use of a large number of hidden layers of the CNN to
improve the efficiency of training modern neural networks. At the same time the paper [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] describes the
model and principles of operation of small CNN, which is important for the efficiency of relatively
lowpowered mobile devices, especially when processing a video stream.
      </p>
      <p>
        An analysis of the current state of the issue allows us to conclude that the use of CNN is relevant for
cars classification [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], road signs [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], lung diseases on x-rays [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], products in warehouses and in
electronic stores [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], gestures [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] and in many other applications [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14-16</xref>
        ].
      </p>
      <p>
        At the same time, the effectiveness of CNN applications is determined by the quality of network
training. In such a situation, a reasonable choice of a machine learning model comes to the foreground
[17, 18] for efficient tuning of neural network parameters. In this respect, a lot of recent research has been
devoted to the development of combined learning technologies that are associated with freezing, training
and retraining of unfrozen layers of a neural network [
        <xref ref-type="bibr" rid="ref7">7, 17, 19</xref>
        ].
      </p>
      <p>In certain applications, research is being carried out related to federated learning, as well as edge
computing, which is relevant for solving problems of centralized data processing. At the same time, great
attention is paid to research on the effectiveness of training deep neural networks, as the most common
architecture. Many aspects of the solution of these issues are provided in the work [20]; in particular,
promising developments in related areas of development of the architecture of deep neural networks and
deep learning methods for such neural networks are presented.</p>
      <p>Minimizing the computational complexity of deep learning algorithms is just as important as ensuring
high accuracy. The paper [21] describes the use of the Broad Learning System (BLS) as an alternative
method of machine learning, which leads to a significant reduction in the amount of calculations and the
duration of training.</p>
      <p>
        In recent years, more and more attention has been paid to the use of Transfer Learning, which allows
you to adapt pre-trained neural networks to new classes of objects by training only classification layers.
Such algorithms work an order of magnitude or more faster than algorithms with a full learning cycle [
        <xref ref-type="bibr" rid="ref7">7,
22</xref>
        ] and, most often, give higher classification accuracy [23, 24].
      </p>
      <p>Recently, a lot of different techniques and methods of machine learning have appeared. And
theoretically it is simply impossible to determine the best method. To solve this problem, it is planned to
develop a plan and conduct a series of experiments on training ResNet and EfficientNet V2 convolutional
neural networks, perform a comparative analysis of efficiency, and evaluate machine learning methods, all
other things being equal. Based on the results, it is planned to formulate recommendations on the practical
application of machine learning methods of modern convolutional neural networks.</p>
      <p>All these experiments are planned to be carried out on the basis of consideration “300 Bird Species”
dataset [25], because it has a lot of important features: a large number of classes, some of which are
similar to each other; a large number of diverse images within the same class; different shooting angles
and bird backgrounds.</p>
      <p>At the same time, for an objective assessment of the effectiveness of neural networks, metrics such as
Precision, Recall and F1 Score type [26] are evaluated, as well as training time.</p>
      <p>Successful solution of this problem through the transfer of machine learning technologies will improve
the efficiency of the use of neural networks in existing video data processing services, as well as in the
creation of promising multimedia traffic processing services in computer networks [27], for image
analysis by robots and drones [28], in many other relevant applications [29].</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods and Materials</title>
      <p>Consider the data (dataset) that will be used in further methods and experiments, data analysis methods
and materials, as well as metrics that will be used for effectiveness evaluation.
3.1.</p>
    </sec>
    <sec id="sec-4">
      <title>Dataset Description</title>
      <p>Methods and experiments are based on the use of Dataset “300 Bird Species” [25]. The dataset
contains 42622 training images (approximately 130-170 images in each class), 1500 test images (5 images
for each class) and 1500 test images (5 images for each class). A total of 300 bird classes were assigned
for training, testing and validation. The partitioning of the original array of images differs significantly
from the standard partitioning (train 70%, validation 10%, test 20%) and is inefficient. Therefore, the
images of each class were first combined (train + test). All images of the dataset are presented in jpg
format and are standardized - these are color images 224х224х3. Detailed information about the Dataset
“300 Bird Species” can be obtained at the resource [25]. Examples of images are shown in Figure 1.
32]. For the purity of experiments and the possibility of adequate comparative analysis with other sources,
minimal preprocessing was used in the work to bring the grayscale to the required format for a given
neural network.
3.2.</p>
    </sec>
    <sec id="sec-5">
      <title>Efficiency Indicators</title>
      <p>To evaluate the quality of classification for each class, we calculate the number   , , ( ,  = 1, … ,  ),
hits of objects of class  in class  . After that for each class  evaluate three main indicators of quality [26]


 1</p>
      <p>(=   ) = 
 (=   ) = 
 (=   ) =
∑ =1   ,</p>
      <p>,
∑ =1   ,
  ,</p>
      <p>,
2∙  ∙  ,
  +</p>
      <p>,</p>
      <sec id="sec-5-1">
        <title>Predict</title>
        <p>Class 1
 1 =  1,1
…
  ,1
…
…
…
…
…
 
 
 
= 
= 
= 
{  } ,  
{  } ,  
{  } ,  
=
=
=


1
1

1
∑

 =1   ;

∑</p>
        <p>∑
 =1   ;
 =1   .</p>
        <p>Class n
 1,
…
  =   ,</p>
        <p>1/ ∑ =1  1, .
…</p>
        <p>/ ∑ =1   , .</p>
        <p>(1)
(2)
(3)
(4)
(5)
(6)
where the coefficients are described in the following Figure 2.</p>
        <p>A
c
t
u
a
l</p>
      </sec>
      <sec id="sec-5-2">
        <title>Class 1</title>
        <p>Class n</p>
        <p>…


 1/ ∑ =1   ,1.</p>
        <p>/ ∑ =1   , .
effectiveness of the application of the machine learning model. Indicators (1) – (6) assess the quality of
the classification. For the purposes of practical application, it is also important to know the training time.
Therefore, to these estimates in experiments, we add the training time.
3.3.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Main Methods and Techniques</title>
      <p>To conduct research, we selected actual convolutional neural networks widely used for image
classification: ResNet and EfficientNet V2 [33, 34]. For these models, relatively small types of models
have been selected CNN ResNet 50 and CNN EfficientNet V2 type B0; these networks are approximately
equivalent in terms of the number of parameters and have the same input data format PT, which is a
tensor ready to be submitted for training with completed preprocessing. This is done in order to:
 create approximately equivalent conditions for use;
 maximize the learning rate in order to devote more time to the comparative analysis of
effectiveness, which is of the greatest interest for the purposes of the work.</p>
      <p>Each neural network is trained using a full training cycle from scratch, as well as using Transfer
Learning technology on frozen and free weights. Thus, for each network, 3 experiments are set to train it.</p>
      <p>After that, the trained networks are tested and performance indicators are evaluated (indicators (1) –
(3) and training time), summary tables of results for all classes are given, recommendations are given for
the practical application of machine learning models. Since there are quite a lot of classes, the results for a
certain number of classes are reflected directly in the work at the beginning and at the end of the list, and
integral estimates of efficiency are also given.. Full results tables are available at the link [35].</p>
    </sec>
    <sec id="sec-7">
      <title>4. Experiment</title>
      <p>Consider the experiments using convolutional neural networks mentioned above.
4.1.</p>
    </sec>
    <sec id="sec-8">
      <title>Experiment Using CNN ResNet</title>
      <p>The PyTorch Framework was used for the experiment. Therefore, the model for the experiment was
taken from torchvision, which is essentially its package. In terms of size, it was agreed to take ResNet50,
because it is the minimum model with available for use Transfer Learning, which best suits the smallest
EfficientNet V2 model of type B0 by the number of network parameters.</p>
      <p>Before training, standard neural network data preparation was performed: 3-channel integer array
(components of the color) in the range [0; 255] converted to the corresponding tensor of floating numbers
in the range [0.0; 1.0]. For better use of Transfer Learning, input tensors were also log-normalized,
because the previous training on ImageNet was also performed with such a transformation.</p>
      <p>Since the calculations on the video card are much faster than on the processor, the NVidia Tesla P100
video card on the Kaggle platform, which has 16 GB of internal memory, was used for training. It is the
main indicator that has affected the size of tensor batches that are simultaneously submitted for training –
128. Using a higher value creates an overflow of memory, and a lower one potentially reduces accuracy.</p>
      <p>Under these conditions, the neural network ResNet50 was used for 3 experiments:
 the first had a model with random weights and passed a maximum of 100 epochs, the coefficient
of learning speed was 0.0001;
 the second used the principle of Transfer Learning with a pre-trained model, but the weights were
frozen before (the calculation of gradients is disabled) in all layers, except the classification and
training of a maximum of 100 epochs with the coefficient of learning speed equals 0.0001, and then
the weights are unfrozen and a maximum of 100 epochs are performed, the coefficient of learning
speed is 0.00001 (10 times less);
 the third experiment also used the principle of Transfer Learning, but without manipulating the
weights of the model and also training during a maximum of 100 epochs with the coefficient of
learning speed of 0.0001.
4.2.</p>
    </sec>
    <sec id="sec-9">
      <title>Experiment Using CNN EfficientNet V2</title>
      <p>In the second part of the experiment, we used EfficientNet V2 B0 network to classify the images of
birds. Unfortunately, the weights of this network had not been ported yet by the time we were performing
the experiment, so we decided to use the TensorFlow implementation of this model. We also chose the B0
version of EfficientNet V2 because our input images have the 224 pixel width and height, which is the
recommended input size for the EfficientNet V2 B0 model. The TensorFlow implementation of the model
uses the Keras library, which already includes the training loop and this makes the experiment iteration
speed much faster.</p>
      <p>
        The preprocessing that we did for the images before the training is rather simple. We divided the pixel
values by 255 making them fit inside the [
        <xref ref-type="bibr" rid="ref1">0, 1</xref>
        ] interval, and after that we normalized them to have zero
mean and standard deviation of 1 using the ImageNet means and standard deviations.
      </p>
      <p>We performed the training using the computational resources provided by Kaggle. Kaggle provides
NVidia Tesla P100 GPU with 16 GB of VRAM. This was enough to use the batch size of 64 images per
batch. We performed three experiments. In the first experiment, we trained the model from scratch using
random weights initialization. The batch normalization layers were in training mode while training. We
trained the model using the early stopping callback with the patience of 3, and the training stopped after
10 epochs. The experiment took 27 minutes and 42 seconds to complete.</p>
      <p>In the second experiment, we firstly trained the model using the pre-trained weights with all of the
layers frozen except for the last fully connected layer. We also used the early stopping callback. After the
end of the first phase of the training, we unfroze all layers and performed the fine-tuning using the small
learning rate. As in the first phase of the experiment, we also used the early stopping callback. The
experiment took 1 hour and 35 minutes to complete.</p>
      <p>In the third experiment, we also used pre-trained weights, but this time we trained the entire model,
without freezing any layers. The experiment took 17 minutes and 38 seconds to complete.</p>
      <p>In all three experiments we used the Adam optimizer with the learning rate of 1e-3.</p>
      <p>To evaluate the performance of the models, we used the precision, recall and f1 score metrics. We used
the scikit-learn implementation of these metrics. All metrics were calculated on the test subset of the
dataset.</p>
    </sec>
    <sec id="sec-10">
      <title>5. Results</title>
      <p>Consider the experiment results using convolutional neural networks mentioned above.
5.1.</p>
    </sec>
    <sec id="sec-11">
      <title>Experiment Results Using CNN ResNet</title>
      <p>The results of the experiments are shown in Figure 3 – Figure 6, respectively. Each table occupies 7
pages, so in Figure 3 – Figure 6 only the initial and final parts of each table of results are shown. In order
not to overload the article with data, these three tables are completely given in electronic form [35].</p>
      <p>In three experiments, the classifier layer in models was previously replaced in order to give
probabilities of 300 classes instead of 1000 (2048), which is the default.</p>
      <p>The quality of the model was characterized by the function of cross-entropy losses, and the use of
gradients relied on the Adam optimizer with a learning speed coefficient of 0.0001. It was decided not to
use the scheduler due to the availability of appropriate internal mechanisms in Adam.</p>
      <p>The training procedure was performed on only 70% of the data set to obtain real indicators during
model validation and final testing. It was also planned to stop training prematurely based on changes in
the indicators of the loss function over the results of the validation part of the entire data set (10%): if a
lower value of loss function is not achieved within three epochs, the training ends with the restoration of
the model with the lowest value of the loss function on validation. During the final test, the quality
characteristics of the models were no longer determined by the loss function, but a classification report
containing metric values f1 (macro and weighted) and accuracy. Also for each class you can see precision,
recall and f1-score on the support pictures of the class in the test dataset. As a result of experiments, you
can see the effectiveness of the Transfer Learning approach. It is necessary to follow the procedure of
initial training of one classifier of the model, and then perform training after unfreezing all weights.</p>
      <p>Duration of training in the first experiment is 30 minutes, the second before unfreezing continued for
37 minutes and after thawing another 16 minutes used (53 minutes in total for the experiment), and the
third experiment lasted for 23 minutes. The model trained for 3 minutes per epoch, but when all layers
except the classifier were frozen, it was reduced to 70 seconds. In order to compare speed and efficiency
of different approaches we can see loss function value by epoch on Figure 7 – Figure 9.
Figure 4: The results of the experiment with frozen weights before thawing
Figure 5: The results of the experiment with frozen weights after training after thawing
Figure 6: The results of the experiment with free weights</p>
    </sec>
    <sec id="sec-12">
      <title>Experiment Results Using CNN EfficientNet V2</title>
      <p>Similar to the previous subsection, the results of the experiments are shown in next figures. Below we
present the ML curves and metrics of the first experiment based on model EfficientNet V2 (Figure 10).</p>
      <p>Below we present the learning curves and the metrics of the first stage of the second experiment based
on model EfficientNet V2 (Figure 11).</p>
      <p>a) b)
Figure 11: ML Result: The learning curves of the EfficientNet V2 B0 first stage model training with
transfer learning (a), metrics of the EfficientNet V2 B0 model with transfer learning after the first stage
of training (b)</p>
      <p>Below we present the learning curves and the metrics of the second stage of the second experiment
based on model EfficientNet V2 (Figure 12).</p>
      <p>Below we present the learning curves and the metrics of the third experiment based on model
EfficientNet V2 (Figure 13).</p>
    </sec>
    <sec id="sec-13">
      <title>6. Discussions</title>
      <p>After analyzing the results (metrics) listed above, we can make the conclusion that the second
experiment provides the best result. The average precision and recall metrics are considerably better than
the metrics of the first experiment’s model. The second experiment also demonstrates better results than
the third experiment, although the third experiment took much less time to complete (Figure 14). We can
make a conclusion that freezing all layers but the last one helps with preventing overfitting. However,
unfreezing all layers and performing fine-tuning produces little improvement. The overall metrics are
improved by 1%. This may help in competitions, but in real world applications this improvement could be
considered negligible and not worth the extra computing power spent on the model fine tuning. This effect
(conclusion) is valid for the convolutional neural networks ResNet and EfficientNet V2 considered in the
work. An analysis of the current state of the issue in relation to the results of machine learning of a
number of other modern convolutional neural networks allows us to generalize the above conclusion.
a)
b)
c)</p>
    </sec>
    <sec id="sec-14">
      <title>7. Conclusions</title>
      <p>The results of the experiments showed the high efficiency of learning convolutional neural networks
based on the Transfer Learning approach. At the same time, two approaches proved to be the best:
Approach 1: convolutional layers are frozen, only the classification layers of the neural network are
trained; Approach 2: after the implementation of the first approach, the unfrozen convolutional layers are
retrained. Comparing these two approaches, it can be noted that the first approach gives high accuracy and
requires approximately 3 times less time. The second approach gives a slight increase in accuracy (within
1%) and requires time up to 3 times more to train the model. In this regard, for application development
purposes, including training ensembles on conventional equipment, the first approach is sufficient.</p>
      <p>The conducted experiments and analysis of the available open experimental data allow us to conclude
that for the purposes of image classification, the efficiency of EffNet networks is significantly higher than
the most well-known analogues.</p>
      <p>Experiments with TensorFlow and PyTorch have shown that training the same models in different
frameworks can give significantly different results. This may be due to the implementation of different
versions of the network. In the experiments performed, this was largely observed for the ResNet 50
network. The accuracy differed by 0.1.</p>
      <p>In general, based on the results of the experiments, we can conclude that for the maximum efficiency
of training a given network, from a practical point of view, it is necessary to:
 split the dataset according to the 70-20-10 standards (because the splitting of many datasets does
not meet these standards, which leads to a decrease in learning efficiency);
 perform standard data preparation for the network in relation to debasing, normalization and
normalization of data;
 to train a neural network based on approach 1, or approach 2 (by default, it is enough to train the
classification layers by freezing the convolutional layers) on TensorFlow and PyTorch to understand
where the accuracy is higher; choose the best option.</p>
      <p>Experiments have shown that for maximum learning efficiency, it is necessary to supply images with
the recommended linear dimensions for a network of this type to the input of the neural network..</p>
    </sec>
    <sec id="sec-15">
      <title>8. References</title>
      <p>[17] J. B. Cloete, T. Stander and D. N. Wilke, "Parametric Circuit Fault Diagnosis Through
OscillationBased Testing in Analogue Circuits: Statistical and Deep Learning Approaches," in IEEE Access,
vol. 10, pp. 15671-15680, 2022, doi: 10.1109/ACCESS.2022.3149324.
[18] J. Sresakoolchai and S. Kaewunruen, "Integration of Building Information Modeling and Machine
Learning for Railway Defect Localization," in IEEE Access, vol. 9, pp. 166039-166047, 2021, doi:
10.1109/ACCESS.2021.3135451.
[19] A. L. C. Ottoni and M. S. Novo, "A Deep Learning Approach to Vegetation Images Recognition in
Buildings: a Hyperparameter Tuning Case Study," in IEEE Latin America Transactions, vol. 19, no.
12, pp. 2062-2070, Dec. 2021, doi: 10.1109/TLA.2021.9480148.
[20] R. Chellappa, S. Theodoridis and A. van Schaik, "Advances in Machine Learning and Deep Neural
Networks," in Proceedings of the IEEE, vol. 109, no. 5, pp. 607-611, May 2021, doi:
10.1109/JPROC.2021.3072172.
[21] F. Yang, "A CNN-Based Broad Learning System," 2018 IEEE 4th International Conference on
Computer and Communications (ICCC), 2018, pp. 2105-2109, doi:
10.1109/CompComm.2018.8780984.
[22] W. Wang et al., "Anomaly detection of industrial control systems based on transfer learning," in
Tsinghua Science and Technology, vol. 26, no. 6, pp. 821-832, Dec. 2021, doi:
10.26599/TST.2020.9010041.
[23] G. Schwartz and K. Nishino, "Recognizing Material Properties from Images," in IEEE Transactions
on Pattern Analysis and Machine Intelligence, vol. 42, no. 8, pp. 1981-1995, 1 Aug. 2020, doi:
10.1109/TPAMI.2019.2907850.
[24] Bodyanskiy Ye., Peleshko D., Setlak G., Mulesa P. Adaptive multivariate hybrid neuro-fuzzy system
and its on-board fast learning, in Neurocomputing, 2017, vol. 230, pp. 409-416
http://dx.doi.org/10.1016/j.neucom.2016.12.042.
[25] Dataset: 300 Bird Species. URL: https://www.kaggle.com/gpiosenka/100-bird-species.
[26] Comprehensive Guide to MCM. URL:
https://towardsdatascience.com/comprehensive-guide-onmulticlass-classification-metrics-af94cfb83fbd.
[27] O. Lemeshko, O. Yeremenko and A. M. Hailan, "Two-level method of fast ReRouting in
softwaredefined networks," 2017 4th International Scientific-Practical Conference Problems of
Infocommunications. Science and Technology (PIC S&amp;T), 2017, pp. 376-379, doi:
10.1109/INFOCOMMST.2017.8246420.
[28] G. Churyumov, V. Tokarev, V. Tkachov and S. Partyka, "Scenario of Interaction of the Mobile
Technical Objects in the Process of Transmission of Data Streams in Conditions of Impacting the
Powerful Electromagnetic Field," 2018 IEEE Second International Conference on Data Stream
Mining &amp; Processing (DSMP), 2018, pp. 183-186, doi: 10.1109/DSMP.2018.8478539.
[29] Shubin, I., Kyrychenko, I., Goncharov, P., Snisar, S., "Formal representation of knowledge for
infocommunication computerized training systems," 2017 IEEE 4th International Scientific-Practical
Conference Problems of Infocommunications, Science and Technology (PIC S&amp;T), 2017, pp. 287–
291, doi: 10.1109/INFOCOMMST.2017.8246399.
[30] K. Smelyakov, A. Chupryna, M. Hvozdiev and D. Sandrkin, "Gradational Correction Models
Efficiency Analysis of Low-Light Digital Image," 2019 Open Conference of Electrical, Electronic
and Information Sciences (eStream), 2019, pp. 1-6, doi: 10.1109/eStream.2019.8732174.
[31] K. Smelyakov, M. Shupyliuk, V. Martovytskyi, D. Tovchyrechko and O. Ponomarenko, "Efficiency
of image convolution," 2019 IEEE 8th International Conference on Advanced Optoelectronics and
Lasers (CAOL), 2019, pp. 578-583, doi: 10.1109/CAOL46282.2019.9019450.
[32] Rafael C. Gonzalez, Richard E. Woods Digital Image Processing, 4th. ed., Pearson/Prentice Hall,
2018, 1168p. DOI/ISBN: 9780133356724.
[33] ResNet and ResNetV2. URL: https://keras.io/api/applications/resnet.
[34] EffNet. URL: https://tfhub.dev/google/imagenet/efficientnet_v2_imagenet21k_b2/feature_vector/2.
[35] Result. URL: https://drive.google.com/drive/folders/1vW77McxfK56OmLeH4BRaT5t65CxJ8p_8.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G. R.</given-names>
            <surname>Djavanshir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>W.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>"A Review of Artificial Intelligence's Neural Networks (Deep Learning) Applications in Medical Diagnosis and Prediction,"</article-title>
          <source>in IT Professional</source>
          , vol.
          <volume>23</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>58</fpage>
          -
          <issue>62</issue>
          , 1 May-June 2021, doi: 10.1109/MITP.
          <year>2021</year>
          .
          <volume>3073665</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Arredondo-Velázquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Diaz-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Barranco-Gutiérrez</surname>
          </string-name>
          and
          <string-name>
            <given-names>C.</given-names>
            <surname>Torres-Huitzil</surname>
          </string-name>
          ,
          <article-title>"Review of prominent strategies for mapping CNNs onto embedded systems,"</article-title>
          <source>in IEEE Latin America Transactions</source>
          , vol.
          <volume>18</volume>
          , no.
          <issue>05</issue>
          , pp.
          <fpage>971</fpage>
          -
          <lpage>982</lpage>
          , May
          <year>2020</year>
          , doi: 10.1109/TLA.
          <year>2020</year>
          .
          <volume>9082927</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z. -Q.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zheng</surname>
          </string-name>
          , S. -T. Xu and
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>"Object Detection With Deep Learning: A Review,"</article-title>
          <source>in IEEE Transactions on Neural Networks and Learning Systems</source>
          , vol.
          <volume>30</volume>
          , no.
          <issue>11</issue>
          , pp.
          <fpage>3212</fpage>
          -
          <lpage>3232</lpage>
          , Nov.
          <year>2019</year>
          , doi: 10.1109/TNNLS.
          <year>2018</year>
          .
          <volume>2876865</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Lai</surname>
          </string-name>
          , T. Liu,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          and
          <string-name>
            <given-names>S.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <article-title>"</article-title>
          <source>The Cloud Images Classification Based on Convolutional Neural Network," 2019 International Conference on Meteorology Observations (ICMO)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          , doi: 10.1109/ICMO49322.
          <year>2019</year>
          .
          <volume>9026121</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Frniak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kamencay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Markovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dubovan</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Dado</surname>
          </string-name>
          ,
          <article-title>"Comparison of Vehicle Categorisation by Convolutional Neural Networks using MATLAB,"</article-title>
          <source>2020 ELEKTRO</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>4</lpage>
          , doi: 10.1109/ELEKTRO49696.
          <year>2020</year>
          .
          <volume>9130238</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Luk</surname>
          </string-name>
          and
          <string-name>
            <given-names>X.</given-names>
            <surname>Niu</surname>
          </string-name>
          ,
          <article-title>"Towards Efficient Convolutional Neural Network for DomainSpecific Applications on</article-title>
          FPGA,
          <article-title>"</article-title>
          <source>2018 28th International Conference on Field Programmable Logic and Applications (FPL)</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>147</fpage>
          -
          <lpage>1477</lpage>
          , doi: 10.1109/FPL.
          <year>2018</year>
          .
          <volume>00033</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Ian</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          , Yoshua Bengio,
          <source>Aaron Courville Deep Learning</source>
          , 3rd. ed., MIT Press,
          <year>2016</year>
          , 787p. DOI/ISBN: 0262035618.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Tripathi</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>"Image Classification using small Convolutional Neural Network,"</article-title>
          <source>2019 9th International Conference on Cloud Computing, Data Science &amp; Engineering (Confluence)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>483</fpage>
          -
          <lpage>487</lpage>
          , doi: 10.1109/CONFLUENCE.
          <year>2019</year>
          .
          <volume>8776982</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Stanford</given-names>
            <surname>Cars</surname>
          </string-name>
          <article-title>Dataset</article-title>
          . URL: https://www.kaggle.com/jessicali9530/stanford-cars-dataset.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Traffic</given-names>
            <surname>Signs</surname>
          </string-name>
          <article-title>Preproc</article-title>
          . URL: https://www.kaggle.com/valentynsichkar/traffic-signs-preprocessed.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[11] COVID-19 Detection and Classification</source>
          . URL: https://www.kaggle.com/c/siim-covid19-detection.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>Fashion</given-names>
            <surname>Product</surname>
          </string-name>
          . URL: https://www.kaggle.com/paramaggarwal/fashion-product
          <article-title>-images-small.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>ASL</given-names>
            <surname>Alphabet</surname>
          </string-name>
          . URL: https://www.kaggle.com/grassknoted/asl-alphabet.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>"A Lightweight Eagle-Eye-Based Vision System for Target Detection and Recognition,"</article-title>
          <source>in IEEE Sensors Journal</source>
          , vol.
          <volume>21</volume>
          , no.
          <issue>22</issue>
          , pp.
          <fpage>26140</fpage>
          -
          <lpage>26148</lpage>
          , 15 Nov.
          <volume>15</volume>
          ,
          <year>2021</year>
          , doi: 10.1109/JSEN.
          <year>2021</year>
          .
          <volume>3120922</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>C.</given-names>
            <surname>Ditzel</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>Dietmayer</surname>
          </string-name>
          ,
          <article-title>"GenRadar: Self-Supervised Probabilistic Camera Synthesis Based on Radar Frequencies,"</article-title>
          <source>in IEEE Access</source>
          , vol.
          <volume>9</volume>
          , pp.
          <fpage>148994</fpage>
          -
          <lpage>149042</lpage>
          ,
          <year>2021</year>
          , doi: 10.1109/ACCESS.
          <year>2021</year>
          .
          <volume>3120202</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>K.</given-names>
            <surname>Smelyakov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tovchyrechko</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Ruban</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chupryna</surname>
          </string-name>
          and
          <string-name>
            <given-names>O.</given-names>
            <surname>Ponomarenko</surname>
          </string-name>
          ,
          <article-title>"Local Feature Detectors Performance Analysis on Digital Image,"</article-title>
          2019 IEEE International Scientific-Practical Conference Problems of Infocommunications, Science and
          <string-name>
            <surname>Technology (PIC S&amp;T)</surname>
          </string-name>
          ,
          <year>2019</year>
          , pp.
          <fpage>644</fpage>
          -
          <lpage>648</lpage>
          , doi: 10.1109/PICST47496.
          <year>2019</year>
          .
          <volume>9061331</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>