<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>December</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Use of Convolutional Neural Networks for Identifying Additional Features on a Digital Image of Human Face</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kateryna Merkulova</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bohdan Pavliukh</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>Volodymyrs'ka str. 64/13, Kyiv, 01601</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>0</volume>
      <fpage>5</fpage>
      <lpage>07</lpage>
      <abstract>
        <p>This article is devoted to the study of the use of CNNs (convolutional neural networks) for the tasks of recognizing certain additional features on digital images of human faces. Recognition of additional features in human face images has a wide range of applications, including photo and video analysis, security systems, collection of useful statistical data, and user comfort in various areas of life. This technology contributes to the improvement of safety and convenience in various life situations. Thanks to numerous studies conducted by scientists on various image analysis methods that can be used to recognize features in an image (for example, histogram-based methods, feature allocation, contour analysis, color analysis), it was found that although the effectiveness of the method is also may depend on specific features of the task, such as lighting conditions, distance to objects and the presence of curtains or other obstacles, one of the most effective solutions is the use of machine learning methods, in particular the use of convolutional neural networks. This paper describes the study of the effectiveness of using convolutional neural networks to recognize additional features on a digital image of a human face, namely the presence of a headdress, glasses, a medical mask, and a beard. Also, the gender of a person was determined as an additional auxiliary feature.</p>
      </abstract>
      <kwd-group>
        <kwd>Keywords1</kwd>
        <kwd>identification</kwd>
        <kwd>image classification</kwd>
        <kwd>convolutional neural network</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Nowadays, technologies for recognition and identification are very widely used in all areas. The
technology designed to identify additional features on a human face, such as the presence of glasses, a
headdress, a medical mask and a beard, can be used to solve a number of important problems. The most
obvious task that the technology under study can handle is the collection of statistical information. Such
technology can be used in access control systems to increase security, for example by requiring users
to remove a mask or glasses for identification, which can help avoid unauthorized access or
circumvention [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Also, the technology can help identify people even when their appearance changes
(that is, when using something that a system that uses recognition technology as an additional feature,
for example, when wearing a glasses) [2; 3].
      </p>
      <p>The tool for recognizing additional features on a digital image of a human face can also be used for
narrowly focused purposes, for example, to track the observance of mask mode by visitors to a
supermarket (or any other crowded place) [4; 5], and in this case it will be enough to determine only
one of the additional features that the system can recognize, i.e. recognition of a medical mask on a
human face (this example is especially relevant against the background of the recent pandemic). That
is, there are such possibilities of application of the researched technology, in which even a limited part
of the functionality of the technology can fully cope with the tasks.</p>
      <p>In addition to the above-described possibilities of using the technology to recognize additional
features in the image of a human face, it is worth paying attention to the non-obvious way of using the</p>
      <p>2023 Copyright for this paper by its authors.
CEUR
technology, which gives commercial value. For example, facial recognition software can be used to
gather statistical information about the number of bearded men passing through checkpoints at various
subway stations, which can then be used to make decisions about the best location for a barbershop
because the collected information will allow us to draw conclusions about the largest "places of
concentration" of potential customers.</p>
      <p>In the course of the research, a search and analysis of systems that determine additional features on
people's faces was carried out, but no software systems were found that comprehensively approach the
task of identifying additional features on a human face and recognize the presence or absence of all the
above-mentioned features: headdress, glasses, medical mask, beard. On the Internet, in public access,
you can find only information about software in which the developed functionality is partially
implemented (for example, software systems designed to recognize the presence of a medical mask) or
software that is able to recognize only some of the above features or others that are similar in essence
additional characteristics of a human face, for example, sex, race, age, emotions, direction of gaze.</p>
      <p>Therefore, the technology for recognizing additional features on images of a human face has a wide
range of applications, contributes to improving safety and convenience in various life scenarios.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Definition, Solution Methods and Technologies</title>
      <p>For the task of recognizing additional features in human face images, the first step is to choose the
image analysis method that will be used for recognition, of which there are many these days. The most
popular of them, which can be used to solve the given problem, are contour analysis, the use of image
segmentation, the analysis of brightness histograms, and the use of neural networks.</p>
      <p>Nowadays, the best "tool" for image analysis is still the visual cortex of the human brain. Computer
analysis tools are already used to analyze images in the fields of medicine, security, and remote sensing,
but they will not be able to replace people for a long time due to the capabilities of the human brain,
because the process of information processing by the human brain is non-linear and extremely complex.
That is why the most popular technologies for image analysis are those that developers try to get as
close as possible to models of human visual perception. Artificial neural networks are endowed with
such technologies. Convolutional neural networks are especially adapted for computer vision tasks.</p>
      <p>Convolutional neural networks usually contain convolutional layers, subsampling (aggregation)
layers, fully connected and normalizing layers in their structure [6]. Convolutional layers apply a set of
filters to the input image, applying element-by-element multiplication of the filter values to the original
pixel values of the image. After that, all the products are added, that is, the image is convoluted. As a
result of applying a convolutional layer, we get several feature maps, the number of which is usually
equal to the number of filters. Each filter determines a certain property on the image, by means of
initialization with random</p>
      <p>values of the filter kernel. The following formula demonstrates the
calculation of the output of a convolutional layer using filters and an activation function:</p>
      <p>=1  =1
 ( ,  ) =  (∑</p>
      <p>∑( ( +  ,  +  ) ∙  ( ,  ) +  )),
where O(x,y) - is the output pixel at the position (x,y), I(x+i,y+j) - is the pixel of the input image at the
where P(x,y) - output of the pooling layer in position (x,y).</p>
      <p>In a fully connected layer, the input neurons are connected to all neurons (activations) of the previous
layer. Usually these layers are used for the output layer. This layer calculates the class score, and outputs
a vector whose dimension is equal to the number of classes. The ordinal number of the element of the
vector with the largest value means the most probable belonging of the input image to the corresponding
class. Therefore, a convolutional neural network is a convenient tool for image classification [8]. The
identification of additional features on a human face belongs precisely to the tasks of image
classification, because we assign a person detected in the image to one of the classes for each feature:
for example, to identify a headdress on a human head, we have 2 classes - the presence of a headdress
and its absence. To train a neural network to recognize a certain number of states of the presence of an
additional feature, it is necessary to prepare an appropriate number of training data sets. The developed
CNN will receive an image as an input, and the output layer will consist of the number of neurons,
which is equal to the number of possible classes (states of the presence of an additional feature).</p>
      <p>Transfer learning is an effective technology in the development of CNN. It is a task in machine
learning that focuses on storing the knowledge gained during solving one problem and applying it to
another, but related problem, i.e. transfer learning allows you to use the accumulated knowledge in
solving one problem experience to solve another, similar problem [9]. That is, a ready-made trained
neural network, which, for example, identifies the type of animal in the image, can be used to recognize
additional features on the human face [10]. To do this, you can remove the last layer of the ready-made
model (CNN) and add one or more own layers, the last of which, as already indicated above, must
contain the number of neurons that will correspond to the number of states of the presence of an
additional feature. The advantages of using transfer training are that there is no need to create a large
number of layers, and that the network will already be trained on a large amount of data, after which
only layers that were not part of the "original" CNN can be trained on the target set.</p>
      <p>For research and testing of convolutional neural networks, it was decided to develop software that
works with streaming video and has the following working principle: after the user launches the main
function of the application, the software in real time receives an image from the camera of the device
on which it is running, further on using the computer vision library OpenCV separates images of human
faces [11], processes them, identifies the presence or absence of additional features with the help of five
developed models and displays images with text labels of the result of their identification.</p>
      <p>To interact with neural networks, the developed software uses Keras, an open-source library written
in the Python programming language, designed to interact with neural networks, including
convolutional and recurrent NNs (neural networks). Keras is one of the most popular tools used when
deep learning is involved in a project. Keras offers consistent and simple APIs, minimizing the number
of user actions required for typical use cases [12].</p>
      <p>Keras includes a number of pre-trained networks that you can download and use right away. One of
the most famous such networks (models) is MobileNetV2, which was trained for image classification.</p>
      <p>The image processing after detecting human faces includes the following image processing
operations:</p>
      <p>1) resizing the image to 224×224 pixels. This is due to the fact that images of this size were used for
training the MobileNetV2 model;</p>
      <p>2) converting the image to array view. For each image, the size of the array will have a dimension
of 224×224×3, that is, for each pixel, 3 values corresponding to the 3 RGB color channels of the model
are stored;</p>
      <p>3) normalization of color channel values. The standard range of pixel values for color channels is
from 0 to 255. For the correct operation of the MobileNetV2 model, this range must be changed to the
range from -1 to 1. The same image processing operations are applied to the images from the training
set immediately before they are used to train the CNNs. The same method of processing training images
and images to which the model is applied ensures the achievement of the maximum efficiency of the
model. The block diagram shown in Figure 1, demonstrates the sequence of actions performed for, as
part of the training of CNNs to identify additional features on a human face.</p>
      <p>Other imported Python programming language libraries besides Keras also play an important role
in the software. The most significant for the developed software are the following used libraries:
matplotlib for plotting graphs of the accuracy of CNNs, numpy for storing multidimensional arrays [13]
containing image information in a convenient form, imutils for obtaining images from a camera,
OpenCV for computer vision [14], Tkinter for creating GUI.</p>
      <p>Therefore, using the methods and technologies described above, it was decided to conduct a study
to determine the effectiveness of the application of convolutional neural networks to recognize
additional features (the presence of a headdress, glasses, a medical mask, a beard and additionally
determine the gender) on images of people's faces obtained from streaming video.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Research</title>
      <p>The research covers the description of the process of collecting input data sets, the details of the
design of convolutional neural networks for the identification of additional features on a digital image
of a human face. As part of the research, software was also developed. This section also presents the
results of the developed software and the achieved accuracy rates for each of the trained convolutional
neural networks responsible for identifying a certain additional feature.
3.1.</p>
    </sec>
    <sec id="sec-4">
      <title>Data Collection</title>
      <p>The developed software must recognize 4 additional features on the face (presence of glasses,
medical mask, headdress and beard) and additionally the gender of the person, accordingly, 5
convolutional neural networks will be developed, for which 5 training data sets are required, and the
images in these sets must additionally be divided into a number of groups corresponding to the number
of possible resulting states recognized. Let's consider the list of signs and possible resulting states that
we will determine:
 Gender – male or female;
 Glasses – sunglasses, glasses for vision or no glasses;
 Medical mask – presence, absence or incorrect wearing (for example, if the mask does not cover
the person's nose);
 Headdress - presence or absence;
 Beard - presence or absence.</p>
      <p>So, we need to form 5 sets of training data, for which a total of 12 groups of training data (images)
need to be formed. To facilitate the process of forming training sets, before "presenting" the image to
the neural network, we will programmatically process the image in the form of reducing the image size
to the required values, this will allow the size of the found images to be neglected. It is worth paying
attention to the format in which the image is saved: it should be suitable for the software tools used.
The most convenient (and at the same time the most popular) formats for working with images are
JPEG and PNG.</p>
      <p>Since the identification of additional features is performed on the image of the human face itself, it
is necessary to choose images for the training sets that cover only a small area around the human face.</p>
      <p>Images of human faces can be used in different training datasets, but the same image, of course,
cannot be in different groups of the same training set. Let's consider several examples of suitable images
and determine to which groups of training data they can be assigned.</p>
      <p>The image of a person's face in Figure 2.a can be assigned to all training sets: to determine gender
to the group "male", to determine the presence of glasses - to the group "with glasses for vision", to
determine the presence of a medical mask - to the group "without medical masks", to determine the
presence of a headdress in the "without headdress" group, to determine the presence of a beard - in the
"with a beard" group.</p>
      <p>The image in Figure 2.b can be assigned to the following training sets: to determine gender - to the
group "female", to determine the presence of glasses - to the group "without glasses", to determine the
presence of a medical mask - to the group "with a medical mask", to determine the presence of the main
recruitment to the group "with headdress". The images cannot be assigned only to the training set for
determining the presence of a beard - there should be only images of men, because the developed
software, when determining the gender of a person as "female", will not use CNN to try to detect a
beard in the image. For training data sets, you can use images in which additional features are artificially
added to the image of a person's face using software such as Adobe Photoshop.</p>
      <p>To the image in Figure 3, a medical mask has been artificially added with the help of software, which
looks incorrectly fitted. Such an image can be used without problems for all training sets: to determine
the gender - to the group "male", to determine the presence of glasses - to the group " with sunglasses",
to determine the presence of a medical mask - to the group "with an incorrectly worn medical mask",
to determine the presence of a headdress - to the group "without a headdress", to determine the presence
of a beard - to the group "without a beard".</p>
      <p>In total, the prepared training data sets for all planned convolutional neural networks have about
10,750 images of people, whose detailed breakdown by set is shown in Table 1.</p>
      <p>It is noticeable that some of the used sets have significantly fewer images, this is due to the fact that
for the beard and headdress it was not possible to find ready-made datasets, and it was necessary to
create datasets for these features, for other features it was possible to find ready-made datasets in public
access.
3.2.</p>
    </sec>
    <sec id="sec-5">
      <title>CNN Designing</title>
      <p>The biggest influence on the design of convolutional neural networks was the decision to use the
technology of transfer (transfer) learning, which allows us to use a ready-made neural network (with
defined weights for neurons) that was created to perform a similar task as the basis for the creation of
convolutional neural networks according to the essence of the problem. This approach allows you to
avoid the independent creation of a large number of layers for CNN, and the neural network that will
be used will already be trained on a large amount of training data (images) [15].</p>
      <p>As already mentioned above, it was decided to use the MobileNetV2 model as such a basis for new
convolutional networks. The MobileNetV2 artificial intelligence system was created for image
recognition of various objects, mainly different types of animals (about 1000 classes of objects) [16].
All layers taken from the MobileNetV2 model will not be retrained. It was decided for simplification
to make the same structure for all CNNs, that is, to add the same layers, namely:</p>
      <p>1) A fully connected layer with 128 neurons and a ReLU (rectified linear node) activation function.
Its goal is to find an additional characteristic to increase the accuracy of the model;
2) An exclusion layer used for regularization to reduce overtraining of the neural network by
preventing complex co-adaptations of the training data (some neurons will always return 0 as a result);
3) A fully connected layer with a softmax activation function (returns a probability distribution).
The neuron with the highest activity (highest value) indicates that the most likely outcome corresponds
to the class represented by that neuron.</p>
      <p>It was decided to train the models in 20 epochs (iterations). During training, as already mentioned,
the input images are pre-processed. As an optimization algorithm that adjusts the weights of the neural
network, we choose one of the most common for this type of problem, the Adam algorithm [17; 18]. A
decision was made to use the sparse categorical crossentropy algorithm as a loss function.</p>
      <p>The input data sets are divided into a training set and a validation set to test how well CNNs performs
on images that were not used for training, this allows for a more adequate accuracy rate and prevents
the model from being overtrained. For validation, we select 15% of the input data set.</p>
      <p>After completing the training of the CNN, the software, using the capabilities of the
matplotlib.pyplot library, plots the dependence of the accuracy and loss of the CNN on the iteration
number (for both training and validation input data sets). The loss value is calculated as follows:

 ( ,  ̂) = − ∑(  ∙ log(̂ )),
(4)
prediction.
where L( ,  ̂) - is the value of the loss function,   is the true answer,  ̂ is the answer of the model</p>
      <p>The developed CNNs consists of 158 layers and 2,422,210 parameters for CNNs with two resulting
classes and 2,422,339 parameters for CNNs with three resulting classes, of which 162,266 parameters
for CNNs with two resulting classes and 162,355 parameters for CNNs with three are subject to training.
resulting classes.
3.3.</p>
    </sec>
    <sec id="sec-6">
      <title>Research Results</title>
      <p>software is shown in Figure 4.</p>
      <p>A demonstration of the identification of additional features on people's faces by the developed
a man with a beard in a headdress and a woman in glasses; b) a man with an improperly worn medical
mask and sunglasses and a man without additional features on his face; c) a man with a beard and a
headdress and a woman with a headdress; d) a man in a medical mask and a man with a beard and
glasses.</p>
      <p>To confirm the correct operation of the program, the operation of the program is demonstrated with
two people in the image (a photo of a person is also suitable for demonstrating the operation of the
program), and in different images, there are different states of the presence of glasses, a medical mask,
a headdress, a beard (and different glasses and masks of different colors are used) , as well as different
sex
loss
genders of people, which allows you to check the correctness of the work of all trained convolutional
neural networks.</p>
      <p>Tables 2-4 contain the obtained results of accuracy and loss (for training and validation data sets)
for all 5 created CNNs for identifying additional features.</p>
      <p>The accuracy graph, which visualizes the achieved accuracy and loss rates for the training set and
the validation set depending on the training iteration, is shown in Figure 5.</p>
      <p>Table 5 shows the values of accuracy and loss for the validation set obtained by convolutional neural
networks after 20 iterations (epochs) of training.</p>
      <p>It is worth noting that the training process was started several times, the difference between the
results was very insignificant.</p>
      <p>loss</p>
    </sec>
    <sec id="sec-7">
      <title>4. Conclusion</title>
      <p>This article is devoted to the study of the effectiveness of using convolutional neural networks for
the task of recognizing additional features on a digital image of a human face. For this purpose, CNNs
was developed to recognize the following additional features: the presence of glasses, a beard, a
headdress, a medical mask, and a person's gender.</p>
      <p>All the developed models showed a good result both on training and validation data sets (which is
indicated by the presented accuracy tables and graphs), and during their experimental application for
streaming video. The final values of accuracy and loss on the validation data sets for developed CNNs
are as follows: for the detection of glasses, the accuracy is 99.72%, the loss is 1.39%; for detecting a
medical mask, the accuracy is 96.14%, the loss is 15.37%; for headdress detection, the accuracy is
100%, the loss is 0.05% (such a high result is probably related to the not very good content of the input
data set); for detecting a beard, the accuracy is 85%, the loss is 44.28%; for gender determination, the
accuracy is 98.62%, the loss is 3.77%. Among the shortcomings of the developed models, we can
highlight a decrease in the accuracy of work when the image was taken in poor lighting conditions
(which was expected).</p>
      <p>One of the interesting results obtained is the fact that the fewest images in the training sets had a
beard and a headdress, and, at the same time, their models gave the opposite result - based on the
obtained accuracy results, the model for recognizing the beard has the largest loss, and the model for
recognizing headdress is the smallest. This result is most likely related to the small number of images
in the training sets.</p>
      <p>For further research on this topic, models can be developed to identify other additional features, for
example, a mustache, as well as such features of a person as age, race, emotions. It is also possible to
increase the number of resulting classes for CNNs, for example, to define different types of headdress
or to define more types of glasses.</p>
      <p>To improve the quality of the obtained results, it is possible to expand data sets, as well as apply
additional methods of image preprocessing, which would help solve some difficulties, for example,
reduce the influence of the level of illumination on the result of the work of CNNs.</p>
    </sec>
    <sec id="sec-8">
      <title>5. References</title>
      <p>[5] V. Petrivskyi, V. Shevchenko, S. Yevseiev, O. Milov, O. Laptiev, O. Bychkov, V. Fedoriienko,
M. Tkachenko, O. Kurchenko and I. Opirskyy, "Development of a Modification of the Method for
Constructing Energy-Efficient Sensor Networks Using Static and Dynamic Sensors",
EasternEuropean Journal of Enterprise Technologies, vol. 1 (9 (115)), 2022, pp. 15–23, doi:
https://doi.org/10.15587/1729-4061.2022.252988.
[6] D. Bhatt, C. Patel, H. Talsania, J. Patel, R. Vaghela, S. Pandya, K. Modi, H. Ghayvat, “CNN
Variants for Computer Vision: History, Architecture, Application, Challenges and Future Scope”,
India, 2021. URL: https://www.mdpi.com/2079-9292/10/20/2470.
[7] N. Ketkar, J. Moolayil, “Convolutional neural network”, Berkeley, CA, USA, April 10, 2021, pp.</p>
      <p>197–242. URL: https://doi.org/10.1007/978-1-4842-5364-9_6.
[8] S. Mathot, “Introduction to deep learning”, 2021. URL:
https://pythontutorials.eu/deeplearning/introduction.
[9] “Face Mask Detection and Correct Mask Wearing Recognition Software. How to Save Your
Business from Quarantine and Closure”, SYTOSS, 2021. URL:
https://www.sytoss.com/blog/face-mask-detection-and-correct-mask-wearing-recognitionsoftware-how-to-save-your-business-from-quarantine-and-closure.
[10] A. Hossain, S. Sajib, “Classification of Image using Convolutional Neural Network (CNN)”,</p>
      <p>Pabna University of Science &amp; Technology, 2019.
[11] S. Aparna, “Face Recognition using OpenCV”, Dublin Business School, 2020, pp. 10-12.
[12] “Keras. Simple. Flexible. Powerful”, Keras. URL: https://keras.io.
[13] S. Shell, “Introduction to Numpy and Scipy”, San Francisco, 2019, pp. 7-11.
[14] M. Khan, S. Chakraborty, R. Astya, S. Khepra, “Face Detection and Recognition Using OpenCV”,
International Conference on Computing, Communication, and Intelligent Systems (ICCCIS),
India, October 18-19, 2019. URL: https://ieeexplore.ieee.org/abstract/document/8974493.
[15] Kutyrev A., Kiktev N., Kalivoshko O., Rakhmedov R. Recognition and Classification Apple Fruits
Based on a Convolutional Neural Network Model. (2022) CEUR Workshop Proceedings, 3347,
pp. 90 – 101. https://ceur-ws.org/Vol-3347/Paper_8.pdf
[16] S. Mathot, “Classifying images with MobileNetV2”, 2021. URL:
https://pythontutorials.eu/deeplearning/image-classification.
[17] Shatyrko A., Khusainov D. On the Interval Stability of Weak-Nonlinear Control Systems with
Aftereffect // Open Source Journal. The Scientific World Journal, vol. 2016, Article ID 6490826,
8 pages, 2016. doi:10.1155/2016/6490826 https://www.hindawi.com/journals/tswj/2016/6490826.
[18] S. Chaganti, I. Nanda, K. Pandi, T. Prudhvith, N. Kumar, “Image Classification using SVM and
CNN”, International Conference on Computer Science, Engineering and Applications (ICCSEA),
Gunupur, India, 2020. URL: https://ieeexplore.ieee.org/abstract/document/9132851.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          , “
          <article-title>Information Technology for Person Identification by Occluded Face Image</article-title>
          ,”
          <source>2022 IEEE 16th International Conference on Advanced Trends in Radioelectronics</source>
          , Telecommunications and Computer Engineering (TCSET),
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Shatyrko, “
          <article-title>Development of information technology for person identification in video stream,” Proceedings of the II International Scientific Symposium “Intelligent Solutions” (IntSol-</article-title>
          <year>2021</year>
          ),
          <source>CEUR Workshop Proceedings</source>
          ,
          <volume>3018</volume>
          , pp.
          <fpage>70</fpage>
          -
          <lpage>80</lpage>
          , Kyiv - Uzhhorod, Ukraine,
          <source>September 28-30</source>
          ,
          <year>2021</year>
          . URL: http://ceur-ws.org/Vol3018/Paper_7.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Martsenyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          ,
          <article-title>"Exploring Image Unified Space for Improving Information Technology for Person Identification,"</article-title>
          <source>in IEEE Access</source>
          , vol.
          <volume>11</volume>
          , pp.
          <fpage>76347</fpage>
          -
          <lpage>76358</lpage>
          ,
          <year>2023</year>
          , doi: 10.1109/ACCESS.
          <year>2023</year>
          .
          <volume>3297488</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Face</surname>
            <given-names>SDK</given-names>
          </string-name>
          , Regula. URL: https://regulaforensics.com/products/face
          <article-title>-recognition-sdk.</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>