<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>During the COVID-19 Pandemics</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Giorgio De Magistris</string-name>
          <email>demagistris@diag.uniroma1.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emanuele Iacobelli</string-name>
          <email>iacobelli@diag.uniroma1.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rafał Brociek</string-name>
          <email>Rafal.Brociek@polsl.pl</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christian Napoli</string-name>
          <email>cnapoli@diag.uniroma1.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>COVID-19</institution>
          ,
          <addr-line>Face Mask, CNN, ResNet50</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer, Control and Management Engineering, Sapienza University of Rome</institution>
          ,
          <addr-line>Via Ariosto 25, Roma, 00185</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Mathematics Applications and Methods for Artificial Intelligence, Faculty of Applied Mathematics, Silesian University of</institution>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Institute for Systems Analysis and Computer Science, Italian National Research Council</institution>
          ,
          <addr-line>Via dei Taurini 19, Roma, 00185</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Technology</institution>
          ,
          <addr-line>Gliwice, 44-100</addr-line>
          ,
          <country country="PL">Poland</country>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Workshop Proce dings</institution>
        </aff>
      </contrib-group>
      <fpage>36</fpage>
      <lpage>41</lpage>
      <abstract>
        <p>The ongoing COVID-19 pandemic has highlighted the importance of wearing face masks as a preventive measure to reduce the spread of the virus. In medical settings, such as hospitals and clinics, healthcare professionals and patients are required to wear surgical masks for infection control. However, the use of masks can hinder facial recognition technology, which is commonly used for identity verification and security purposes. In this paper, we propose a convolutional neural network (CNN) based approach to detect faces covered by surgical masks in medical settings. We evaluated the proposed CNN model on a test set comprising of masked and unmasked faces. The results showed that our model achieved an accuracy of over 96% in detecting masked faces. Furthermore, our model demonstrated robustness to diferent mask types and fit variations commonly encountered in medical settings. Our approaches reaches state of the art results in terms of accuracy and generalization.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>COVID-19 Pandemics</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <sec id="sec-2-1">
        <title>The use of face masks has become a critical preventive</title>
        <p>particularly in the context of the ongoing COVID-19
pandemic. In medical settings, such as hospitals and
clinics, healthcare professionals and patients are required to
wear surgical masks to minimize the risk of transmission.</p>
      </sec>
      <sec id="sec-2-2">
        <title>However, the use of masks can hinder facial recognition</title>
        <p>object detection, image classification, and image
segmentation. They have the potential to learn complex patterns
and representations from large datasets, which can aid</p>
        <p>In this paper, we propose a CNN-based approach for
detecting faces covered by surgical masks in medical
settings. We aim to develop a model that can accurately and
robustly identify masked faces, considering the unique
challenges posed by diferent mask types, colors, and
ifcation and security purposes. Accurate and eficient
technology, which is commonly used for identity veri- fit variations commonly encountered in medical
envidetection of faces covered by surgical masks is thus cru- individuals wearing surgical masks, and fine-tuned a
ronments. We collect a dataset of facial images with</p>
      </sec>
      <sec id="sec-2-3">
        <title>ResNet50 model on the specific task. The contributions</title>
        <p>of our work include the development of a CNN-based</p>
        <p>Traditional approaches for face detection and recog- approach tailored for face mask detection in medical
setcial for maintaining security measures while adhering to
infection control protocols.
nition may face challenges in the presence of masks, as
masks can alter facial features, obstructing key facial
landmarks and reducing facial visibility. To address this
challenge, convolutional neural networks (CNNs), a type
ically learn hierarchical features from images, have been
proposed as a promising solution. CNNs have shown
great success in various computer vision tasks, including
(C. Napoli)
CEUR
htp:/ceur-ws.org
ISN1613-073</p>
        <p>CEUR
generalization in the presence of diferent mask types
and fit variations. The proposed approach has the
potential to enhance security measures while maintaining
applications in various real-world scenarios. The rest of
the paper is organized as follows: in Section 2, we review
related works in the field; in Section</p>
      </sec>
      <sec id="sec-2-4">
        <title>3, we describe the</title>
        <p>dataset used in our study; in Section 4, we introduce the
proposed method; in Section 5 we present the
experimental results and discuss the findings; and finally, in Section</p>
      </sec>
      <sec id="sec-2-5">
        <title>6, we conclude the paper and outline future directions of research in this area.</title>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>2. Related Works</title>
      <p>
        learning algorithms. The latest algorithms for face
recognition are based on deep learning architectures, 3D face
Machine learning and convolutional neural networks recognition, and hybrid approaches. As face recognition
have been widely applied to the general field of face recog- technology becomes more prevalent, there is a growing
nition. Face recognition is a rapidly developing field that need to address privacy and security concerns. ArcFace
has seen significant advancements in recent years, driven [10] is a face recognition algorithm that uses a
marginby the increasing availability of large-scale datasets and based softmax loss function to optimize feature
reprepowerful machine learning algorithms. Deep learning sentations for face recognition, similarly CosFace [11],
has become the dominant approach for face recognition another deep learning-based face recognition algorithm,
due to its superior performance on large-scale datasets. uses a cosine-based loss function to improve the
discrimSome of the most widely used deep learning architec- inability of the learned features. Like ArcFace, it has
tures for face recognition include Convolutional Neural achieved state-of-the-art performance on several
benchNetworks (CNNs), Siamese Networks, Triplet Networks, mark datasets. Another algorithm, named SphereFace
and Deep Belief Networks (DBNs). These models are [12], uses a angular-based softmax loss function to
imtrained using large-scale datasets such as VGGFace [
        <xref ref-type="bibr" rid="ref1 ref4">1</xref>
        ], prove the discriminability of the learned features. While
FaceNet [2], and IE-CNN models [3], which contain mil- the said algorithm have many similarities, another
syslions of face images. Deep learning models have achieved tem named DeepID [
        <xref ref-type="bibr" rid="ref7">13, 14</xref>
        ], has beeen developed to ofer
state-of-the-art results on various benchmarks such as a multi-task learning approach to learn multiple levels
the Labeled Faces in the Wild (LFW) [4] and MegaFace [5] of features for face recognition. However it has been
datasets. Many other applications have been developed surpassed by more recent approaches such as VGGFace2
using machine learning and face recognition algorithms [
        <xref ref-type="bibr" rid="ref1 ref4">1</xref>
        ]. VGGFace2 is a large-scale face recognition dataset
[
        <xref ref-type="bibr" rid="ref5">6, 7, 8, 9, 8</xref>
        ] 3D face recognition is an emerging field that has been used to train several deep learning-based
that uses 3D information to improve the accuracy of face face recognition algorithms, including some of the
aprecognition systems. 3D face models can capture addi- proaches mentioned above. It contains over 3 million
tional facial details such as the depth of the facial features, face images of over 9,000 subjects, making it one of the
which are not present in 2D images. Some of the popular largest face recognition datasets available. Another
conapproaches for 3D face recognition include the use of 3D volutional model is FaceNet [2] that, diferently from the
morphable models (3DMM), depth-based methods, and previous approaches, uses a triplet loss function to learn
multi-view based methods. Face recognition can be di- discriminative features for face recognition, being widely
vided into two categories: verification and identification. adopted in industry. With the outbreak of the
COVIDVerification aims to determine if two face images belong 19 pandemic, the use of face recognition algorithms has
to the same person, while identification aims to identify been applied to the recognition of people wearing (or
a person from a set of images. Face verification systems not wearing) face masks. One common approach is to
are commonly used for security applications, while face use deep learning-based object detection methods to
deidentification systems are used in large-scale surveillance tect the presence of a face and then classify whether the
applications. Recent advancements in face recognition face is wearing a mask or not. This approach typically
have focused on improving the accuracy of both veri- involves training a CNN on a dataset of masked and
unifcation and identification systems. Hybrid approaches masked faces. The CNN learns to extract features from
combine multiple techniques to improve the performance the face images and use them to classify whether a mask
of face recognition systems. For example, a hybrid ap- is present or not. Several studies have reported high
accuproach may combine deep learning models with 3D face racy rates for face mask detection using deep learning
alrecognition techniques to improve accuracy. Another gorithms. In fatct during the COVID-19 pandemic the use
approach is to use facial landmark detection algorithms of convolutional neural networks (CNNs) for face mask
to improve the alignment of face images before recog- detection has gained significant attention in literature.
nition. As face recognition technology becomes more Several studies have proposed CNN-based approaches
prevalent, there are growing concerns about privacy and for detecting masked faces. In [
        <xref ref-type="bibr" rid="ref9">15</xref>
        ] the authors appliy
security. Several approaches have been proposed to ad- a Long Short-Term Memory (LSTM) network to model
dress these concerns, such as anonymization techniques, the time-dependencies in order detect whether a person
which modify the facial features to protect the privacy of wears a face mask while speaking. In [
        <xref ref-type="bibr" rid="ref11">16</xref>
        ] the authors
the individuals in the images. Other approaches include are able to detect if face masks are worn by people in a
adversarial attacks, which aim to fool the face recognition closed enviroinment. Overall, deep learning algorithms
system into misidentifying a person. Face recognition have shown promise for detecting whether a person is
is a rapidly developing field that has seen significant wearing a face mask [17], and their use could help to
advancements in recent years, driven by the increasing improve public health measures during the COVID-19
availability of large-scale datasets and powerful machine pandemic. For example, the authors of [18] propose a
slightly modified version of LeNet [ 19] to detect masked than other facial image datasets such as the popular
Lafaces in the wild. Similarly the authors of [
        <xref ref-type="bibr" rid="ref15">20</xref>
        ] propose beled Faces in the Wild (LFW) dataset. The FFHQ dataset
a novel CNN architecture tailored for the specific task also includes annotations for facial landmarks, which can
and evaluate their model on a custom dataset. For more be used for tasks such as face alignment and face tracking.
information about the existing methods of covered face The FFHQ dataset has been used in a range of computer
detection we refer the reader to the recent survey [
        <xref ref-type="bibr" rid="ref19">21</xref>
        ]. vision research projects, including the development of
Our method difers from the other works in literature generative models for face synthesis and style transfer,
both in the dataset used and the training strategy. In as well as the development of deep learning-based
modparticular we will address the task as a multi-label clas- els for facial expression recognition and emotion
detecsification problem. Previous works [
        <xref ref-type="bibr" rid="ref22 ref24">22, 23</xref>
        ] have shown tion. The availability of high-quality facial images in the
that this approach allows a better understanding of the FFHQ dataset has helped to advance the state-of-the-art
context. We will see that this approach allows to reduce in these and other computer vision tasks, and it is likely
overfitting and consequently to generalize better on un- to continue to be an important resource for researchers
seen data. working in the field of computer vision. In this study
1000 images of faces uncovered and covered with masks
have been selected from the FFHQ dataset and merged,
3. Dataset to preserve generality, with a large portion of Kaggle face
mask dataset[
        <xref ref-type="bibr" rid="ref31">25</xref>
        ] while others were added manually in
order to increase the variability in the masks types and
colors; 160 images of masks with no faces scraped from
the web and 150 images containing diferent classes of
objects all unrelated to faces and to masks also scraped
from the web.
      </p>
      <sec id="sec-3-1">
        <title>To train our model we created a custom dataset con</title>
        <p>
          taining 151 images of uncovered faces extracted from
the Flickr-Faces-HQ dataset[
          <xref ref-type="bibr" rid="ref29">24</xref>
          ]. The Flickr-Faces-HQ
(FFHQ) dataset is a large-scale dataset of high-quality
facial images collected from the photo-sharing website
Flickr. The dataset was created by NVIDIA Research in
2019 and contains 70,000 images of 1,024x1,024
resolution, with a diverse range of ages, ethnicities, and genders. 4. Method
The images in the FFHQ dataset are highly curated and
ifltered for quality, ensuring that they are of high fidelity, For the classification task we used ResNet50 [ 26] a
dReshigh resolution, and well-lit. The dataset is designed to Net50 is a 50-layer deep neural network that is based
be used for training and evaluating machine learning on residual connections, which allow for the reuse of
models for various computer vision tasks, including face features from previous layers in the network. The
archirecognition, facial expression analysis, and face synthe- tecture includes several blocks of convolutional layers,
sis. One of the key features of the FFHQ dataset is that batch normalization, and pooling layers, as well as
shortit includes a wide range of facial expressions, poses, and cut connections that bypass one or more layers. These
lighting conditions, which makes it more challenging shortcut connections enable the network to learn
residual functions, which can be more easily optimized during
ResNet50 pretrained on ImageNet and we added a single
training and help to mitigate the vanishing gradient
probclassification layer with sigmoid activation (remember
lem that occurs in very deep networks. ResNet50 has
that the labels are not mutually exclusive). We used the
been trained by means of ImageNet [27]. ImageNet is a
standard binary cross entropy loss: considering the batch
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Results</title>
      <p>We fine-tuned the network for 200 epochs on images with
size 224x224 pixels. We used the feature extractor of the
size B, the predicted value  ̂ and the true value  we get:
( ,  )̂ = −</p>
      <p>3
1
 =1 =1
∑
∑  
where</p>
      <p>=   [] log( ̂ []) + (1 −   []) log(1 −  ̂ [])</p>
      <p>With this training strategy we obtained an accuracy
of 96% on a balanced testset, against the 90% of accuracy
obtained by the network trained to classify only covered
versus uncovered faces. To make a fair comparison
between the two approaches, the accuracy in the multilabel
classification problem is computed considering only the
label that indicates if the face is covered or not, which is
in fact a binary classification problem. Figure</p>
      <sec id="sec-4-1">
        <title>2 shows some samples along with the network predictions.</title>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. Conclusion</title>
      <p>Our findings suggest that CNNs can efectively detect
faces covered by surgical masks in medical settings,
which can be beneficial for enhancing security measures
while maintaining infection control protocols. The
proposed model has potential applications in healthcare
facilities, airports, and other settings where face mask
detection is critical for security and safety purposes. Future
work can explore additional data sources and further
optimization techniques to improve the model’s performance
and real-world applicability.</p>
      <p>815–823.
Vggface2: A dataset for recognising faces across
pose and age, in: 2018 13th IEEE international
conference on automatic face &amp; gesture recognition
(FG 2018), IEEE, 2018, pp. 67–74.
[2] F. Schrof, D. Kalenichenko, J. Philbin, Facenet: A
unified embedding for face recognition and
clustering, in: Proceedings of the IEEE conference on
computer vision and pattern recognition, 2015, pp.</p>
      <sec id="sec-5-1">
        <title>Similar face recognition using the ie-cnn model,</title>
        <p>IEEE Access 8 (2020) 45244–45253.
[4] G. B. Huang, E. Learned-Miller, Labeled faces in
the wild: Updates and new reporting procedures,</p>
      </sec>
      <sec id="sec-5-2">
        <title>Dept. Comput. Sci., Univ. Massachusetts Amherst, Amherst, MA, USA, Tech. Rep 14 (2014).</title>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Parkhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          , [3]
          <string-name>
            <given-names>A.-P.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.-H.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.-Y.</given-names>
            <surname>Di</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.-H.</given-names>
            <surname>Song</surname>
          </string-name>
          , [5]
          <string-name>
            <given-names>I.</given-names>
            <surname>Kemelmacher-Shlizerman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Seitz</surname>
          </string-name>
          ,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <article-title>Miller, tic scene understanding for context-aware humanE. Brossard, The megaface benchmark: 1 million robot interaction, Lecture Notes in Computer Scifaces for recognition at scale</article-title>
          ,
          <source>in: Proceedings of the ence (including subseries Lecture Notes in ArtiIEEE conference on computer vision and pattern ifcial Intelligence and Lecture Notes in Bioinforrecognition</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>4873</fpage>
          -
          <lpage>4882</lpage>
          . matics) 13196
          <source>LNAI</source>
          (
          <year>2022</year>
          )
          <fpage>310</fpage>
          -
          <lpage>325</lpage>
          .
          <source>doi:1 0 . 1 0</source>
          <volume>0 7</volume>
          / [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Marcotrigiano</surname>
          </string-name>
          , G. Stingi,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fregnan</surname>
          </string-name>
          , P. Maga-
          <volume>9 7 8 - 3 - 0 3 1 - 0 8 4 2 1 - 8</volume>
          _
          <fpage>2</fpage>
          1 .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>relli</surname>
            , P. Pasquale,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Russo</surname>
            , G. Orsi,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Montagna</surname>
            , [10]
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Deng</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Guo</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Xue</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Zafeiriou</surname>
            , Arcface: Addi
            <given-names>C.</given-names>
            Napoli, C.
          </string-name>
          <string-name>
            <surname>Napoli</surname>
          </string-name>
          ,
          <article-title>An integrated control plan tive angular margin loss for deep face recognition, in primary schools: Results of a field investiga- in: Proceedings of the IEEE/CVF conference on tion on nutritional and hygienic features in the computer vision</article-title>
          and pattern recognition,
          <year>2019</year>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <article-title>apulia region (southern italy)</article-title>
          ,
          <source>Nutrients</source>
          <volume>13</volume>
          (
          <year>2021</year>
          ).
          <fpage>4690</fpage>
          -
          <lpage>4699</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <source>doi:1 0 . 3 3 9 0 / n u 1 3</source>
          <volume>0 9 3 0 0 6</volume>
          . [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          , [7]
          <string-name>
            <given-names>G.</given-names>
            <surname>Capizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Woźniak</surname>
          </string-name>
          , Lessen-
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          , W. Liu, Cosface:
          <article-title>Large margin cosine loss ing stress and anxiety-related behaviors by means for deep face recognition, in: Proceedings of the of ai-driven drones for aromatherapy</article-title>
          ,
          <source>in: CEUR IEEE conference on computer vision and pattern Workshop Proceedings</source>
          , volume
          <volume>2594</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>recognition</fpage>
          ,
          <year>2018</year>
          , pp.
          <fpage>5265</fpage>
          -
          <lpage>5274</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          7-
          <fpage>12</fpage>
          . [12]
          <string-name>
            <given-names>W.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Raj</surname>
          </string-name>
          , L. Song, [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Ponzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Bianco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , A. Wa- Sphereface:
          <article-title>Deep hypersphere embedding for face jda, Psychoeducative social robots for an healthier recognition</article-title>
          ,
          <source>in: Proceedings of the IEEE conference lifestyle using artificial intelligence: a case-study, on computer vision and pattern recognition</source>
          ,
          <year>2017</year>
          , in: CEUR Workshop Proceedings, volume
          <volume>3118</volume>
          , pp.
          <fpage>212</fpage>
          -
          <lpage>220</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <year>2021</year>
          , pp.
          <fpage>26</fpage>
          -
          <lpage>33</lpage>
          . [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Tang</surname>
          </string-name>
          , Deepid3: Face [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>De Magistris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Caprari</surname>
          </string-name>
          , G. Castro,
          <string-name>
            <surname>S.</surname>
          </string-name>
          <article-title>Russo, recognition with very deep neural networks</article-title>
          , arXiv L.
          <string-name>
            <surname>Iocchi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Nardi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Napoli</surname>
          </string-name>
          ,
          <article-title>Vision-based holis-</article-title>
          preprint
          <source>arXiv:1502.00873</source>
          (
          <year>2015</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>W.</given-names>
            <surname>Ouyang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zeng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Qiu</surname>
          </string-name>
          , P. Luo, gle.com/datasets/andrewmvd/face-mask-detection, Y. Tian,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Li</surname>
          </string-name>
          , et al., Deepid-
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>net: Object detection with deformable part based</article-title>
          [26]
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , S. Ren,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Deep residual learnconvolutional neural networks, IEEE Transactions ing for image recognition</article-title>
          ,
          <source>in: Proceedings of the on Pattern Analysis and Machine Intelligence</source>
          <volume>39</volume>
          IEEE conference
          <article-title>on computer vision and pattern (</article-title>
          <year>2016</year>
          )
          <fpage>1320</fpage>
          -
          <lpage>1334</lpage>
          . recognition,
          <year>2016</year>
          , pp.
          <fpage>770</fpage>
          -
          <lpage>778</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mallol-Ragolta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Qian</surname>
          </string-name>
          , E. Parada- [27]
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>FeiCabaleiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. W.</given-names>
            <surname>Schuller</surname>
          </string-name>
          ,
          <article-title>Capturing time Fei, Imagenet: A large-scale hierarchical imdynamics from speech using neural networks for age database</article-title>
          , in: 2009 IEEE conference
          <article-title>on comsurgical mask detection</article-title>
          ,
          <source>IEEE Journal of Biomedical puter vision and pattern recognition, Ieee</source>
          ,
          <year>2009</year>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>and Health Informatics</source>
          <volume>26</volume>
          (
          <year>2022</year>
          )
          <fpage>4291</fpage>
          -
          <lpage>4302</lpage>
          .
          <fpage>248</fpage>
          -
          <lpage>255</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sang</surname>
          </string-name>
          ,
          <article-title>Face-mask recognition for fraud</article-title>
          [28]
          <string-name>
            <given-names>N.</given-names>
            <surname>Brandizzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brociek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wajda</surname>
          </string-name>
          ,
          <article-title>First prevention using gaussian mixture model, Journal studies to apply the theory of mind theory to green of Visual Communication and Image Representa- and smart mobility by using gaussian area clustertion 55 (</article-title>
          <year>2018</year>
          )
          <fpage>795</fpage>
          -
          <lpage>801</lpage>
          . ing,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3118</volume>
          , [17]
          <string-name>
            <given-names>R.</given-names>
            <surname>Brociek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Magistris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Cardia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Coppa</surname>
          </string-name>
          ,
          <year>2021</year>
          , pp.
          <fpage>71</fpage>
          -
          <lpage>76</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          , Contagion prevention of covid-19 by [29]
          <string-name>
            <given-names>V.</given-names>
            <surname>Ponzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Russo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wajda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Brociek</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Napoli, means of touch detection for retail stores</article-title>
          ,
          <source>in: Analysis pre and post covid-19 pandemic rorschach CEUR Workshop Proceedings</source>
          , volume
          <volume>3092</volume>
          ,
          <year>2021</year>
          ,
          <article-title>test data of using em algorithms and gmm modpp</article-title>
          .
          <volume>89</volume>
          -
          <fpage>94</fpage>
          . els,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3360</volume>
          , [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Cai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ji</surname>
          </string-name>
          , Masked face detection
          <year>2022</year>
          , pp.
          <fpage>55</fpage>
          -
          <lpage>63</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <article-title>via a modified lenet</article-title>
          ,
          <source>Neurocomputing</source>
          <volume>218</volume>
          (
          <year>2016</year>
          ) [30]
          <string-name>
            <given-names>G.</given-names>
            <surname>Magistris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rametta</surname>
          </string-name>
          , G. Capizzi,
          <string-name>
            <given-names>C.</given-names>
            <surname>Napoli</surname>
          </string-name>
          , Fpga
          <volume>197</volume>
          -
          <fpage>202</fpage>
          .
          <article-title>implementation of a parallel dds for wide-band ap</article-title>
          [19]
          <string-name>
            <given-names>Y.</given-names>
            <surname>LeCun</surname>
          </string-name>
          , L. Bottou,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hafner</surname>
          </string-name>
          , Gradient- plications, in: CEUR Workshop Proceedings, volbased learning applied to document recognition,
          <source>ume 3092</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>12</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <source>Proceedings of the IEEE</source>
          <volume>86</volume>
          (
          <year>1998</year>
          )
          <fpage>2278</fpage>
          -
          <lpage>2324</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>W.</given-names>
            <surname>Bu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Peng</surname>
          </string-name>
          , A cascade
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <article-title>framework for masked face detection</article-title>
          , in: 2017 IEEE
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <article-title>automation and mechatronics (RAM)</article-title>
          , IEEE,
          <year>2017</year>
          ,
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          pp.
          <fpage>458</fpage>
          -
          <lpage>462</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Chen</surname>
          </string-name>
          , A survey on masked
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <source>against covid-19</source>
          , IEEE Transactions on Artificial
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <source>Intelligence</source>
          <volume>3</volume>
          (
          <year>2021</year>
          )
          <fpage>323</fpage>
          -
          <lpage>343</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          , W. Xu,
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <year>2016</year>
          , pp.
          <fpage>2285</fpage>
          -
          <lpage>2294</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [23]
          <string-name>
            <surname>G. De Magistris</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Caprari</surname>
          </string-name>
          , G. Castro, S. Russo,
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <article-title>robot interaction</article-title>
          , in: AIxIA 2021-Advances in
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Artificial</surname>
            <given-names>Intelligence:</given-names>
          </string-name>
          20th International Confer-
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>gence</surname>
          </string-name>
          ,
          <source>Virtual Event, December 1-3</source>
          ,
          <year>2021</year>
          , Revised
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Selected</surname>
            <given-names>Papers</given-names>
          </string-name>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>310</fpage>
          -
          <lpage>325</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>T.</given-names>
            <surname>Karras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Laine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Aila</surname>
          </string-name>
          ,
          <article-title>A style-based generator</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <article-title>computer vision</article-title>
          and pattern recognition,
          <year>2019</year>
          , pp.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Kaggle</surname>
          </string-name>
          , Face mask detection, https://www.kag-
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>