<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>International Workshop of IT-professionals on Artificial Intelligence, October</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Analysis of the Influence of Cosine Distance Threshold Values in a Real-time Face Recognition System</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Maksym Holikov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Volodymyr Donets</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Viktoriia Strilets</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kyryl Korobchynskyi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Institute of Computer Science and Artificial Intelligence, V.N. Karazin Kharkiv National University</institution>
          ,
          <addr-line>4, Svobody, Sq., Kharkiv, 61022</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>1</volume>
      <fpage>5</fpage>
      <lpage>17</lpage>
      <abstract>
        <p>The quality of a face recognition system largely depends on the correct choice of threshold value when comparing vector features (embeddings). This paper investigates the impact of cosine distance thresholds on the performance of a real-time face recognition system. The proposed approach combines face detection using MediaPipe FaceMesh and feature extraction using the ArcFace model. A series of experiments with different threshold values was conducted, the results of which were evaluated using the following metrics: Accuracy, Precision, Recall, False Accept Rate (FAR), and False Reject Rate (FRR). The results show that the choice of threshold directly determines the trade-off between security and convenience of the system. The optimal range of cosine distance threshold values was established as 0.050.07, which minimizes both FAR and FRR, which is important for practical use in video surveillance, access control, and user authentication systems.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Computer vision</kwd>
        <kwd>deep learning</kwd>
        <kwd>face recognition</kwd>
        <kwd>cosine distance</kwd>
        <kwd>real-time systems</kwd>
        <kwd>MediaPipe</kwd>
        <kwd>FaceMesh</kwd>
        <kwd>ArcFace 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Biometric face recognition technologies are one of the most widespread and most researched areas
in the field of computer vision. As a result of the development of deep neural networks, modern
face recognition models have achieved a level of accuracy that is close to human perception on
well-known test sets (LFW, MegaFace, IJB-C) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Such systems are already widely used in video
surveillance, access control, banking security, and mobile authentication.
      </p>
      <p>
        A key step in the task of face recognition is calculating the distance between two embedding
vectors representing faces in recognition models. Most often, cosine distance or its variations are
used for this purpose [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. However, the final decision on whether two samples match or differ
depends on a threshold value that determines the balance between the False Accept Rate (FAR) and
False Reject Rate (FRR) indicators [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. A threshold value that is too low leads to frequent false
rejections (failure to recognize even a real user), while a threshold value that is too high leads to
false acceptances (identifying different people as the same person). Thus, the correct choice of
threshold is crucial for the system's reliability.
      </p>
      <p>
        The literature notes that optimal threshold values can vary significantly depending on the data
set, shooting conditions, and even personal characteristics (race, age, gender) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. This creates a
problem of threshold inconsistency, where the model demonstrates different effectiveness at the
same threshold on different samples. For practical applications, a universal threshold is usually
chosen, for example, one that provides an Equal Error Rate (EER) or a fixed FAR level, but accuracy
may be reduced [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        This work focused on an experimental study of the impact of cosine distance thresholds on face
recognition quality. To this end, a pipeline was used that combines face detection with MediaPipe
FaceMesh [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and feature extraction using the ArcFace model [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. A series of experiments with
different thresholds was conducted, and the results were evaluated using standard metrics
(Accuracy, Precision, Recall, FAR, FRR) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. The results obtained allow determining the optimal
range of thresholds for practical use and show the compromise between security and the
convenience of the system.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>In face recognition systems, two images are converted into embedding vectors (e.g., using FaceNet,
ArcFace, CosFace, etc.), and the similarity between them is usually assessed using cosine similarity
(or the corresponding cosine distance) [8]. A pair is considered a ‘match’ (the same identifier) if the
similarity value exceeds a certain threshold; otherwise, it is considered a ‘non-match’.
Classification errors are defined as False Accept (FA) – mistakenly accepting different individuals
as one, and False Reject (FR) – mistakenly rejecting one individual as two. The corresponding FAR
(False Accept Rate) and FRR (False Reject Rate) indicators depend on the selected threshold. The
EER (Equal Error Rate) point corresponds to the threshold at which FAR = FRR [8]. The ROC
(TPR–FPR) curve and the construction of DET nomograms are also obtained by searching through
thresholds. Therefore, the choice of threshold value directly determines the quality indicators
(Accuracy, Precision, Recall, etc.) [8].</p>
      <p>Lowering the threshold (a more ‘lenient’ matching criterion) leads to an increase in FAR, when
more impostors are mistakenly accepted, and a decrease in FRR, and vice versa when the threshold
is raised. For example, when evaluating on LFW or other benchmarks, the standard technique is
10fold cross-validation to select the optimal threshold, and most often it is the threshold value of
cosine similarity [9]. However, many studies note that a threshold selected for one sample may not
be suitable for another with a different origin, lighting, racial composition, etc. [9]. During training,
the model can be optimised based on both the distance between feature vectors and the angle
between them. In the case of the ArcFace and CosFace approaches, an additional angular or cosine
shift is introduced, which increases the resolution of the feature space. However, at the verification
stage, it is still necessary to determine the threshold value for making a decision. This threshold is
often set based on the desired level of FAR or EER. The literature states that traditional approaches
compare both classes (mated/impostor) with a fixed threshold, but ‘the best threshold for different
classes is often different’ [9]. It is argued that the optimal threshold is usually specific to a
particular dataset – the best thresholds for different datasets often differ. In practice, it is difficult to
find the optimal threshold without access to test data [9].</p>
      <p>
        Some works propose adaptive or coordinated threshold selection strategies that implement
adaptive thresholding for each registered face in the database: instead of a single global threshold,
they store a separate threshold for each sample, which leads to a significant improvement in
accuracy (up to a 22% increase on LFW in their protocol) [9]. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the authors highlight the
discrepancy in thresholds across different domains and propose a new protocol called
‘OneThreshold-for-All’, which utilizes a single fixed threshold (referred to as the Calibration Threshold)
for evaluating multiple datasets simultaneously. They show that the traditional approach of
selecting a separate threshold for each dataset is inconsistent with the practical scenario of a single
threshold and slows down the implementation of models. In [10], the concept of threshold
inconsistency is introduced: even if the model is very accurate, different thresholds may be
required for different classes to maintain the same FAR/FRR level, and the OPIS metric is proposed
to measure the discrepancy of thresholds across classes, showing that solutions optimized for
accuracy alone often worsen threshold consistency [10].
      </p>
      <p>Other approaches focus on the selected operating mode. For example, in [11], the focus is on a
fixed FAR level (‘Anchor FAR’) as a key criterion for practical FR systems: they optimize the goal of
maximising TAR (True Accept Rate) at a given FAR, showing that different models are optimally
different at different target FAR values [11]. Thus, the choice of threshold (and, accordingly, FAR)
determines which model gives the best result.</p>
      <p>A number of studies specifically investigate the role of cosine similarity in verification tasks.
Work [12] analyses the distribution of cosine distances between positive and negative pairs in the
complex DFW2019 dataset: it turns out that many ‘real’ pairs have low cosine similarity (due to
face masking), which complicates threshold selection. This demonstrates that a fixed threshold on
hard data can generate false rejections. Similarly, work [13] indicates that after training, the model
has a fixed ‘cut-off level’ of cosine similarity, which is not entirely consistent with the testing
procedure (where the threshold is strictly fixed). New loss functions also explicitly take the
threshold into account: for example, [14] introduces USS Loss, which trains a single unified
threshold for all pairs. Using 20 random identities as an example, it is shown that the optimal
thresholds for them almost coincide with a single ‘unified’ threshold (about 0.4896). This indicates
that this approach provides a more compact distribution of threshold values and simplifies the
decision on a match.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Problem statement</title>
      <p>The task of face verification is formulated as determining whether two images belong to the same
person. To do this, each image is converted into a vector embedding using a neural network
f ( x )∈ Rd, where d - the dimensionality of features (e.g., 512 in the case of ArcFace). For a pair of
embeddings f ( xi), f ( x j) the cosine distance is calculated using the formula:</p>
      <p>D ( f ( xi) , f ( x j))=1−</p>
      <p>n
∑ ( xi yi)
i=1
√∑ ( xi2)⋅√∑ ( yi2)
n n
i=1 i=1
.</p>
      <p>The system decides on a ‘match’ (1) or ‘no match’ (0) based on the threshold value τ :
(1)
(2)
decision={0 , D ( f ( xi) , f ( x j))≥τ</p>
      <p>1 , D ( f ( xi) , f ( x j))&lt; τ .</p>
      <p>The problem lies in choosing the optimal threshold value τ . Depending on its value, the system
indicators change:
• False Accept Rate (FAR) – the proportion of cases in which different people are mistakenly
identified as the same person;
• False Reject Rate (FRR) – the proportion of cases where a genuine user is rejected by the
system;
• Accuracy, Precision, Recall – integral quality indicators;
• Equal Error Rate (EER) – the point where FAR = FRR, often used as a benchmark for
selecting τ .</p>
      <p>In practical systems, it is impossible to find a universal threshold that will be equally effective
for all scenarios and data sets. Too low a value reduces FAR but increases FRR, which degrades
usability. Conversely, too high a value reduces FRR but increases FAR. Thus, the balance depends
directly on the chosen threshold.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Face recognition system model</title>
      <p>The input data for the face recognition system is a video stream from a camera. Each frame is sent
to a pre-processing module, where image normalisation and basic transformations are performed to
improve the stability of subsequent analysis stages. In particular, procedures for conversion to
standard colour space, scaling, and noise filtering are applied.</p>
      <p>The sequence diagram for face identification is shown inFigure 1:</p>
      <p>As can be seen in Figure 1, the system's operation sequence unfolds in several main stages.
Preprocessing of the frame is implemented in steps 2–3, where conversion to the standard colour
space and key point detection are performed. Further face region selection (step 4) ensures the
formation of an ROI for each face found. In step 4.2.1, the ROI is transferred to the vector
representation formation service, where a 512-dimensional embedding is calculated using the
ArcFace model. Next, depending on the state of the user base (whether the database is empty or the
database is not empty branch), the embedding is either added as a new record or compared with
existing vectors based on cosine distance. The final step is to return the closest user or add a new
one, which is consistent with the subsequent text description.</p>
      <p>Face detection in a frame is performed using key point analysis and facial geometry methods.
The use of a topological grid allows not only to highlight the region of interest (ROI), but also to
increase accuracy by taking into account variations in poses, expressions, and partial overlaps.</p>
      <p>This method provides significantly greater accuracy and stability compared to classic trackers
such as CSRT, MOSSE, or KCF. Traditional trackers focus on local patterns or pixel movement
between frames and work well in relatively static conditions. However, they have a number of
limitations in dynamic environments: with partial overlaps, sudden changes in head position or
facial expressions, trackers can lose the object, and the bounding box ‘slides’, leading to incorrect
ROI formation and reduced accuracy of subsequent identification.</p>
      <p>The use of a topological grid allows for the identification of over 400 key facial points, including
the contours of the eyes, nose, mouth, and outer contour of the head. This ensures accurate ROI
selection, which includes only the most relevant facial pixels, avoiding the background, hair, or
other extraneous elements. High ROI detail directly affects the quality of the resulting vector
representations (embeddings), increasing the accuracy of comparison and verification.</p>
      <p>Another advantage of this approach is that classic trackers are often prone to ‘shifts’ when the
user moves closer to or further away from the camera or when the lighting angle changes. Face
Mesh, on the other hand, provides stable ROI detection regardless of such variations, and
subsequent normalization of the region of interest ensures uniformity of vector representations for
all frames.</p>
      <p>In addition, this method of determining ROI does not depend on the initial frame or prior
initialization. Unlike trackers, which lose the object when it disappears from the frame and require
re-initialization, Face Mesh processes each frame independently. This makes the system more
reliable in dynamic environments where users appear or disappear from the camera's field of view.</p>
      <p>Thus, the use of a topological grid to determine ROI provides more accurate face detection, high
resistance to changes in pose, lighting, and partial overlaps, as well as stability of the resulting
embeddings. Compared to classical trackers, this approach increases the reliability of the face
verification system and improves the quality of the final result, which is critical in real-time tasks
and interactive user monitoring.</p>
      <p>After ROI selection, the face image is converted into a compact vector representation —
embedding. This is a multidimensional vector that encodes the most important features for
identifying a person. The vector space is chosen so that the distances between points correspond to
semantic proximity: two images of the same person are located close to each other, while images of
different people are located at a relatively large distance.</p>
      <p>In the proposed system, embeddings are used as a universal format that allows comparisons to
be made regardless of lighting conditions, head position, or changes in appearance. This approach
makes the method more generalised and less dependent on a specific data set, which is especially
important for systems that need to work with new users without retraining.</p>
      <p>Cosine distance is used to assess the degree of similarity between face embeddings. Cosine
distance was chosen because it is invariant to the absolute length of feature vectors and can more
accurately reflect the similarity between multidimensional representations of faces. Unlike
Euclidean metrics, which can be sensitive to scale variations, cosine distance only evaluates the
angle between vectors, making it more reliable in conditions of changing lighting or small
variations in facial expressions.</p>
      <p>Thanks to this approach, verification is reduced to the task of comparing numerical values and
can be performed at high speed, which meets the requirements for real-time systems.</p>
      <p>The final stage is the integration of verification results into the video stream. For each face
detected in the frame, the system applies a corresponding label with the user ID or a ‘new’ mark.</p>
      <p>The proposed approach has the following advantages:
• resistance to environmental dynamics;
• the algorithm operates in real time using vector distance calculations.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and discussion</title>
      <p>The experiments were conducted on the LFW (Labelled Faces in the Wild) dataset [15], which is
widely used to evaluate face verification algorithms. The dataset contains over 13,000 photographs
of people taken in uncontrolled conditions, allowing for the simulation of real-life scenarios. The
images show significant variations in lighting, head position, accessories (glasses, headwear), and
0.02
0.03
0.04
0.05
0.06
0.07
0.08
image quality, making LFW one of the most widely used standards for evaluating face recognition
algorithms.</p>
      <p>To construct test pairs, both positive examples (images of one person) and negative examples
(different people) were selected.</p>
      <p>The results for different threshold values of the cosine distance are given in Table 1.</p>
      <p>As can be seen from the table, the choice of threshold value directly affects the accuracy of the
system and the FAR/FRR error ratio.</p>
      <p>Low threshold values (0.02 – 0.03) are too strict. At a threshold of 0.02, the FRR value reaches
34.2%, which means that one-third of genuine users fail the verification. Despite the absence of
false acceptances (FAR = 0), this mode is unsuitable for real-world applications due to its low
usability. Raising the threshold to 0.03 significantly reduces the FRR to 13.2% and increases
accuracy to 86.8%, but the number of false rejections is still too high.</p>
      <p>In the middle range (0.04–0.05), performance gradually stabilises. FRR decreases to 10.5% and
7.9% respectively, while accuracy increases to 92.1%. This already makes the system suitable for use
in relatively controlled scenarios (e.g., office entrance with regular users). However, there is still a
risk that some users will be falsely rejected.</p>
      <p>At a threshold of 0.06, the system demonstrates very high performance: Accuracy = 97.4%, FRR
= 2.6%, FAR = 0. This means that only 1 in 38 genuine users may be rejected, with no false
acceptances recorded. This result is the most balanced and practically significant: the system
becomes user-friendly while maintaining a high level of security.</p>
      <p>Starting from a threshold of 0.07, the system achieves perfect results — Accuracy, Precision, and
Recall are 100%, and FAR and FRR are zero. From a technical point of view, this means that no
errors were recorded in the test sample. However, as previous research shows, achieving ‘perfect’
results is often explained by the limited or homogeneous nature of the sample. In real-life scenarios
— with different lighting, poses, accessories (glasses, masks) — it is practically impossible to avoid
errors. Therefore, such a result should be considered more as an artefact of a specific experiment
rather than a guarantee of absolute reliability.</p>
      <p>However, the obtained ‘ideal zone’ (0.07–0.08) can be partially explained by the specifics of the
test data set, since in practice there is always noise, variations in lighting, poses, and appearance,
which make it impossible to achieve absolute indicators. This is consistent with the well-known
FAR
0.0
0.0
0.0
0.0
0.0
0.0
0.0</p>
      <p>FRR
0.3421
0.1316
0.1053
0.0789
0.0263
0.0000
0.0000
problem of threshold inconsistency, where the optimal threshold depends on the conditions of use
and sample characteristics.</p>
      <p>Thus, the optimal operating range for the system can be defined as 0.05–0.07, where the best
compromise between minimizing FRR and maintaining zero FAR is achieved. This result is
important for practical real-time applications such as video surveillance, where even a single
impostor acceptance error can have critical consequences.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>This paper investigated the impact of cosine distance threshold values on the performance of
realtime face recognition systems. Experimental results showed that the choice of threshold directly
determines the balance between FAR and FRR metrics, as well as the overall accuracy of the
system.</p>
      <p>It was established that:
• thresholds that are too low (0.02–0.03) result in high FRR, which reduces usability;
• in the range of 0.05–0.06, the system demonstrates an optimal compromise between
security and accessibility, ensuring high Accuracy values and zero FAR;
• starting from a threshold of 0.07, the test set shows perfect results (100% Accuracy,
Precision, and Recall), but this result is likely due to the characteristics of the sample and
requires additional verification on more heterogeneous data.</p>
      <p>Thus, the optimal operating threshold for the system under study can be determined as 0.05–
0.07, which minimises the number of false rejections without the risk of accepting an imposter. The
conclusions obtained are important for the practical implementation of face verification
technologies in video surveillance, access control, and user authentication tasks.</p>
      <p>Further research could focus on testing the stability of the optimal threshold on different data
sets, developing adaptive thresholding methods for specific users, and integrating additional factors
(lighting conditions, dynamic scenes, changes in appearance) that affect the accuracy of the system.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The authors have not used Generative AI tools and services.
[8] S. Kilany and A. Mahfouz, “A comprehensive survey of deep face verification systems
adversarial attacks and defense strategies”, Scientific Reports, vol. 15, no. 1, Aug. 2025. doi:
10.1038/s41598-025-15753-8.
[9] H.-R. Chou, J.-H. Lee, Y.-M. Chan, and C.-S. Chen, “Data-specific adaptive threshold for face
recognition and authentication”, 2019 IEEE Conference on Multimedia InformationProcessing
and Retrieval (MIPR), Mar 2019. doi: 10.1109/MIPR.2019.00034.
[10] Q. Zhang, L. Xu, Q. Tang, J. Fang, Y. N. Wu, J. Tighe, and Y. Xing, “Threshold-consistent
margin loss for open-world deep metric learning”, arXiv, 2023. doi: 10.48550/arXiv.2307.04047.
[11] Liu, J.; Qin, H.; Wu, Y.; Liang, D. Anchorface: Boosting tar@far for practical face recognition.</p>
      <p>In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 22
February–1 March 2022.
[12] Deng, J., Zafeririou, S.: ArcFace for disguised face recognition. In: 2019 IEEE/CVF International</p>
      <p>Conference on Computer Vision Workshop (ICCVW), pp. 485–493. IEEE, Piscataway (2019).
[13] Liu, L.; Chen, M.; Chen, X.; Zhu, S.; Tan, P. GB-CosFace: Rethinking softmax-based face
recognition from the perspective of openset classification. arXiv 2021, arXiv:2111.11186.
[14] Q. Li, X. Jia, J. Zhou, L. Shen, and J. Duan, “UniTSFace: Unified threshold integrated
sample-tosample loss for face recognition”, arXiv, 2023. doi: 10.48550/arXiv.2311.02523.
[15] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller. Labeled Faces in the Wild: A
Database for Studying Face Recognition in Unconstrained Environments. University of
Massachusetts, Amherst, Technical Report 07-49, 2007.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wu</surname>
          </string-name>
          , “
          <article-title>Rotation consistent margin loss for efficient low-bit face recognition,”</article-title>
          <source>in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>6866</fpage>
          -
          <lpage>6876</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Datacamp</surname>
          </string-name>
          , “What is Cosine Distance?”,
          <year>2024</year>
          . URL: https://www.datacamp.com/tutorial/cosinedistance.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Zaidi</surname>
            ,
            <given-names>A.Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chong</surname>
            ,
            <given-names>C.Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jin</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parthiban</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Sadiq</surname>
            ,
            <given-names>A.S.</given-names>
          </string-name>
          ,
          <article-title>Touch-based continuous mobile device authentication: State-of-the-art, challenges and opportunities</article-title>
          ,
          <source>J. Network Comput. Appl.</source>
          ,
          <year>2021</year>
          , vol.
          <volume>191</volume>
          ,p.
          <fpage>103162</fpage>
          . https://doi.org/10.1016/j.jnca.
          <year>2021</year>
          .
          <volume>103162</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Cavazos</surname>
            ,
            <given-names>J. G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Phillips</surname>
            ,
            <given-names>P. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castillo</surname>
            ,
            <given-names>C. D.</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>O</given-names>
            <surname>'Toole</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. J.</surname>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Accuracy comparison across face recognition algorithms: where are we on measuring race bias? arXiv preprint</article-title>
          arXiv:
          <year>1912</year>
          .07398. doi:
          <volume>10</volume>
          .48550/arXiv.
          <year>1912</year>
          .
          <volume>07398</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Jiaheng</given-names>
            <surname>Liu</surname>
          </string-name>
          , Zhipeng Yu, Haoyu Qin, Yichao Wu, Ding Liang,
          <string-name>
            <given-names>Gangming</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>and Ke</given-names>
            <surname>Xu</surname>
          </string-name>
          .
          <year>2022</year>
          .
          <article-title>OneFace: One Threshold for All</article-title>
          .
          <source>In Eur. Conf. Comput. Vis</source>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6] GitHub,
          <string-name>
            <surname>Google AI Edge MediaPipe Face Mesh</surname>
          </string-name>
          ,
          <year>2020</year>
          . URL: https://github.com/google-aiedge/mediapipe/wiki/MediaPipe-Face-Mesh.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>Y. H.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>H. H.</given-names>
          </string-name>
          (
          <year>2022</year>
          ).
          <article-title>Deep face recognition for dim images</article-title>
          .
          <source>Pattern Recognition</source>
          ,
          <volume>126</volume>
          ,
          <fpage>108580</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>