Mathematical Methods for Information Technology of Biometric Identification in Conditions of Incomplete Data Oleksii Bychkov, Oleksii Ivanchenko, Kateryna Merkulova and Yelyzaveta Zhabska Taras Shevchenko National University of Kyiv, Volodymyrs’ka str. 64/13, Kyiv, 01601, Ukraine Abstract The purpose of this research is to develop mathematical methods for information technology of biometric identification which will allow to recognize person’s face in conditions of incomplete data such as wearing a medical face masks during the pandemic. During the ongoing pandemic researchers focus on quick and effective solutions to develop technologies that handle this problem. This research concentrates on the analysis of the already existing solutions and proposes a mathematical method of face identification for information technology based on wavelet transform under the condition of wearing masks by people. During this research, the experiments with face detection and recognition have been performed with the constraint information of covered face. There is no database of face images with masks, therefore a new database was created. This database contains 820 images of 40 people, whose faces was limited only by top part of the face (forehead, eyes). First experimental part was performed with the use of standard Python library face_recognition, which allows to perform face recognition from Python or from the command line with one of the simplest face recognition libraries. It built based on dlib face recognition toolkit with deep learning – it is a ResNet network that contains 29 conv layers. In the second set of experiments FaceNet system was used. It is a system that after high-quality features extracting from the face containing image creates a face embedding and predicts these features representation in a form of 128-element vector. Third part of experiments was performed to analyze the efficiency of three well-known face recognition methods: Eigenfaces, Fisherfaces and LBPH. Eigenfaces algorithm considers that different face parts are not identically significant in a face recognition process. It processes all the training images of all the people as a whole and extracts the components which are relevant and useful. Fisherfaces algorithm extracts principal components that differentiate one person from the others, so an individual's components become more useful over the others. The of LBPH algorithm is to find face local structure by comparing each pixel to its neighboring pixel, forming a list of local binary patterns, that can be converted into a decimal number. Mathematical methods proposed in this research based on wavelet transform, that is widely used in the image processing tasks. Wavelet transform provides processing of patterns hidden in the data performing data analysis in general as well as in the detail. To compare the results of the commonly used algorithms with wavelet transform there was developed algorithm with the use of Daubechies wavelets and reverse biorthogonal wavelets. The results of experimentation analysis indicate that the popular and commonly used methods of face identification do not demonstrate high efficiency results. Proposed mathematical methods for information technology based on wavelet transform improves the face recognition and identification process under the condition of faces covered with mask. Specifically, the most accurate identification rate of 77,5% was obtained with the use of Daubechies wavelets. Keywords Biometric identification, face recognition, wavelet transform 1 IT&I-2020 Information Technology and Interactions, December 02–03, 2020, KNU Taras Shevchenko, Kyiv, Ukraine EMAIL: : bos.knu@gmail.com (O. Bychkov); ivanchenko.oleksii@gmail.com (O. Ivanchenko); kate.don11@gmail.com (K. Merkulova), y.zhabska@gmail.com (Y. Zhabska) ORCID: 0000-0002-9378-9535 (O. Bychkov); 0000-0002-8526-8211 (O. Ivanchenko); 0000-0001-6347-5191 (K. Merkulova), 0000-0002-9917-3723 (Y. Zhabska) ©️ 2020 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings (CEUR-WS.org) 336 1. Introduction COVID-19 is an unprecedented crisis that has caused a huge number of casualties and security concerns. One of the methods which are using to reduce the spread of the coronavirus, is to wear masks in all public places to protect yourself and the others from being infected. So protective masks in these conditions have become an attribute of everyday life. It makes face recognition very challenging because some parts of the face are hidden. But face is the most common pattern that people use to identify each other in everyday life. Also, governmental identity documents, such as passport or driver’s license, contains face image. Even in ideal conditions, face recognition technologies often struggle with accuracy. Nowadays wearing medical masks caused increasing of the probability of false identification results of recognition systems, as the US National Institute of Standards and Technology (NIST) concluded. The National Institute of Standards and Technology conducted the research in which 89 face recognition algorithms with error rate of 0.3% were tested. And researchers found, after applying of those algorithms on images of persons with face masks, that the error rate increased from 5% to 50% [1]. Mask-wearing is now recommended as a measure to provide the spread of COVID-19. Therefore, the government, that uses face recognition algorithms to track and identify people across the US, conducted the research performed by the NIST in collaboration with the US Customs and Border Protection and Department of Homeland Security, both of which apply face recognition methods in their work. The research suggests that face mask wearing would decrease a person identification rate by face recognition. The result of a face recognition depends on mathematical models of the relative positions of face features. Anything that reduces the visibility of key characteristics of a face (such as the nose, mouth and chin) obstructs the positive outcome of face recognition. Findings also suggest that dark masks make it more difficult to recognize a face than blue surgical masks, and that wide masks that cover a person's entire face prevent the recognition more than round N95-style masks [2]. Notably, the report only lists a type of face recognition known as one-to-one matching. This is a procedure used at border crossings and at passport control, where an algorithm checks if a person matches their ID. However, facial recognition systems of mass surveillance scan the crowd to find matches with faces in the database, which is called a one-to-many system. And while the NIST report does not cover one-to-many systems, they are generally considered the most error-prone. The process of identifying a face in a crowd is more difficult, because of the different image scaling, angle of face position and lightning. Thus, wearing masks is likely to seriously interfere with one-to-many algorithms as well. Many companies that work on face recognition technologies have claimed they can identify people with high accuracy rate even in conditions of face masks wearing, but the latest research results show that the face coverings significantly increase face recognition error rates [3, 4]. During the ongoing pandemic researchers focus on quick and effective solutions to develop technologies that handle problem of face identification accuracy decreasing. This paper concentrates on the analysis of the already existing solutions and proposes a method of face identification based on wavelet transform under the condition of wearing masks by people, considering different image scaling, angle of face position and lightning. Experiments with face detection and recognition have been performed during the research with the incomplete data such as covered with mask face. Since there is no database which contains face images with masks, a new database was created. This database contains 9 images of 80 people, whose faces was limited only by top part of the face (forehead, eyes). These images prepared with different image scaling, angle and lightning. 2. Task solution methods 2.1. Face_recognition method Python face_recognition library allows to perform face recognition and manipulation on images from Python or from the command line with one of the simplest face recognition libraries, built based 337 on dlib face recognition toolkit with the use of deep learning. Applying of this model on the Labeled Faces in the Wild database indicated an accuracy rate of 99.38% [5]. Dlib's face recognition library performs conversion of a face image to a 128-dimensional vector space where images of the same person have a small Euclidean distance between them and images of different persons are, on the contrary, long-distance by Euclidean distance metric calculation. Dlib model is based on the ResNet network that contains 29 conv layers. It is a version of the ResNet-34 network from the research study [6] with an exclusion of a few layers and filters per layer reduction by half. The network training was performed on a 3 million face images dataset, that includes several datasets (the VGG dataset [7], face scrub dataset [8]) and a large amount number of images derived from the internet. The total number of face images in the dataset is 7485 images. Also, the face_recognition library method is using the k-nearest-neighbors (KNN) algorithm for face recognition [9]. It is useful when it is needed to recognize a large set of people, whose face images are stored in the database, and make a prediction for an unknown person in a feasible computation time. KNN algorithm provided by Python data science library - scikit-learn. Neighbors-based classification based on prediction of a class labels using majority rule voting for the nearest neighbors of each point: the data class that has the most representatives within the nearest neighbors of the point defines as the query point. Scikit-learn provides two different nearest neighbors classifiers: KNeighborsClassifier and RadiusNeighborsClassifier. The most used method in KNeighborsClassifier is k-neighbors classification. Each query point can be obtained with the use of k nearest neighbors learning, where k is an integer value specified by the user. The determination of k value depends on the data meaning that a larger k reduces the noise effects but, at the same time, makes the classification boundaries less distinct. For the classification based on nearest neighbors there are commonly used uniform weights. But some conditions require to use the neighbors weighting so that nearer neighbors contribute more to the fit. For the purpose of this research it is better to use distance weights which are proportional to the inverse of the distance from the query point. The KNN classifier is initially trained on a set of face images stored in the database that were labeled. Then it can identify the person in an input image by finding the k most similar faces by calculation of the closet face-features under Euclidean distance in its training set, and performing a majority vote (possibly weighted) on their label. For example, if k=3, and the three closest face images to the inputed image in the training set are one image of class1 and two images of class2, the result would be class2. 2.2. FaceNet method FaceNet is a face recognition system that was presented in the paper [10]. It creates a mapping from face images to Euclidean space where, on the basis of a deep convolutional network computation, calculated distances correspond to a measure of face similarity. System extracts high-quality features from the face image, given as an input, and transform they to a 128-element vector representation, performing a face embedding. This compact 128-D embedding uses a LMNN based triplet loss function [11]. Triplet loss function uses to determine similarity of the vectors, that can be defined by the calculation of the distance between vectors. Vectors for image of the same person is more similar (distance between those vectors is smaller) and vectors for the images of different persons is less similar (distance between those vectors is larger). It means that an embedding f(x), from an image x into a feature space Rd, such that, independently of image quality, the squared distance between all face images of the same person is small, and the squared distance between face images of different persons is large. Triplet loss function can be described as following. Image embedding representation expression is: 𝑓 (𝑥 ) ∈ 𝑅 𝑑 . (1) It means that an image x is being converted into a d-dimensional Euclidean space. Then it is necessary to determine that an input image xi a of a person is closer to other images of the same person xi p than all images of different persons xi n. It can be expressed with the next: ‖𝑓(𝑥𝑖 𝑎 ) − 𝑓(𝑥𝑖 𝜌 )‖22 + 𝛼 < ‖𝑓(𝑥𝑖 𝑎 ) − 𝑓(𝑥𝑖 𝑛 )‖22 , (2) 338 ∀(𝑓 (𝑥𝑖 𝑎 ), (𝑓𝑥𝑖 𝜌 ), 𝑓(𝑥𝑖 𝑛 )) ∈ 𝑇, (3) where α – is a limited space between images of the same person and all images of any other person; T represents the training set of all possible triplets with N cardinality. The loss is needed to minimize, that can be described with the following: 𝑁 (4) ∑[‖𝑓(𝑥𝑖 𝑎 ) − 𝑓(𝑥𝑖 𝜌 )‖22 − ‖𝑓(𝑥𝑖 𝑎 ) − 𝑓(𝑥𝑖 𝑛 )‖22 + 𝛼 ]. 𝑖 As the basis for classifier training system uses obtained face embeddings. FaceNet model is provided by the Python deep learning API – Keras, that contains the pretrained Inception ResNet v1 model. This model has certain restrictions concerning input data. Expected input images must be colored, with pixel values standardized, and to have a square shape with the size of 160×160 pixels. Face detection process is based on the use of Multi-Task Cascaded Convolutional Neural Network (MTCNN) [12]. It performs face finding and extracting from images with solution of 3 tasks: face classification, bounding box regression and facial landmark localization. For each sample of classification xi the cross-entropy loss can be calculated: 𝐿𝑑𝑒𝑡 𝑖 = −(𝑦𝑖𝑑𝑒𝑡 log(𝜌𝑖 ) + (1 − 𝑦𝑖𝑑𝑒𝑡 )(1 − log(𝜌𝑖 ))), (5) where pi is the network probability of a sample being a face; yidet∈ {0, 1} is the correct label. For each window, that possibly contains face, the offset between it and the nearest correct label can be predicted. Therefore, the research objective can be expressed as a regression problem, so the Euclidean loss for each sample xi can be applied: 2 𝐿𝑏𝑜𝑥 = ‖𝑦̂𝑖𝑏𝑜𝑥 − 𝑦𝑖𝑏𝑜𝑥 ‖2 , (6) 𝑖 where 𝑦̂𝑖𝑏𝑜𝑥 regression target obtained from the network and 𝑦𝑖𝑏𝑜𝑥 is the correct coordinate. There are 4 coordinates, including left top, height and width, and thus 𝑦𝑖𝑏𝑜𝑥 ∈ R4. After face detection and feature vector extraction, face classification takes part. For this purpose, Linear Support Vector Machine [13] can be used. The SVM algorithm is implemented using a kernel. Linear SVM can be presented with the use of the inner product between any two observations, rather than the observations themselves. The inner product between two vectors is the sum of the multiplication of each pair of input values. A prediction for an input can be expressed with the following equation that uses the dot product between the input (x) and each support vector xi: 𝑓 (𝑥 ) = 𝐵0 + 𝑠𝑢𝑚(𝑎𝑖 ∗ (𝑥, 𝑥𝑖 )), (7) where the coefficients 𝐵0 and 𝑎𝑖 (for each input) must be obtained after applying of the learning algorithm to the training data. Thereby, this face classification system uses 3 methods: an MTCNN model for face detection, FaceNet model to create a feature vectors for each detected face and Linear Support Vector Machine (SVM) classifier model to predict the identity of a given face. 2.3. OpenCV methods: Eigenfaces, Fisherfaces, and LBPH OpenCV (Open Source Computer Vision Library) is a computer vision and machine learning library, that provides commonly used tools for computer vision software. It contains well-known face recognition methods, such as Eigenfaces method, Fisherfaces method, and Local Binary Patterns method. The Eigenfaces method [14] can be described with the following. Let 𝑋 = {𝑥1 , 𝑥2 , … , 𝑥𝑛 } be a random vector with observations 𝑥𝑖 ∈ 𝑅𝑑 . First step is to compute the mean value μ of these observations: 𝑛 1 𝜇 = ∑ 𝑥𝑖 . (8) 𝑛 𝑖=1 Then the covariance matrix S can be obtained: 339 𝑛 1 𝑆 = ∑(𝑥𝑖 − 𝜇)(𝑥𝑖 − 𝜇)𝑇 . (9) 𝑛 𝑖=1 With the values of covariance matrix S the eigenvalues λi and eigenvectors vi can be expressed with the next formula: 𝑆𝑣𝑖 = 𝜆𝑖 𝑣𝑖 , 𝑖 = 1,2, … , 𝑛. (10) After that it is needed to order the eigenvectors descending by their eigenvalue. The core of the Eigenfaces method is Principal Component Analysis (PCA) [15]. It computes a features linear combination that maximizes the total variance in data. The k principal components are the eigenvectors corresponding to the k largest eigenvalues. The k principal components of the observed vector x are: 𝑦 = 𝑊 𝑇 (𝑥 − 𝜇), (11) where 𝑊 = (𝑣1 , 𝑣2 , … , 𝑣𝑘 ). The reconstruction with the use of the PCA basis is provided by: 𝑥 = 𝑊𝑦+𝜇 , (12) where 𝑊 = (𝑣1 , 𝑣2 , … , 𝑣𝑘 ). Mathematical description of the Fisherfaces [16] method is provided with the following. Let X be a random vector with samples drawn from c classes: 𝑋 = {𝑋1 , 𝑋2 , … , 𝑋𝑐 }, (13) 𝑋𝑖 = {𝑥1 , 𝑥2 , … , 𝑥𝑛 }. (14) From this representation the scatter matrices SB and Sw can be calculated as: 𝑐 𝑆𝐵 = ∑ 𝑁𝑖 (𝜇𝑖 − 𝜇)(𝜇𝑖 − 𝜇)𝑇 , (15) 𝑖=1 𝑐 𝑆𝑊 = ∑ ∑ (𝑥𝑗 − 𝜇𝑖 )(𝑥𝑗 − 𝜇𝑖 )𝑇 , (16) 𝑖=1 𝑥𝑗∈𝑋𝑖 where μ is the total mean, that expresses as: 𝑁 1 𝜇 = ∑ 𝑥𝑖 . (17) 𝑁 𝑖=1 Mean value μi of class i ∈ {1, …, c} is: 1 𝜇𝑖 = ∑ (𝑥𝑗 ). (18) |𝑋𝑖 | 𝑥𝑗∈𝑋𝑖 Fisherfaces algorithm performs search of a projection W, that maximizes the class separability criterion: |𝑊 𝑇 𝑆𝐵 𝑊 | 𝑊𝑜𝑝𝑡 = 𝑎𝑟𝑔 𝑚𝑎𝑥𝑊 . (19) |𝑊 𝑇 𝑆𝑊 𝑊 | Following [17], this optimization problem can be solved by finding a solution of the general eigenvalue problem: 𝑆𝐵 𝑣𝑖 = 𝜆𝑖 𝑆𝑤 𝑣𝑖 , (20) 𝑆𝑤 −1 𝑆𝐵 𝑣𝑖 = 𝜆𝑖 𝑣𝑖 . (21) Eigenfaces and Fisherfaces methods based on a holistic approach to face recognition that involves more efficient concentrating on extracting image local features. The basic idea of Local Binary Patterns (LBP) method [18] is to summarize the local structure in an image by comparing each pixel with its neighbors. Description of the LBP operator expressed with the following: 𝑃−1 𝐿𝐵𝑃(𝑥𝑐 , 𝑦𝑐 ) = ∑ 2𝑝 𝑠(𝑖𝑝 − 𝑖𝑐 ), (22) 𝑝=0 340 where (xc, yc) is a central pixel with intensity ic; in is intensity of the neighbor pixel; s is the sign function defined as: 1 𝑖𝑓 𝑥 ≥ 0 𝑠 (𝑥 ) = { , (23) 0 𝑒𝑙𝑠𝑒 Position of the neighbor (xp, yp), p ∈ P for a pixel point (xc, yc) can be calculated by: 2𝜋𝑝 𝑥𝑝 = 𝑥𝑐 + 𝑅𝑐𝑜𝑠( ), (24) 𝑃 2𝜋𝑝 𝑦𝑝 = 𝑦𝑐 − 𝑅𝑠𝑖𝑛( ), (25) 𝑃 where R is the radius of the circle and P is the number of sample points. 2.4. Daubechies wavelets transform The technology proposed in this research based on wavelet transform, that is widely used in the image processing tasks. Wavelet transform provides processing of patterns hidden in the data performing data analysis in general as well as in the detail. Wavelets apply in image processing when it is necessary for the result of analysis to contain information about location of characteristic frequencies and scales. Daubechies wavelets [19] are the type of basic wavelets, that orthonormal basis defined as: 𝑗 𝜙𝑟,𝑗,𝑘 (𝑥 ) = 22 𝜙𝑟 (2𝑗 𝑥 − 𝑘), 𝑗, 𝑘 ∈ 𝑍, (26) where function {𝜙𝑟 (𝑥 − 𝑘)|𝑘 ∈ 𝑍}, j is the scaling index, k is the displacement index, and r is the filter index. To analyze these equations in more detail at a certain scale, it is necessary to define an orthonormal basis 𝜓𝑟 (𝑥) with similar properties of 𝜙𝑟 (𝑥) 𝜙 (𝑥 ) = √2 ∑ ℎ𝑘 𝜙(2𝑥 − 𝑘), (27) 𝑘 𝜓(𝑥 ) = √2 ∑ 𝑔𝑘 𝜙(2𝑥 − 𝑘), (28) 𝑘 where ∑𝑘|ℎ𝑘 |2 < ∞. Property of the scaled functions orthogonality allows to determine the coefficients: ∑ ℎ𝑘 𝑔𝑘+2𝑚 = 𝛿0𝑚 . (29) 𝑘 Wavelets are orthogonal to scalable functions, thus wavelet coefficients gk are depend on the scaling function coefficients hk: 𝑔𝑘 = (−1)𝑘 ℎ2𝑀−1−𝑘 . (30) Due to these properties, Daubechies wavelets provide good results of the image processing. 2.5. Reverse biorthogonal wavelets transform Linear phase of the biorthogonal wavelets [20] allows to use them in image processing as well. It is maintained by the filter coefficients symmetry. Comparing to the orthogonal wavelets, biorthogonal wavelets involve higher degree of freedom and contain a symmetrical compact support. Let’s denote through f ( x) and y ( x) bases double to f ( x) and y ( x). , expressing it with the formula: [〈𝜙(𝑥 )|𝜙̅(𝑥)〉] = 𝐼, (31) [〈𝜓(𝑥 )|𝜓̅(𝑥)〉] = 𝐼, (32) Basic scaling functions in biorthogonal wavelet basis are orthogonal to double wavelets and scaling functions. This can be expressed with the following: [〈𝜙(𝑥 )|𝜓̅(𝑥)〉] = 0. (33) 341 Reverse biorthogonal system uses wavelet and scaling functions separately for signal analysis and design both time and frequency domains. Problem of time and frequency resolution in reverse biorthogonal wavelet function solves with the property that it is always equal to the viewport range. 3. Experimental research and analysis First set of experiments was performed with the use of standard Python library face_recognition. The training of KNN classifier was performed on the set of the images of 80 people, who are wearing masks, 8 images for each person. Images are not stable in scaling, position of a face and contain different levels of lightning. The total number of training set – 640 images. The dataset for the identification experiment was prepared with other images of same people. Obtained results indicate accuracy rate of identification of 55%. In the second set of experiments FaceNet system was used. The dataset for the identification is the same. In this case system output provides an information of the predicted class of the input image and probability with which input image belongs to the class. FaceNet method performance on the set of masked face images indicated the identification accuracy rate of 72,5%. Third part of experiments was performed to analyze the efficiency of three popular facial recognition methods: eigenface, fisherfaces, and LBPH. The accuracy of identification by eigenface method on the set of masked face images is 12,5%. Accuracy rate of the fisherfaces method is 25%. LBPH method provided accuracy rate of 5% of correctly identified images. Results of performed experiments are presented in the Table 1 and Table 2. Table 1 Results of experiments performed with face_recognition and FaceNet methods on the dataset of masked images Face_recognition library FaceNet Correctly Incorrectly Correctly Incorrectly identified images identified images identified images identified images Total number 80 80 of images Number 44 36 58 22 Percentage 55% 45% 72,5% 27,5% Table 2 Results of experiments performed with Eigenfaces, Fisherfaces and LBPH methods on the dataset of masked images Eigenfaces Fisherfaces LBPH Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number 80 80 80 of images Number 10 70 20 60 4 76 Percentage 12,5% 87,5% 25% 75% 5% 95% As can be seen in a Figure 1, Eigenface, Fisherface and LBPH algorithms performed identification with a small part of a dataset, therefore percentage of incorrectly identified images is much greater than identification accuracy rate. On the other hand, face_recognition and FaceNet methods do not tend to have such an issue, but still the difference between correctly and incorrectly identified images is decreasing. 342 Figure 1: Results of identification with the use of face_recognition, FaceNet, Eigenfaces, Fisherfaces and LBPH methods on the dataset of masked images Results of the experimentation with the same methods on the dataset of unmasked face images are presented in Table 3 and Table 4. Table 3 Results of experiments performed with face_recognition and FaceNet methods on the dataset of unmasked images Face_recognition library FaceNet Correctly Incorrectly Correctly Incorrectly identified images identified images identified images identified images Total number 80 80 of images Number 78 2 79 1 Percentage 97,5% 2,5% 98,75% 1,25% Table 4 Results of experiments performed with Eigenfaces, Fisherfaces and LBPH methods on the dataset of unmasked images Eigenfaces Fisherfaces LBPH Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number 80 80 80 of images Number 60 20 74 6 48 32 Percentage 75% 25% 92,5% 7,5% 60% 40% Figure 2 depicts percentage of correctly and incorrectly identified images. Comparative diagram of the results between methods applied to the dataset of unmasked face images and to the dataset of masked face images is depicted on Figure 3. As can be seen, presence of the mask on the images considerably decrease identification accuracy rate from 26% to 67%. 343 Figure 2: Results of identification with the use of face_recognition, FaceNet, Eigenfaces, Fisherfaces and LBPH methods on the dataset of unmasked images Figure 3: Comparative diagram of the results between face_recognition, FaceNet, Eigenfaces, Fisherfaces and LBPH methods applied to the dataset of unmasked face images and to the dataset of masked face images During the analysis of experimentation results of the research [21] it was found that the system provided the highest accuracy of identification (92,5%) on unmasked face images with the use of Daubechies wavelet transform for image processing, standard deviation and variance methods for feature extraction stage and image classification by this vector by calculating distance with Euclidean, quadratic Euclidean and Canberra metrics [22]. Therefore, in this research it was decided to use the same methods for the dataset of face images with protective masks. Results of experiments performed with Daubechies wavelet transform, standard deviation and variance calculation, and Euclidean, quadratic Euclidean, Canberra metrics presented in Tables 5 and 6. Figure 4 depicts results of the Daubechies wavelet transform and standard deviation calculation on the dataset of masked images. Figure 5 depicts results of the Daubechies wavelet transform and variance calculation on the dataset of masked images. Analyzing the results of Daubechies wavelet transform, the highest identification accuracy rate result was obtained with the use of standard deviation calculation and Euclidean distance metric (77,5%). Accuracy rates for Daubechies wavelets and other methods are following: standard deviation calculation and Canberra metric (65%), standard deviation calculation and quadratic Euclidean metric (70%), standard deviation calculation and quadratic Euclidean metric (57,5%), variance calculation and Canberra metric (65%), variance calculation and quadratic Euclidean metric (57,5%). 344 Table 5 Results of experiments performed with Daubechies wavelet transform and standard deviation calculation methods on the dataset of masked images Standard deviation Euclidean distance Canberra distance Squared Euclidean distance Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number of 80 80 80 images Number 62 18 52 28 56 24 Percentage 77,5% 22,5% 65% 35% 70% 30% Table 6 Results of experiments performed with Daubechies wavelet transform and variance calculation methods on the dataset of masked images Variance Euclidean distance Canberra distance Squared Euclidean distance Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number of 80 80 80 images Number 46 34 52 46 34 52 Percentage 57,5% 42,5% 65% 57,5% 42,5% 65% Figure 4: Results of the Daubechies wavelet transform and standard deviation calculation on the dataset of masked images In the research [23] it was experimentally obtained the identification accuracy rate of 97,5% on face images without wearing mask. This result was provided by the method of image processing based on reverse biorthogonal wavelets, standard deviation calculation and variance methods of feature vector extraction and Bray- Curtis, Canberra, and Manhattan distance metrics of image feature vector classification. Experimentation results of the same methods on the face images with mask dataset are provided next. 345 Figure 5: Results of the Daubechies wavelet transform and variance calculation on the dataset of masked images Results of experiments performed with reverse biorthogonal wavelets, standard deviation and variance calculation and Bray-Curtis, Canberra, Manhattan distance metrics are presented in Table 7 and Table 8, accordingly. Table 7 Results of experiments performed with reverse biorthogonal wavelet transform and standard calculation methods on the dataset of masked images Standard deviation Bray-Curtis distance Canberra distance Manhattan distance Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number of 80 80 80 images Number 52 28 52 52 28 52 Percentage 60% 40% 60% 60% 40% 60% Reverse biorthogonal wavelets transform based experiments provided the highest identification accuracy rate of 65% with combination of the following methods: standard deviation calculation and Bray-Curtis distance metric, standard deviation calculation and Bray-Curtis distance metric, variance calculation and Manhattan distance metric. The results of other methods are next: variance calculation and Bray-Curtis metric – 52.5%, standard deviation calculation and Manhattan metric – 55%, variance calculation and Canberra distance metric – 55%. Figure 6 depicts results of the reverse biorthogonal wavelet transform and standard deviation. Analysis between standard deviation and variance calculation methods for feature vector extraction indicated that the use of the first method is more effective during the experiments on both wavelet transforms. 346 Table 8 Results of experiments performed with reverse biorthogonal wavelet transform and standard calculation methods on the dataset of masked images Variance Bray-Curtis distance Canberra distance Manhattan distance Correctly Incorrectly Correctly Incorrectly Correctly Incorrectly identified identified identified identified identified identified images images images images images images Total number of 80 80 80 images Number 42 38 44 36 44 36 Percentage 52,5% 47,5% 55% 45% 55% 45% Figure 6: Results of the reverse biorthogonal wavelet transform and standard deviation calculation on the dataset of masked images Figure 7 depicts results of the reverse biorthogonal wavelet transform and variance calculation. Figure 7: Results of the reverse biorthogonal wavelet transform and variance calculation on the dataset of masked images 347 4. Conclusion During the research experiments were performed on the dataset of faces covered with mask in a conditions of different image scaling, face position and level of lightning with the use of face_recognition, FaceNet, Eigenfaces, Fisherfaces and LBPH methods. Results of the experimentation research are the following: face_recognition method – 55% of correctly identified images, FaceNet method – 72,5% of correctly identified images, Eigenfaces method – 12,5%, Fisherfaces method – 25% of correctly identified images, LBPH method – 5% of correctly identified images. The obtained results indicate that commonly used face recognition methods identification rates are decreasing in a range from 26% to 67% in conditions of incomplete data, such as face covered with protective masks, and different image scaling, face position and level of lightning. Appliance of Daubechies wavelet transform method indicated the highest identification accuracy rate result with the use of standard deviation calculation and Euclidean distance metric - 77,5% of correctly identified images, that was obtained on the same dataset of faces covered with mask in a conditions of different image scaling, face position and level of lightning. Accuracy rates for Daubechies wavelets and other methods are following: standard deviation calculation and Canberra metric - 65% of correctly identified images, standard deviation calculation and quadratic Euclidean metric - 70% of correctly identified images, standard deviation calculation and quadratic Euclidean metric - 57,5% of correctly identified images, variance calculation and Canberra metric - 65% of correctly identified images, variance calculation and quadratic Euclidean metric - 57,5% of correctly identified images. Appliance of reverse biorthogonal wavelet transform method indicated the highest identification accuracy rate of 65% on the dataset of faces covered with mask in a conditions of different image scaling, face position and level of lightning. This result was obtained with combination of the following methods: standard deviation calculation and Bray-Curtis distance metric, standard deviation calculation and Bray-Curtis distance metric, variance calculation and Manhattan distance metric. Results of the other methods are the following: variance calculation and Bray-Curtis metric – 52.5% of correctly identified images, standard deviation calculation and Manhattan metric – 55% of correctly identified images, variance calculation and Canberra distance metric – 55% of correctly identified images. Summarizing the foregoing conclusions, mathematical methods for information technology of biometric identification, proposed in this research, can be applied for face recognition in conditions of masked face images. The highest accuracy rate (77,5% of correctly identified images) of the identification performed on the dataset of masked face images was obtained during the experiments based on Daubechies wavelet transform as image processing method, standard deviation calculation as feature extraction method and Euclidean distance metric as image vector classification method. 5. References [1] M. Ngan, P. Grother, K. Hanaoka. National Institute of Standards and Technology Interagency or Internal Report 8311 Natl. Inst. Stand. Technol. Interag. Intern. Rep. 8311, 58 pages (July 2020). doi: 10.6028/NIST.IR.8311. [2] W. Yan. Face-mask recognition has arrived—for better or worse. National Geographic, September 11, 2020. URL: https://www.nationalgeographic.com/science/2020/09/face-mask- recognition-has-arrived-for-coronavirus-better-or-worse-cvd/ [3] A. Ng. Facial recognition designed to detect around face masks is failing, study finds. CNET, August 26, 2020. URL: https://www.cnet.com/health/facial-recognition-designed-to-detect- around-face-masks-is-failing-study-finds/ [4] A. Hern. Face masks give facial recognition software an identity crisis. The Guardian. August 21, 2020. URL: https://www.theguardian.com/technology/2020/aug/21/face-masks-give-facial- recognition-software-identity-crisis 348 [5] D. King. High Quality Face Recognition with Deep Metric Learning. Dlib C++ Library, February 12, 2017. URL: http://blog.dlib.net/2017/02/high-quality-face-recognition-with- deep.html [6] K. He, X. Zhang, S. Ren and J. Sun, "Deep Residual Learning for Image Recognition," 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, 2016, pp. 770-778, doi: 10.1109/CVPR.2016.90. [7] O. M. Parkhi, A. Vedaldi and A. Zisserman. Deep Face Recognition. In Xianghua Xie, Mark W. Jones, and Gary K. L. Tam, editors, Proceedings of the British Machine Vision Conference (BMVC), pages 41.1-41.12. BMVA Press, September 2015. doi: 10.5244/C.29.41. [8] H. Ng and S. Winkler, "A data-driven approach to cleaning large face datasets," 2014 IEEE International Conference on Image Processing (ICIP), Paris, 2014, pp. 343-347, doi: 10.1109/ICIP.2014.7025068. [9] scikit-learn. Machine Learning in Python. URL: https://scikit- learn.org/stable/modules/neighbors.html#classification [10] F. Schroff, D. Kalenichenko and J. Philbin, "FaceNet: A unified embedding for face recognition and clustering," 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, 2015, pp. 815-823, doi: 10.1109/CVPR.2015.7298682. [11] K. Q. Weinberger, L. K. Saul. 2009. Distance Metric Learning for Large Margin Nearest Neighbor Classification. J. Mach. Learn. Res. 10 (12/1/2009), 207–244. [12] K. Zhang, Z. Zhang, Z. Li and Y. Qiao, "Joint Face Detection and Alignment Using Multitask Cascaded Convolutional Networks," in IEEE Signal Processing Letters, vol. 23, no. 10, pp. 1499-1503, Oct. 2016, doi: 10.1109/LSP.2016.2603342. [13] scikit-learn user guide. Release 0.23.2, August 4, 2020. URL: https://scikit- learn.org/stable/_downloads/scikit-learn-docs.pdf [14] M. Turk and A. Pentland, "Eigenfaces for Recognition," in Journal of Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, Jan. 1991, doi: 10.1162/jocn.1991.3.1.71. [15] R. Mejía-Campos, D. Néjer-Haro, S. Recalde-Avincho, P. Rosero-Montalvo and D. Peluffo- Ordóñez, "Face Detection and Classification Using Eigenfaces and Principal Component Analysis: Preliminary Results," 2017 International Conference on Information Systems and Computer Science (INCISCOS), Quito, 2017, pp. 309-315, doi: 10.1109/INCISCOS.2017.59. [16] P. N. Belhumeur, J. P. Hespanha and D. J. Kriegman, "Eigenfaces vs. Fisherfaces: recognition using class specific linear projection," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 711-720, July 1997, doi: 10.1109/34.598228. [17] P. N. Belhumeur, J. P. Hespanha and D. J. Kriegman, "Eigenfaces vs. Fisherfaces: recognition using class specific linear projection," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 7, pp. 711-720, July 1997, doi: 10.1109/34.598228. [18] T. Ahonen, A. Hadid, M. Pietikäine. Face Recognition with Local Binary Patterns. In: Pajdla T., Matas J. (eds) Computer Vision - ECCV 2004. ECCV, 2004. Lecture Notes in Computer Science, vol 3021. Springer, Berlin, Heidelberg. Doi: 10.1007/978-3-540-24670-1_36. [19] I. Daubechies. Ten lectures on wavelets. Society for Industrial and Applied Mathematics, USA, 1992. [20] C. Stolojescu, I. Railean, S. Moga and A. Isar, "Comparison of wavelet families with application to WiMAX traffic forecasting," 2010 12th International Conference on Optimization of Electrical and Electronic Equipment, Basov, 2010, pp. 932-937, doi: 10.1109/OPTIM.2010.5510403. [21] G. Dimitrov, O. Bychkov, P. Petrova, K. Merkulova, Y. Zhabska, E. Zaitseva, L. Adaryukova, P. Petrov, "Creation of Biometric System of Identification by Facial Image", Third International Colloquium on Intelligent Grid Metrology, SMAGRIMET 2020, Zagreb, Croatia, 20-23 October. [22] O. Bychkov, K. Merkulova and Y. Zhabska, "Information Technology of Person’s Identification by Photo Portrait," 2020 IEEE 15th International Conference on Advanced Trends in Radioelectronics, Telecommunications and Computer Engineering (TCSET), Lviv-Slavske, Ukraine, 2020, pp. 786-790, doi: 10.1109/TCSET49122.2020.235542. [23] O. Bychkov, K. Merkulova and Y. Zhabska, "Software Application for Biometrical Person’s Identification by Portrait Photograph Based on Wavelet Transform," 2019 IEEE International Conference on Advanced Trends in Information Theory (ATIT), Kyiv, Ukraine, 2019, pp. 253- 256, doi: 10.1109/ATIT49449.2019.9030462. 349