Development of Information Technology for Person Identification in Video Stream Oleksii Bychkov, Kateryna Merkulova, Yelyzaveta Zhabska and Andriy Shatyrko Taras Shevchenko National University of Kyiv, Volodymyrs’ka str. 64/13, Kyiv, 01601, Ukraine Abstract This paper presents the research of methods for development of information technology for person identification in video stream. For the research conduction following methods were selected: anisotropic diffusion as an image preprocessing method, Gabor wavelet transform as an image processing method, histogram of oriented gradients (HOG) and local binary patterns in 1-dimensional space (1DLBP) as the methods of feature vector extraction from the images, square Euclidean distance metric for vector classification. The purpose of the work is to analyze and test these methods in order to develop an algorithm that will form the basis of the information technology of person identification in video stream. Experimental research of the methods was performed using well-known databases, such as The Database of Faces, The FERET Database and SCface Database. The obtained results of experiments indicate that proposed information technology provides the highest identification accuracy rate of 97.5% on the images with low quality and resolution. That result means that the developed information technology can be applied for person identification in real-world conditions when it is necessary to identify person from video stream. Keywords 1 Biometric identification, face recognition, wavelet transform. 1. Introduction Face recognition technology has long been in use at law enforcement processes, state borders and on smartphones. Nowadays it becomes a part of public and private areas of living. Hundreds of municipalities all over the world have installed cameras equipped with face recognition technology, sometimes promising to send data to central command centers as part of the programs that improves crime investigation cases. The COVID-19 pandemic catalyzed the fast spread of such solutions. In China more than 100 cities were equipped with surveillance systems based on face recognition technology last year. Now municipal network of cameras, installed on the streets, monitors the movement of people in residential areas of the city, when entering the offices and shops, as well as when using vehicles. Presumably, information from cameras can be used by the city police and the data will also be transferred to a special operational center, where computer algorithms will verify the faces of citizens and check their status in the database of the Ministry of Health. During the pandemic in January 2020 it was started to use citywide video surveillance system in Moscow, using software developed and supplied by NtechLab. During the first few weeks of COVID-19 lockdown company reported that the system found 200 quarantine violators [1]. Face recognition is non-invasive biometric technique, so it is the area of interest for small surveillance systems, as well as for the national security purposes. Face recognition is one of the most II International Scientific Symposium «Intelligent Solutions» IntSol-2021, September 28–30, 2021, Kyiv-Uzhhorod, Ukraine EMAIL: bos.knu@gmail.com (O. Bychkov); kate.don11@gmail.com (K. Merkulova); y.zhabska@gmail.com (Y. Zhabska); shatyrko.a@gmail.com (A. Shatyrko) ORCID: 0000-0002-9378-9535 (O. Bychkov); 0000-0001-6347-5191 (K. Merkulova); 0000-0002-9917-3723 (Y. Zhabska); 0000-0002-5648-2999 (A. Shatyrko) ©️ 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings (CEUR-WS.org) 70 important law enforcement techniques when video or crime scene images are available. Automatic face recognition technologies have improved the efficiency of the judiciary and simplified the comparison process. Modern face recognition techniques have achieved impressive results with face images of medium and high quality, but the performance is not satisfactory with low quality images. The main difficulty in recognizing low-resolution face images is the lack of facial details that distinguish it from the background. Another problem is that modern methods of face recognition are based on convolutional neural networks and use convolutional maps of features with a low sampling rate and a large step to represent a face. It leads to the information lose and inaccurate description of low-quality images. In March 2021 analytical the company Mordor Intelligence presented the report [2], where the global face recognition market was valued at USD 3.72 billion in 2020 and will surpass to USD 11.62 billion in 2026. Another forecasting company Grand View Research published the report on face recognition market in May 2021. According to this report surveillance and security segments are expected to significantly increase a compound annual growth rate (CAGR) through all analyzed period. The rate will grow till face recognition technology had become steadily used high-security areas. For example, security and surveillance systems are used by the law enforcement authorities with the purpose of criminals’ identification or search for missing persons [3]. The purpose of this work is to develop an information technology of person identification in video stream based on the algorithm that provides high identification results being applied on images with low quality and resolution. 2. Task solution methods It is well-known that information technology of face recognition and identification can be claimed efficient and reliable only when it thoroughly tested and validated, preferably on real data sets. Thus, it is necessary to research the methods, which will be used to develop the algorithm underlying in information technology under development, in order to define their efficiency on the datasets of low- quality images. In the context of this paper low-quality images define as small-sized, blurry, indistinct, pixelated, noised images, and high-quality images vice versa define as more distinct, low-compressed and noise-reduced. During the research it was decided to use anisotropic diffusion as an image preprocessing method, Gabor wavelet transform for image processing, histogram of oriented gradients (HOG) and local binary patterns in 1-dimensional space (1DLBP). 2.1. Anisotropic diffusion The idea of using anisotropic diffusion as a preprocessing technique before wavelet transform have been proposed in [4]. Authors of this paper performed experiments on high-quality images. To develop an information technology for person identification in video stream, it was decided to apply this method to the images of low quality and resolution. In image processing the concept of anisotropic diffusion is a process of successive diffusion based on more and more blurring of images in a scale space. Firstly, it was presented by Perona and Malik in [6]. As a preprocessing step it makes thinning and linking of the edges unnecessary, because the resulting images preserve the lines, edges and other important properties also leading to smoothing This method is based on the approximation of a parameterized version of the above defined process obtained by applying anisotropic diffusion equation, results in the space variant filters being anisotropic, close to lines and edges. The attributes of anisotropic diffusion are the anisotropic smoothing and iterative diffusion for the processing of each image pixel [4]. Anisotropic diffusion equation can be expressed as following: (1) 𝐼𝑡 = 𝑑𝑖𝑣(𝑐 (𝑥, 𝑦, 𝑡)𝛻𝐼) = 𝑐 (𝑥, 𝑦, 𝑡)𝛥𝐼 + 𝛻𝑐 ∙ 𝛻𝐼, where div is the divergence operator, ∇ is the gradient operator, Δ is the Laplacian operator, I0(x, y) is the input image, t is the Gaussian kernel variance G(x, y, t0) and I(x, y, t) is a family of derived images 71 obtained by convolving the original image I0(x, y) with a Gaussian kernel G(x, y, t0). Anisotropic diffusion helps successfully remove noise and preserve image edges and small structures if the diffusion coefficient or edge stopping function c(∇I) is estimated correctly. If c(x, y, t) is a constant the equation can be reduced to the equation of isotropic heat diffusion 𝐼𝑡 = 𝑐𝛥𝐼 [6, 7]. 2.2. Gabor wavelets transform The usage of the combination of anisotropic diffusion and Gabor wavelet transform was described in [4]. Although authors concluded that combination of these methods provides high recognition rates, however it wasn’t tested on low-quality images. During this research, it was decided to explore if these methods are applicable to development of information technology of person identification in video stream. Gabor wavelet transform is widely used in pattern recognition field because of its biological significance and technical properties. The complex Gabor function in the spatial domain is defined by the following: (2) 𝑔(𝑥, 𝑦) = 𝑠(𝑥, 𝑦)𝑤𝑟 (𝑥, 𝑦). s ( x, y ) – is a complex sinewave (carrier), defined as: (3) 𝑠(𝑥, 𝑦) = exp(j(2π(𝑢0 𝑥 + 𝑣0 𝑦) + 𝑃 )), where (u0 , v0 ) and P determine the spatial frequency and sinewave phase respectively; wr ( x, y ) – Gaussian 2D-function (envelope function). The Gaussian envelope is represented as the following: 2 (4) 𝑤𝑟 (𝑥, 𝑦) = K ∙ exp(−𝜋(𝛼 2 (𝑥 − 𝑥0 )𝑟 + 𝑏2 (𝑦−𝑦0 )𝑟 2 , where ( x0 , y0 ) is the peak of the function, a and b are the Gaussian scaling parameters, and the index r denotes the rotation operation, which can be described this way: (5) (𝑥 − 𝑥0 )𝑟 = (𝑥 − 𝑥0 ) cos 𝜃 + (𝑦 − 𝑦0 ) sin 𝜃, (6) (𝑦 − 𝑦0 )𝑟 = −(𝑥 − 𝑥0 ) sin 𝜃 +(𝑦 − 𝑦0 ). Therefore, the complex Gabor function in the spatial domain can be written as follows [8]: 2 𝑔(𝑥, 𝑦) = K ∙ exp(−𝜋(𝛼 2 (𝑥 − 𝑥0 )𝑟 + 𝑏2 (𝑦−𝑦0 )𝑟 2 ))exp⁡(𝑗(2𝜋 (𝑢0 𝑥 + 𝑣0 𝑦) + 𝑃 ). (7) The function is determined by the following parameters: K - scales the value of the Gaussian envelope; (a, b) - scales two axes of the Gaussian envelope; ( x0 , y0 ) - peak coordinates of the Gaussian envelope; (u0 , v0 ) - spatial frequencies of the sinusoidal carrier in Cartesian coordinates; P is the phase of the sinusoidal carrier [9]. 2.3. Histogram of Oriented Gradients (HOG) In [5] it was presented the algorithm that uses Haar wavelet decomposition as a processing technique together with BSIF (binarized statistical image features) and HOG (histogram of oriented gradients) as methods of feature extraction. Method proposed in this work solves the problem of palmprint features extraction. Based on this research it was decided to apply similar algorithm with other methods for the task of face recognition. The Histogram of Oriented Gradients (HOG) method can be applied to images that were processed by wavelet transform to extract important features of image shape. Histogram of oriented gradients calculates with the execution of the following steps. Firstly, before building of the orientation histogram for each cell, there must be obtained the value of the gradient. After that, obtained histograms that grouped by the individual cell are normalized. This process can be described mathematically: 72 1 𝐷𝑥 = [−1 0 1], 𝐷𝑦 = [ 0 ]. (8) −1 Dx is a 1-dimensional horizontal discrete derivative mask, Dy is a 1-dimensional vertical discrete derivative mask. The gradient histogram retrieves with the equations: 𝐼 (𝑥) = 𝐼 ∙ 𝐷𝑥 , 𝐼 (𝑦) = 𝑖 ∙ 𝐷𝑦 . (9) Gradient size is calculated as: |𝐺 | = √𝐼2 (𝑥) + 𝐼2 (𝑦). (10) Orientation is calculated with the following equation: 𝐼𝑦 𝜃 = tan−1 . (11) 𝐼𝑥 Then orientation distributes equally in a range of 0 to 180 deg (for an unsigned gradient) or 0 to 360 deg (for a signed gradient). Each pixel computes a channel weighted vote based on the gradient values. After that cells groups into blocks that are connected in space. It allows to obtain the vector of the normalized histogram elements. 2.4. Local Binary Patterns (LBP) in 1-Dimensional Space For the palmprint recognition task authors of [5] used BSIF method (binarized statistical image features) as feature extraction technique. BSIF is an LBP-based algorithm. For the task of face recognition there are more efficient algorithms based on LBP like 1DLBP and I1DLBP. Local Binary Patterns in 1-Dimensional Space (1DLBP) firstly was presented in [10] and tested on high-quality images. In this work 1DLBP will be explored with the appliance on low-quality images. The purpose of the 1DLBP method is to describe the local agitation of 1-dimensional signal segment in binary code. That description can be obtained by comparing of the neighbor pixel values with the central pixel value. All neighbor elements get the value of 1 if they are greater or equal to the current element and value of 0 if they are less than the current element. Then, each element of the obtained vector is multiplied by a weight according to its position. Finally, the current element is replaced by the sum of the values of obtained vector. Described process can be expressed as follows: 𝑁−1 1𝐷𝐿𝐵𝑃 = ∑ 𝑆(𝑔𝑛 − 𝑔0 ) ∙ 2𝑛 . (12) 𝑛=0 S(x) function defines as S(x)={1 if x⩾0;⁡0⁡otherwise}; g0 and gn are the values of the central element and its 1-dimensional neighbors, respectively. The index n changes its value increasingly from the left to the right in the 1-dimensional string. The histogram of the 1-dimensional pattern defines the 1DLBP descriptor. The idea of the I1DLBP is similar to the idea of the 1DLBP - the local patterns are extracted by thresholding the linear neighbors of each pixel, from the projected image, with the mean value of the summed neighbors: 𝑁−1 𝑛 𝐼1𝐷𝐿𝐵𝑃 = ∑ 𝑆(𝑔𝑛 − 𝑔́ ) ∙ 2 , (13) 𝑛=0 where S(x) function defines as S(x)={1 if x⩾0;⁡0⁡otherwise}; g’ and gn are the mean value of the linear neighborhood and the values of its 1-dimensional neighbors, respectively. The 1DLBP algorithm applies on all blocks of the split image with different resolution. The extracted histograms are concatenated in one global 1-dimensional vector that represents one face image [11]. Summarizing the foregoing description of 1DLBP method, the following stages of image processing can be distinguished to obtain 1DLBP feature vector:  Input image preprocesses. 73  Image splits into multiple blocks.  Each decomposed block projects in 1-dimensional space.  1DLBP algorithm processes each projected block.  The vectors that obtained after each block processing concatenate in one global feature vector [12]. 3. Experimental research and analysis Experimental research of the selected methods was performed using software created with Python. The research was conducted on three different face databases: The Database of Faces, Facial Recognition Technology (FERET) database and Surveillance Cameras Face Database (SCface). The purpose of experimental research is to establish the most efficient combination of methods by comparing results of their work with the use of one feature extraction method (HOG or 1DLBP) and the use of combination of feature extraction methods (HOG and 1DLBP). For the classification of the feature vector, obtained as a result, square Euclidean distance metric was used in all sets of experiments. The examples of images obtained as a result of applying described methods are depicted on Figure 1. Figure 1: Examples of images: a) original image; b) image processed with anisotropic diffusion; c) image processed with Gabor wavelet transform; d) histogram of gradients of given image. 3.1. The Database of Faces First set of experiments was performed using The Database of Faces that was created by AT&T Laboratories Cambridge during the work on face recognition project. Database presented with a set of face images of 40 different individuals. It is organized in 40 directories with 10 PGM format images in each directory. The size of each image is 92x112 pixels, with 256 grey levels per pixel. Images were taken with the person in frontal position with lighting, facial expression and facial details varying. The database was retrieved openly from official site of AT&T Laboratories Cambridge [13]. Results of experiments performed on The Database of Faces are presented in the Table 1. Table 1 Results of experiments with The Database of Faces AD + Gabor + HOG AD + Gabor + 1DLBP AD + Gabor + HOG + 1DLBP Identification False Identification False Identification False accuracy recognition accuracy recognition accuracy recognition rate rate rate rate rate rate Total number 40 40 40 of images Number 26 14 17 23 28 12 Identification 65% 35% 42.5% 57.5% 70% 30% accuracy rate As can be concluded from the Figure 2, when only one feature extraction method is used, HOG or 1DLBP separately, the percentage of correctly identified images varying from 42.5% to 65%. And the highest rate of identification accuracy of 70% was obtained during the usage of combination of feature 74 vector extraction methods. It confirms the reasonability of the usage of two feature extraction methods through one iteration of the algorithm work. Therefore, false recognition rate varies from 30 to 57.5%. 3.2. Facial Recognition Technology (FERET) database The FERET Database is a part of the Face Recognition Technology program [14] that researches automatic face recognition capabilities to develop new algorithms for the automatic recognition of human faces. As Technical Agent for distribution of the database serves The National Institute of Standards and Technology (NIST). The FERET database contains 14126 high-resolution images of 1199 individuals with the resolution of 256x384. To obtain database it was necessary to request an account for downloading the FERET database. Corresponding request was sent according to the instruction given by the owner [15]. As far as first set of experiments was conducted on the database that contains images of total 40 people, it was decided to use images of 40 people as well to conduct the experiments on the FERET database. Figure 2: Results of identification with the use of The Database of Faces Experimental results for The FERET Database are presented in Table 2. Table 2 Results of experiments performed with The FERET Database AD + Gabor + HOG AD + Gabor + 1DLBP AD + Gabor + HOG + 1DLBP Identification False Identification False Identification False accuracy recognition accuracy recognition accuracy recognition rate rate rate rate rate rate Total number 40 40 40 of images Number 29 11 26 14 29 11 Identification 72.5% 27.5% 65% 35% 72.5% 27.5% accuracy rate As can be seen from Fig. 3, on high resolution images the usage of 1DLBP method is not necessary, because the identification accuracy rate is the same during the experiment performance with HOG feature vector extraction method only, as well as with combination of HOG and 1DLBP methods, which is 72.5%. For the FERET database the usage of two feature extraction methods is not justified. As a conclusion, false recognition rate obtained during the experiments with The FERET Database varies from 27.5 to 35.5%. 75 Figure 3: Results of identification with the use of The FERET Database 3.3. SCface database SCface database was created by researchers from University of Zagreb [16] for testing face recognition algorithms in conditions of the real world. Images of the database was taken on surveillance cameras of varying quality and resolution, therefore images of this database are vary-resolution. Also, the database contains frontal mug shot images for the scenario when a person needed to be recognized comparing mug shot image to a low-quality video surveillance image. Database contains 4160 static images of 130 individuals. More technical details can be found in [17]. For the purpose of this research, low quality video surveillance images were used to test the selected methods and explore the small sample size problem. As far as first set of experiments was conducted on the database that contains images of total 40 people, it was decided to use images of 40 people as well to conduct the experiments on the SCface database. Obtained results of the experiments performed on SCface database are presented in Table 3. Table 3 Results of experiments performed with the SCface Database AD + Gabor + HOG AD + Gabor + 1DLBP AD + Gabor + HOG + 1DLBP Identification False Identification False Identification False accuracy recognition accuracy recognition accuracy recognition rate rate rate rate rate rate Total number 40 40 40 of images Number 37 3 31 9 39 1 Identification 92.5% 7.5% 77.5% 22.5% 97.5% 2.5% accuracy rate From Fig. 4 it can be concluded that using 1DLBP method for feature extraction process by itself provides only 77.5% of correctly identified face images. Usage of HOG method improves algorithm work results within the order of 15% and provides the identification accuracy rate of 92.5%. But fusing of HOG and 1DLBP feature vectors escalate the correctness of algorithm to 97.5%. Results of this set of experiments confirms that the usage of combination of two feature vector extraction methods is justified on low quality and resolution image, and the false recognition rate in this case is the lowest among all the experiments – 2.5%. 76 3.4. Analysis of obtained results Comparative diagram on Fig. 5 presents the results of identification with the use of selected methods on different databases. Taking into consideration the fact that facial images from SCface database, that were used for experimentation performance, are low-quality video surveillance images, it can be concluded that combination of all researched methods provides high identification accuracy rate on low-resolution facial images. Figure 4: Results of identification with the use of SCface Database After the analysis of all experimentational results it can be concluded that the combination of such methods as anisotropic diffusion, Gabor wavelet transform, histogram of oriented gradients and local binary patterns in 1-dimensional space demonstrates the highest identification results being applied to all researched databases. Figure 5: Comparative diagram of the experimentational results of the researched methods appliance to different databases 77 4. Development of image processing algorithm From the perspective of analysis of obtained research results, combination of methods with the highest identification accuracy rate can be used to develop the algorithm underlying in information technology for person identification in video stream. In general, the person identification process consists of two stages. The first is the determination of face location in the image taken from video stream. The original image is scanned with a smaller window, and every time there is a determination of the similarity degree between the image in the window with a human face. Formally, the face image can be structurally defined, statistically or by sample face images list [18]. After the window, which most likely contains just a person's face, has been identified, the second stage starts- identification. The purpose of person identification is to define a unique identifier for each input biometric parameter or to determine it as unknown if such sample is not in the database [19]. The algorithm for information technology of person identification must be applicable to facial images in aim to extract their feature vectors for further classification in order to identify a person. Data flow diagram of a proposed algorithm is presented on Fig. 6. Figure 6: Data flow diagram of the algorithm 78 The algorithm uses a portrait image of a person as input data. Step-by-step algorithm execution can be described as follows: 1. Face detection and localization on the input image. Only face image of a person is used for further processing because it is the area of interest for person identification process. 2. Image pre-processing using anisotropic diffusion that allows to preserve and enhance edge information and remove noise. 3. The image of the face, after applying anisotropic diffusion to it, is processed by Gabor wavelets with various changes in the parameters of the wavelet function so that 16 wavelet-transformed variations of the face image can be obtained. After that, the wavelet-transformed images are summed to form a global image, which is processed by the following methods. 4. The global image formed as a result of the Gabor wavelet transform is simultaneously fed to the input of two independent methods - HOG and 1DLBP. As a result of the operation of each of these methods, two separate 512-value feature vectors are formed. 5. HOG feature vector and 1DLBP feature vector normalization. Because of the deviation in vector distributions and ranges, feature vectors extracted separately from 1DLBP and the HOG features are incompatible. There are some methods of vector normalization, that can improve compatibility. For example, min–max normalization that transforms the feature vectors in the range [0, 1]. If X = [x1; x2; x3; ...; xn] is the feature vector, the normalized feature vector can be represented using min–max normalization [5]: 𝑥𝑖 − 𝑀𝑖𝑛(𝑋) 𝑥′ = . (14) 𝑀𝑎𝑥 (𝑋) − 𝑀𝑖𝑛(𝑋) 6. The HOG feature vector and 1DLBP feature vector concatenates to form a 1024-value global feature vector of a face image. The global feature vector of the image can be obtained by concatenating into a single feature vector the normalized feature vectors of 1DLBP and HOG features. Let the normalized feature vectors be D = [d1; d2; d3 ; ...; dn] for 1DLBP and h = [h1; h2; h3; ...; hn] for HOG extraction. The global vector can be represented as [5]: 𝑉𝑔 = [𝑑1 , 𝑑2 , 𝑑3 , … , 𝑑𝑛 , ℎ1 , ℎ2 , ℎ3 , … , ℎ𝑛 ]. (15) The obtained global feature vector is using for further classification. The algorithm for information technology of person identification in video stream that includes such methods as anisotropic diffusion, Gabor wavelet transform, histogram of oriented gradients and local binary patterns in 1-dimensional space proposed in this paper firstly. 5. Conclusion This paper describes the research of methods to develop the algorithm underlying in information technology of person identification in video stream. During the research, experiments were performed on the testing of methods being applicable in face recognition in order to establish the most efficient combination of methods by comparing results of their work with the use of one feature extraction method (HOG or 1DLBP) and the use of combination of feature extraction methods (HOG and 1DLBP). The obtained results indicated that the highest identification accuracy rate of 97.5% was obtained with both feature extraction methods (HOG and 1DLBP) on the images from SCface database, which are low-quality video surveillance images. Experimental results on the Database of Faces and FERET database provides from 70 to 72.5% of correctly identified images. Proposed algorithm is based on anisotropic diffusion as an image preprocessing method, Gabor wavelet transform as an image processing method, histogram of oriented gradients (HOG) and local binary patterns in 1-dimensional space (1DLBP) as the methods of feature vector extraction from the images, square Euclidean distance metric for vector classification. Summarizing the foregoing conclusions, proposed in this work algorithm for information technology of person identification based on anisotropic diffusion, Gabor wavelet transform, histogram of oriented gradients (HOG) and local binary patterns in 1-dimensional space (1DLBP) can be applied for face recognition and person identification in conditions of face images with low quality and resolution, that means it partially solves small sample size problem. For future research it is aimed to improve the 79 proposed algorithm for the scenario when low quality video surveillance images compare with frontal mug shot image from law enforcement or national security databases. 6. References [1] B. Thompson and R. Van Noorden. The troubling rise of facial recognition technology. Nature. November 18, 2020. URL: https://www.nature.com/articles/d41586-020-03271-8 [2] Facial recognition market - growth, trends, Covid-19 impact, and forecasts (2021 - 2026). Mordor Intelligence. March 2021. URL: https://www.mordorintelligence.com/industry-reports/facial- recognition-market [3] Facial Recognition Market Size, Share & Trends Analysis Report, 2021 – 2028. Grand View Research. May 2021. 92 p. Report ID: 978-1-68038-311-9. URL: https://www.grandviewresearch.com/industry-analysis/facial-recognition-market/segmentation [4] T.M. Abhishree, J. Latha, K. Manikantan, S. Ramachandran. Face Recognition Using Gabor Filter Based Feature Extraction with Anisotropic Diffusion as a Pre-processing Technique. Procedia Computer Science, Volume 45, 2015, pp. 312-321, ISSN 1877-0509, doi: 10.1016/j.procs.2015.03.149. [5] B. Attallah, A. Serir, Y. Chahir, A. Boudjelal. Histogram of gradient and binarized statistical image features of wavelet subband-based palmprint features extraction. J. Electron. Imag. 26(6) 063006, November 8, 2017, doi: 10.1117/1.JEI.26.6.063006. [6] P. Perona and J. Malik. Scale-Space and Edge Detection Using Anisotropic Diffusion. IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 12, No. 7, pp. 629-639. July 1990. [7] C. Yu, and Y. Jia. Anisotropic Diffusion-based Kernel Matrix Model for Face Liveness Detection. Image Vis. Comput., 89, 2019, pp. 88-94, doi: 10.1016/J.IMAVIS.2019.06.009 [8] J.R. Movellan. Tutorial on Gabor, 2002. URL: https://inc.ucsd.edu/mplab/tutorials/gabor.pdf [9] G. P. Dimitrov et al. Creation of Biometric System of Identification by Facial Image. 2020 3rd International Colloquium on Intelligent Grid Metrology (SMAGRIMET), 2020, pp. 29-34, doi: 10.23919/SMAGRIMET48809.2020.9263995. [10] A. Benzaoui and A. Boukrouche. Face Recognition Using 1DLBP Texture Analysis. Future Computing 2013: The Fifth International Conference on Future Computational Technologies and Applications, 2013, pp. 14-19, ISBN: 978-1-61208-272-1. [11] A. Benzaoui and A. Boukrouche. 1DLBP and PCA for face recognition. 2013 11th International Symposium on Programming and Systems (ISPS), 2013, pp. 7-11, doi: 10.1109/ISPS.2013.6581486. [12] A. Benzaoui and A. Boukrouche. Face Analysis, Description and Recognition using Improved Local Binary Patterns in One Dimensional Space. Journal of Control Engineering and Applied Informatics, Vol. 16, No. 4, pp. 52-60, 2014. [13] The Database of Faces. URL: https://cam-orl.co.uk/facedatabase.html [14] Face Recognition Technology (FERET). URL: https://www.nist.gov/programs-projects/face- recognition-technology-feret [15] Color FERET database. URL: https://www.nist.gov/itl/products-and-services/color-feret-database [16] SCface - Surveillance Cameras Face Database. URL: https://www.scface.org [17] M. Grgic, K. Delac and S. Grgic. SCface - surveillance cameras face database. Multimedia Tools and Applications Journal, Vol. 51, No. 3, February 2011, pp. 863-879. [18] O. Bychkov, K. Merkulova and Y. Zhabska, "Information Technology of Person’s Identification by Photo Portrait," 2020 IEEE 15th International Conference on Advanced Trends in Radioelectronics, Telecommunications and Computer Engineering (TCSET), 2020, pp. 786-790, doi: 10.1109/TCSET49122.2020.235542. [19] O. Bychkov, O. Ivanchenko, K. Merkulova and Y. Zhabska. Mathematical Methods for Information Technology of Biometric Identification in Conditions of Incomplete Data. Proceedings of the 7th International Conference "Information Technology and Interactions" (IT&I-2020), Kyiv, Ukraine, December 02-03, 2020, pp. 336-349. 80