=Paper= {{Paper |id=Vol-3018/Paper_7 |storemode=property |title=Development of Information Technology for Person Identification in Video Stream |pdfUrl=https://ceur-ws.org/Vol-3018/Paper_7.pdf |volume=Vol-3018 |authors=Oleksii Bychkov,Kateryna Merkulova,Yelyzaveta Zhabska,Andriy Shatyrko |dblpUrl=https://dblp.org/rec/conf/intsol/BychkovMZS21 }} ==Development of Information Technology for Person Identification in Video Stream == https://ceur-ws.org/Vol-3018/Paper_7.pdf
Development of Information Technology for Person Identification in
Video Stream
Oleksii Bychkov, Kateryna Merkulova, Yelyzaveta Zhabska and Andriy Shatyrko
Taras Shevchenko National University of Kyiv, Volodymyrs’ka str. 64/13, Kyiv, 01601, Ukraine


                Abstract
                This paper presents the research of methods for development of information technology for
                person identification in video stream. For the research conduction following methods were
                selected: anisotropic diffusion as an image preprocessing method, Gabor wavelet transform as
                an image processing method, histogram of oriented gradients (HOG) and local binary patterns
                in 1-dimensional space (1DLBP) as the methods of feature vector extraction from the images,
                square Euclidean distance metric for vector classification. The purpose of the work is to
                analyze and test these methods in order to develop an algorithm that will form the basis of the
                information technology of person identification in video stream.
                Experimental research of the methods was performed using well-known databases, such as The
                Database of Faces, The FERET Database and SCface Database.
                The obtained results of experiments indicate that proposed information technology provides
                the highest identification accuracy rate of 97.5% on the images with low quality and resolution.
                That result means that the developed information technology can be applied for person
                identification in real-world conditions when it is necessary to identify person from video
                stream.

                Keywords 1
                Biometric identification, face recognition, wavelet transform.

1. Introduction
   Face recognition technology has long been in use at law enforcement processes, state borders and
on smartphones. Nowadays it becomes a part of public and private areas of living. Hundreds of
municipalities all over the world have installed cameras equipped with face recognition technology,
sometimes promising to send data to central command centers as part of the programs that improves
crime investigation cases. The COVID-19 pandemic catalyzed the fast spread of such solutions. In
China more than 100 cities were equipped with surveillance systems based on face recognition
technology last year. Now municipal network of cameras, installed on the streets, monitors the
movement of people in residential areas of the city, when entering the offices and shops, as well as
when using vehicles. Presumably, information from cameras can be used by the city police and the data
will also be transferred to a special operational center, where computer algorithms will verify the faces
of citizens and check their status in the database of the Ministry of Health. During the pandemic in
January 2020 it was started to use citywide video surveillance system in Moscow, using software
developed and supplied by NtechLab. During the first few weeks of COVID-19 lockdown company
reported that the system found 200 quarantine violators [1].
   Face recognition is non-invasive biometric technique, so it is the area of interest for small
surveillance systems, as well as for the national security purposes. Face recognition is one of the most


II International Scientific Symposium «Intelligent Solutions» IntSol-2021, September 28–30, 2021, Kyiv-Uzhhorod, Ukraine
EMAIL: bos.knu@gmail.com (O. Bychkov); kate.don11@gmail.com (K. Merkulova); y.zhabska@gmail.com (Y. Zhabska);
shatyrko.a@gmail.com (A. Shatyrko)
ORCID: 0000-0002-9378-9535 (O. Bychkov); 0000-0001-6347-5191 (K. Merkulova); 0000-0002-9917-3723 (Y. Zhabska);
0000-0002-5648-2999 (A. Shatyrko)
           ©️ 2021 Copyright for this paper by its authors.
           Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
           CEUR Workshop Proceedings (CEUR-WS.org)



                                                                                                                     70
important law enforcement techniques when video or crime scene images are available. Automatic face
recognition technologies have improved the efficiency of the judiciary and simplified the comparison
process.
    Modern face recognition techniques have achieved impressive results with face images of medium
and high quality, but the performance is not satisfactory with low quality images. The main difficulty
in recognizing low-resolution face images is the lack of facial details that distinguish it from the
background. Another problem is that modern methods of face recognition are based on convolutional
neural networks and use convolutional maps of features with a low sampling rate and a large step to
represent a face. It leads to the information lose and inaccurate description of low-quality images.
    In March 2021 analytical the company Mordor Intelligence presented the report [2], where the global
face recognition market was valued at USD 3.72 billion in 2020 and will surpass to USD 11.62 billion
in 2026.
    Another forecasting company Grand View Research published the report on face recognition market
in May 2021. According to this report surveillance and security segments are expected to significantly
increase a compound annual growth rate (CAGR) through all analyzed period. The rate will grow till
face recognition technology had become steadily used high-security areas. For example, security and
surveillance systems are used by the law enforcement authorities with the purpose of criminals’
identification or search for missing persons [3].
    The purpose of this work is to develop an information technology of person identification in video
stream based on the algorithm that provides high identification results being applied on images with
low quality and resolution.

2. Task solution methods
    It is well-known that information technology of face recognition and identification can be claimed
efficient and reliable only when it thoroughly tested and validated, preferably on real data sets. Thus, it
is necessary to research the methods, which will be used to develop the algorithm underlying in
information technology under development, in order to define their efficiency on the datasets of low-
quality images. In the context of this paper low-quality images define as small-sized, blurry, indistinct,
pixelated, noised images, and high-quality images vice versa define as more distinct, low-compressed
and noise-reduced.
    During the research it was decided to use anisotropic diffusion as an image preprocessing method,
Gabor wavelet transform for image processing, histogram of oriented gradients (HOG) and local binary
patterns in 1-dimensional space (1DLBP).

2.1.    Anisotropic diffusion
    The idea of using anisotropic diffusion as a preprocessing technique before wavelet transform have
been proposed in [4]. Authors of this paper performed experiments on high-quality images. To develop
an information technology for person identification in video stream, it was decided to apply this method
to the images of low quality and resolution.
    In image processing the concept of anisotropic diffusion is a process of successive diffusion based
on more and more blurring of images in a scale space. Firstly, it was presented by Perona and Malik in
[6]. As a preprocessing step it makes thinning and linking of the edges unnecessary, because the
resulting images preserve the lines, edges and other important properties also leading to smoothing
    This method is based on the approximation of a parameterized version of the above defined process
obtained by applying anisotropic diffusion equation, results in the space variant filters being anisotropic,
close to lines and edges. The attributes of anisotropic diffusion are the anisotropic smoothing and
iterative diffusion for the processing of each image pixel [4].
    Anisotropic diffusion equation can be expressed as following:
                                                                                                       (1)
                          𝐼𝑡 = 𝑑𝑖𝑣(𝑐 (𝑥, 𝑦, 𝑡)𝛻𝐼) = 𝑐 (𝑥, 𝑦, 𝑡)𝛥𝐼 + 𝛻𝑐 ∙ 𝛻𝐼,
where div is the divergence operator, ∇ is the gradient operator, Δ is the Laplacian operator, I0(x, y) is
the input image, t is the Gaussian kernel variance G(x, y, t0) and I(x, y, t) is a family of derived images

                                                                                                         71
obtained by convolving the original image I0(x, y) with a Gaussian kernel G(x, y, t0). Anisotropic
diffusion helps successfully remove noise and preserve image edges and small structures if the diffusion
coefficient or edge stopping function c(∇I) is estimated correctly. If c(x, y, t) is a constant the equation
can be reduced to the equation of isotropic heat diffusion 𝐼𝑡 = 𝑐𝛥𝐼 [6, 7].

2.2.    Gabor wavelets transform
    The usage of the combination of anisotropic diffusion and Gabor wavelet transform was described
in [4]. Although authors concluded that combination of these methods provides high recognition rates,
however it wasn’t tested on low-quality images. During this research, it was decided to explore if these
methods are applicable to development of information technology of person identification in video
stream.
    Gabor wavelet transform is widely used in pattern recognition field because of its biological
significance and technical properties. The complex Gabor function in the spatial domain is defined by
the following:
                                                                                                   (2)
                                  𝑔(𝑥, 𝑦) = 𝑠(𝑥, 𝑦)𝑤𝑟 (𝑥, 𝑦).
   s ( x, y ) – is a complex sinewave (carrier), defined as:
                                                                                                      (3)
                            𝑠(𝑥, 𝑦) = exp(j(2π(𝑢0 𝑥 + 𝑣0 𝑦) + 𝑃 )),
where (u0 , v0 ) and P determine the spatial frequency and sinewave phase respectively; wr ( x, y ) –
Gaussian 2D-function (envelope function). The Gaussian envelope is represented as the following:
                                                         2                                      (4)
                    𝑤𝑟 (𝑥, 𝑦) = K ∙ exp(−𝜋(𝛼 2 (𝑥 − 𝑥0 )𝑟 + 𝑏2 (𝑦−𝑦0 )𝑟 2 ,
where ( x0 , y0 ) is the peak of the function, a and b are the Gaussian scaling parameters, and the index
r denotes the rotation operation, which can be described this way:
                                                                                                     (5)
                           (𝑥 − 𝑥0 )𝑟 = (𝑥 − 𝑥0 ) cos 𝜃 + (𝑦 − 𝑦0 ) sin 𝜃,
                                                                                                      (6)
                            (𝑦 − 𝑦0 )𝑟 = −(𝑥 − 𝑥0 ) sin 𝜃 +(𝑦 − 𝑦0 ).
   Therefore, the complex Gabor function in the spatial domain can be written as follows [8]:
                                        2
     𝑔(𝑥, 𝑦) = K ∙ exp(−𝜋(𝛼 2 (𝑥 − 𝑥0 )𝑟 + 𝑏2 (𝑦−𝑦0 )𝑟 2 ))exp⁡(𝑗(2𝜋 (𝑢0 𝑥 + 𝑣0 𝑦) + 𝑃 ).     (7)
   The function is determined by the following parameters: K - scales the value of the Gaussian
envelope; (a, b) - scales two axes of the Gaussian envelope; ( x0 , y0 ) - peak coordinates of the
Gaussian envelope; (u0 , v0 ) - spatial frequencies of the sinusoidal carrier in Cartesian coordinates; P
is the phase of the sinusoidal carrier [9].

2.3.    Histogram of Oriented Gradients (HOG)

   In [5] it was presented the algorithm that uses Haar wavelet decomposition as a processing technique
together with BSIF (binarized statistical image features) and HOG (histogram of oriented gradients) as
methods of feature extraction. Method proposed in this work solves the problem of palmprint features
extraction. Based on this research it was decided to apply similar algorithm with other methods for the
task of face recognition.
   The Histogram of Oriented Gradients (HOG) method can be applied to images that were processed
by wavelet transform to extract important features of image shape. Histogram of oriented gradients
calculates with the execution of the following steps. Firstly, before building of the orientation histogram
for each cell, there must be obtained the value of the gradient. After that, obtained histograms that
grouped by the individual cell are normalized. This process can be described mathematically:




                                                                                                         72
                                                            1
                              𝐷𝑥 = [−1      0 1],    𝐷𝑦 = [ 0 ].                                (8)
                                                           −1
   Dx is a 1-dimensional horizontal discrete derivative mask, Dy is a 1-dimensional vertical discrete
derivative mask.
   The gradient histogram retrieves with the equations:
                                𝐼 (𝑥) = 𝐼 ∙ 𝐷𝑥 ,    𝐼 (𝑦) = 𝑖 ∙ 𝐷𝑦 .                                 (9)
   Gradient size is calculated as:
                                     |𝐺 | = √𝐼2 (𝑥) + 𝐼2 (𝑦).                                       (10)
   Orientation is calculated with the following equation:
                                                      𝐼𝑦
                                          𝜃 = tan−1      .                                          (11)
                                                      𝐼𝑥
   Then orientation distributes equally in a range of 0 to 180 deg (for an unsigned gradient) or 0 to 360
deg (for a signed gradient). Each pixel computes a channel weighted vote based on the gradient values.
   After that cells groups into blocks that are connected in space. It allows to obtain the vector of the
normalized histogram elements.

2.4.    Local Binary Patterns (LBP) in 1-Dimensional Space

   For the palmprint recognition task authors of [5] used BSIF method (binarized statistical image
features) as feature extraction technique. BSIF is an LBP-based algorithm. For the task of face
recognition there are more efficient algorithms based on LBP like 1DLBP and I1DLBP.
   Local Binary Patterns in 1-Dimensional Space (1DLBP) firstly was presented in [10] and tested on
high-quality images. In this work 1DLBP will be explored with the appliance on low-quality images.
   The purpose of the 1DLBP method is to describe the local agitation of 1-dimensional signal segment
in binary code. That description can be obtained by comparing of the neighbor pixel values with the
central pixel value. All neighbor elements get the value of 1 if they are greater or equal to the current
element and value of 0 if they are less than the current element. Then, each element of the obtained
vector is multiplied by a weight according to its position. Finally, the current element is replaced by the
sum of the values of obtained vector. Described process can be expressed as follows:
                                             𝑁−1

                                1𝐷𝐿𝐵𝑃 = ∑ 𝑆(𝑔𝑛 − 𝑔0 ) ∙ 2𝑛 .                                        (12)
                                             𝑛=0

    S(x) function defines as S(x)={1 if x⩾0;⁡0⁡otherwise}; g0 and gn are the values of the central element
and its 1-dimensional neighbors, respectively. The index n changes its value increasingly from the left
to the right in the 1-dimensional string. The histogram of the 1-dimensional pattern defines the 1DLBP
descriptor.
    The idea of the I1DLBP is similar to the idea of the 1DLBP - the local patterns are extracted by
thresholding the linear neighbors of each pixel, from the projected image, with the mean value of the
summed neighbors:
                                              𝑁−1
                                                                𝑛
                                 𝐼1𝐷𝐿𝐵𝑃 = ∑ 𝑆(𝑔𝑛 − 𝑔́ ) ∙ 2 ,                                       (13)
                                              𝑛=0

where S(x) function defines as S(x)={1 if x⩾0;⁡0⁡otherwise}; g’ and gn are the mean value of the linear
neighborhood and the values of its 1-dimensional neighbors, respectively.
   The 1DLBP algorithm applies on all blocks of the split image with different resolution. The extracted
histograms are concatenated in one global 1-dimensional vector that represents one face image [11].
   Summarizing the foregoing description of 1DLBP method, the following stages of image processing
can be distinguished to obtain 1DLBP feature vector:
        Input image preprocesses.

                                                                                                        73
          Image splits into multiple blocks.
          Each decomposed block projects in 1-dimensional space.
          1DLBP algorithm processes each projected block.
          The vectors that obtained after each block processing concatenate in one global feature
           vector [12].

3. Experimental research and analysis
   Experimental research of the selected methods was performed using software created with Python.
The research was conducted on three different face databases: The Database of Faces, Facial
Recognition Technology (FERET) database and Surveillance Cameras Face Database (SCface). The
purpose of experimental research is to establish the most efficient combination of methods by
comparing results of their work with the use of one feature extraction method (HOG or 1DLBP) and
the use of combination of feature extraction methods (HOG and 1DLBP). For the classification of the
feature vector, obtained as a result, square Euclidean distance metric was used in all sets of experiments.
   The examples of images obtained as a result of applying described methods are depicted on Figure 1.




    Figure 1: Examples of images: a) original image; b) image processed with anisotropic diffusion;
c) image processed with Gabor wavelet transform; d) histogram of gradients of given image.

3.1.    The Database of Faces

   First set of experiments was performed using The Database of Faces that was created by AT&T
Laboratories Cambridge during the work on face recognition project. Database presented with a set of
face images of 40 different individuals. It is organized in 40 directories with 10 PGM format images in
each directory. The size of each image is 92x112 pixels, with 256 grey levels per pixel. Images were
taken with the person in frontal position with lighting, facial expression and facial details varying. The
database was retrieved openly from official site of AT&T Laboratories Cambridge [13]. Results of
experiments performed on The Database of Faces are presented in the Table 1.

Table 1
Results of experiments with The Database of Faces
                      AD + Gabor + HOG            AD + Gabor + 1DLBP          AD + Gabor + HOG + 1DLBP
                  Identification    False      Identification    False      Identification     False
                    accuracy     recognition     accuracy     recognition      accuracy     recognition
                      rate          rate           rate          rate            rate          rate
 Total number
                              40                           40                            40
   of images
    Number             26            14             17             23            28              12
 Identification
                      65%            35%          42.5%          57.5%          70%             30%
 accuracy rate

   As can be concluded from the Figure 2, when only one feature extraction method is used, HOG or
1DLBP separately, the percentage of correctly identified images varying from 42.5% to 65%. And the
highest rate of identification accuracy of 70% was obtained during the usage of combination of feature

                                                                                                          74
vector extraction methods. It confirms the reasonability of the usage of two feature extraction methods
through one iteration of the algorithm work. Therefore, false recognition rate varies from 30 to 57.5%.

3.2.    Facial Recognition Technology (FERET) database
   The FERET Database is a part of the Face Recognition Technology program [14] that researches
automatic face recognition capabilities to develop new algorithms for the automatic recognition of
human faces. As Technical Agent for distribution of the database serves The National Institute of
Standards and Technology (NIST). The FERET database contains 14126 high-resolution images of
1199 individuals with the resolution of 256x384. To obtain database it was necessary to request an
account for downloading the FERET database. Corresponding request was sent according to the
instruction given by the owner [15]. As far as first set of experiments was conducted on the database
that contains images of total 40 people, it was decided to use images of 40 people as well to conduct
the experiments on the FERET database.




   Figure 2: Results of identification with the use of The Database of Faces

   Experimental results for The FERET Database are presented in Table 2.

Table 2
Results of experiments performed with The FERET Database
                      AD + Gabor + HOG            AD + Gabor + 1DLBP          AD + Gabor + HOG + 1DLBP
                  Identification    False      Identification    False      Identification     False
                    accuracy     recognition     accuracy     recognition      accuracy     recognition
                      rate          rate           rate          rate            rate          rate
 Total number
                             40                           40                             40
   of images
    Number             29            11             26            14             29             11
 Identification
                     72.5%         27.5%           65%           35%           72.5%          27.5%
 accuracy rate

    As can be seen from Fig. 3, on high resolution images the usage of 1DLBP method is not necessary,
because the identification accuracy rate is the same during the experiment performance with HOG
feature vector extraction method only, as well as with combination of HOG and 1DLBP methods, which
is 72.5%. For the FERET database the usage of two feature extraction methods is not justified. As a
conclusion, false recognition rate obtained during the experiments with The FERET Database varies
from 27.5 to 35.5%.

                                                                                                          75
   Figure 3: Results of identification with the use of The FERET Database

3.3.    SCface database
   SCface database was created by researchers from University of Zagreb [16] for testing face
recognition algorithms in conditions of the real world. Images of the database was taken on surveillance
cameras of varying quality and resolution, therefore images of this database are vary-resolution. Also,
the database contains frontal mug shot images for the scenario when a person needed to be recognized
comparing mug shot image to a low-quality video surveillance image. Database contains 4160 static
images of 130 individuals. More technical details can be found in [17]. For the purpose of this research,
low quality video surveillance images were used to test the selected methods and explore the small
sample size problem. As far as first set of experiments was conducted on the database that contains
images of total 40 people, it was decided to use images of 40 people as well to conduct the experiments
on the SCface database. Obtained results of the experiments performed on SCface database are
presented in Table 3.

Table 3
Results of experiments performed with the SCface Database
                      AD + Gabor + HOG            AD + Gabor + 1DLBP          AD + Gabor + HOG + 1DLBP
                  Identification    False      Identification    False      Identification     False
                    accuracy     recognition     accuracy     recognition      accuracy     recognition
                      rate          rate           rate          rate            rate          rate
 Total number
                             40                           40                             40
   of images
    Number             37            3              31            9              39              1
 Identification
                     92.5%          7.5%          77.5%         22.5%          97.5%           2.5%
 accuracy rate

   From Fig. 4 it can be concluded that using 1DLBP method for feature extraction process by itself
provides only 77.5% of correctly identified face images. Usage of HOG method improves algorithm
work results within the order of 15% and provides the identification accuracy rate of 92.5%. But fusing
of HOG and 1DLBP feature vectors escalate the correctness of algorithm to 97.5%. Results of this set
of experiments confirms that the usage of combination of two feature vector extraction methods is
justified on low quality and resolution image, and the false recognition rate in this case is the lowest
among all the experiments – 2.5%.



                                                                                                          76
3.4. Analysis of obtained results

   Comparative diagram on Fig. 5 presents the results of identification with the use of selected methods
on different databases. Taking into consideration the fact that facial images from SCface database, that
were used for experimentation performance, are low-quality video surveillance images, it can be
concluded that combination of all researched methods provides high identification accuracy rate on
low-resolution facial images.




   Figure 4: Results of identification with the use of SCface Database

    After the analysis of all experimentational results it can be concluded that the combination of such
methods as anisotropic diffusion, Gabor wavelet transform, histogram of oriented gradients and local
binary patterns in 1-dimensional space demonstrates the highest identification results being applied to
all researched databases.




   Figure 5: Comparative diagram of the experimentational results of the researched methods
appliance to different databases


                                                                                                     77
4. Development of image processing algorithm
   From the perspective of analysis of obtained research results, combination of methods with the
highest identification accuracy rate can be used to develop the algorithm underlying in information
technology for person identification in video stream.
   In general, the person identification process consists of two stages. The first is the determination of
face location in the image taken from video stream. The original image is scanned with a smaller
window, and every time there is a determination of the similarity degree between the image in the
window with a human face. Formally, the face image can be structurally defined, statistically or by
sample face images list [18]. After the window, which most likely contains just a person's face, has
been identified, the second stage starts- identification. The purpose of person identification is to define
a unique identifier for each input biometric parameter or to determine it as unknown if such sample is
not in the database [19]. The algorithm for information technology of person identification must be
applicable to facial images in aim to extract their feature vectors for further classification in order to
identify a person. Data flow diagram of a proposed algorithm is presented on Fig. 6.




   Figure 6: Data flow diagram of the algorithm


                                                                                                        78
    The algorithm uses a portrait image of a person as input data. Step-by-step algorithm execution can
be described as follows:
    1. Face detection and localization on the input image. Only face image of a person is used for further
processing because it is the area of interest for person identification process.
    2. Image pre-processing using anisotropic diffusion that allows to preserve and enhance edge
information and remove noise.
    3. The image of the face, after applying anisotropic diffusion to it, is processed by Gabor wavelets
with various changes in the parameters of the wavelet function so that 16 wavelet-transformed
variations of the face image can be obtained. After that, the wavelet-transformed images are summed
to form a global image, which is processed by the following methods.
    4. The global image formed as a result of the Gabor wavelet transform is simultaneously fed to the
input of two independent methods - HOG and 1DLBP. As a result of the operation of each of these
methods, two separate 512-value feature vectors are formed.
    5. HOG feature vector and 1DLBP feature vector normalization.
    Because of the deviation in vector distributions and ranges, feature vectors extracted separately from
1DLBP and the HOG features are incompatible. There are some methods of vector normalization, that
can improve compatibility. For example, min–max normalization that transforms the feature vectors in
the range [0, 1]. If X = [x1; x2; x3; ...; xn] is the feature vector, the normalized feature vector can be
represented using min–max normalization [5]:
                                               𝑥𝑖 − 𝑀𝑖𝑛(𝑋)
                                      𝑥′ =                   .                                       (14)
                                             𝑀𝑎𝑥 (𝑋) − 𝑀𝑖𝑛(𝑋)
    6. The HOG feature vector and 1DLBP feature vector concatenates to form a 1024-value global
feature vector of a face image. The global feature vector of the image can be obtained by concatenating
into a single feature vector the normalized feature vectors of 1DLBP and HOG features.
    Let the normalized feature vectors be D = [d1; d2; d3 ; ...; dn] for 1DLBP and h = [h1; h2; h3; ...; hn]
for HOG extraction. The global vector can be represented as [5]:
                            𝑉𝑔 = [𝑑1 , 𝑑2 , 𝑑3 , … , 𝑑𝑛 , ℎ1 , ℎ2 , ℎ3 , … , ℎ𝑛 ].                   (15)
   The obtained global feature vector is using for further classification.
   The algorithm for information technology of person identification in video stream that includes such
methods as anisotropic diffusion, Gabor wavelet transform, histogram of oriented gradients and local
binary patterns in 1-dimensional space proposed in this paper firstly.

5. Conclusion
    This paper describes the research of methods to develop the algorithm underlying in information
technology of person identification in video stream. During the research, experiments were performed
on the testing of methods being applicable in face recognition in order to establish the most efficient
combination of methods by comparing results of their work with the use of one feature extraction
method (HOG or 1DLBP) and the use of combination of feature extraction methods (HOG and 1DLBP).
    The obtained results indicated that the highest identification accuracy rate of 97.5% was obtained
with both feature extraction methods (HOG and 1DLBP) on the images from SCface database, which
are low-quality video surveillance images. Experimental results on the Database of Faces and FERET
database provides from 70 to 72.5% of correctly identified images.
    Proposed algorithm is based on anisotropic diffusion as an image preprocessing method, Gabor wavelet
transform as an image processing method, histogram of oriented gradients (HOG) and local binary patterns in
1-dimensional space (1DLBP) as the methods of feature vector extraction from the images, square Euclidean
distance metric for vector classification.
    Summarizing the foregoing conclusions, proposed in this work algorithm for information technology
of person identification based on anisotropic diffusion, Gabor wavelet transform, histogram of oriented
gradients (HOG) and local binary patterns in 1-dimensional space (1DLBP) can be applied for face
recognition and person identification in conditions of face images with low quality and resolution, that
means it partially solves small sample size problem. For future research it is aimed to improve the

                                                                                                         79
proposed algorithm for the scenario when low quality video surveillance images compare with frontal
mug shot image from law enforcement or national security databases.

6. References
[1] B. Thompson and R. Van Noorden. The troubling rise of facial recognition technology. Nature.
     November 18, 2020. URL: https://www.nature.com/articles/d41586-020-03271-8
[2] Facial recognition market - growth, trends, Covid-19 impact, and forecasts (2021 - 2026). Mordor
     Intelligence. March 2021. URL: https://www.mordorintelligence.com/industry-reports/facial-
     recognition-market
[3] Facial Recognition Market Size, Share & Trends Analysis Report, 2021 – 2028. Grand View
     Research.      May       2021.     92     p.     Report     ID:     978-1-68038-311-9.          URL:
     https://www.grandviewresearch.com/industry-analysis/facial-recognition-market/segmentation
[4] T.M. Abhishree, J. Latha, K. Manikantan, S. Ramachandran. Face Recognition Using Gabor Filter
     Based Feature Extraction with Anisotropic Diffusion as a Pre-processing Technique. Procedia
     Computer Science, Volume 45, 2015, pp. 312-321, ISSN 1877-0509, doi:
     10.1016/j.procs.2015.03.149.
[5] B. Attallah, A. Serir, Y. Chahir, A. Boudjelal. Histogram of gradient and binarized statistical image
     features of wavelet subband-based palmprint features extraction. J. Electron. Imag. 26(6) 063006,
     November 8, 2017, doi: 10.1117/1.JEI.26.6.063006.
[6] P. Perona and J. Malik. Scale-Space and Edge Detection Using Anisotropic Diffusion. IEEE
     Transactions on Pattern Analysis and Machine Intelligence, vol. 12, No. 7, pp. 629-639. July 1990.
[7] C. Yu, and Y. Jia. Anisotropic Diffusion-based Kernel Matrix Model for Face Liveness Detection.
     Image Vis. Comput., 89, 2019, pp. 88-94, doi: 10.1016/J.IMAVIS.2019.06.009
[8] J.R. Movellan. Tutorial on Gabor, 2002. URL: https://inc.ucsd.edu/mplab/tutorials/gabor.pdf
[9] G. P. Dimitrov et al. Creation of Biometric System of Identification by Facial Image. 2020 3rd
     International Colloquium on Intelligent Grid Metrology (SMAGRIMET), 2020, pp. 29-34, doi:
     10.23919/SMAGRIMET48809.2020.9263995.
[10] A. Benzaoui and A. Boukrouche. Face Recognition Using 1DLBP Texture Analysis. Future
     Computing 2013: The Fifth International Conference on Future Computational Technologies and
     Applications, 2013, pp. 14-19, ISBN: 978-1-61208-272-1.
[11] A. Benzaoui and A. Boukrouche. 1DLBP and PCA for face recognition. 2013 11th International
     Symposium on Programming and Systems (ISPS), 2013, pp. 7-11, doi: 10.1109/ISPS.2013.6581486.
[12] A. Benzaoui and A. Boukrouche. Face Analysis, Description and Recognition using Improved
     Local Binary Patterns in One Dimensional Space. Journal of Control Engineering and Applied
     Informatics, Vol. 16, No. 4, pp. 52-60, 2014.
[13] The Database of Faces. URL: https://cam-orl.co.uk/facedatabase.html
[14] Face Recognition Technology (FERET). URL: https://www.nist.gov/programs-projects/face-
     recognition-technology-feret
[15] Color FERET database. URL: https://www.nist.gov/itl/products-and-services/color-feret-database
[16] SCface - Surveillance Cameras Face Database. URL: https://www.scface.org
[17] M. Grgic, K. Delac and S. Grgic. SCface - surveillance cameras face database. Multimedia Tools
     and Applications Journal, Vol. 51, No. 3, February 2011, pp. 863-879.
[18] O. Bychkov, K. Merkulova and Y. Zhabska, "Information Technology of Person’s Identification
     by Photo Portrait," 2020 IEEE 15th International Conference on Advanced Trends in
     Radioelectronics, Telecommunications and Computer Engineering (TCSET), 2020, pp. 786-790,
     doi: 10.1109/TCSET49122.2020.235542.
[19] O. Bychkov, O. Ivanchenko, K. Merkulova and Y. Zhabska. Mathematical Methods for
     Information Technology of Biometric Identification in Conditions of Incomplete Data.
     Proceedings of the 7th International Conference "Information Technology and Interactions"
     (IT&I-2020), Kyiv, Ukraine, December 02-03, 2020, pp. 336-349.




                                                                                                      80