<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>December</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>The System for Recognizing Useful Information of the Client's ID-Card Based on Machine Learning Technologies</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oleksii Bychkov</string-name>
          <email>oleksiibychkov@knu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Liudmyla Zubyk</string-name>
          <email>zubyk.liudmyla@knu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dmytro Gololobov</string-name>
          <email>gololobov.dma@meta.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yaroslav Isaienkov</string-name>
          <email>yisaienkov@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ganna</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Grynkevych</string-name>
          <email>Ggrynkevych@ukr.net</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anastasiia Ivanytska</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jimi IoT Technology Co.</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>National Aviation University</institution>
          ,
          <addr-line>1 ave. Huzar Lubomyr, Kyiv, 03058</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>60 Volodymyrska str., Kyiv, 01601</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Vinnytsia National Technical University</institution>
          ,
          <addr-line>95 Khmelnytske highway, Vinnytsia, 21021</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>1</volume>
      <fpage>9</fpage>
      <lpage>21</lpage>
      <abstract>
        <p>Existing approaches to recognition of text information from user documents in client-server systems were analyzed in order to solve the problem of user identification. The comparative analysis of the Optical Character Recognition library using Tesseract Engine and Paddle Optical Character Recognition was carried out. The feasibility and effectiveness of using Paddle Optical Character Recognition for analyzing documents with Chinese characters was substantiated. The text processing model was proposed to highlight valuable information and form the dictionary, which will be transmitted to the system server. Its effectiveness was verified on test data and the adequacy of the model was assessed based on Character Error Rate and Word Error Rate (CER and WER). In contrast to existing approaches, the model showed the increase in text recognition accuracy by 1.8-11.3%, depending on the quality of the source image. The result is implemented in client-server applications for product solutions of Zhejiang Optical character recognition, machine learning, client-server systems.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Identification of relevant information from the documents of the customer system and further
identification of the system is the important part of the online applications of accounts and services.</p>
      <p>
        Proposed extracting useful data from customer IDs card using Optical Character Recognition
(OCR). After analyzing the literature [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1-3</xref>
        ], it was revealed that the Tesseract Engine library was used
to recognize the texts of past documents (papers, articles) from popular libraries, tax card.
      </p>
      <p>
        Tesseract, the open-source package implemented in Python, was used to process digitized images in
the issue [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The OCR output was processed using Python modules applying localization and text
detection followed by classification, but without the use of machine learning/deep learning/natural
language processing techniques, which are quite complex and time and data intensive. However, the
average accuracy of the obtained result was quite low (34.60%).
      </p>
      <p>
        ID verification is undoubtedly one of the most difficult steps in the Know Your Customer (KYC)
process, requiring the lot of effort, time and money to implement as usually. A revolutionary solution
to ID verification problems involving machine learning and deep neural network techniques was
proposed in the study [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        The study [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] looked at the OCR model (Tesseract) built on CNN that extracted data from the tax
card image. Good results were obtained, providing an accuracy level of 80%.
      </p>
      <p>2023 Copyright for this paper by its authors.
CEUR</p>
      <p>ceur-ws.org</p>
      <p>
        OCR is the technology to identify characters with the highest possible accuracy through the use of
appropriate pre-processing, processing and post-processing refinements. The significant contribution to
the final result is provided by the avoidance of noise in the original image, which is emphasized in the
study [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Automatic information extraction from scanned documents significantly increases efficiency,
accuracy and speed in all those business processes where data collection from documents plays an
important role. Such documents are often digitized as images and then converted to text format. A text
recognition method based on transfer learning and scene text recognition (STR) networks is presented
in [
        <xref ref-type="bibr" rid="ref6 ref7">6-7</xref>
        ]. The formalization model of recommendations that based on technology of machine learning
was proposed. After analyzing of different methods, the product-based collaborative filtering (CF) was
chosen to solve the research problem [
        <xref ref-type="bibr" rid="ref8 ref9">8-9</xref>
        ].
      </p>
      <p>
        For experimental research the algorithm was used that underlies in information technology of person
identification in video stream. The algorithm performance provided various identification accuracy rate
results with the difference, which amount to 20% on average [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The block diagram and techniques
for optimal implementation of the improved gradient method in device control for unmanned aerial
vehicle are available in the paper [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
      <p>
        In this paper [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], was proposed the new original architecture of a model based on an artificial
convolutional neural network and semantic segmentation approach for the recognition and detection of
identity documents in images. Authors [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] proposed a new pre-processing approach consists of
DeblurGAN (Generative Adversarial Network for deburring image), shadow removal, and binarization
to pre-process the image for Tesseract-OCR, achieved an average Character Error Rate of 18.82%
which is better compared to without pre-processing which is 38.13%.
      </p>
      <p>However, existing approaches to text recognition often require image pre-processing, which requires
the development of additional models for processing the resulting text or input image. The main
problem in text recognition is the quality of the input image, its brightness, position relative to the
camera and size. Thus, today an urgent scientific task is to find approaches to image recognition of user
documents, which provides the convenient format for the resulting data. To solve this problem, it is
necessary to analyze different approaches to recognize documents with Chinese characters and develop
a new model for forming the dictionary of Chinese user’s information, which can be verified on test
images.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Purpose of the study</title>
      <p>The purpose of this research is to develop the model for forming the dictionary based on text
recognition technologies from document images and increase the accuracy of information transfer in
the client-service system. To achieve this goal, it is necessary to solve the following tasks:
 analysis of text recognition technologies on documents;
 select the optimal module from two approaches for recognizing text images
 determine the stages of developing the model of the dictionary formation which based of text
recognition technologies, which was selected;</p>
      <p> evaluate the adequacy of the proposed model on test data.</p>
    </sec>
    <sec id="sec-3">
      <title>2.1 Text recognition tools based on machine learning technologies</title>
      <p>
        Initially, Tesseract technology was used during the research to recognize documents with Chinese
characters. It can be used directly or through the API to extract written text from images with different
languages. Tesseract may be used in conjunction with the existing layout analysis to recognize text
inside a big document, or with an external text detector to recognize text from an image of a single text
line. Pillow and Leptonica imaging libraries, including jpeg, png, gif, bmp, tiff, and others formats of
file [
        <xref ref-type="bibr" rid="ref14 ref15 ref16 ref7">7, 14-16</xref>
        ].
      </p>
      <p>
        The first approach, Tesseract technology was used in the next steps, which presented in fig. 1. Test
images were uploaded to the recognition system and recognition results were obtained. For good quality
images taken directly, the system produced recognition results. To track the quality of the
recommendations of viewed methods, it is proposed to choose metrics from the following: root mean
square error (RMSE), mean absolute error (MAE), normalized value of the mean absolute error
(NMAE) [
        <xref ref-type="bibr" rid="ref17 ref18 ref19 ref20 ref21">17-21</xref>
        ]. But for evaluating the quality of recognition according to follow the parameters CER
and WER, calculated according to expressions 1-2, the following results were obtained:
(1)
(2)
=
 +  +
      </p>
      <p>=
  +   +</p>
      <p />
      <p>
        The formula for WER is the same as that of CER, but WER operates at the word level instead. It
represents the number of word substitutions, deletions, or insertions needed to transform one sentence
into another [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
      <p>The CER becomes in interval [25%, 33.33%] and the WER value of 50% is clearly understood since
2 out of 4 words in the sentence were wrongly transcribed. The obtained results do not make it possible
to build a text processing model for dictionary formation. Therefore, a decision was made to choose
other technologies for conducting research, namely Paddle OCR.</p>
      <p>Paddle OCR is the open-source OCR toolkit developed by PaddlePaddle, an advanced AI model
based on the powerful GPT-3.5 architecture, developed by Open AI. As the cutting-edge OCR
technology, Paddle OCR excels in converting images containing textual content into editable,
searchable, and machine-readable text. It efficiently analyzes images and recognizes characters,
where  – number of substitutions;
 – number of deletions;
 – number of insertions;
 – number of characters in reference text.</p>
      <p>
        =  +  +  (where  − number of correct characters).
numbers, and symbols, enabling seamless image-to-text conversion. Paddle OCR consists of an
ultralightweight and general OCR model, integrating OCR algorithms like:
 Text detection models: EAST, DB, SAST
 Text recognition models: CRNN, Rosetta, STAR-Net, RARE, SRN [
        <xref ref-type="bibr" rid="ref23 ref24 ref25 ref8">8, 23-25</xref>
        ].
      </p>
      <p>The next machine learning technologies are used for the proposed text recognition system, such as
Convolutional Recurrent Neural Network (CRNN), sequence recognition network (SRN). The network
architecture of CRNN consists of three components, including the convolutional layers, the recurrent
layers, and a transcription layer, from bottom to top. At the bottom of CRNN, the convolutional layers
automatically extract a feature sequence from each input image. It can be applied to problems – Chinese
character recognition, which involve sequence prediction in images. In this case, the following steps
were used to find useful information from user documents:
 image loading;
 download of all libraries, packages and language configuration (chinese_cht.ttf);
 setting the recognition threshold to 0.1;
 text recognition from user documents.</p>
      <p>Next, the Paddle OCR also was tested on 10 test images, and the system produced recognition
results. For 6 images, the recognition result was obtained with the recognition accuracy in the range
from 0.835 to 0.999.</p>
      <p>For images of low quality, the system produces result that was difficult to process and reveal
valuable information for further dictionary formation. After evaluating the quality of recognition
according to the CER and WER parameters, calculated according to expressions 1-2, the next values
were obtained, which are presented in Table 1.
乡 四街四巷七横巷2号</p>
      <p>乡 四街四巷七横巷2号
广州市番禺区洛溪新东</p>
      <p>广州市番禺区洛溪新城
吉 祥道一幢之二601房</p>
      <p>吉 祥道一幢之二601房
安徽省涡阳县西阳镇大</p>
      <p>安徽省涡阳县西阳镇文
庙
行政村王大庄自然村
庙 行政村王大庄自然村</p>
      <p>CER, %
98.0
98.7
96.7
95.8</p>
      <p>WER, %
2.0
1.8
3.3
4.3
049 号
平庄村049号
河北省庄市郭村乡北自
河北省河间市郭村乡北
89.5</p>
      <p>11.3
049 号
太 平庄村049号</p>
      <p>As can be seen from the table, the result of text recognition for images of good and normal quality
(look at fig. 2) is sufficient for forming the text processing model and forming the dictionary for the
client-server system. Characters that were incorrectly recognized are highlighted in bold. Also, the
structure of the source text corresponds to the structure of the document, which greatly complicates the
process of developing the dictionary formation model.</p>
      <p>After a comparative analysis of the two approaches, Paddle OCR was selected for further research
based on the text recognition quality scores for 10 images of varying quality.</p>
    </sec>
    <sec id="sec-4">
      <title>2.2 Development of the text processing model for dictionary formation</title>
      <p>The object of the study is the image of the client's ID card, which is presented in figure 2, after text
recognition by the Paddle OCR system, the processing model was developed, which will allow the
formation of the dictionary of useful information, which looks like this:
"name":"张*舒",
"sex":"女",
"nation":"汉",
"birth":"197***512",
"address":"广州市番**洛溪新城吉祥道十幢之二601房",
"idcard":"441*******0427</p>
      <p>After text recognition, the text information processing model is proposed and includes the next
stages of forming the dictionary of useful information:
1. Combining the recognition result into one string
2. Function to find part of text
3. Find id_number after the characters '公民身份号码' in the client's passport.
4. Using of regular expression for search id_number
5. Search Data of birthday after symbols '出生','住址'
6. Find address after characters '住址', '公民身份号码'
7. Find id_number after '住址'
8. Formation of the result table styling.</p>
      <p>The developed text processing model made it possible to obtain dictionaries containing the user's
personal data, so they will not be included in the research results. This made it possible to proceed to
the development of the client-service system for user interaction and a user application.</p>
    </sec>
    <sec id="sec-5">
      <title>3. Structure of the project</title>
      <p>The project contains various components, to use the HTTP protocol for interaction. To initiate the
request to identify an ID card, the following is the sample Python request code. The responder sends
the generated dictionary to the server if the text recognition module, the post OCR model for processing
useful information, or the error in recognition are successful. Then the user can upload the better quality
image again and try identification again. To form the responder sends for the client-server system, the
next steps are proposed:
1. Read the input image.
2. Run the text recognition algorithm.
3. Use text processing model for dictionary formation.
5. Send valuable information to the server.</p>
      <p>6. If the result of forming the dictionary is error, then send the result – error to the document.</p>
      <p>The result was implemented in client-server applications for product solutions of Zhejiang Jimi IoT
Technology Co., Ltd.</p>
    </sec>
    <sec id="sec-6">
      <title>4. Conclusions</title>
      <p>As the result of the research, the model of forming dictionary basing on text recognition
technologies, that uses machine learning technologies, was developed and implemented. The following
stages of research were performed.</p>
      <p>Firstly, existing approaches to recognition of text information from user documents in client-server
systems were analyzed in order to solve the problem of user identification. Two approaches were
selected for text recognition algorithm.</p>
      <p>Secondly, the comparative analysis of the Optical Character Recognition library using Tesseract
Engine and Paddle Optical Character Recognition was held. The feasibility and effectiveness of using
Paddle OCR for analyzing documents with Chinese characters was substantiated.</p>
      <p>Thirdly, the model of the dictionary formation which based of text recognition technologies was
developed and tested.</p>
      <p>Finally, the adequacy of the model of the dictionary formation was evaluated. Its effectiveness was
verified on test data and the adequacy of the model was assessed based CER and WER. In contrast to
existing approaches, the model showed the increase in text recognition accuracy by 1.8-11.3%,
depending on the quality of the source image. The result is implemented in client-server applications
for product solutions of Zhejiang Jimi IoT Technology Co., Ltd.</p>
    </sec>
    <sec id="sec-7">
      <title>5. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Qamar</given-names>
            <surname>Uddin</surname>
          </string-name>
          ,
          <article-title>Features Extraction of Tax Card by Using OCR Based Deep Learning Techniques</article-title>
          ,
          <source>Master's thesis</source>
          , University of Eastern Finland,
          <source>Faculty of Science and Forestry, Joensuu School of Computing. Computer Science</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Tuan</given-names>
            <surname>Bui</surname>
          </string-name>
          ,
          <article-title>Applications of Machine Learning in EKYC'S Identity Document recognition</article-title>
          .
          <source>Bachelor's thesis</source>
          , South-Eastern Finland University of Applied Sciences,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jatowt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Coustaty</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Doucet</surname>
          </string-name>
          ,
          <article-title>Deep statistical analysis of OCR errors for effective post-OCR processing</article-title>
          ,
          <source>ACM/IEEE Joint Conference on Digital Libraries (JCDL)</source>
          ,
          <year>Jun</year>
          .
          <year>2019</year>
          ,
          <fpage>29</fpage>
          -
          <lpage>38</lpage>
          . doi:
          <volume>10</volume>
          .1109/JCDL.
          <year>2019</year>
          .
          <volume>00015</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Ch. Padole</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Sh. Verma</surname>
          </string-name>
          , Pr. Gujral, Mr. Kumar etc.,
          <article-title>Information Extraction from Visiting Cards Using OCR and Post-Processing in Python</article-title>
          ,
          <source>International Journal of Scientific and Technical Research in Engineering (IJSTRE)</source>
          , Vol.
          <volume>7</volume>
          ,
          <string-name>
            <surname>ISSUE</surname>
          </string-name>
          05(
          <year>2022</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C.</given-names>
            <surname>Madan Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Brindha</surname>
          </string-name>
          ,
          <article-title>Text Extraction from Business Cards and Classification of Extracted Text Into Predefined Classes</article-title>
          ,
          <source>Proceedings of International Conference on Computational Intelligence &amp; IoT (ICCIIoT)</source>
          ,
          <year>2018</year>
          ,
          <fpage>595</fpage>
          -
          <lpage>602</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Arulanantham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Snekha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Logeshwaran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Nishanth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lavanya</surname>
          </string-name>
          , Utilizing OCR to Retrieve Text from Identity Documents,
          <source>International Journal for Research in Applied Science &amp; Engineering Technology (IJRASET)</source>
          , Vol.
          <volume>11</volume>
          ,
          <string-name>
            <surname>Issue</surname>
            <given-names>IV</given-names>
          </string-name>
          ,
          <year>Apr 2023</year>
          ,
          <fpage>349</fpage>
          -
          <lpage>363</lpage>
          . doi:
          <volume>10</volume>
          .22214/ijraset.
          <year>2023</year>
          .
          <volume>50090</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>James</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seth</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mukhopadhyay</surname>
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Ch</surname>
          </string-name>
          .
          <source>Iot Systems Design : Projects Based Approach</source>
          , Springer,
          <year>2022</year>
          ,
          <fpage>227</fpage>
          -
          <lpage>279</lpage>
          . doi :
          <volume>10</volume>
          .1007/978-3-
          <fpage>030</fpage>
          -85863-6_
          <fpage>12</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>What is Paddle</surname>
            <given-names>OCR</given-names>
          </string-name>
          ? URL: https://www.plugger.ai/blog/what-is
          <string-name>
            <surname>-</surname>
          </string-name>
          paddle-ocr
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Ivanytska</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubyk</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubyk</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ivanov</surname>
            <given-names>D.</given-names>
          </string-name>
          <article-title>The advertising predictivon model based on machine learning technologies</article-title>
          <source>// CEUR Workshop Proceedings of the 8th International Conference "Information Technology and Interactions" (IT&amp;I-2021)</source>
          ,
          <fpage>1</fpage>
          -2 Desember,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>V.</given-names>
            <surname>Martsenyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Bychkov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Merkulova</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhabska</surname>
          </string-name>
          ,
          <article-title>Exploring Image Unified Space for Improving Information Technology for Person Identification /</article-title>
          / IEEE Access, vol.
          <volume>11</volume>
          , pp.
          <fpage>76347</fpage>
          -
          <lpage>76358</lpage>
          ,
          <year>2023</year>
          , doi: 10.1109/ACCESS.
          <year>2023</year>
          .
          <volume>3297488</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Dakhno</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barabash</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shevchenko</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leshchenko</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Dudnik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <article-title>Integro-differential models with a K-symmetric operator for controlling unmanned aerial vehicles using a improved gradient method</article-title>
          . // 2021 IEEE 6th International Conference on Actual Problems of Unmanned Aerial Vehicles
          <string-name>
            <surname>Development (APUAVD) October</surname>
          </string-name>
          ,
          <year>2021</year>
          . pp.
          <fpage>61</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Kozlenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sendetskyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Simkiv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Savchenko</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Bosyi</surname>
          </string-name>
          ,
          <article-title>Identity Documents Recognition and Detection using Semantic Segmentation with Convolutional Neural Network</article-title>
          ,
          <source>Cybersecurity Providing in Information and Telecomunicationg Systems</source>
          ,
          <year>2021</year>
          , Kyiv, Ukraine,
          <fpage>234</fpage>
          -242 doi :10.5281/zenodo.5758182
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>E.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. A.</given-names>
            <surname>Putra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. P.</given-names>
            <surname>Kusuma</surname>
          </string-name>
          ,
          <article-title>Improving Optical Character Recognition Accuracy for Indonesia Identification Card Using Generative Adversarial Network</article-title>
          ,
          <source>Journal of Theoretical and Applied Information Technology</source>
          ,
          <year>2022</year>
          , Vol.
          <source>No</source>
          <volume>8</volume>
          ,
          <fpage>2424</fpage>
          -
          <lpage>2437</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. R. M. Ribeiro</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Julio</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Abelha</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Abelha</surname>
            , and
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Machado</surname>
          </string-name>
          ,
          <article-title>A comparative study of optical character recognition in health information system</article-title>
          , International Conference in Engineering Applications (ICEA),
          <year>Jul</year>
          .
          <year>2019</year>
          , 1-
          <fpage>5</fpage>
          . doi:
          <volume>10</volume>
          .1109/CEAP.
          <year>2019</year>
          .
          <volume>8883448</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Ivanytska</surname>
            <given-names>А.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubyk</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ivanov</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Domracheva</surname>
            <given-names>K.</given-names>
          </string-name>
          <article-title>Study of methods of complex data analysis based on machine learning technologies</article-title>
          <source>IEEE Internation Conference on Advanced Trends in Information Theory, ATIT</source>
          ,
          <year>2019D</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Phan Van Hoai</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.-Th. Duong</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Th</surname>
          </string-name>
          . Hoang,
          <article-title>Text recognition for Vietnamese identity card based on deep features network</article-title>
          ,
          <source>International Journal on Document Analysis and Recognition (IJDAR)</source>
          ,
          <volume>24</volume>
          (
          <issue>1-2</issue>
          ),
          <year>2021</year>
          . doi :
          <volume>10</volume>
          .1007/s10032-021-00363-7
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Ivanytska</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubyk</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dudnik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurchenko</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berestov</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <article-title>Evaluation of the</article-title>
          <source>Effectiveness of Information Technology Methods for Processing Diagnostic Information Based on Complex Data // 2022 IEEE 3rd International Conference on System Analysis and Intelligent Computing, SAIC 2022 - Proceedings</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Kutyrev</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kiktev</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kalivoshko</surname>
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rakhmedov</surname>
            <given-names>R</given-names>
          </string-name>
          .
          <article-title>Recognition and Classification Apple Fruits Based on a Convolutional Neural Network Model</article-title>
          . (
          <year>2022</year>
          ) CEUR Workshop Proceedings,
          <volume>3347</volume>
          , pp.
          <fpage>90</fpage>
          -
          <lpage>101</lpage>
          . https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3347</volume>
          /Paper_8.pdf
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N. S.</given-names>
            <surname>Yusman</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Ibrahim</surname>
          </string-name>
          ,
          <article-title>Extracting information from identity card into electronic form using image processing technique</article-title>
          ,
          <source>Proceedings of Innovation and Technology Competition (INOTEK)</source>
          , Vol.
          <volume>1</volume>
          (
          <issue>2021</issue>
          ),
          <year>2021</year>
          , Melaka, Malaysia,
          <fpage>135</fpage>
          -
          <lpage>136</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Kr</surname>
            . Olejniczak,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Sulc</surname>
          </string-name>
          ,
          <string-name>
            <surname>Text Detection Forgot About Document</surname>
            <given-names>OCR</given-names>
          </string-name>
          , 26th Computer Vision Winter Workshop, Robert Sablatnig and Florian Kleber (eds.), Krems. Lower Austria, Feb.
          <fpage>15</fpage>
          -
          <lpage>17</lpage>
          ,
          <year>2023</year>
          ,
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>D.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shen</surname>
          </string-name>
          and
          <string-name>
            <given-names>D.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <source>Functional Structure Recognition of Scientific Documents in Information Science, Join Workshop of the 4th Extraction</source>
          and
          <article-title>Evaluation of Knowledge Entities from Scientific Documents and the 3rd AI + Informatics (EEKE-AII2023)</article-title>
          , June 26,
          <year>2023</year>
          ,
          <string-name>
            <given-names>Santa</given-names>
            <surname>Fe</surname>
          </string-name>
          , New Mexico, USA and Online,
          <volume>10</volume>
          -
          <fpage>12</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>K.</given-names>
            <surname>Leung</surname>
          </string-name>
          ,
          <article-title>Evaluate OCR Output Quality with Character Error Rate (CER) and Word Error Rate (WER)</article-title>
          . URL: https://towardsdatascience.com
          <article-title>/evaluating-ocr-output-quality-with-charactererror-rate-cer-and-word-error-rate-wer-853175297510</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>F.</given-names>
            <surname>Borisyuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gordo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sivakumar</surname>
          </string-name>
          , Rosetta:
          <article-title>Large scale system for text detection and recognition in images</article-title>
          ,
          <source>KDD'</source>
          <year>2018</year>
          , London, United Kingdom. doi.org/10.475/123_
          <fpage>4</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Deli</surname>
            <given-names>Yu</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Xuan</given-names>
            <surname>Li</surname>
          </string-name>
          , Chengquan Zhang, Tao Liu, Junyu Han, Jingtuo Liu, Errui Ding,
          <article-title>Towards Accurate Scene Text Recognition with Semantic Reasoning Networks</article-title>
          ,
          <source>2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <fpage>13</fpage>
          -
          <lpage>19</lpage>
          June,
          <year>2020</year>
          , DOI:10.1109/cvpr42600.
          <year>2020</year>
          .01213
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Xinyan</surname>
            <given-names>Zu</given-names>
          </string-name>
          , Haiyang Yu,
          <string-name>
            <given-names>Bin</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Xianguang Xue, Towards Accurate Video Text Spotting with Text-wise Semantic Reasoning</article-title>
          ,
          <source>Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence (IJCAI-23)</source>
          ,
          <year>2023</year>
          ,
          <fpage>1858</fpage>
          -
          <lpage>1866</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>