<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>COLINS-</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Methodological Foundations of an Information System Construction for the Recognition of Ukrainian Sign Language</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Taras Basyuk</string-name>
          <email>Taras.M.Basyuk@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrii Vasyliuk</string-name>
          <email>Andrii.S.Vasyliuk@lpnu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Lviv Polytechnic National University</institution>
          ,
          <addr-line>Bandera str.12, Lviv, 79013</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>8</volume>
      <fpage>12</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>The article analyzes existing methods and known systems that provide means of recognizing Ukrainian sign language and describes the mechanisms of their implementation. Technologies and software tools for sign language recognition were analyzed, which made it possible to identify the main shortcomings of existing approaches and showed the relevance of the research. The diagram reflecting the main stages that must be implemented in the process of gesture recognition has been finalized. The structural design of the software system was carried out with the display of created diagrams in accordance with the IDEF0 standard. The article presents a context diagram and a decomposition diagram, which created the basis for the study of features and the formation of methodological foundations for the construction of an information system. The main stages of gesture recognition are highlighted and described, namely: transformation of the input image, its filtering and actual recognition. The justification of the choice of methods for displaying contours and recognizing gestures in the incoming information message was made and their analysis was carried out. The constructed prototype of the system for recognizing Ukrainian sign language consists of four main modules: HandGesturesRecognitionForm, NeuralNetwork, CsvManager, TrainingImageDataManager, which provide basic functionality. At the current stage, it can be useful as an additional communication tool for people with special needs. Further research will be aimed at testing and improving systems, eliminating conflicts and expanding functionality in accordance with the specified requirements.</p>
      </abstract>
      <kwd-group>
        <kwd>1 Ukrainian sign language</kwd>
        <kwd>pattern recognition</kwd>
        <kwd>learning process</kwd>
        <kwd>communication</kwd>
        <kwd>information system</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Today, computer technologies are involved in almost all spheres of human life. With the help of
various technical solutions, a person is able to solve daily tasks with greater simplicity and
efficiency. If at the end of the 20th century computer technologies were primarily associated with
the scientific and military spheres, then in the second decade of the 21st century they are
associated with almost all spheres of human life. It is quite natural that various computer
solutions are widely used in the field of communication between individuals with special needs.
At the same time, one of the communication devices is sign language [1]. Sign language is a type
of speech that makes it possible to express thoughts using facial expressions, emotions, and hand
gestures that correspond to letters, words, or individual phrases. Despite the large number of
people who suffer from hearing or speech impairments, sign language has received little attention
from linguistics. In the world, the share of people with hearing problems is about 5% or 430
million [2] of the total population. Sign languages are not universal in all countries, as they arise
and develop naturally in different territories and change over time with the emergence of new
vocabulary. The debate about sign language has been going on for about half a century. Until
recently, the attitude towards it in different countries ranged from introducing it for learning in
Paragraph text. Paragraph text. Paragraph text. Paragraph text. Paragraph text. Paragraph text.
educational institutions for children with hearing impairments to ignoring the existence of the
language and even completely banning it [3].</p>
      <p>The hearing-impaired community is often embarrassed by their difficulties in communicating
with the rest of the world. Although sign language is used as a means of conveying its message,
there are still problems in communication because there are few people who are familiar with
this type of language. In addition, the number of available translators is insufficient to solve the
problem. This motivated scientists from different countries to study this problem and work on it.
In general, this issue can be divided into two parts: the first involves the development of
automatic sign language synthesizers that allow people with hearing impairments to understand
messages transmitted by people who can hear; the second part - the opposite - concerns the
development of sign language interpreters that allow the hearing community to understand sign
language [4]. In view of the mentioned patterns, the urgent task is to develop an information
system for recognizing Ukrainian sign language, which will provide additional means of
overcoming the language barrier between communication subjects.</p>
      <sec id="sec-1-1">
        <title>1.1. Analysis of recent researches and publications</title>
      </sec>
      <sec id="sec-1-2">
        <title>1.1.1. Analysis of known research</title>
        <p>Although sign language has been a subject of study for centuries, it was only at the end of the
last century that the subject became the focus of linguistic research. This was facilitated by the
publication of William K. Stoke "The Structure of Sign Language", which marked the beginning of
the linguistics of sign language. The proposed structure consisted of 55 symbols, which formed
three groups according to the parameters (place of execution of the gesture, nature of movement
and shape of the hand), which Stoke considered relevant in determining the structure of the
gesture. Stoke's notation formed the basis of the organizing principles of the first dictionary of
American Sign Language [5].</p>
        <p>As the analysis showed, the existing methods of gesture recognition in computer systems are
divided into two types - recognition based on the creation of a 3D model and methods built on
the principle of feature selection [6,7]. The first class of methods is based on the creation of a
kinematic model. This model must take into account each of the possible degrees of freedom.
When building such a model, hand gestures are evaluated using a comparison of hand coordinates
on the image. Methods of this type make it possible to recognize a significant number of gestures,
but when implementing them, you need to create a large-scale database with images. Images from
the database will also be used to resolve conflicts during feature selection that arise due to
various shapes and sizes of recognition objects. The second class of methods is based on the
processing of details of the input data stream, which are designed to determine the coordinates
of the object of recognition. This method can be applied only if it is possible to determine
characteristic anchor points or features on the images of objects. Then the object itself can be
defined as a combination of these points or planes that they form. In this case, instead of creating
a complete object, a subset of its characteristic points or areas is created. This approach is
resistant to deformations and changes in input sequences. In the presence of characteristic
features, the object can always be unambiguously classified [8]. A separate approach to gesture
recognition is a method based on artificial neural networks [9]. Convolutional neural networks
can successfully identify individual dactyls, but this applies only to static gestures, the analysis of
dynamic movements based on images is too cumbersome and resource-consuming [10].</p>
        <p>In general, scientific research in this direction can be presented in the form of the following
publications:
• "Sign language recognition using Microsoft Kinect" [11] authors developed a method for
recognizing sign language using depth images from the Kinect sensor. The depth and motion
profile are calculated from the generated images and used to construct a feature matrix for
each gesture. Recognition is performed on the basis of a linear classifier based on the method
of support vectors.</p>
        <p>In the work "Multi-sensor data fusion for sign language recognition based on dynamic
Bayesian network and convolutional neural network" [12], a multi-sensor fusion structure
based on convolutional neural network and dynamic bayesian network for sign language
recognition is proposed. In this framework, Microsoft Kinect, which is an RGB-D sensor, is used
as a human-computer interaction tool. In particular, in the proposed approach, data is first
collected using Kinect, then all features of the image sequence are extracted using a
convolutional neural network. Sequences of color and depth features are input to the DBN as
observation data. The maximum level of recognition of dynamic isolated sign language is
calculated based on the union of the graph model.
• In the work "A Real-time Hand Gesture Recognition System for Human-Computer and
Human-Robot Interaction" [13], the proposed gesture recognition system is designed to
improve human-computer interaction and human-robot interaction. As the authors of the
study assure, such interaction ensures natural and intuitive communication between people
and technology using gestures.
• The robot "3D Dynamic Hand Gesture Recognition with Fused RGB and Depth Images"
[14] offers dynamic gesture recognition technology. In order to solve the existing technology
problems, the authors propose to use a network model of three-dimensional dynamic gesture
recognition, which uses CNN and LSTM networks and can combine information about RGB and
image depth.
• In the work "Hand gesture recognition using convolutional neural network and histogram
of oriented gradients features" [15], the authors emphasize that gesture recognition is the
main part of creating a sign language recognition system for people with hearing impairments
and is widely used in human-computer interaction. The selected dataset for building the
gesture recognition system model is based on American Sign Language using pre-trained
AlexNet Convolutional Neural Network and Oriented Gradient Histogram.
• In the work "Mid-air Gesture Recognition by Ultra-Wide Band Radar Echoes" [16], the
authors propose the technology of using microwave radar sensors for human-computer
interaction. The peculiarity is that the raw signals generated by such radars have a large
dimension and are very difficult to process and interpret for gesture recognition. For these
reasons, machine learning techniques are mainly used for gesture recognition, but require
numerous gesture patterns for training and calibration, which are specific to each radar [17].</p>
        <p>The given list of studies that are used in the process of sign language recognition is not
exhaustive. But the conducted analysis shows that the ideal method does not exist and is unlikely
to exist. Therefore, it can be concluded that the mentioned approaches can be ensured by their
further adaptation for Ukrainian-language content.</p>
      </sec>
      <sec id="sec-1-3">
        <title>1.1.2. Analysis of the Ukrainian sign language development</title>
        <p>In many countries of the world, the possibility of creating and popularizing translators from
audio language to sign language and vice versa is being investigated. However, the problem of
translating sign language into Ukrainian-language audio content still remains unresolved. It is
worth noting that Ukrainian sign language, like any other sign language, has its own rules and
grammar, which in turn does not allow the use of existing dictionaries of foreign sign languages.
On the territory of modern Ukraine, sign language began to develop in the 19th century - the time
of the founding of the first communities, that is, Ukrainians have been creating their own sign
language for about two centuries. In 1830, the Lviv school for hearing-impaired children was
opened, and in 1843 - in Odesk - these are the approximate dates of the beginning of the
development of Ukrainian sign language [18].</p>
        <p>It was only recently that sign language was recognized and equated with verbal language. UN
General Assembly Resolution 48/96 of December 20, 1993, "Standard Rules for Ensuring Equal
Opportunities for Persons with Disabilities," stated that care should be taken to ensure that sign
language is used in the education of deaf children, in their families and communities, and it was
also recommended to provide sign language interpretation services to facilitate the
communication of sign language people with other people. Subsequently, the issue of using sign
language became more active in Ukraine, but the use of Ukrainian sign language in education in
independent Ukraine was not introduced until 2006 [19].</p>
        <p>The study of sign language linguistics in Ukraine was started by R. Kraevskyi. The
speechlanguage pathologist worked on the study of sign language, thus carried out its linguistic
description on the basis of Ukrainian studies material and created a unique sign dictionary in the
form of a manual "Sign Language of the Deaf" [20]. For each gesture, the spatial position and way
of movement of the hands are described. In the 21st century in Ukraine, N. Adamyuk, O. Drobot,
S. Kulbida, O. Lozynska, M. Davydov are engaged in studying the peculiarities of the syntax of
Ukrainian sign language.</p>
        <p>Most of N. Adamyuk's scientific works are aimed at studying the peculiarities and linguistic
didactic technologies of teaching Ukrainian sign language to deaf and hard-of-hearing children,
studying the linguistic features of Ukrainian sign language, as well as studying the basic
requirements for teachers of sign language in higher educational institutions and an innovative
model their training and retraining [21]. The works of O. Drobot are devoted to the formation of
communication skills and comprehensive development of preschool children with hearing
impairment [22]. S. Kulbida's research is related to deaf pedagogy of the socio-cultural direction
and the conceptual foundations of the development of Ukrainian sign language as a means of
learning and a subject of study [23]. The researches of O. Lozynska and M. Davydov are related in
their main emphasis on the translation of Ukrainian sign language based on ontology [24, 25].</p>
        <p>The analysis of the completed work shows significant progress in popularizing the study of
Ukrainian sign language, but the lack of problem-oriented software solutions makes its further
research an urgent task.</p>
      </sec>
      <sec id="sec-1-4">
        <title>1.2. The main tasks of the research and their significance</title>
        <p>The purpose of the research is to develop an information system for the recognition of
Ukrainian sign language. The conducted research will provide means for creating on its basis
software for managing information and reference content, generating/transforming elements of
sign language and forming an individual learning environment for people with special needs. To
achieve the goal, the following tasks must be solved: analyze the existing approaches, methods
and software tools used in the field of Ukrainian sign language recognition; to determine the main
tasks that arise at the same time; analyze the methods and algorithms of sign language
recognition that can be adapted during system development; implement a prototype system for
recognizing Ukrainian sign language.</p>
        <p>The results of the study solve the actual scientific and practical problem of recognizing
Ukrainian sign language and will provide the means to open up additional opportunities for
individualizing the educational process for people with special needs.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Major research results</title>
      <p>People are constantly faced with the task of object recognition. Namely, the human brain
processes the information received from the senses, on the basis of which an appropriate decision
is made. After that, thanks to the transmission of electrochemical impulses, certain organs carry
out the decision made. The above process will occur every time there is a change in the
environment. A key stage in this process is the recognition and classification of the surrounding
environment, which will help to make the right decision. Given the current development of
computer technology, pattern recognition tasks have become the beginning of an independent
field and a multitude of tasks that can be solved using gesture recognition.</p>
      <p>In order to present the main aspects of the studied subject area, a scheme was finalized that
reflects the main stages that must be implemented in the gesture recognition system (Fig. 1).</p>
      <p>As can be seen from Figure 1, the main tasks in the process of gesture recognition are:
• Obtaining an image - usually, this process is implemented using two or more
synchronized infrared cameras or smartphone cameras, which continuously transmit a video
stream to the system in real time (25-30 frames/sec.);
• Localization of the hand area in the image – on each frame (or series of frames) obtained
from the video stream of the camera, the area on which the hand is located is determined. This
procedure mainly consists of two stages. The first stage is segmentation (selection) and
analysis from the received data of the hand area. This process is performed to remove artifacts
from the image and separate the hand region in the image from the background region. As a
result of this stage, a selected image of the hand suitable for further processing is formed in
the system.
• Gesture recognition – at this stage, the contours of the hand and its characteristics are
determined on the image obtained as a result of localization of the hand region. Based on the
received data, the gesture is classified.</p>
      <p>Let's consider in more detail the main tasks and stages of gesture recognition. The first step is
to obtain images that are processed to separate the hand region from the background [26]. This
phase is called object localization in the image. After collecting the information, it becomes
possible to apply the primary information about the hand in order to filter the data and remove
noise from the image. Noises can appear, for example, due to changes in lighting. Artifacts (such
as the presence of tattoos, jewelry, etc. on the hand) are also removed. This procedure is very
important, considering the set of gestures that need to be distinguished. At the phase of
recognition of hand gestures, feature selection is performed [27]. This stage is an important part
of the recognition process because hand movements have a significant number of shapes and
textures. To recognize a static image of a hand, geometric features are used, among them: the
location of the fingertips and their direction. The problem is that these features are not always
available due to self-shading and lighting features. Next comes the stage of identifying specific
gestures using methods of analyzing filtered data that carry information about hand movements.
For this, the classification procedure is used. Before this stage, it is necessary to carry out a
process of training the system to enable it to respond to gestures, and to carry out their
adaptation for the correct detection of movements [28]. To create a comfortable environment for
the user, all processes of capturing, classifying and transforming gestures into text instructions
must be performed in real time with an update rate of 25-30 frames per second.</p>
      <p>Further work was aimed at conducting a systematic analysis of the subject area using the
methodology of functional modeling and graphic description of processes. For these purposes, a
structural approach and the IDEF0 standard, which is intended for the formalization and
description of business processes, were used. A context diagram showing the process of
recognizing Ukrainian sign language is presented in Fig. 2.</p>
      <p>In the specified model, the input receives information about the gesture, the value of which
must be displayed on the screen. The gesture can be transferred in the form of a photo or a video
sequence. The output data of the system is the recognized gesture and its text value of the gesture.
The driving influences are: image capture methods (methods and algorithms needed to capture
an image and localize a gesture on the captured image. These methods are based on the analysis
of external features of the gesture); image processing methods (methods and algorithms needed
to process the image and extract the outline of the gesture for further analysis. A contour is a
curve of a function of two variables along which the function has a constant value. Contours are
straight or curved lines that describe sharp changes in brightness in the image [ 29]. There is a
high probability of obtaining more than one contour, which is formed in the image due to the
presence of noise in the background. Methods for processing the image are necessary in order to
remove excess noise from the image and select a clean contour of the gesture for further analysis);
the rules of the Ukrainian sign language (information about the gestures of the Ukrainian
language. As it is known [30,31], the Ukrainian sign language differs from other sign languages.
The rules of the Ukrainian sign language are necessary to highlight the specific features of each
gesture. This information is used during image analysis and the formation of the original result).
Smartphone cameras or computer web cameras act as mechanisms.</p>
      <p>For a more detailed understanding of the logic of the processes taking place in the gesture
recognition system, the developed context diagram was decomposed into several sub-processes.
The decomposition diagram is presented in Figure 3. As can be seen from the decomposition
diagram, the entire process of gesture recognition has been broken down into several
subprocesses for greater detail and understanding. Each of the sub-processes has its own input data,
output data, control influences and mechanisms necessary for the operation of the process. The
entire gesture recognition system is divided into the following three sub-processes: image
capture process; image processing process; image analysis process. Image capture is the first
subprocess of the entire system. At this stage, the input device converts the gesture into digital form
and transfers it to the image processing unit. Image processing is the second sub-process in the
entire system. At this stage, the image is processed in such a way that the outline of the hand is
clearly visible. The result of the recognition of the gesture of the Ukrainian sign language depends
on the result of this process. An incorrectly selected processing algorithm or an incorrectly set
parameter of the selected algorithm (for example, the binarization threshold for the image
binarization algorithm) will lead to a poor-quality selection of the hand contour, which in turn
will not allow accurate identification of the gesture. After successful image processing, it is time
for the final stage of gesture analysis. It is here that the image of the gesture is translated into text.
Considering the described stages, the most important stages of the work are contour selection
and actual recognition of gestures.</p>
      <sec id="sec-2-1">
        <title>2.1. Contour selection methods</title>
        <p>To highlight the contour, you can use a number of methods: image binarization, wavelet
transformation, Canny edge detecto algorithm. In order to choose the optimal one for this task,
we will analyze them. The process of binarization is the conversion of a color image or a grayscale
image into two-color black and white. The main parameter of this transformation is the threshold,
with the value of which the brightness is then compared. After comparing a single image pixel, it
is assigned one of two possible values: 0 - "object boundary" or 1 - "another area" [32]. The main
goal of binarization is to reduce the amount of information you have to work with. Successful
binarization greatly simplifies further work with the image. There are various methods of
binarization, which can be conditionally divided into two groups: global (threshold); local
(adaptive). Global binarization methods work with the entire image at once. Threshold methods
of binarization include: binarization by the lower threshold; upper threshold binarization;
double-constrained binarization; incomplete threshold processing; multilevel boundary
transformation [33].</p>
        <p>
          0, F(m,n) ≥ t,
F'(m, n)= # 1, F(m,n) &lt; t
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
        </p>
        <p>If the first condition is fulfilled for the image point in the given formula 1, then such a point is
an object point, if the second condition is fulfilled, then the point will be a background point. In
some cases, you can use a variant of the binarization method with a lower threshold [34], which
results in a negative of the original image. This method is called binarization with an upper
threshold and is represented by the formula:</p>
        <p>0, F(m,n) ≤ t,</p>
        <p>
          F'(m, n)= # 1, F(m,n) &gt; t (
          <xref ref-type="bibr" rid="ref2">2</xref>
          )
        </p>
        <p>
          If it is necessary to highlight certain areas in which the brightness values of pixels can vary in
a certain range, then the binarization method with a double constraint is used [35]. This method
is called binarization with an upper threshold and is represented by the formula:
0, F(m,n)≥ t1,
F'(m, n)= $1,t1&lt; F(m,n) ≤ t2 (
          <xref ref-type="bibr" rid="ref3">3</xref>
          )
        </p>
        <p>0, F(m,n)&gt;t2</p>
        <p>If it is necessary to obtain the simplest image for further analysis, then it is worth applying the
incomplete threshold processing algorithm, during which the image is deprived of the
background with all its details that were in the original photo. Incomplete threshold binarization
is represented by the formula:</p>
        <p>F(m,n), F(m,n) &gt; t,</p>
        <p>
          F'(m, n)= # 0, F(m,n) ≤ t (
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
        </p>
        <p>If you need to get an image that contains segments with different brightness, you can apply
the method of multi-level threshold transformation. However, at the same time, the image
obtained during the transformation will no longer be binary [36].</p>
        <p>The formula for this transformation is presented below:</p>
        <p>
          1, F(m,n) ϵ D1,
⎧⎪ 2, F(m,n) ϵ D2,
F'(m, n)= … (
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
⎨ n,F(m,n) ϵ Dn
⎩⎪0, в усіх інших випадках
        </p>
        <p>The conducted analysis showed that, taking into account the peculiarities of the input
information, it is advisable to use a single binarization threshold, which is used to divide into
black and white. The result of the threshold binarization method is shown in Figure 4.</p>
        <p>
          Wavelet transforms are effectively used in signal compression and spectrum analysis [37].
Virtually all wavelets are traditionally defined as functions of a single real variable. Depending on
the mathematical model (the structure of the domain of definition, the structure of the domain of
possible values and the type of transformations), discrete and continuous wavelets are
distinguished. Since the decomposition of wavelets is carried out using floating-point arithmetic,
inaccuracies may occur, the magnitude of which is affected by the degree of approximation of the
signal. Taking into account the specifics of the subject area, it is possible to use the Haar wavelet
[38]. Its technical drawback is that it is not continuous, and therefore not differentiable. However,
this property is an advantage when analyzing signals with sudden transitions (discrete signals)
that are inherent in this area. In the traditional setting, the wavelet transformation in the Haar
basis consists in the linear transformation of a vector of even dimension into another vector of
the same dimension. Each pixel of the image can be represented in a binary number system. This
decomposition determines the number of bits (N, usually N = 1, 8, 24) and their specific values
for storing each pixel [39].
(
          <xref ref-type="bibr" rid="ref6">6</xref>
          )
        </p>
        <p>
          To apply the wavelet transformation over the field GF (p), each pixel of the image must be
represented in some number system. This decomposition determines the number of digits of the
number system and their specific values that are used in the wavelet transform. The algorithm of
order wavelet-transformation of the image is carried out according to the following stages: each
pixel of the image is decomposed according to (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ) into the digits of a certain numbering system
p. A transformation is applied to all p digits with the same numbers. The digits of a certain
        </p>
        <p>
          J= ∑ kN=-10 Jk∙ 2k
numbering system p of the transformation result are folded into one number according to (
          <xref ref-type="bibr" rid="ref6">6</xref>
          ). In
fig. 5 presents the variants of the initial image and the image after wavelet transformation by
rows in the Haar basis over the field GF (
          <xref ref-type="bibr" rid="ref3">3</xref>
          ) and GF (13).
        </p>
        <p>The Canny edge detector algorithm was developed taking into account such criteria as fast
detection and good contour localization. Based on these criteria, an objective function of the cost
of errors was constructed, the minimization of which is the "optimal" linear operator for image
convolution [40]. In general, Kenny's algorithm consists of five stages.</p>
        <p>1. Smoothing. At this stage, the image is blurred using a Gaussian filter for localization and
noise removal [41].
2. Search for gradients. Boundaries are searched - where the gradient reaches its maximum
value, the boundaries are there:</p>
        <p>The angle of the direction of the gradient vector is rounded and can take the following values:
0, 45, 90, 135. If the angle is from 1 to 20, then it refers to the value 0, and if it is greater than 20,
then to the value 45, etc.</p>
        <p>3. Muting the lows. Only local maxima are marked as limits.
4. Double threshold filtering. Potential limits are determined by thresholds.</p>
        <p>5. Tracing the area of ambiguity. End boundaries are defined by muting all edges not connected
to certain (strong) boundaries.</p>
        <p>Before applying the detector, the image is usually converted to shades of gray to reduce
computational losses. The contour detector algorithm is not limited to calculating the gradient of
the smoothed image. Only the points of maximum gradient of the image remain in the contour of
the border, and all others lying next to the border are removed. The inclusion of noise suppression
in the Kenny algorithm, on the one hand, increases the stability of the results, and on the other
hand, increases the computational costs and leads to distortion and even loss of contour accuracy.
The result of the Kenny algorithm is shown in Figure 6.</p>
        <p>
          T = +Gx2+ Gy2
θ=arctan( GGyx )
(
          <xref ref-type="bibr" rid="ref8">8</xref>
          )
(
          <xref ref-type="bibr" rid="ref9">9</xref>
          )
        </p>
        <p>Comparing the results of image processing by the mentioned algorithms showed that the
image binarization method works faster than the wavelet transformation and the Kenny
algorithm. However, it should be noted that clearer borders of objects in the image are obtained
during processing based on the application of the Kenny algorithm. However, to detect a
highquality contour of the palm, the image binarization method is quite sufficient. In view of that, the
image binarization method will be applied in further work.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Gesture recognition</title>
        <p>There are many methods that can be used to recognize gestures, among the most common are
methods based on the hidden Markov model and neural networks. The hidden Markov model
[42] is a statistical model in which the system for which it is created is represented as a Markov
process with invisible states. The model can also be represented as the simplest Bayesian
network. The main application of hidden Markov models was in the field of recognition of images
(gestures), speech, writing and bioinformatics. In addition, they are used in cryptanalysis,
machine translation. The simplified structure of the hidden Markov model is represented by the
following elements: ovals (these are variables that have random values, namely, the random
variable x(t) is the value of the hidden variable at the time t, and the random variable y(t) is the
value of the observed variable at the time t); arrows (indicate conditional dependencies).</p>
        <p>
          The probability of finding a sequence Y = y(0), y(
          <xref ref-type="bibr" rid="ref1">1</xref>
          ), … , y(L-1) of length L is determined by the
dependence:
        </p>
        <p>P(Y)= ∑ ! ( | ) () (10)</p>
        <p>This modeling technology gained considerable popularity as a result of its successful
application and further development in the field of automatic recognition of speech and gestures.
Research on hidden Markov models has outperformed all competing approaches, and is the
dominant processing paradigm. Their ability to describe processes or signals has been
successfully studied for a long time. The reason for this, in particular, is that the technology of
building artificial neural networks is rarely used for gesture recognition and similar
segmentation problems. However, there are a number of hybrid systems that consist of a
combination of hidden Markov models and artificial neural networks, in which the advantages of
both modeling methods are used [43].</p>
        <p>In general, hidden Markov models describe a two-stage stochastic process. The first stage
consists of a discrete stochastic process that is static, causal, and simple. The state space is
considered finite. Thus, the process probabilistically describes the state of transition to
discreteness, a finite space of states. It can be visualized as a finite automaton with transitions
between any pairs of states that are denoted by the transition probability. The behavior of the
process at the current moment of time t depends only on the immediate state of the previous
element and can be determined by the dependence:</p>
        <p>P ("|#, $…"%# ) =P ("|"%#) (11)</p>
        <p>At the second stage, for each moment of time t, additionally, by derivation or on the basis of
output data, Оt is generated. The associative probability distribution depends only on the current
state St, not on any previous states or inputs.</p>
        <p>P ("|#… "%#, #…" ) =P ("|") (12)</p>
        <p>The specified sequence of output data is the only thing that can be observed in the behavior of
the model. On the other hand, the sequence state assumed during data generation cannot be
examined. This is the so-called "hiddenness" from which the definition of hidden Markov models
is derived. If you look at the model from the outside - that is, observe its behavior - quite often
there are references to the sequence of initial states O1, O2 ... Ot, as the reason for observing the
sequence. Individual elements of this sequence are called observation results [44].</p>
        <p>In the literature, behavior recognition patterns of the hidden Markov model are always
considered at a certain time interval t. To initialize the model at the beginning of this period,
additional probabilities are used to describe the probability distribution of states at time t = 1. An
equivalent final state criterion is generally absent. Thus, the action of the model enters the final
state as soon as an arbitrary state is reached at the time t. As for gesture recognition, in order to
reliably determine the semantics of the movement, it is necessary to allocate it to one of the
classes of gestures. Next comes the stage of calculating the probability of receiving a read gesture
from models of available gestures. Then the received gesture is classified using the Bayesian
classifier. Based on the classification, the gesture can be recognized as one of the available
options.</p>
        <p>The task of determining the end of a gesture is also not easy. For this, edge cases are considered
[43]. When using this classification algorithm, it is highly undesirable to obtain unclear values
(data about movements that cannot be clearly attributed to a certain class of gestures). To reduce
the number of errors, in the algorithm described above, when situations arise that cannot be
unambiguously attributed to a certain class, a weighted sum of the consequences of performing
all the gestures classified by the algorithm should be used, or one of the gestures classified with
the highest probability should be selected.</p>
        <p>As for neural networks, the main research and scientific results obtained in the field of their
application for gesture recognition include various methods and architectures that allow to
perform this task effectively. In general, since an artificial neural network usually learns with a
teacher, this means the presence of a training set (dataset). Ideally, this set contains examples
with true values: tags, classes, metrics. An artificial neural network consists of three components:
an input layer; hidden (computing) layers; source layer [45]. Neural network training takes place
in two stages: direct error propagation; error back propagation. During direct error propagation,
a response prediction is made. In backpropagation, the error between the actual response and
the predicted one is minimized. Initial weights are randomly assigned. Next, the input data are
multiplied by weights to form a hidden layer [46]:
h1=(x1∙w1)+(x2∙w1) (13)
h2=(x1∙w2)+(x2∙w2) (14)
h3=(x1∙w3)+(x2∙w3) (15)</p>
        <p>The output data from the hidden layer is passed through a nonlinear function (activation
function) to obtain the output of the network:</p>
        <p>y=f(h1,h2,h3) (16)</p>
        <p>During the backpropagation of the error, the total error is calculated as the difference between
the expected value from the training set and the obtained value (calculated at the stage of forward
error propagation), passing through the loss function. The derivative of the error is calculated for
each weight (these differentials reflect the contribution of each weight to the total error). These
differentials are then multiplied by the learning rate number. The obtained result is then
subtracted from the corresponding weights. As a result, the following updated weights will be
obtained:
w1=w1-(η∙ ∂(err) )</p>
        <p>∂(w1)
w2=w2-(η∙ ∂(err) )
∂(w2)
(17)</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. System design</title>
        <p>The next stage was the construction of the system, using modern software tools. To implement
the software product, it was decided to use the C# programming language and the .NET
crossplatform technology. We will use Visual Studio as a development environment. To work with a
single-camera system and process the image for further analysis, the OpenCV library is used, or
rather, the C# version of EmguCV. The Math.NET library is used to perform matrix operations.
The constructed prototype system for recognizing Ukrainian sign language can be conditionally
divided into several main independent parts: HandGesturesRecognitionForm, NeuralNetwork,
CsvManager, TrainingImageDataManager.</p>
        <p>HandGesturesRecognitionForm is the main class of the program, it contains methods for
working with the form and analyzing and processing images. The constructor of the
HandGesturesRecognitionForm class initializes all components located on the form: fields,
buttons, menus, switches. Next, an object of the VideoCapture class is created, which is a class
from the EmguCV library designed to capture an image from the device's camera. The Rectangle
object is used to position and size the red rectangle in the video from which the gesture will be
read for recognition. The result of the recognition zone reproduction is presented in Fig. 7.</p>
        <p>NeuralNetwork – a class that represents a neural network, contains information about the
number of nodes of the input, output and hidden layers, the learning coefficient, the matrix of
weights between the input and hidden layers, the matrix of weights between the hidden and
output layers. To create it, you need to set the following mandatory parameters:
• inputLayerNodesCount – the number of input layer nodes. In this case, a value of 4096 is
passed, which corresponds to the value of each pixel of the 64 by 64 binary image.
• hiddenLayerNodesCount – the number of hidden layer nodes.
• outputLayerNodesCount – the number of output layer nodes.
• learningRate is a learning rate, a parameter of gradient learning methods of neural
networks, which allows you to control the amount of weight correction at each iteration.
• epochs – the number of steps (epochs) required to find the optimal value.</p>
        <p>CsvManager is a class responsible for saving the training set to a csv file. Contains private file
name and save path information.</p>
        <p>TrainingImageDataManager is a class that is responsible for saving pictures for neural
network training. It contains private information about how photos are saved.</p>
        <p>Let's take a closer look at the implementation of some key methods of the class. The EmguCV
Threshold library method is used for binarization. The method accepts a binarization threshold
that is obtained from a control of the form binaryImageThresholdTrackBar. The result of image
binarization is shown in Figure 8.</p>
        <p>The DrawContours library method is used to select the contour. When the method is called,
the contour color and its thickness are set. The result of the method is shown in Figure 9.</p>
        <p>The NeuralNetwork class is designed to create a neural network object, train it, and poll it. The
class contains information about the nodes of the input, output, and hidden layers, the values of
certain learning coefficients, and methods for training and polling a neural network. The
structure of the NeuralNetwork class is shown in Figure 11.</p>
        <p>When creating an object of the neural network class, the weight matrices are initialized. Large
values should be avoided when initializing the initial values of the weights, as using the activation
function in this range of values may reduce the ability of the network to learn to better values.
Therefore, the weights are selected from a normal distribution centered at zero and with a
standard deviation whose value is inversely proportional to the square root of the number of
input nodes. The Query method accepts the input data of the neural network as an argument and
returns its output data. To do this, signals from the input layer nodes must be passed through the
hidden layer to the output layer nodes to receive the output data. At the same time, as the signals
spread, it is necessary to smooth them using the weighting coefficients of the connections
between the relevant nodes, and also to apply the sigmoid to reduce the output signals of the
nodes. To obtain the output signals of the hidden layer, it is necessary to apply them to each
sigmoid value. Training includes two phases: the first is the calculation of the output signal, which
is what the Query function does, and the second is the backpropagation of errors, which informs
what the corrections to the weighting factors should be.</p>
        <p>The first part is the calculation of output signals for a given training example. The second part
is a comparison of the calculated output signals with the desired response and updating the
weighting coefficients of connections between nodes based on the differences found.</p>
        <p>As a result of the work, a prototype of the application was developed, which is able to recognize
the gesture of the alphabet of the Ukrainian sign language. For clarity, the program outputs the
result at each iteration, starting with the raw video and ending with the recognition result in the
form of a gesture value. It all starts with video capture. An example of a frame from the original,
unprocessed video stream and its binarized version is shown in Fig. 12.</p>
        <p>By capturing a binarized image, you can shape the outline of the palm. The output contour
together with the selected region of the largest closed contour highlighted by a rectangle is shown
in Figure 13.</p>
        <p>After selecting the working surface, it is necessary to reduce it to a square image, since the
selected largest contour is not always a square, as shown in Figure 13. Since the neural network
contains 4096 input layers, the final image is reduced to a size of 64 by 64 pixels. After capturing
the image and clicking on the "Recognize the gesture" button, the recognition settings panel looks
like this:</p>
        <p>The text value of the gesture is displayed in the "Recognition result" text field. The program
successfully recognized the demonstrated gesture and displayed its explanation on the screen.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Conclusion</title>
      <p>As a result of the conducted research, the existing methods and known systems that provide
means of recognizing Ukrainian sign language and describe the mechanisms of their
implementation were analyzed. Technologies and software tools for sign language recognition
were analyzed, which made it possible to identify the features of existing approaches. As the
analysis showed, today there are many software systems, but all of them are characterized by
certain shortcomings, from the commerciality of the application to the impossibility of the
application for the recognition of Ukrainian-language content, which makes the task of
constructing an information system for the recognition of Ukrainian sign language urgent. In
order to present the main aspects of the studied subject area, a scheme was finalized that reflects
the main stages that must be implemented in the gesture recognition system. The next stage was
the design of the software system using a structural approach and displaying the created
diagrams in accordance with the IDEF0 standard. The study presents a context diagram and its
decomposition, which created the basis for the study of features and the formation of
methodological foundations for the construction of an information system. The analysis and
justification of the choice of methods for the selection of contours and recognition of gestures in
the incoming information message was carried out. The developed prototype is characterized by
modular construction, the ability to recognize gestures of the Ukrainian alphabet and can be
useful as an additional communication tool. The conducted research provides methodological
and algorithmic foundations for building a communication environment for people with special
needs.</p>
      <p>Further research will be directed to testing and improving the system, eliminating conflicts
and expanding functionality in accordance with the specified requirements.
International conference on computational linguistics and intelligent systems (COLINS
2022). Vol. 1 : Main conference, Gliwice, Poland, May 12-13, 2022, pp. 1282–1296.
[10] D. Zhu, V. Czehmann, E.Avramidis, Neural Machine Translation Methods for Translating Text
to Sign Language Glosses. Proceedings of the 61st Annual Meeting of the Association for
Computational Linguistics. Vol.1: Long Papers, 2023, Toronto, Canada, pp. 12523–12541.
[11] A. Agarwal, M. Thakur, Sign language recognition using Microsoft Kinect, Proceedings of the
sixth International Conference on Contemporary Computing (IC3), Noida, India, 2013, pp.
181-185.
[12] Q. Xiao, Y. Zhao, W. Huan, Multi-sensor data fusion for sign language recognition based on
dynamic Bayesian network and convolutional neural network. Multimed Tools Appl. 2019.</p>
      <p>Vol. 78, pp. 15335–15352.
[13] V. Ponzi, E. Iacobelli, C. Napoli, J. Starczewski, A Real-time Hand Gesture Recognition System
for Human-Computer and Human-Robot Interaction. Proceedings of the International
Conference of Yearly Reports on Informatics, Mathematics, and Engineering, Catania, Italy,
August 26-29, 2022, pp.52-58.
[14] Y. Qingshan, B. Yong, C. Lu, J. Wenjie, 3D Dynamic Hand Gesture Recognition with Fused RGB
and Depth Images. Proceedings of the 2022 3rd International Conference on Big Data &amp;
Artificial Intelligence &amp; Software Engineering, Virtual Event, Guangzhou, China, October
2123, 2022, pp.38-44.
[15] A. Kika, A.Koni, Hand gesture recognition using convolutional neural network and histogram
of oriented gradients features. Proceedings of the 3rd International Conference on Recent
Trends and Applications in Computer Science and Information Technology Tirana, Albania,
November 23rd to 24th, 2018, pp.75-79.
[16] A. Sluÿters, Mid-air Gesture Recognition by Ultra-Wide Band Radar Echoes. Proceedings of
the Workshops on Engineering Interactive Computing Systems (EICS-WS 2022) co-located
with teh 14th ACM SIGCHI Symposium on Engineering Interactive Computing Systems
(SIGCHI 2022). Sophia Antipolis, France, June 21, 2022, pp.28-39.
[17] A. Vasyliuk, T. Basyuk, V. Lytvyn. Specialized interactive methods for using data on radar
application models, Proceedings of the 2nd International workshop on modern machine
learning technologies and data science (MoMLeT+DS 2020). Vol. I: Main conference,
LvivShatsk, Ukraine, June 2-3, 2020, Vol. 2631: pp. 1-11.
[18] Association of deaf teachers. Educational institutions for the deaf in the pre-revolutionary
period. URL.
https://onp.ucoz.ua/news/navchalni_zaklady_dlja_gluxyx_v_dorevoljuciyniy_period/201306-13-51. (In Ukrainian)
[19] S. Kulbida, Gesture bilingual approach in the practice of special institutions of Ukraine.</p>
      <p>Special child: training and education. - 2022. - N 3. - P. 7-18. (in Ukrainian).
[20] R. Kraevskyi, Sign Language of the Deaf. – K. 1964. P.220. (in Ukrainian).
[21] N. Adamyuk, Features of socio-cultural communication of sign language people in the
educational process. Abstracts of XXX International Scientific and Practical Conference
Interaction Of Society And Science: Problems And Prospects. London, England June 15 – 18,
2021, pp. 307-311. (in Ukrainian).
[22] O. Drobot, Psychophysiological features of the formation of the lexical competence of
national verbal languages among students with hearing impairments. Education of persons
with special needs: ways of development, 2019, № 15. pp.67-77. (in Ukrainian).
[23] S. Kulbida, A competent approach in the training of deaf-pedagogical personnel. Modern
technologies for the development of professional skills of future teachers: coll. of science
Proceedings of the First International Internet Conference, October 26, 2017. Uman: FOP,
2017, pp. 122-124. (in Ukrainian).
[24] O. Lozynska, M. Davydov, Information technology for Ukrainian Sign Language translation
based on ontologies. Econtechmod. An International Quarterly Journal. 2015, vol. 04, No. 2,
pp.13–18.
[25] T. Basyuk, A. Vasyliuk, Approach to a Subject Area Ontology Visualization System Creating,
Proceedings of the 5rd International Conference on Computational Linguistics and
Intelligent Systems (COLINS-2021). Volume I: Main Conference, Kharkiv, Ukraine, April
2223, 2021, Vol-2870, pp. 528–540.
[26] D. Khurana, A. Koli, K. Khatter, Natural language processing: state of the art, current trends
and challenges. Multimed Tools Appl 82, 2023, pp.3713–3744.
[27] D. Bragg, O. Koller, M. Bellard, L. Berke, P. Boudrealt, A. Braffort, N. Caselli, M. Huenerfauth,
H. Kacorri, T. Verhoef et al., Sign language recognition, generation, and translation: An
interdisciplinary perspective, arXiv preprint arXiv:1908.08597, 2019, pp16-31.
[28] T. Adugna, A. Ramu, A. Haldorai, A Review of Pattern Recognition and Machine Learning.</p>
      <p>Journal of Machine and Computing. 2024, pp.210-220.
[29] M. Baker, U. Solanki, Artificial Intelligence Models in Pattern Recognition. In: Handbook of</p>
      <p>Artificial Intelligence Applications for Industrial Sustainability. CRC Press. 2024, pp. 18-36.
[30] A. Zamsha, The category of quantity in signs of Ukrainian Sign Language. Proceedings of the
International scientific conference “Current trends and fields of philological studies in the
challenging reality”. Riga, the Republic of Latvia, July 29–30, 2022, pp.268-270.
[31] T. Basyuk, A. Vasyliuk, Peculiarities of an Information System Development for Studying
Ukrainian Language and Carrying out an Emotional and Content Analysis // CEUR Workshop
Proceedings. – 2023. – Vol. 3396: Computational Linguistics and Intelligent Systems 2023:
Proceedings of the 7th International Conference on Computational Linguistics and Intelligent
Systems. Volume II: Computational Linguistics Workshop, Kharkiv, Ukraine, April 20-21,
2023.pp. 279–294.
[32] M. Prodan, C.-A. Boiangiu, Document Image Binarization Process. BRAIN. Broad Research in</p>
      <p>
        Artificial Intelligence and Neuroscience, 2023. Vol. 14(
        <xref ref-type="bibr" rid="ref2">2</xref>
        ), pp.93-114.
[33] F. Kasmin, A. Abdullah, A. Prabuwono, Ensemble of Steerable Local Neighbourhood
Greylevel Information for Binarization. Pattern Recognition Letters. 2017. Vol. 98, pp.8-15.
[34] S. Abdullah, S. Ismail, M. Hasan, P. Shivakumara, Novel Adaptive Binarization Method for
Degraded Document Images. Computers, Materials &amp; Continua 2021, Vol. 67(
        <xref ref-type="bibr" rid="ref3">3</xref>
        ),
pp.38153832.
[35] J. Wu, Z. Li, Y. Liu, Double-Constraint Inpainting Model of a Single-Depth Image, Data, Signal
and Image Processing and Applications in Sensors. 2020, Vol. 20(
        <xref ref-type="bibr" rid="ref6">6</xref>
        ), pp. 345-364.
[36] A. Abubakar, Multilevel Thresholding for Image Segmentation Using Mean Gradient, Journal
of Electrical and Computer Engineering. Vol. 2022 (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), pp.1-9.
[37] A. Osadchiy, A. Kamenev, V. Saharov, S. Chernyi, Signal Processing Algorithm Based on
      </p>
      <p>Discrete Wavelet Transform. Designs 2021, Vol 5(41), pp. 1-13.
[38] X. Guoping, L. Wentao, Z. Xuan, L. Chang, H. Xinwei, W. Xinglong. Haar Wavelet
Downsampling: A Simple but Effctive Downsampling Module for Semantic Segmentation.</p>
      <p>Pattern Recognition. 2023. Vol. 143, pp.678-689
[39] P. Fleet, The Haar Wavelet Transformation. Journal of Computer Engineering. 2019, Vol. 34,
pp.125-181.
[40] X. Qin, A modified Canny edge detector based on weighted least squares. Computational</p>
      <p>Statistics. Issue 1. 2021. pp 641–659.
[41] J. Patel, J. Patwardhan, K. Sankhe, R. Kumbhare, Fuzzy inference based edge detection system
using Sobel and Laplacian of Gaussian operators. ICWET '11: Proceedings of the
International Conference &amp; Workshop on Emerging Trends in TechnologyFebruary. 2011,
pp.694–697.
[42] M. Franzese, A. Iuliano, Hidden Markov Models. Encyclopedia of Bioinformatics and</p>
      <p>Computational Biology. Vol. 1, 2019, pp.753-762.
[43] T. Hiraoka, S. Takase, K. Uchiumi, A. Keyaki, N. Okazaki, Recurrent Neural Hidden Markov
Model for High-order Transition. ACM Transactions on Asian and Low-Resource Language
Information Processing. 2021.Vol. 21, Issue 2. pp 1–15
[44] A. Tur, H. Keles, Evaluation of hidden Markov models using deep CNN features in isolated
sign recognition. Multimed Tools Appl. 2021. Vol. 80, pp 19137–19155.
[45] H. Ahn, J. Kim, J. Shim, J. Kim, Hand Gesture Recognition for Doors with Neural Network.</p>
      <p>Proceedings of the International Conference on Research in Adaptive and Convergent
Systems (RACS '17). 2017, pp.15–18.
[46] G. Murthy, R. Jadon, Hand Gesture Recognition using Neural Networks. Advance Computing</p>
      <p>Conference (IACC). 2019 IEEE 2nd International, pp.134-138.
[47] Z. Bao, T. Liu, Radar micro moving gesture recognition method based on multi-scale fusion
deep network. Proceedings of the 2022 5th International Conference on Artificial
Intelligence and Pattern Recognition (AIPR '22). 2022, pp. 657–663.
[48] H. Alimam,W. Mohamed, A. Selmy, Deep Recurrent Neural Network Approach with LSTM
Structure for Hand Movement Recognition Using EMG Signals. Proceedings of the 2023 12th
International Conference on Software and Information Engineering (ICSIE '23). 2023, pp 58–
65.
[49] L. Li, W. Wei, D. Chen, W. Yang, H. Jiang, Gesture Recognition with Complex Background Based
on Improved Convolutional Neural Network. Proceedings of the 2021 5th International
Conference on Electronic Information Technology and Computer Engineering (EITCE '21).
2021, pp.1345–1349.
[50] L. Gao, L. Zhu, S. Xue, L. Wan, P. Li, W. Feng, Multi-View Fusion for Sign Language Recognition
through Knowledge Transfer Learning. Proceedings of the 18th ACM SIGGRAPH
International Conference on Virtual-Reality Continuum and its Applications in Industry.
(VRCAI '22). 2022, pp 1–9.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Coster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Shterionov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Herreweghe</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. Dambre,</surname>
          </string-name>
          <article-title>Machine translation from signed to spoken languages: state of the art and challenges</article-title>
          .
          <source>Universal Access in the Information Society</source>
          .
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2] World Health Organization.
          <article-title>Deafness and hearing loss, URL</article-title>
          . https://www.who.int/newsroom/fact-sheets/detail/deafness-and
          <string-name>
            <surname>-</surname>
          </string-name>
          hearing-loss
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Núñez-Marcos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Perez-de-Viñaspre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Labaka</surname>
          </string-name>
          ,
          <article-title>A survey on Sign Language machine translation</article-title>
          .
          <source>Expert Systems with Applications</source>
          .
          <year>2023</year>
          , Vol.
          <volume>213</volume>
          : pp.
          <fpage>1</fpage>
          -
          <lpage>28</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Adaloglou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Chatzis</surname>
          </string-name>
          ,
          <string-name>
            <surname>I.Papastratis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Stergioulas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Papadopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Zacharopoulou</surname>
          </string-name>
          ,
          <article-title>A comprehensive study on sign language recognition methods</article-title>
          , arXiv:
          <year>2007</year>
          .12530,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>S. McBurney</surname>
          </string-name>
          ,
          <source>Sign Language: History of Research. Encyclopedia of Language &amp; Linguistics</source>
          ,
          <year>2006</year>
          , pp.
          <fpage>310</fpage>
          -
          <lpage>318</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Quek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>McNeill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bryll</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Duncan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kirbas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>McCullough</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ansari</surname>
          </string-name>
          .
          <article-title>Multimodal Human Discourse: Gesture and Speech, ACM Transactions on Computer-Human Interaction</article-title>
          , vol.
          <volume>9</volume>
          , no.
          <issue>3</issue>
          ,
          <year>2002</year>
          . pp.
          <fpage>171</fpage>
          -
          <lpage>193</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <article-title>Skeleton aware multi-modal sign language recognition</article-title>
          .
          <source>Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2021</year>
          . pp.
          <fpage>3413</fpage>
          -
          <lpage>3423</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Minu</surname>
          </string-name>
          ,
          <string-name>
            <surname>A Extensive</surname>
          </string-name>
          <article-title>Survey on Sign Language Recognition Methods</article-title>
          .
          <source>In Proceedings of the 2023 7th International Conference on Computing Methodologies and Communication (ICCMC)</source>
          , Erode, India,
          <fpage>23</fpage>
          -
          <lpage>25</lpage>
          February 2023; pp.
          <fpage>613</fpage>
          -
          <lpage>619</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>O.</given-names>
            <surname>Mediakov</surname>
          </string-name>
          , T. Basyuk,
          <article-title>Specifics of Designing and Construction of the System for Deep Neural Networks Generation /</article-title>
          / CEUR Workshop Proceedings. -
          <year>2022</year>
          . - Vol.
          <volume>3171</volume>
          :
          <string-name>
            <surname>Computational</surname>
            <given-names>Linguistics</given-names>
          </string-name>
          <source>and Intelligent Systems 2022 : Proceedings of the 6th</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>