<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>December</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Aarti</string-name>
          <email>aarti.1208@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Swathi Gowroju</string-name>
          <email>swathigowroju@sreyas.ac.in</email>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Raju Pal</string-name>
          <email>raju3131.pal@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vaddiraju Swathi</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sirisha Yerraboina</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Jaypee Institute of Information Technology</institution>
          ,
          <addr-line>Noida, Uttar Pradesh</addr-line>
          ,
          <country country="IN">INDIA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Lovely Professional University</institution>
          ,
          <addr-line>Punjab</addr-line>
          ,
          <country country="IN">INDIA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Matrusri Engineering College</institution>
          ,
          <addr-line>Hyderabad, Telangana</addr-line>
          ,
          <country country="IN">INDIA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Sreyas Institute of Engineering and Technology</institution>
          ,
          <addr-line>Hyderabad, Telangana</addr-line>
          ,
          <country country="IN">INDIA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>2</volume>
      <fpage>9</fpage>
      <lpage>30</lpage>
      <abstract>
        <p>A machine and a person can interact through hand gestures by using a hand gesture identification device. Real Time Hand Gesture Recognition (RTHGR) is discussed in this work in order to carry out system control actions as intended. With the help of this application, the user's hand gestures may be detected by the webcam and basic actions can be taken as a result. The user must make a distinct gesture. The webcam records this, recognizing the gesture and carrying out the action in accordance with a list of recognized gestures. This process requires a binary threshold value to recognize gestures. A neural network is employed in the proposed classification process. The efectiveness of this technique for operating various systems will be assessed, and other hand recognition techniques will be compared.</p>
      </abstract>
      <kwd-group>
        <kwd>Gesture recognition</kwd>
        <kwd>CNN</kwd>
        <kwd>YOLO</kwd>
        <kwd>Region of Interest</kwd>
        <kwd>deep learning</kwd>
        <kwd>object recognition</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The world is going through a technological revolution. The rapid advancement in the
technological aspect of human beings is growing faster than ever which has allowed the various
computer systems to indulge in our everyday lives as an integral part of our daily activities.
Various computer systems have diferent methods of interaction with the users. This is known
as HCI in technical term. It is a very important aspect of intractability, usefulness, practicality,
and overall user experience on a computer system. In the past few years, the user experience
is focused more than anything on a system made to be used by humans. This is because
the efectiveness of a system is greatly measured based on how a system interacts with the
user to make the overall experience of using a system easier and more wonderful. In the last
decade, hardware like keyboard, mouse and touch-screen have been crucial in how people
engage with technology. However, new forms of engagement tools have been created as a
result of the quick advancement of technology. In the realm of HCI, technologies like thought
processing, gesture recognition, and speech recognition have advanced significantly. In our
proposed system, gesture recognition is one of these that is covered. In this, hand gestures are
utilized as communication between humans and electronic devices. It difers significantly from
CEUR
Workshop
Proceedings</p>
      <p>ceur-ws.org
ISSN1613-0073
conventional hardware-based techniques, which are capable of achieving human-computer
interaction on a totally other level. The subject of computer vision and image processing has
been substantially altered by convolutional neural networks (CNNs), a significant advancement
in artificial intelligence and machine learning. These neural networks were created expressly
to tackle visual perception, one of the most dificult and inherently human jobs. CNNs have
advanced standard machine learning techniques by imitating the hierarchical and feature-driven
nature of how we, as humans, perceive and recognize patterns in the visual environment. CNNs
were inspired by the complex workings of the human visual system. By doing this, they have
unlocked astonishing ability for machines to understand, categorize, and extract valuable
information from photos in addition to allowing them to ”see” images. CNNs operate on small,
overlapping regions of the image known as receptive fields, which allow the network to capture
local patterns and gradually build a rich hierarchical representation of the input data. CNNs
are a class of deep neural networks distinguished by their unique architecture, which includes
convolutional layers and pooling layers. A wide range of fields have been significantly impacted
by the introduction of CNNs. In addition, they have found use in a variety of fields, including
object identification, facial recognition, autonomous cars, medical picture analysis, and more.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Survey</title>
      <p>
        Since the subject of hand gesture recognition is expanding quickly, as many implementations
employing both deep learning and machine learning algorithms that attempt to identify a
gesture that is exhibited by a human using his/her hand. The study [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] showed the common
and upcoming machine learning designs, CNN, achieved faster rates of successfully perceiving
components at a negligible computational cost. The suggested strategy focused primarily on
instances of movements that existed in pictures with two sets such as with hand gestures
and without hand gestures placement and followed instances of hand obstruction with 24
movements. It used a segmentation algorithm and back propagation algorithm to prepare
the multi-layer propagation, and for the back propagation going backwards from nodes that
produce to input nodes in order to check for faults to have an impact, sorting. Among these
approaches, Hidden Markov Models (HMM) is a well-liked technique that is employed by a
number of other detecting applications. The proposed system in this article refers to checks
and operates with all of the numerous detection versions that are frequently employed by an
application that we have seen by studying and examining other papers such as image, video,
and webcam are all addressed. According to a generic document produced by Francois et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
the posture detection application he refers to employs video and the HMM to do so. The three
aforementioned approaches all identify these features, but whether they are based on CNNs,
RNNs, or some other technique, the primary issue with all of them is that they all employ fitting
techniques, which all make reference to the bounding box [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] that was covered in this study.
The output of what image is being presented is determined by the confidence value that is
the highest, which is derived from the bounding box that represents the data that is detected.
Certain additional tools and methods connected to segmentation, general localization, and
even the union of other diferent areas aid in the accomplishment of the tasks of detecting and
recognizing. Fuzzy based human behavior recognition model was developed based on the body
gestures [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Rahim et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] analysis uses the conversion of a signed language word’s gesture
into text. The skin mask segmentation was used by the authors of this work used basic CNN
model to extract features. The support vector machine algorithm is a kind of supervised learning
method used to address regression and classification problems in machine learning used to
classify the signs’ movements with a 95.28% accuracy using a dataset of 10 movements from one
hand and 8 from both the hands [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Mambou et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] analysis of nighttime both indoors and
outdoors included hand gestures connected to sexual assault. The YOLO CNN architecture was
used to create the gesture recognition system. This architecture extracted hand motions and
then classified bounding box images to provide the assault alert. MOving object classification
was done using eigen faces and optical flow approaches [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. To decode gestures or finger-spelling
from movies that express multiple letters are signed in a series to form meaningful words by
identifying finger spellings in uncut sign language footage is a hard process, Ashiquzzaman et
al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] presented the lightweight spatial prism pooling (SPP) using a CNN model. The model
performed 3 times faster than conventional models and required 65% fewer parameters than
conventional classifiers. A lightweight semantic segmentation, Fast and Acurate Semantic
Segmentation Dilated-Net network was used by Benitez-Garcia et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] in place of Temporal
Segment Networks (TSN), It is predicated on the notion of modeling long-range temporal
structures and Spatiotemporal Shift Modules (SSM). On a dataset of thirteen gestures aimed at
real-time interaction with touch less screens, they proved the efectiveness of the idea. There
are several other CNNs [
        <xref ref-type="bibr" rid="ref12 ref13 ref14 ref15 ref2">12, 13, 14, 2, 15</xref>
        ] that implemented mark-based prediction accurately up
to 98% for various biometric applications. Most of the publications [
        <xref ref-type="bibr" rid="ref16 ref17">16, 17</xref>
        ] concentrate on the
data gathering, surroundings, and hand gesture representation three essential components of
the “vision-based hand-gesture” identification system. We have also evaluated the vision-based
recognition of hand movements system’s performance in terms of recognizing precision. The
prediction accuracy for the signer dependent, CNN was used to train a total of 21 ISL static
alphabets, yielding verification and testing accuracy of 97.34% and training accuracy of 98.50%..
On the other hand, the signer independent’s claimed identification accuracy varies from 50-90%,
with a standard recognition accuracy of 78.2%, according to the studies that were chosen. Musa
et al. developed some models to identify and trace the suspicious activities based on body
movements [
        <xref ref-type="bibr" rid="ref18 ref19">18, 19</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed System</title>
      <p>In this investigation of CNNs, we will enlarge on their structural elements, the guiding principles
for their success, and the various applications that make use of their extraordinary capabilities.
We’ll show how these networks have changed computer vision and ushered in a new era of
artificial intelligence by enabling machines to understand and interact with the visual
environment. CNNs have shown to be quite successful at recognizing hand gestures. In order to
efectively categorize and interpret these motions, CNNs are employed in the context of hand
gesture recognition to automatically extract features from pictures or video frames including
hand gestures. Fig. 1 shows various steps of proposed hand gesture recognition system.</p>
      <sec id="sec-3-1">
        <title>3.1. Data Gathering and Pre-processing:</title>
        <p>A dataset of hand gesture photos or video frames is normally gathered in order to train a CNN
for hand gesture identification. This dataset should contain a variety of hand gestures made
by various people in various situations. Preparing the data is a crucial step before training
and testing. Each image was duplicated for reading and training, for two hands, by flipping
it horizontally, occasionally taking the corresponding image from both hands for making the
set more precise. Proposed system uses YOLO architecture about 230+ images of the dataset,
where 100 images were utilised for the testing by promoting it to 2-fold. 15 more pictures
were also captured and labelled for the testing set. Data pre-processing is essential before
post-processing so that we can identify the kind of data we collected and which parts will be
useful for developing, evaluating, and enhancing accuracy.</p>
        <p>A represents X-Value, B represents Y- Value, C represents Width, and D represents Height.
Table 1 illustration depicts how these files would appear when we label our dataset in order to
train it on the desired model.</p>
        <p>Five diferent traits, each with their own significance, are included in each line. The class ID
is the first item on the left, are the coordinates that define the labelled box around the gesture.
It includes the x-axis and y-axis values, specifying the position of the top-left corner and the
bottom-right corner of the bounding box.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Gesture Segmentation:</title>
        <p>The challenging nature of gesture training increases as a result of the fact that gesture data is
recorded in various places, under various lighting conditions, and at various times of day. The
RGB color space data is transformed into the YCbCr color space in the color image pipeline.
This conversion allows the separation of chroma (Cr) and brightness (Cb), efectively mitigating
interference from brightness characteristics.</p>
        <p>[     ] = [0.210.71230.0692 − 0.1146 − 0.37850.490.48 − 0.4653 − 0.0456][]
more noticeable as the value increases.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Seed Filling Algorithm:</title>
        <p>To separate touching objects in the image the proposed system uses the mark-based algorithm
that assigns pixels to regions based on proximity to markers, resulting in segmented regions
delineated by watershed boundaries is used to segment the data images for the image samples
processed by skin colour detection. The algorithm connects neighbouring pixels with similar
grey values, forming contours for image segmentation. This approach is especially efective for
images with noise and irregular gradients, simplifying the segmentation process by highlighting
distinct intensity patterns in the image. Since the standing level of the component that is
connected can be raised like a dam, the watershed algorithm based on mark may stop the
local smaller edges are merged and inseparable. Through the supervision of the mark-based
watershed algorithm, the gesture features may be eficiently segmented. The 8-connected edge
iflling method is used to sporadic gesture portion once the gesture features have been accurately
segmented. An enhancement on the four-connected filling process is the 8 edge connected seed
iflling algorithm. Unlike the four-connected filling method, which begins the process at the
beginning, the 8-connected seed filling technique spreads in eight directions, speeding up the
process. at the centre of injection in the area and expands in each direction to cover all of the
area’s pixels. The algorithm used to obtain the sign feature data has eight connected seeds.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Normalization and Gesture Labelling:</title>
        <p>After the image data has been filled in and segmented, the scale normalisation operation may
guarantee the reliability of the feature extraction and the processing of labelling the data
with gestures clear features for training the model eficiently. In the proposed method, the
segmented and filled picture data are normalised to 128x128, which efectively increases the
model’s performance, speed and accuracy during training and prevents gradient explosion.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Evaluation and Fine-Tuning:</title>
        <p>Following training, the model is evaluated on a diferent validation or test dataset to gauge its
performance in terms of accuracy and generalization. To attain the appropriate level of accuracy,
the model may need to be fine-tuned and its hyper-parameters may need to be adjusted.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.6. Post-processing:</title>
        <p>Post-processing techniques may be used to improve the precision of gesture detection in
realworld scenarios or to smooth predictions over time.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results &amp; Analysis</title>
      <p>We tested on computer with an AMD Ryzen 5 processor and a 64-bit version of Windows 10,
our experiment is run using OpenCV and PIL installed for image processing, and Anaconda
is installed to create an interactive interface, were used to implement the identification and
classification operations.</p>
      <sec id="sec-4-1">
        <title>4.1. Experimental Analysis:</title>
        <p>The data set is collected from publicly available sources, and each gesture data gathers 250
image data. The case study’s trained model is capable of identifying ten distinct movement
types, ranging from 0 to 9. First, a total of 1980 data samples representing twenty types of
gestures are gathered for each class of 10 individuals over a range of time periods. The training
outcome for gestures 3, 4, and 5 is poor. The training part can be significantly increased by
including sample data. As a result, 100 gestures are added to each category of gestures, which
considerably raises the success rate of recognition.</p>
        <p>A comprehensive data set requirement is necessary to enhance training efect. The position
on the photograph will difer for the same motion when performed by diferent people, at
diferent angles, and with variable lighting conditions. Fig. 3 displays various data for the
same move. In the detection rate test, 150 test specimens were collected for quantifying across
various time periods; 95 of the samples had been successfully identified, 7 of the samples
experienced recognition mistakes, and 1 sample did not exhibit any gesture. The test’s outcomes
in the very dark environment weren’t the best; there were four times as many mistakes in the
prediction of gestures 2 and 3. This is due to the fact that the predicted gesture features cannot
be distinguished in setting and because gestures 4 and 5 are too similar to one another. In a
typical context, the model generated by the data pre-processing method can attain an optimal
prediction rate by improving the gesture characteristic. Data from four groups of samples is
collected and divided into the model’s robustness and stability tests. The analysis result is
shown in Fig. 4.</p>
        <p>We have introduced a real-time hand gesture recognition system in this research that makes
use of cutting-edge computer vision and machine learning algorithms to precisely interpret and
categorize hand gestures in practical settings. Our study has significantly advanced the field of
gesture recognition in a number of ways.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Robustness and Accuracy</title>
        <p>We have shown that our system achieves accuracy in recognizing a wide variety of hand
gestures through thorough experimentation and review. While CNNs excel in extracting spatial
features from grid-like data such as images, RNNs are well-suited for sequential data with
temporal dependencies. The choice between them depends on the nature of the task and
the characteristics of the input data. among other deep learning models, have considerably
increased the system’s capacity to handle complicated and dynamic movements. The main goal
of our research was to create a system that could process data in real-time. Our system can now
process and recognize gestures in real time, enabling technologies or methodologies likely play
a crucial role in improving user experiences and enhancing interactions in these areas. This
indicates that we have successfully accomplished our goal.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Comparison of each group</title>
        <p>In a typical context, the model produced by the data pre-processing approach can reach an
optimal recognition rate by improving the gesture feature. Four sample groups of test data,
each with 30 test data graphs, are randomly chosen from the test data set in order to assess the
model’s stability and robustness. Each data group was recognized and put to the test. The model
exhibits strong resilience and stability after testing. Table 2 presents the comparing findings.</p>
        <p>The model’s convergence rate is accelerated since dropout is used to prevent over-fitting.
After numerous tries, increasing the batch value in the training process can lower the number of
epochs and accelerate training. It could not be able to efectively collect the data characteristics
due to the decreased number of repetitions, which would lower the prediction rate. When the
batch size reaches 35 following testing, the training time and model constancy are well-aligned.
The training is stopped early after the saturation point. The loss curve comparison chart for
various batch values is shown in Fig. 5.</p>
        <p>Three people are chosen at random to participate in the laboratory testing to confirm the
acceptance of prediction of gestures on various participants in proposed system. Each of the 10
motions must be tested 500 times, with each tester testing each position nearly 50 times, by
cross folding into two groups and testing each gesture. If testing is needed during the day, each
group must take it 15 times, and if it is needed at night, each group must take it 250 times. The
three testers’ respective recognition rates for the ten gestures are 94.4%, 94.6%, and 93.8%.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this study, the gesture data is pre-processed using skin colour recognition, marker-based
watershed algorithms, and seed filling algorithms. in order to generate the gesture data with
clear gesture features. Proposed system accurately performed on the test data and then achieves
98.66% under conditions of ordinary light by using the training of 10 diferent types of gesture
data following YOLO convolution neural network pre-processing. The pre-processing technique
applied in this study efectively mitigates the influence of the surrounding background on
gesture recognition and detection. Notably, its implementation does not necessitate additional
training or detection time. Additionally, the post-processing step would depend on the specific
requirements of our gesture recognition task.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. A.</given-names>
            <surname>Zaman</surname>
          </string-name>
          ,
          <article-title>Hand gesture recognition: a literature review</article-title>
          ,
          <source>International journal of artificial Intelligence &amp; Applications</source>
          <volume>3</volume>
          (
          <year>2012</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Swathi</surname>
          </string-name>
          , Aarti,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>A smart application to detect pupil for small dataset with low illumination</article-title>
          ,
          <source>Innov. Syst. Softw. Eng</source>
          .
          <volume>17</volume>
          (
          <year>2021</year>
          )
          <fpage>29</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Caputo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Giachetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Soso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pintani</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. D'Eusanio</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Pini</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Borghi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Simoni</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Vezzani</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cucchiara</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ranieri</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Giannini</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Lupinetti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Monti</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maghoumi</surname>
            ,
            <given-names>J. J. LaViola</given-names>
          </string-name>
          , Jr,
          <string-name>
            <given-names>M.-Q.</given-names>
            <surname>Le</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-D. Nguyen</surname>
            , M.-T. Tran,
            <given-names>SHREC</given-names>
          </string-name>
          <year>2021</year>
          :
          <article-title>Track on skeleton-based hand gesture recognition in the wild (</article-title>
          <year>2021</year>
          ). arXiv:
          <volume>2106</volume>
          .
          <fpage>10980</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Chmurski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mauro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Santra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zubert</surname>
          </string-name>
          , G. Dagasan,
          <article-title>Highly-optimized radar-based gesture recognition system with depthwise expansion module</article-title>
          ,
          <source>Sensors (Basel) 21</source>
          (
          <year>2021</year>
          )
          <fpage>7298</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Madaan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kundu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sethi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <surname>X-</surname>
          </string-name>
          <article-title>hubis: A fuzzy rule based human behaviour identification system based on body gestures</article-title>
          ,
          <source>Indian Journal of Science and Technology</source>
          (
          <year>2016</year>
          )
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Rahim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shin</surname>
          </string-name>
          ,
          <article-title>Non-touch sign word recognition based on dynamic hand gesture using hybrid segmentation and CNN feature fusion</article-title>
          ,
          <source>Appl. Sci. (Basel) 9</source>
          (
          <year>2019</year>
          )
          <fpage>3790</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Mustafa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Jomhari</surname>
          </string-name>
          ,
          <article-title>A review of the hand gesture recognition system: Current progress and future directions</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2021</year>
          )
          <fpage>157422</fpage>
          -
          <lpage>157436</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Mambou</surname>
          </string-name>
          , Krejcar, Maresova, Selamat, Kuca,
          <article-title>Novel hand gesture alert system</article-title>
          ,
          <source>Appl. Sci. (Basel) 9</source>
          (
          <year>2019</year>
          )
          <fpage>3419</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>P.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kaur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Madaan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Babu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Sethi</surname>
          </string-name>
          ,
          <article-title>Moving object detection and recognition using optical flow and eigen face using low resolution video</article-title>
          ,
          <source>Recent Advances in Computer Science and Communications (Formerly: Recent Patents on Computer Science</source>
          )
          <volume>13</volume>
          (
          <year>2020</year>
          )
          <fpage>1180</fpage>
          -
          <lpage>1187</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ashiquzzaman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kim</surname>
          </string-name>
          , H.-Y. Kim,
          <string-name>
            <given-names>J.</given-names>
            <surname>Park</surname>
          </string-name>
          , J. Kim,
          <article-title>Compact spatial pyramid pooling deep convolutional neural network based hand gestures decoder</article-title>
          ,
          <source>Appl. Sci</source>
          . (Basel)
          <volume>10</volume>
          (
          <year>2020</year>
          )
          <fpage>7898</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gibran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Haris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tsuda</surname>
          </string-name>
          ,
          <article-title>Continuous finger gesture spotting and recognition based on similarities between start and end frames</article-title>
          ,
          <source>IEEE Transactions on Intelligent Transportation Systems</source>
          <volume>23</volume>
          (
          <year>2020</year>
          )
          <fpage>296</fpage>
          -
          <lpage>307</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gowroju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>Aarti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ghimire</surname>
          </string-name>
          ,
          <article-title>Deep neural network for accurate age group prediction through pupil using the optimized UNet model</article-title>
          ,
          <source>Math. Probl. Eng</source>
          .
          <year>2022</year>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gowroju</surname>
          </string-name>
          , Aarti,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Robust pupil segmentation using UNET and morphological image processing</article-title>
          , in: 2021
          <source>International Mobile, Intelligent, and Ubiquitous Computing Conference (MIUCC)</source>
          , IEEE,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gowroju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>Aarti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ghimire</surname>
          </string-name>
          ,
          <article-title>Deep neural network for accurate age group prediction through pupil using the optimized UNet model</article-title>
          ,
          <source>Math. Probl. Eng</source>
          .
          <year>2022</year>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>24</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gowroju</surname>
          </string-name>
          , Aarti,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Review on secure traditional and machine learning algorithms for age prediction using IRIS image, Multimed</article-title>
          .
          <source>Tools Appl</source>
          . (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Mustafa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Jomhari</surname>
          </string-name>
          ,
          <article-title>A review of the hand gesture recognition system: Current progress and future directions</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2021</year>
          )
          <fpage>157422</fpage>
          -
          <lpage>157436</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Chmurski</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mauro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Santra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zubert</surname>
          </string-name>
          , G. Dagasan,
          <article-title>Highly-optimized radar-based gesture recognition system with depthwise expansion module</article-title>
          ,
          <source>Sensors (Basel) 21</source>
          (
          <year>2021</year>
          )
          <fpage>7298</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Ben-Musa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <article-title>Object detection and recognition in cluttered scene using harris corner detection</article-title>
          , in: 2014 International Conference on Control,
          <source>Instrumentation, Communication and Computational Technologies (ICCICCT)</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>181</fpage>
          -
          <lpage>184</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICCICCT.
          <year>2014</year>
          .
          <volume>6992953</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>A. S. B. Musa</surname>
            ,
            <given-names>S. K.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <article-title>Suspicious human activity recognition for video surveillance system</article-title>
          ,
          <source>in: IEEE proceedings of 2014 international conference on control, instrumentation, communication and computational technologies</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>214</fpage>
          -
          <lpage>218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>A.</given-names>
            <surname>Swathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Subbamma</surname>
          </string-name>
          .,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Emotion classification using feature extraction of facial expression</article-title>
          ,
          <source>in: 2022 2nd International Conference on Technological Advancements in Computational Sciences (ICTACS)</source>
          , IEEE,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>