<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>J. Mendoza-Montoya);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Design of a Prototype Mechatronic Wheelchair for People with Motor Disabilities</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Javier Mendoza-Montoya</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kevin Enrique Angulo-Oviedo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Elmer Yuver Turpo-Apaza</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Reynaldo Alfonte-Zapana</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Tecnológica del Perú</institution>
          ,
          <addr-line>Av. Parra 203 Arequipa, Arequipa</addr-line>
          ,
          <country country="PE">Peru</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>People with motor disabilities require some mechanism or technological solution to move around autonomously. The most common way is through a voice-controlled electromechanical chair; however, these solutions are expensive and often require expert support in training. This paper aims to design a low-cost and easy-to-use electromechanical chair that allows autonomous navigation controlled by an electronic interface, a joystick, and voice recognition. In the first step, modeling an electrical wheelchair prototype with 3D software is essential; subsequently, a voice command system to control wheelchair navigation is developed based on the speech recognition library of Python. Finally, an electronic interface is designed and integrated into an Arduino Mega embedded module to recognize the voice commands and control the wheelchair's mechanical system navigation. The results of this research are promising, considering the low cost of hardware design, the well-performance prototype, and the easyto-use navigation system. The joystick control was accurate and gave four essential movements to the wheelchair; the voice recognition uses multi-tasking with threading algorithms to interpret and execute wheelchair navigation commands simultaneously. To enhance the prototype performance in future advances, optimizing the time response in command execution and setup system to individual user's requirements is recommendable.</p>
      </abstract>
      <kwd-group>
        <kwd>Electrical wheelchair</kwd>
        <kwd>disabilities person</kwd>
        <kwd>voice recognizing</kwd>
        <kwd>threading programming1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Motor disability is a condition characterized by an individual's inability to independently engage
in physical movement. Peru has a notable prevalence in this particular domain, and despite
ongoing efforts towards the integration of individuals with disabilities, achieving this objective
remains challenging within our societal context. The inclusion of individuals with disabilities in
all economic activities is essential for promoting their independence. However, it is often
observed that these individuals have challenges in securing employment due to their specific
disability-related problems. According to the annual report of the Informatic and Statistic
National Institute of Perú (INEI) in 2013, it was found that a total of 932 thousand individuals
exhibit motor disabilities, which encompass challenges in mobility, ambulation, and the
utilization of limbs [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. According to a recent statistics analysis published by the National Institute
of Statistics and Informatics (INEI) in 2019 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], approximately 10% of the Peruvian population is
affected by impairments. Among this group, 15.1% specifically experience motor disabilities,
which corresponds to a total of 484,598 individuals facing challenges related to mobility and
ambulation. The aforementioned publications provide an analysis of the various constraints
experienced by individuals with disabilities, encompassing physical, occupational, social, and
economic dimensions.
      </p>
      <p>
        In the academic literature, there has been significant emphasis placed on the investigation of
electrical wheelchair navigation. As a result, numerous solutions have been proposed for
controlling an electrical wheelchair through the integration of electronic interfaces with a
joystick [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. This particular control method relies on manual input through the use of hands.
However, in cases where individuals are unable to move their arms or hands due to disabilities,
alternative solutions based on signal processing are required. These solutions involve the
utilization of bio-signals, which are generated from various parts of the human body. For instance,
the bio-signal known as electro-oculography (EOG) can be produced through eye movement and
can be obtained using a system consisting of an operational amplifier, an analog-to-digital
converter (ADC), and a computer. This setup allows for the processing of the bio-signal and the
transmission of commands to the wheelchair [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ]. Another type of biological signal is the
electroencephalogram (EEG), which originates from brain activity. However, the analysis and
processing of this signal are highly intricate and necessitate the use of expensive equipment [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
The electromyogram (EMG) is a cost-effective method for capturing bio-signals associated with
muscle movements. However, its implementation necessitates the use of bio-amplifiers,
analogto-digital converters (ADCs), and computational resources for signal processing [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ]. Voice
recognition technology has emerged as a highly promising solution for individuals with motor
disabilities. To enable this functionality, the use of a headset and microphone is required to
facilitate communication and issue commands to the wheelchair. In order to accurately interpret
and respond to specific verbal instructions for wheelchair control, the integration of dedicated
interpreter software and a microcontroller capable of processing voice signals is imperative [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref13 ref3 ref9">9,
10, 11, 3, 12, 13</xref>
        ]. In conclusion, certain hybrid or dual approaches have been suggested in the
form of voice and vision-based mechanisms [14] or head gesture recognition [15, 16].
Nevertheless, it should be noted that these solutions tend to be expensive and frequently
necessitate the involvement of an expert to facilitate their utilization, upkeep, and instruction.
      </p>
      <p>The objective of this study is to present a conceptual design for an electric wheelchair that can
be controlled using both voice commands and a joystick. The inclusion of these dual control
options is intended to assure the wheelchair's functionality in the event of voice command failure.
In such a scenario, the joystick serves as a backup control mechanism, allowing the wheelchair to
continue running seamlessly. The speech recognition library in Python is employed for voice
recognition, while the Arduino Atmega embedded system is utilized for processing and
recognizing voice commands. Additionally, the Arduino Atmega system is responsible for
controlling the motor that enables the wheelchair's regulated navigation.</p>
      <p>The present study is organized in the following manner: Section 2 provides a comprehensive
account of the materials and methods employed in this study, encompassing the hardware design
and software development aspects. Section 3 of the document provides an exposition of the
experimental findings. Section 4 of the document encompasses the presentation of the
conclusions and recommendations.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Material and method</title>
      <sec id="sec-2-1">
        <title>2.1. Hardware Design</title>
      </sec>
      <sec id="sec-2-2">
        <title>2.1.1. Structural design</title>
        <p>The construction process of the wheelchair prototype commenced by implementing the
measurements specified in the design. This involved cutting the pipes and subsequently joining
them together using appropriate couplings. These steps were undertaken to create the
foundation and subsequently the complete chassis of the prototype, with the aim of assuring its
stability and symmetrical structure. Following this, drillings were conducted in order to install
steel plates that were designed to give support for the motors and provide a level surface for their
attachment using bolts and nuts. This process was carried out to ensure the structural soundness
of the motor and to offer support for the primary axis of the tire. Subsequently, the front tires are
affixed and calibrated in order to finalize the assembly of the chassis. In addition, an assessment
is conducted on the state of the components, ensuring that they do not encounter any obstacles
in the movement of the prototype.</p>
        <p>The wheelchair operates by utilizing batteries that have been conditioned and are housed
within a white box located underneath the seat. The strategic placement of the batteries
contributes to maintaining the stability of the chair. The correctness of the wheelchair's design is
demonstrated through the precise dimensions of its sections, weight support capabilities, and
center of gravity. This particular prototype has been engineered to accommodate individuals
weighing up to 70 kg. After the completion of the structural design, a series of paint layers are
applied to enhance the aesthetic appearance of the chassis, resulting in the final appearance
depicted in Figure 1.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.1.2. Electronic system</title>
        <p>The electronic system comprises two power H-bridges, specifically the LM298n module, which
incorporates buffers and enables the control of two motors. In our situation, these motors are
responsible for controlling the left and right movements of the wheelchair. The control circuit is
regulated by an Arduino embedded system, which incorporates an atmega328 microcontroller.
This microcontroller, with its eight-bit architecture, possesses sufficient computational
capabilities to effectively handle real-time data processing for our specific application. In order
to access the voice recognition library in the cloud, it is necessary for the embedded system to
establish an internet connection.</p>
        <p>The system initiates the activation of two H-bridges to enable continuous control of the left
and right motors. The movement of the wheelchair is determined by six fundamental commands,
Table 1.</p>
        <p>A pulse width modulation (PWM) signal is sent from the microprocessor to each H-bridge that
controls the gearmotors and controls the speed motor as well, Figure 2. It works with the Echo
and Trigger inputs, which are linked to Arduino pins 7 and 6.</p>
        <p>Below is a list of the parts that were used in the system.
• 1 Arduino uno with 5v
• 1 LM298n H-bridge 5v
• 2 12v/730mA gearmotors with 4.2Nm of torque
• 1 ultrasound monitor with 5v
• 2 power inputs (5v to 12v, 20W)
• 1 microphone
• 1 headset</p>
        <p>Python is needed to build the system, along with the speech recognition library and the
algorithm to understand voice directions like "Avanza" for "Forward," "derecha" for "turn right,"
"izquierda" for "turn left," "Retrocede" for "backward," "Para" for "stop," and "Rápido" for "Fast."</p>
      </sec>
      <sec id="sec-2-4">
        <title>2.1.3. Joystick control circuit</title>
        <p>
          In order to operate the wheelchair using a joystick, a logical framework will be implemented,
consisting of four distinct movements: forward, backward, left turn, and right turn. Each of these
movements will be calibrated to accommodate pulse width modulation based on the degree of
depression of the lever positioned at its center. To analyze this, we represent the range of joystick
displacement for each movement on a Cartesian coordinate system. In order to make progress,
we divide the Cartesian region into two zones: Zone "a" represents the left zone, which begins at
0, and Zone "b" represents the right zone, which ends at 1023. The range is measured in bits,
taking into account the sensitivity of the analog-to-digital converter (ADC). In this particular case,
the ADC has 10 bits, resulting in a range of [
          <xref ref-type="bibr" rid="ref1 ref10 ref11 ref12 ref13 ref2 ref3 ref4 ref5 ref6 ref7 ref8 ref9">0-1023</xref>
          ] bits, which is equivalent to a voltage range
of [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">0-5</xref>
          ] Volts. The center zone can be adjusted based on the joystick's sensitivity and individual
requirements.
        </p>
        <p>To facilitate the selection of the mode for joystick or voice command, a red button has been
incorporated to enable the immediate choice of the desired operation mode by simply pressing
the button.</p>
      </sec>
      <sec id="sec-2-5">
        <title>2.2. Software design</title>
      </sec>
      <sec id="sec-2-6">
        <title>2.2.1. Voice control</title>
        <p>In order to facilitate navigation through voice commands, it is necessary to utilize the Python
programming language along with the Speech Recognition library. Additionally, the following
libraries are required: time, serial, pyttxs3, soundfile, and sounddevice. Similarly, it is imperative
to assess the voice recognition service provided by Google for the Spanish language. To
accomplish this, it is proposed to construct a comprehensive table that documents the outcomes
of each command, categorizing them as either successful or erroneous. Additionally, the
effectiveness of the system will be quantified by determining the percentage of successful
commands. In brief, the algorithm will facilitate the evaluation of Spanish voice recognition by
performing the following tasks:</p>
        <p>An example of numbered list is as following.
1. Connect Arduino to a USB port.
2. Make a list of the directions that will be used to tell the actuators what to do through
Arduino: ['forward', 'forward', 'forward', 'backward', 'retro', 'right', 'right', 'left', 'left',
'stop', 'anti', 'stop', 'stop', 'quickly']
3. A sample rate of 44100 Hz is set for recording a ".wav" file. That is, 44,100 pieces of sound
will be recorded for every second. 44100 samples were picked because that's a common
number for playing back audio and it's enough to get most of the details of the recording.
4. We set the time to "X" seconds to give the user more time to record in case the word gets
longer and to take into account how long each Python command takes to run. A person's
time for saying each order will help us figure out what "X" is worth.
5. With the previous settings, the audio is taken, and when it's done, a ".wav" file is made
that will be overwritten over and over again.
6. The recorded sound will be cleaned of noise, and the recognition feature set to use the</p>
        <p>Google service and speak Spanish will be used.
7. The word that was recognized must be saved in a "command" variable. All words that
were recognized will be made lowercase.
8. Any words that could have been recognized will be split up and the space between them
will be taken out, leaving us with just the lowercase word.
9. The information that was gathered will be compared word by word until one fits a
command that was set up.
10. After detecting the first spoken word, a signal will be sent to the Arduino through the
serial port. The Arduino will read it and tell the H-bridge how to control the motors'
rotation.
11. The last order that was spoken will be saved and work until a new command is spoken.
12. Inside the Arduino IDE, a bit will be set aside for each type of movement so that they can
be controlled. In this case, "Forward" will be run if Python sends bit 1, "Backward" will be
run if bit 2 is received, and so on. One problem with this way of moving the wheelchair is
that it throws off the user's timing with the recording. Because this was such a big
problem, an algorithm was made that runs the listener in one thread and the orders in
another. This way, the orders are carried out in the motors while the listener waits for
new directions. In the same way, this new way of working lets us get the best response
time from the motors to the controller's commands.</p>
        <p>This is how the new program is put together.</p>
        <p>1. Threading, a special library, is loaded so that the program can run in one or more
processing threads.
2. The process of establishing serial communication with an Arduino device is initiated.
3. The aforementioned compilation of instructions is designated with a name.
4. Develop a function that encompasses the implementation of speech recognition and
offers formatting capabilities for the transcribed words.
5. Furthermore, this function will specify the commands that are to be transmitted to the</p>
        <p>Arduino.
6. A new loop function is implemented with the sole purpose of actively monitoring and
awaiting incoming commands within a specified time frame of 1.1 seconds.
7. The voice recognition feature is invoked to operate concurrently with the process of
listening to incoming voice commands.</p>
        <p>By employing this novel recognition technique, it is possible to guarantee a decrease in reading
inaccuracies and enhance the efficiency of actuator reaction time.</p>
      </sec>
      <sec id="sec-2-7">
        <title>2.3. Design of functional prototype</title>
      </sec>
      <sec id="sec-2-8">
        <title>2.3.1. Wheelchair security</title>
        <p>The proposed prototype will use a dual security system. To initiate the cancellation of a
command, the user has the option to adjust the selector position. This action effectively halts the
movement of the chair while transitioning between voice navigation mode and joystick
navigation mode, or vice versa. Similarly, the prototype will incorporate ultrasonic sensors
capable of detecting immediate obstructions at a range of less than 40 cm. This feature will enable
the chair to halt its movement prior to any potential collision with an object. In this study, a total
of 256 measurements were conducted using a specific sensor across several materials. The
objective was to assess the potential inaccuracy in measurements at distances of 100 cm, 50 cm,
and 40 cm in diverse materials, including concrete, cardboard, Medium Density Fiberboard
(MDF), mirror plastic, and expanded polystyrene.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>3.1. Voice recognition</title>
        <p>The experimentation involved conducting success and error tests with Google's speech
recognition library, and the outcomes are presented in Table 2.</p>
        <p>Based on these data, it is possible to find out how effective each command is and how effective
all commands are.</p>
        <p>1. “Avanza”: 18 hits were found out of 20 samples. It works 90% of the time when the order
is given.
2. “Retrocede”: Thirteen out of twenty samples gave right answers, which is a 35% error
rate. Sometime this order wouldn't work in an emergency, which could hurt the person
using it if the ultrasonic sensor emergency stop wasn't available. The order is changed to
"retro," and the test is done again.
3. “Izquierda”: Out of 20 samples, 10 had right answers, which means there was a 50% error.</p>
        <p>If we leave this command alone, it wouldn't be very useful in an emergency. But if the user
types the same command more than twice and doesn't get an answer, they might think
that the recognition isn't working and stop using this option. We change this command to
"anti" (which means reverse) and run the test again.
4. “Derecha”: 18 hits were found out of 20 samples. It works 90% of the time when the order
is given.
5. “Para”: Nineteen hits were found out of twenty samples. It works 95% of the time when
the order is given.
6. “Rapido”: Twenty samples were used, and twenty hits were found. It works every time
that the order is given.
7. We'll now change "Izquierda" to "anti" and "Retrocede" to "retro" to get the result shown
in Table 3.</p>
        <p>We will add "anti" and "retro" as key commands for left turn and reverse after this test. This
will also make it faster to say a command.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Voice command control</title>
        <p>A novel software application was developed with the purpose of regulating the functioning of
the motors. In order to supply power to the Arduino and the L298n, a power bank was employed.
Similarly, it was linked to the computer system and initiated the issuance of commands, which
were executed in accordance with anticipated outcomes. An issue that became apparent was the
rather long recognition time of approximately 4 seconds for our initial code implementation.
Table 4 displays the duration it takes for an individual to articulate each of the previously
suggested instructions.</p>
        <p>Consequently, a novel experiment was conducted utilizing the updated orders in order to
enhance the time response. The outcomes of this experiment are outlined in Table 5, presented
as follows:</p>
        <p>Regarding the preceding instructions, we have obtained these novel findings with a reduced
time interval. The new "anti" command for the "Izquierda" function offers enhanced precision
and improved efficiency, resulting in reduced recording time.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Distance ultrasound sensor</title>
        <p>In accordance with our established approach, an assessment was conducted to determine the
precision of the HCSR-04 sensor when measuring various materials, including concrete,
cardboard, MDF, mirror plastic, and technopor. The results of this evaluation are presented in
Table 6.</p>
        <p>In order to have a deeper comprehension of the data, an analysis was conducted on the
dispersion pertaining to the mean. Hence, it is evident that the sensor exhibits a superior
responsiveness when applied to concrete surfaces. However, when utilized with alternative
materials such as mirror plastic and styrofoam, the signal may introduce inaccuracies in distance
estimation. Hence, it is recommended that the wheelchair software control be modified to
incorporate a joystick interface in order to enhance the navigation performance.</p>
        <p>As shown in Figure 4, the HSR-04 ultrasound sensor responds differently to different
materials. Concrete gives more accurate readings, but there isn't much difference between the
readings of other materials (a maximum standard deviation of 0.47 cm), so the system could keep
the wheelchair moving.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Structural analysis</title>
        <p>In the mechanical part, 3D software was used to do the structural analysis. This gave us useful
information about how the chassis's structure would behave in different situations. In our
models, we think that the chassis could have been made of PVC, aluminum, or steel tubes, with
the same material used for the joints between the tubes.</p>
        <p>During the simulations test, we look at the following features:
• A force of 9.81 m/s2 of gravity
• The patient's weight limit is 40 kg
• The motor weighs 1 kg
• The battery weighs 3 kg
• The electrical parts weighs 1 kg</p>
        <p>With these numbers, we'll start the simulations that go with them and show the results in
terms of shrinkage, taking three things into account at the same time for this test.
1. PVC
2. Aluminum
3. Steel</p>
        <p>We want to find the best material out of all of them to use as a base for cheaply putting together
this type of wheelchair.</p>
        <p>With these simulations, we can see that the deformation changes depending on the chassis's
shape and the type of material used. Following these tests, the PVC frame has the most
deformation, going up to a height of 6,307 mm. The next one is made of metal and is 0.282 mm
thick. The last one is made of steel and is 0.239 mm thick. The displacement zone in all of these
simulations is the same. It's in the area where the load acts directly (Table 7).</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Power-Weight relationship</title>
        <p>For a power of 350 and a weight of 45 kg, the power-to-weight ratio is 7.78, which can be seen
by comparing the numbers to Table 8. Additionally, using the information in the same table, we
can say that if we add more mass, the weight-to-power ratio will go down because the engine will
have to make more power to make up for it.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>In summary, we have successfully developed a mechatronics-based prototype wheelchair that
incorporates essential features and is cost-effective. This design enables manual navigation using
a joystick or automated navigation through voice commands issued by the user.</p>
      <p>A cost-effective prototype of a wheelchair was developed using both computational modeling
and physical construction, with the primary material being polyvinyl chloride (PVC). A
comprehensive structural analysis was conducted on the prototype in order to validate its ability
to endure the designated load, despite its relatively modest cost. Furthermore, the final iteration
of the wheelchair prototype incorporated the installation of tires sourced from an authentic
wheelchair, so enhancing its aesthetic appeal and lending it an authentic semblance.</p>
      <p>The development of a control system that effectively coordinates the navigation of the
prototype wheelchair has been successfully achieved. This system demonstrates adequacy,
efficiency, and the potential for optimization. The implementation involved utilizing speech
recognition and threading techniques. Python was employed as the programming language for
voice command control, while the Arduino Software was utilized to facilitate the transmission
and reception of instructions to the actuators.</p>
      <p>Similarly, we have successfully enhanced the efficiency of voice command responses by
reprogramming and organizing the available phrases or commands in a manner that ensures user
comfort. In addition, it is possible to receive these words simultaneously with the transmission
of commands to the actuators. Similarly, a delay of 1.1 seconds was achieved, resulting in an
enhanced efficiency of our workflow that initially experienced a wait of 4 seconds between the
execution of each order. Ultimately, we replaced the lexemes that exhibited the greatest temporal
duration for directives that conveyed an identical directive, yet possessed enhanced efficiency in
terms of articulation.</p>
      <p>Ultimately, we have successfully integrated all the constituent stages or systems comprising
the mechatronic wheelchair prototype in order to validate its functionality and enhance its
efficiency. This phenomenon is evident in the outcomes gained and is demonstrated in the
present research study.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Recommendations</title>
      <p>Based on the findings of our research, we are able to offer the following advice for future
endeavors.</p>
      <p>•
•
•
•</p>
      <p>The prototype needs to be made from something other than PVC so that it can hold a
person who weighs more than 50 kilograms. Aluminum may be a good choice because
it is not more expensive than other materials and can hold more weight than PVC.
Similarly, we suggest getting motors with more power that can move the suggested
load through some kind of extra transmission. Online shops sell these, but they cost a
lot more than the ones we used for this project.</p>
      <p>To use electronic control, we suggest making a PCB card with all the parts that will be
used. In the same way, a Raspberry Pi or an ESP32 card can be used instead of an
Arduino. This makes sending messages to the actuators faster. Also, we believe it
would be more effective to use the VRM v3 Voice Recognition Module version 3. This
module can recognize voice commands without an Internet connection, does not
require a computer for signal processing, and lets you train up to seven commands.
Finally, based on where the prototype will be used, we suggest adding more sensors
to the area around the wheelchair to make the person using it safer.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>We express our gratitude to Universidad Tecnológica del Perú for their valuable logistical
assistance, which has facilitated the execution of this review paper.
[14] M. Anshar, R. Sadjad, D. Dewiani, M. Abry, R. Prayudha, F. Fiqhi, M. Takbir, Control system
design for smart wheelchair robot, in: AIP Conference Proceedings, volume 2543, AIP
Publishing, 2022.
[15] I. K. Somawirata, F. Utaminingrum, Smart wheelchair controlled by head gesture based on
vision, Journal of Physics: Conference Series 2497 (2023) 012011. URL: https://dx.doi.org/
10.1088/1742-6596/2497/1/012011. doi:10.1088/1742-6596/2497/1/012011.
[16] S. Chatzidimitriadis, S. M. Bafti, K. Sirlantzis, Non-intrusive head movement control for
powered wheelchairs: A vision-based approach, IEEE Access 11 (2023) 65663–65674.
doi:10.1109/ACCESS.2023.3275529.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>INEI</surname>
          </string-name>
          ,
          <article-title>En el perú 1 millón 575 mil personas presentan algún tipo</article-title>
          de discapacidad,
          <year>2013</year>
          . URL: https://m.inei.gob.pe/media/MenuRecursivo/noticias/np-178-2013.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>INEI</surname>
          </string-name>
          ,
          <article-title>En el país existen 3 millones 209 mil 261 personas con</article-title>
          discapacidad,
          <year>2019</year>
          . URL: https://m.inei.gob.pe/media/MenuRecursivo/noticias/np_136_2019_inei.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Shifa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. M.</given-names>
            <surname>Mridul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Rafat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Monjur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. R.</given-names>
            <surname>Islam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. K.</given-names>
            <surname>Hassan</surname>
          </string-name>
          ,
          <article-title>Design and implementation of smart wheelchair with advanced control interfaces</article-title>
          ,
          <source>in: 2023 3rd International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>155</fpage>
          -
          <lpage>159</lpage>
          . doi:
          <volume>10</volume>
          .1109/ICREST57604.
          <year>2023</year>
          .
          <volume>10070067</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. P. A.</given-names>
            <surname>Ogenga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. W.</given-names>
            <surname>Njeri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Muguro</surname>
          </string-name>
          ,
          <article-title>Development of a virtual environment-based electrooculogram control system for safe electric wheelchair mobility for individuals with severe physical disabilities</article-title>
          ,
          <source>Journal of Robotics and Control (JRC) 4</source>
          (
          <year>2023</year>
          )
          <fpage>165</fpage>
          -
          <lpage>178</lpage>
          . URL: https://doi.org/10.18196/jrc.v4i2.17165. doi:
          <volume>10</volume>
          .18196/jrc.v4i2.
          <fpage>17165</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <article-title>Eye-gaze controlled wheelchair based on deep learning</article-title>
          ,
          <source>Sensors</source>
          <volume>23</volume>
          (
          <year>2023</year>
          ). URL: https://www.mdpi.com/1424-8220/23/13/6239. doi:
          <volume>10</volume>
          . 3390/s23136239.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Maule</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zanetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Luchetti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Tomasin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dallapiccola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Covre</surname>
          </string-name>
          , G. Guandalini,
          <string-name>
            <surname>M. De Cecco</surname>
          </string-name>
          ,
          <article-title>Wheelchair driving strategies: A comparison between standard joystick and gazebased control</article-title>
          ,
          <source>Assistive technology : the official journal of RESNA</source>
          <volume>35</volume>
          (
          <year>2023</year>
          )
          <fpage>180</fpage>
          -
          <lpage>192</lpage>
          . URL: https://doi.org/10.1080/10400435.
          <year>2021</year>
          .
          <volume>2009593</volume>
          . doi:
          <volume>10</volume>
          .1080/10400435.
          <year>2021</year>
          .
          <volume>2009593</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Gunarathne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Welihinda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Herath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yasakethu</surname>
          </string-name>
          ,
          <article-title>Eeg-assisted emg-controlled wheelchair for improved mobility of paresis patients</article-title>
          ,
          <source>in: 2023 IEEE IAS Global Conference on Emerging Technologies (GlobConET)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/GlobConET56651.
          <year>2023</year>
          .
          <volume>10149923</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gopichand</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Rajeswari</surname>
          </string-name>
          , E. Deepthi,
          <article-title>Human-machine interface for wheelchair control using semg signals</article-title>
          , in: A.
          <string-name>
            <surname>Kumar</surname>
            , G. Ghinea,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Merugu</surname>
          </string-name>
          , T. Hashimoto (Eds.),
          <source>Proceedings of the International Conference on Cognitive and Intelligent Computing</source>
          , Springer Nature Singapore, Singapore,
          <year>2023</year>
          , pp.
          <fpage>395</fpage>
          -
          <lpage>406</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Muthulakshmi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Padmavathy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kirthika</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Vidhya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A. P.</given-names>
            <surname>Manimekalai</surname>
          </string-name>
          ,
          <article-title>Internet of things and smart intelligence-based google assistant voice controller for wheelchair</article-title>
          , in: S. Shakya, G. Papakostas,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>A</article-title>
          .
          <string-name>
            <surname>Kamel</surname>
          </string-name>
          (Eds.),
          <source>Mobile Computing and Sustainable Informatics</source>
          , Springer Nature Singapore, Singapore,
          <year>2023</year>
          , pp.
          <fpage>761</fpage>
          -
          <lpage>774</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>T. K. Hou</surname>
          </string-name>
          , Yagasena, Chelladurai,
          <article-title>Arduino based voice controlled wheelchair</article-title>
          ,
          <source>Journal of Physics: Conference Series</source>
          <volume>1432</volume>
          (
          <year>2020</year>
          )
          <article-title>012064</article-title>
          . URL: https://dx.doi.org/10.1088/
          <fpage>1742</fpage>
          - 6596/ 1432/1/012064. doi:
          <volume>10</volume>
          .1088/
          <fpage>1742</fpage>
          -6596/1432/1/012064.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Omair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. A.</given-names>
            <surname>Syed</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. M. Khan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Shaikh</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Ismail</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Shah</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Irfan</surname>
          </string-name>
          ,
          <article-title>Prototype model of voice activated wheelchair</article-title>
          ,
          <source>in: 2023 Global Conference on Wireless and Optical Technologies (GCWOT)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/GCWOT57803.
          <year>2023</year>
          .
          <volume>10064667</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Murugaiyan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Varatharaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Anandraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Duraisamy</surname>
          </string-name>
          ,
          <article-title>Voice controlled wheelchair</article-title>
          ,
          <source>AIP Conference Proceedings</source>
          <volume>2581</volume>
          (
          <year>2023</year>
          )
          <article-title>030001</article-title>
          . URL: https://doi.org/10.1063/5.0126311. doi:
          <volume>10</volume>
          .1063/5.0126311.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H. F.</given-names>
            <surname>Fakhruldeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Meri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. H.</given-names>
            <surname>Sa</surname>
          </string-name>
          <article-title>'id,</article-title>
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Makttoof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Kadhim</surname>
          </string-name>
          , H. A.
          <string-name>
            <surname>-J. Al-Asady</surname>
          </string-name>
          ,
          <article-title>An arduino-based voice-recognition elevator for special purposes</article-title>
          ,
          <source>Indonesian Journal of Electrical Engineering and Computer Science</source>
          <volume>31</volume>
          (
          <year>2023</year>
          )
          <article-title>828</article-title>
          . URL: https://doi.org/10.11591/ ijeecs.v31.
          <year>i2</year>
          .
          <fpage>pp828</fpage>
          -
          <lpage>834</lpage>
          . doi:
          <volume>10</volume>
          .11591/ijeecs.v31.
          <year>i2</year>
          .
          <fpage>pp828</fpage>
          -
          <lpage>834</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>