0% found this document useful (0 votes)
8 views13 pages

Paper 40

mechatronic sytem balance robot mechatronic sytem balance robot

Uploaded by

hanthiendi115
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views13 pages

Paper 40

mechatronic sytem balance robot mechatronic sytem balance robot

Uploaded by

hanthiendi115
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 13

Design of a Prototype Mechatronic Wheelchair for People

with Motor Disabilities


Javier Mendoza-Montoya1, Kevin Enrique Angulo-Oviedo1, Elmer Yuver Turpo-Apaza1
and Reynaldo Alfonte-Zapana1
1 Universidad Tecnológica del Perú, Av. Parra 203 Arequipa, Arequipa, Peru

Abstract
People with motor disabilities require some mechanism or technological solution to move around
autonomously. The most common way is through a voice-controlled electromechanical chair; however,
these solutions are expensive and often require expert support in training. This paper aims to design a
low-cost and easy-to-use electromechanical chair that allows autonomous navigation controlled by an
electronic interface, a joystick, and voice recognition. In the first step, modeling an electrical wheelchair
prototype with 3D software is essential; subsequently, a voice command system to control wheelchair
navigation is developed based on the speech recognition library of Python. Finally, an electronic
interface is designed and integrated into an Arduino Mega embedded module to recognize the voice
commands and control the wheelchair's mechanical system navigation. The results of this research are
promising, considering the low cost of hardware design, the well-performance prototype, and the easy-
to-use navigation system. The joystick control was accurate and gave four essential movements to the
wheelchair; the voice recognition uses multi-tasking with threading algorithms to interpret and execute
wheelchair navigation commands simultaneously. To enhance the prototype performance in future
advances, optimizing the time response in command execution and setup system to individual user's
requirements is recommendable.

Keywords
Electrical wheelchair, disabilities person, voice recognizing, threading programming1

1. Introduction
Motor disability is a condition characterized by an individual's inability to independently engage
in physical movement. Peru has a notable prevalence in this particular domain, and despite
ongoing efforts towards the integration of individuals with disabilities, achieving this objective
remains challenging within our societal context. The inclusion of individuals with disabilities in
all economic activities is essential for promoting their independence. However, it is often
observed that these individuals have challenges in securing employment due to their specific
disability-related problems. According to the annual report of the Informatic and Statistic
National Institute of Perú (INEI) in 2013, it was found that a total of 932 thousand individuals
exhibit motor disabilities, which encompass challenges in mobility, ambulation, and the
utilization of limbs [1]. According to a recent statistics analysis published by the National Institute
of Statistics and Informatics (INEI) in 2019 [2], approximately 10% of the Peruvian population is
affected by impairments. Among this group, 15.1% specifically experience motor disabilities,
which corresponds to a total of 484,598 individuals facing challenges related to mobility and
ambulation. The aforementioned publications provide an analysis of the various constraints
experienced by individuals with disabilities, encompassing physical, occupational, social, and
economic dimensions.

CITIE 2023: International Congress on Education and Technology in Sciences, December 04–06, 2023, Zacatecas,
Mexico
jmendozamo@utp.edu.pe (J. Mendoza-Montoya); u18100130@utp.edu.pe (K. E. Angulo-Oviedo);
u18103301@utp.edu.pe (E. Y. Turpo-Apaza); c19206@utp.edu.pe (R. Alfonte-Zapana)
0000-0001-9365-1723 (J. Mendoza-Montoya); 0000-0002-5602-201X (R. Alfonte-Zapana)
© 2023 Copyright for this paper by its authors.
Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
CEUR Workshop Proceedings (CEUR-WS.org)
CEUR
ht
tp:
//
ceur
-ws
.or
g
Works
hop I
SSN1613-
0073
Pr
oceedi
ngs

CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
In the academic literature, there has been significant emphasis placed on the investigation of
electrical wheelchair navigation. As a result, numerous solutions have been proposed for
controlling an electrical wheelchair through the integration of electronic interfaces with a
joystick [3]. This particular control method relies on manual input through the use of hands.
However, in cases where individuals are unable to move their arms or hands due to disabilities,
alternative solutions based on signal processing are required. These solutions involve the
utilization of bio-signals, which are generated from various parts of the human body. For instance,
the bio-signal known as electro-oculography (EOG) can be produced through eye movement and
can be obtained using a system consisting of an operational amplifier, an analog-to-digital
converter (ADC), and a computer. This setup allows for the processing of the bio-signal and the
transmission of commands to the wheelchair [4, 5, 6]. Another type of biological signal is the
electroencephalogram (EEG), which originates from brain activity. However, the analysis and
processing of this signal are highly intricate and necessitate the use of expensive equipment [7].
The electromyogram (EMG) is a cost-effective method for capturing bio-signals associated with
muscle movements. However, its implementation necessitates the use of bio-amplifiers, analog-
to-digital converters (ADCs), and computational resources for signal processing [7, 8]. Voice
recognition technology has emerged as a highly promising solution for individuals with motor
disabilities. To enable this functionality, the use of a headset and microphone is required to
facilitate communication and issue commands to the wheelchair. In order to accurately interpret
and respond to specific verbal instructions for wheelchair control, the integration of dedicated
interpreter software and a microcontroller capable of processing voice signals is imperative [9,
10, 11, 3, 12, 13]. In conclusion, certain hybrid or dual approaches have been suggested in the
form of voice and vision-based mechanisms [14] or head gesture recognition [15, 16].
Nevertheless, it should be noted that these solutions tend to be expensive and frequently
necessitate the involvement of an expert to facilitate their utilization, upkeep, and instruction.
The objective of this study is to present a conceptual design for an electric wheelchair that can
be controlled using both voice commands and a joystick. The inclusion of these dual control
options is intended to assure the wheelchair's functionality in the event of voice command failure.
In such a scenario, the joystick serves as a backup control mechanism, allowing the wheelchair to
continue running seamlessly. The speech recognition library in Python is employed for voice
recognition, while the Arduino Atmega embedded system is utilized for processing and
recognizing voice commands. Additionally, the Arduino Atmega system is responsible for
controlling the motor that enables the wheelchair's regulated navigation.
The present study is organized in the following manner: Section 2 provides a comprehensive
account of the materials and methods employed in this study, encompassing the hardware design
and software development aspects. Section 3 of the document provides an exposition of the
experimental findings. Section 4 of the document encompasses the presentation of the
conclusions and recommendations.

2. Material and method


2.1. Hardware Design

2.1.1. Structural design


The construction process of the wheelchair prototype commenced by implementing the
measurements specified in the design. This involved cutting the pipes and subsequently joining
them together using appropriate couplings. These steps were undertaken to create the
foundation and subsequently the complete chassis of the prototype, with the aim of assuring its
stability and symmetrical structure. Following this, drillings were conducted in order to install
steel plates that were designed to give support for the motors and provide a level surface for their
attachment using bolts and nuts. This process was carried out to ensure the structural soundness
of the motor and to offer support for the primary axis of the tire. Subsequently, the front tires are
affixed and calibrated in order to finalize the assembly of the chassis. In addition, an assessment
is conducted on the state of the components, ensuring that they do not encounter any obstacles
in the movement of the prototype.
The wheelchair operates by utilizing batteries that have been conditioned and are housed
within a white box located underneath the seat. The strategic placement of the batteries
contributes to maintaining the stability of the chair. The correctness of the wheelchair's design is
demonstrated through the precise dimensions of its sections, weight support capabilities, and
center of gravity. This particular prototype has been engineered to accommodate individuals
weighing up to 70 kg. After the completion of the structural design, a series of paint layers are
applied to enhance the aesthetic appearance of the chassis, resulting in the final appearance
depicted in Figure 1.

Figure 1: Design and completion of the wheelchair prototype's assembly

2.1.2. Electronic system

The electronic system comprises two power H-bridges, specifically the LM298n module, which
incorporates buffers and enables the control of two motors. In our situation, these motors are
responsible for controlling the left and right movements of the wheelchair. The control circuit is
regulated by an Arduino embedded system, which incorporates an atmega328 microcontroller.
This microcontroller, with its eight-bit architecture, possesses sufficient computational
capabilities to effectively handle real-time data processing for our specific application. In order
to access the voice recognition library in the cloud, it is necessary for the embedded system to
establish an internet connection.
The system initiates the activation of two H-bridges to enable continuous control of the left
and right motors. The movement of the wheelchair is determined by six fundamental commands,
Table 1.
Table 1
Six voice commands
English Spanish
Fordward Avanza
Turn right Derecha
Turn left Izquierda
Backward Retrocede
Fast Rápido
Stop Para

A pulse width modulation (PWM) signal is sent from the microprocessor to each H-bridge that
controls the gearmotors and controls the speed motor as well, Figure 2. It works with the Echo
and Trigger inputs, which are linked to Arduino pins 7 and 6.
Below is a list of the parts that were used in the system.
• 1 Arduino uno with 5v
• 1 LM298n H-bridge 5v
• 2 12v/730mA gearmotors with 4.2Nm of torque
• 1 ultrasound monitor with 5v
• 2 power inputs (5v to 12v, 20W)
• 1 microphone
• 1 headset
Python is needed to build the system, along with the speech recognition library and the
algorithm to understand voice directions like "Avanza" for "Forward," "derecha" for "turn right,"
"izquierda" for "turn left," "Retrocede" for "backward," "Para" for "stop," and "Rápido" for "Fast."

Figure 2: Arduino-H-bridge L298n two-gearmotor control


(https://naylampmechatronics.com/)

2.1.3. Joystick control circuit

In order to operate the wheelchair using a joystick, a logical framework will be implemented,
consisting of four distinct movements: forward, backward, left turn, and right turn. Each of these
movements will be calibrated to accommodate pulse width modulation based on the degree of
depression of the lever positioned at its center. To analyze this, we represent the range of joystick
displacement for each movement on a Cartesian coordinate system. In order to make progress,
we divide the Cartesian region into two zones: Zone "a" represents the left zone, which begins at
0, and Zone "b" represents the right zone, which ends at 1023. The range is measured in bits,
taking into account the sensitivity of the analog-to-digital converter (ADC). In this particular case,
the ADC has 10 bits, resulting in a range of [0-1023] bits, which is equivalent to a voltage range
of [0-5] Volts. The center zone can be adjusted based on the joystick's sensitivity and individual
requirements.
To facilitate the selection of the mode for joystick or voice command, a red button has been
incorporated to enable the immediate choice of the desired operation mode by simply pressing
the button.

2.2. Software design

2.2.1. Voice control


In order to facilitate navigation through voice commands, it is necessary to utilize the Python
programming language along with the Speech Recognition library. Additionally, the following
libraries are required: time, serial, pyttxs3, soundfile, and sounddevice. Similarly, it is imperative
to assess the voice recognition service provided by Google for the Spanish language. To
accomplish this, it is proposed to construct a comprehensive table that documents the outcomes
of each command, categorizing them as either successful or erroneous. Additionally, the
effectiveness of the system will be quantified by determining the percentage of successful
commands. In brief, the algorithm will facilitate the evaluation of Spanish voice recognition by
performing the following tasks:
An example of numbered list is as following.
1. Connect Arduino to a USB port.
2. Make a list of the directions that will be used to tell the actuators what to do through
Arduino: ['forward', 'forward', 'forward', 'backward', 'retro', 'right', 'right', 'left', 'left',
'stop', 'anti', 'stop', 'stop', 'quickly']
3. A sample rate of 44100 Hz is set for recording a ".wav" file. That is, 44,100 pieces of sound
will be recorded for every second. 44100 samples were picked because that's a common
number for playing back audio and it's enough to get most of the details of the recording.
4. We set the time to "X" seconds to give the user more time to record in case the word gets
longer and to take into account how long each Python command takes to run. A person's
time for saying each order will help us figure out what "X" is worth.
5. With the previous settings, the audio is taken, and when it's done, a ".wav" file is made
that will be overwritten over and over again.
6. The recorded sound will be cleaned of noise, and the recognition feature set to use the
Google service and speak Spanish will be used.
7. The word that was recognized must be saved in a "command" variable. All words that
were recognized will be made lowercase.
8. Any words that could have been recognized will be split up and the space between them
will be taken out, leaving us with just the lowercase word.
9. The information that was gathered will be compared word by word until one fits a
command that was set up.
10. After detecting the first spoken word, a signal will be sent to the Arduino through the
serial port. The Arduino will read it and tell the H-bridge how to control the motors'
rotation.
11. The last order that was spoken will be saved and work until a new command is spoken.
12. Inside the Arduino IDE, a bit will be set aside for each type of movement so that they can
be controlled. In this case, "Forward" will be run if Python sends bit 1, "Backward" will be
run if bit 2 is received, and so on. One problem with this way of moving the wheelchair is
that it throws off the user's timing with the recording. Because this was such a big
problem, an algorithm was made that runs the listener in one thread and the orders in
another. This way, the orders are carried out in the motors while the listener waits for
new directions. In the same way, this new way of working lets us get the best response
time from the motors to the controller's commands.

This is how the new program is put together.


1. Threading, a special library, is loaded so that the program can run in one or more
processing threads.
2. The process of establishing serial communication with an Arduino device is initiated.
3. The aforementioned compilation of instructions is designated with a name.
4. Develop a function that encompasses the implementation of speech recognition and
offers formatting capabilities for the transcribed words.
5. Furthermore, this function will specify the commands that are to be transmitted to the
Arduino.
6. A new loop function is implemented with the sole purpose of actively monitoring and
awaiting incoming commands within a specified time frame of 1.1 seconds.
7. The voice recognition feature is invoked to operate concurrently with the process of
listening to incoming voice commands.
By employing this novel recognition technique, it is possible to guarantee a decrease in reading
inaccuracies and enhance the efficiency of actuator reaction time.

2.3. Design of functional prototype

2.3.1. Wheelchair security


The proposed prototype will use a dual security system. To initiate the cancellation of a
command, the user has the option to adjust the selector position. This action effectively halts the
movement of the chair while transitioning between voice navigation mode and joystick
navigation mode, or vice versa. Similarly, the prototype will incorporate ultrasonic sensors
capable of detecting immediate obstructions at a range of less than 40 cm. This feature will enable
the chair to halt its movement prior to any potential collision with an object. In this study, a total
of 256 measurements were conducted using a specific sensor across several materials. The
objective was to assess the potential inaccuracy in measurements at distances of 100 cm, 50 cm,
and 40 cm in diverse materials, including concrete, cardboard, Medium Density Fiberboard
(MDF), mirror plastic, and expanded polystyrene.

3. Results
3.1. Voice recognition
The experimentation involved conducting success and error tests with Google's speech
recognition library, and the outcomes are presented in Table 2.

Table 2
Voice command hits and misses list
Command Success Error
“Avanza” 18 (90%) 2 (10%)
“Retrocede” 13 (65%) 7 (35%)
“Izquierda” 10 (50%) 10 (50%)
“Derecha” 18 (90%) 2 (10%)
“Para” 19 (95%) 1 (5%)
“Rapido” 20 (100%) 0 (0%)
Based on these data, it is possible to find out how effective each command is and how effective
all commands are.
1. “Avanza”: 18 hits were found out of 20 samples. It works 90% of the time when the order
is given.
2. “Retrocede”: Thirteen out of twenty samples gave right answers, which is a 35% error
rate. Sometime this order wouldn't work in an emergency, which could hurt the person
using it if the ultrasonic sensor emergency stop wasn't available. The order is changed to
"retro," and the test is done again.
3. “Izquierda”: Out of 20 samples, 10 had right answers, which means there was a 50% error.
If we leave this command alone, it wouldn't be very useful in an emergency. But if the user
types the same command more than twice and doesn't get an answer, they might think
that the recognition isn't working and stop using this option. We change this command to
"anti" (which means reverse) and run the test again.
4. “Derecha”: 18 hits were found out of 20 samples. It works 90% of the time when the order
is given.
5. “Para”: Nineteen hits were found out of twenty samples. It works 95% of the time when
the order is given.
6. “Rapido”: Twenty samples were used, and twenty hits were found. It works every time
that the order is given.
7. We'll now change "Izquierda" to "anti" and "Retrocede" to "retro" to get the result shown
in Table 3.

Table 3
Recognizing commands after making changes
Retrocede Retro Izquierda anti
13 19 10 17
65 % 95 % 50 % 85 %

We will add "anti" and "retro" as key commands for left turn and reverse after this test. This
will also make it faster to say a command.

3.2. Voice command control

A novel software application was developed with the purpose of regulating the functioning of
the motors. In order to supply power to the Arduino and the L298n, a power bank was employed.
Similarly, it was linked to the computer system and initiated the issuance of commands, which
were executed in accordance with anticipated outcomes. An issue that became apparent was the
rather long recognition time of approximately 4 seconds for our initial code implementation.
Table 4 displays the duration it takes for an individual to articulate each of the previously
suggested instructions.

Table 4
Command recognizing delay
N Avanza Retrocede Izquierda Derecha Para Rápido
1 0.567 0.940 0.827 0.626 0.379 0.673
2 0.520 0.809 0.860 0.580 0.423 0.455
3 0.529 0.831 0.714 0.639 0.401 0.586
4 0.473 1.094 0.682 0.637 0.410 0.629
5 0.540 0.930 0.721 0.584 0.334 0.719
6 0.814 0.824 0.752 0.620 0.379 0.694
7 0.700 0.853 0.601 0.586 0.401 0.738
8 0.669 0.924 0.850 0.599 0.360 0.607
Mean 0.602 0.901 0.751 0.609 0.386 0.638

Consequently, a novel experiment was conducted utilizing the updated orders in order to
enhance the time response. The outcomes of this experiment are outlined in Table 5, presented
as follows:

Table 5
Command recognizing delay after change order
Retrocede Retro Izquierda Anti
0.901 0.523 0.751 0.524

Figure 3: Time response in seconds before and after optimizing commands

Regarding the preceding instructions, we have obtained these novel findings with a reduced
time interval. The new "anti" command for the "Izquierda" function offers enhanced precision
and improved efficiency, resulting in reduced recording time.

3.3. Distance ultrasound sensor

In accordance with our established approach, an assessment was conducted to determine the
precision of the HCSR-04 sensor when measuring various materials, including concrete,
cardboard, MDF, mirror plastic, and technopor. The results of this evaluation are presented in
Table 6.
In order to have a deeper comprehension of the data, an analysis was conducted on the
dispersion pertaining to the mean. Hence, it is evident that the sensor exhibits a superior
responsiveness when applied to concrete surfaces. However, when utilized with alternative
materials such as mirror plastic and styrofoam, the signal may introduce inaccuracies in distance
estimation. Hence, it is recommended that the wheelchair software control be modified to
incorporate a joystick interface in order to enhance the navigation performance.
Table 6
Standard deviation of how far the HC-SR04 can measure
Material Mean Standard deviation
Concrete 100.0636 0.1236
Cardboard 100.0377 0.1880
MDF 99.6463 0.1874
Plastic 100.6274 0.4727
Technopor 99.9567 0.2714

Figure 4: Data spread over a one-meter distance measured with the HC-SR04

As shown in Figure 4, the HSR-04 ultrasound sensor responds differently to different


materials. Concrete gives more accurate readings, but there isn't much difference between the
readings of other materials (a maximum standard deviation of 0.47 cm), so the system could keep
the wheelchair moving.

3.4. Structural analysis


In the mechanical part, 3D software was used to do the structural analysis. This gave us useful
information about how the chassis's structure would behave in different situations. In our
models, we think that the chassis could have been made of PVC, aluminum, or steel tubes, with
the same material used for the joints between the tubes.
During the simulations test, we look at the following features:
• A force of 9.81 m/s2 of gravity
• The patient's weight limit is 40 kg
• The motor weighs 1 kg
• The battery weighs 3 kg
• The electrical parts weighs 1 kg
With these numbers, we'll start the simulations that go with them and show the results in
terms of shrinkage, taking three things into account at the same time for this test.
1. PVC
2. Aluminum
3. Steel
We want to find the best material out of all of them to use as a base for cheaply putting together
this type of wheelchair.

Figure 5: Analysis of structure in three materials: A) Steel B) Aluminum C) PVC

With these simulations, we can see that the deformation changes depending on the chassis's
shape and the type of material used. Following these tests, the PVC frame has the most
deformation, going up to a height of 6,307 mm. The next one is made of metal and is 0.282 mm
thick. The last one is made of steel and is 0.239 mm thick. The displacement zone in all of these
simulations is the same. It's in the area where the load acts directly (Table 7).

Table 7
Deformation in different materials
Material Max. deformation (mm)
PVC 6.307
Aluminum 0.282
Steel 0.239

3.5. Power-Weight relationship


For a power of 350 and a weight of 45 kg, the power-to-weight ratio is 7.78, which can be seen
by comparing the numbers to Table 8. Additionally, using the information in the same table, we
can say that if we add more mass, the weight-to-power ratio will go down because the engine will
have to make more power to make up for it.

Table 8
Relationship of power and weight
Power 45kg 50kg 55kg 60kg 65kg 70kg
120W 2.7 2.4 2.2 2 1.8 1.7
150W 3.3 3 2.7 2.5 2.3 2.1
180W 4 3.6 3.3 3 2.8 2.6
210W 4.7 4.2 3.8 3.5 3.2 3
240W 5.3 4.8 4.4 4 3.7 3.4
270W 6 5.4 4.9 4.5 4.1 3.8
300W 6.7 6 5.4 5 4.6 4.3
330W 7.3 6.6 6 5.5 5 4.7

4. Conclusions
In summary, we have successfully developed a mechatronics-based prototype wheelchair that
incorporates essential features and is cost-effective. This design enables manual navigation using
a joystick or automated navigation through voice commands issued by the user.
A cost-effective prototype of a wheelchair was developed using both computational modeling
and physical construction, with the primary material being polyvinyl chloride (PVC). A
comprehensive structural analysis was conducted on the prototype in order to validate its ability
to endure the designated load, despite its relatively modest cost. Furthermore, the final iteration
of the wheelchair prototype incorporated the installation of tires sourced from an authentic
wheelchair, so enhancing its aesthetic appeal and lending it an authentic semblance.
The development of a control system that effectively coordinates the navigation of the
prototype wheelchair has been successfully achieved. This system demonstrates adequacy,
efficiency, and the potential for optimization. The implementation involved utilizing speech
recognition and threading techniques. Python was employed as the programming language for
voice command control, while the Arduino Software was utilized to facilitate the transmission
and reception of instructions to the actuators.
Similarly, we have successfully enhanced the efficiency of voice command responses by
reprogramming and organizing the available phrases or commands in a manner that ensures user
comfort. In addition, it is possible to receive these words simultaneously with the transmission
of commands to the actuators. Similarly, a delay of 1.1 seconds was achieved, resulting in an
enhanced efficiency of our workflow that initially experienced a wait of 4 seconds between the
execution of each order. Ultimately, we replaced the lexemes that exhibited the greatest temporal
duration for directives that conveyed an identical directive, yet possessed enhanced efficiency in
terms of articulation.
Ultimately, we have successfully integrated all the constituent stages or systems comprising
the mechatronic wheelchair prototype in order to validate its functionality and enhance its
efficiency. This phenomenon is evident in the outcomes gained and is demonstrated in the
present research study.

5. Recommendations
Based on the findings of our research, we are able to offer the following advice for future
endeavors.
• The prototype needs to be made from something other than PVC so that it can hold a
person who weighs more than 50 kilograms. Aluminum may be a good choice because
it is not more expensive than other materials and can hold more weight than PVC.
• Similarly, we suggest getting motors with more power that can move the suggested
load through some kind of extra transmission. Online shops sell these, but they cost a
lot more than the ones we used for this project.
• To use electronic control, we suggest making a PCB card with all the parts that will be
used. In the same way, a Raspberry Pi or an ESP32 card can be used instead of an
Arduino. This makes sending messages to the actuators faster. Also, we believe it
would be more effective to use the VRM v3 Voice Recognition Module version 3. This
module can recognize voice commands without an Internet connection, does not
require a computer for signal processing, and lets you train up to seven commands.
• Finally, based on where the prototype will be used, we suggest adding more sensors
to the area around the wheelchair to make the person using it safer.
Acknowledgements
We express our gratitude to Universidad Tecnológica del Perú for their valuable logistical
assistance, which has facilitated the execution of this review paper.

References

[1] INEI, En el perú 1 millón 575 mil personas presentan algún tipo de discapacidad, 2013. URL:
https://m.inei.gob.pe/media/MenuRecursivo/noticias/np-178-2013.pdf.
[2] INEI, En el país existen 3 millones 209 mil 261 personas con discapacidad, 2019. URL:
https://m.inei.gob.pe/media/MenuRecursivo/noticias/np_136_2019_inei.pdf.
[3] S. M. Shifa, T. M. Mridul, S. S. Rafat, M. H. Monjur, M. R. Islam, M. K. Hassan, Design and
implementation of smart wheelchair with advanced control interfaces, in: 2023 3rd
International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST),
2023, pp. 155–159. doi:10.1109/ICREST57604.2023.10070067.
[4] J. P. A. Ogenga, P. W. Njeri, J. K. Muguro, Development of a virtual environment-based
electrooculogram control system for safe electric wheelchair mobility for individuals with
severe physical disabilities, Journal of Robotics and Control (JRC) 4 (2023) 165–178. URL:
https://doi.org/10.18196/jrc.v4i2.17165. doi:10.18196/jrc.v4i2.17165.
[5] J. Xu, Z. Huang, L. Liu, X. Li, K. Wei, Eye-gaze controlled wheelchair based on deep learning,
Sensors 23 (2023). URL: https://www.mdpi.com/1424-8220/23/13/6239. doi:10.
3390/s23136239.
[6] L. Maule, M. Zanetti, A. Luchetti, P. Tomasin, M. Dallapiccola, N. Covre, G. Guandalini, M. De
Cecco, Wheelchair driving strategies: A comparison between standard joystick and gaze-
based control, Assistive technology : the official journal of RESNA 35 (2023) 180—192. URL:
https://doi.org/10.1080/10400435.2021.2009593. doi:10.1080/10400435.
2021.2009593.
[7] L. Gunarathne, D. Welihinda, H. Herath, S. Yasakethu, Eeg-assisted emg-controlled
wheelchair for improved mobility of paresis patients, in: 2023 IEEE IAS Global Conference
on Emerging Technologies (GlobConET), 2023, pp. 1–6. doi:10.1109/GlobConET56651.
2023.10149923.
[8] M. Gopichand, K. Rajeswari, E. Deepthi, Human–machine interface for wheelchair control
using semg signals, in: A. Kumar, G. Ghinea, S. Merugu, T. Hashimoto (Eds.), Proceedings of
the International Conference on Cognitive and Intelligent Computing, Springer Nature
Singapore, Singapore, 2023, pp. 395–406.
[9] K. Muthulakshmi, C. Padmavathy, N. Kirthika, B. Vidhya, M. A. P. Manimekalai, Internet of
things and smart intelligence-based google assistant voice controller for wheelchair, in: S.
Shakya, G. Papakostas, K. A. Kamel (Eds.), Mobile Computing and Sustainable Informatics,
Springer Nature Singapore, Singapore, 2023, pp. 761–774.
[10] T. K. Hou, Yagasena, Chelladurai, Arduino based voice controlled wheelchair, Journal of
Physics: Conference Series 1432 (2020) 012064. URL: https://dx.doi.org/10.1088/1742-
6596/ 1432/1/012064. doi:10.1088/1742-6596/1432/1/012064.
[11] S. M. Omair, S. A. Syed, M. M. Khan, A. Shaikh, R. Ismail, H. Shah, Z. Irfan, Prototype model of
voice activated wheelchair, in: 2023 Global Conference on Wireless and Optical Technologies
(GCWOT), 2023, pp. 1–6. doi:10.1109/GCWOT57803.2023.10064667.
[12] S. Murugaiyan, J. Varatharaj, K. Anandraj, M. Duraisamy, Voice controlled wheelchair, AIP
Conference Proceedings 2581 (2023) 030001. URL: https://doi.org/10.1063/5.0126311.
doi:10.1063/5.0126311.
[13] H. F. Fakhruldeen, A. A. Meri, A. H. Sa'id, A. N. Makttoof, M. A. Kadhim, H. A.-J. Al-Asady, An
arduino-based voice-recognition elevator for special purposes, Indonesian Journal of
Electrical Engineering and Computer Science 31 (2023) 828. URL:
https://doi.org/10.11591/ ijeecs.v31.i2.pp828-834. doi:10.11591/ijeecs.v31.i2.pp828-834.
[14] M. Anshar, R. Sadjad, D. Dewiani, M. Abry, R. Prayudha, F. Fiqhi, M. Takbir, Control system
design for smart wheelchair robot, in: AIP Conference Proceedings, volume 2543, AIP
Publishing, 2022.
[15] I. K. Somawirata, F. Utaminingrum, Smart wheelchair controlled by head gesture based on
vision, Journal of Physics: Conference Series 2497 (2023) 012011. URL: https://dx.doi.org/
10.1088/1742-6596/2497/1/012011. doi:10.1088/1742-6596/2497/1/012011.
[16] S. Chatzidimitriadis, S. M. Bafti, K. Sirlantzis, Non-intrusive head movement control for
powered wheelchairs: A vision-based approach, IEEE Access 11 (2023) 65663–65674.
doi:10.1109/ACCESS.2023.3275529.

You might also like