1. Introduction
Since entering the 21st century, China’s population aging problem has been deepening [
1,
2], and the number of people with disabilities is also on the rise, increasing the demand for wheelchairs year by year [
3]. Ground mobility robots such as robotic wheelchairs can significantly improve people’s comfort and quality of life [
4,
5]. The traditional wheel + chair, electric wheel + chair + control handle is the mainstream of the current wheelchair, which focuses on the “human-chair” motor function interaction to ensure the elderly and people with disabilities’ essential travel. For people with visual, auditory, tactile, and physical disabilities, the current wheelchair has apparent deficiencies in achieving independent travel for this group. In addition, the traditional wheelchair does not have the function of autonomous navigation. In the process of using it, it is often necessary to have auxiliary personnel to ensure the travel safety of the wheelchair occupants.
To make up for the shortcomings of traditional wheelchairs and ensure the safety of wheelchair occupants, it is crucial to combine the relevant technologies in the field of robotics, such as SLAM (Simultaneous Localization and Mapping), environment perception, motion control, path planning and multimodal human–computer interaction, etc., to upgrade and transform the traditional wheelchair, and to develop an intelligent wheelchair with multimodal human–computer interaction and autonomous navigation functions, which is of excellent research significance and practical value.
Intelligent wheelchair mobility can be categorized into manual, electric, state-detecting, and autonomous navigation mobility, in which state-detecting mobility is to control the movement of the wheelchair by detecting the wheelchair occupant’s electroencephalogram information, electromyography information, pupil position, gesture, or head posture [
6,
7,
8,
9]. Autonomous navigation is to sense the wheelchair by installing LiDAR, inertial sensors and vision sensors, etc., to sense the wheelchair environment around the wheelchair, thus realizing the autonomous movement of the wheelchair. Li et al. proposed a navigation method combining differential GNSS (Global Navigation Satellite System) and LiDAR SLAM, improving localization and navigation accuracy [
10]. Ferracuti et al. proposed the human-in-the-loop framework, which is used to navigate a wheelchair to its destination in indoor scenarios according to the EEG signals generated by the person after the path-planning error. EEG signal serves as an additional input for navigation, thus realizing real-time path modification [
11]. Li et al. studied a computer vision-based wheelchair following system, which utilizes camera visual information to achieve target tracking, position prediction, and localization functions [
12]. Wang et al. proposed a path-planning method for robotic wheelchair navigation, which combines the utility function of human comfort and path cost for path optimization through navigation map modeling to improve the efficiency of optimal trajectory search and ensure navigation safety [
13]. Maksud researched and designed a brain–computer interface-based smart wheelchair, which acquires attention and blinking signals through a wearable device and then controls the wheelchair through virtual maps, completes destination mapping, and autonomously reaches the desired location [
14].
The human–computer interaction of the intelligent wheelchair can be divided into two aspects: (1) two-way human–computer interaction between the occupant and the wheelchair, where the occupant can check the status information of the wheelchair through the display screen and control the movement of the wheelchair through gestures or voice; and (2) two-way human–computer interaction between the guardian and the wheelchair. With the help of a cloud server, 5G, and other technologies, the guardian can interact with the wheelchair remotely through the cell phone APP, check the parameters of the wheelchair, or control the wheelchair.
In terms of human–computer interaction between the occupant and the wheelchair, Xu et al. designed an intelligent wheelchair with eye-movement control, which acquires the occupant’s eye image through a camera and uses deep learning to determine the direction of eye movement, as well as to establish a motion acceleration model for the intelligent wheelchair to improve motion smoothness [
15]. Cui et al. proposed an intelligent wheelchair posture adjustment method based on action intent recognition, which adjusts the wheelchair posture by investigating the relationship between the force changes on the contact surface between the human body and the wheelchair and the action intent [
16]. Wanluk proposed the concept of an eye-tracking intelligent wheelchair, which can control the movement of the wheelchair through the eyes and remotely control some electrical devices [
17]. Aktar developed an intelligent wheelchair based on a speech recognition system, where the user can use voice to control the wheelchair’s movement and speed and use infrared sensors to ensure that the wheelchair is within a safe distance from obstacles [
18]. Dey proposed an intelligent wheelchair system for head posture navigation, which integrates modules such as acceleration, ultrasound, and photoresistor, and the system can realize the wheelchair movement in five directions according to different head postures [
19]. Welihinda et al. proposed a hybrid control system to operate a powered wheelchair using a combination of EEG and EMG, and developed an EEG-based user attention detection system and an EMG-based navigation system [
20]. Regarding human–computer interaction between the guardian and the wheelchair, Lu et al. used Internet of Things (IoT) technology to monitor the data of the position of the electric wheelchair, the tire wear, the battery power level, etc., and the guardian can know the user’s operation status [
21]. Li et al. designed a mobile terminal APP based on virtual reality and IoT technology, sent commands to a remote server through the app, and completed rehabilitation training with the help of virtual reality technology [
22]. Cui et al. developed an intelligent wheelchair with multimodal sensing and control, which utilizes LiDAR and temperature and humidity sensors to sense environmental information and achieve wheelchair control with the help of sensors such as gestures and handles [
23].
In this paper, the human–computer interaction and assisted driving technology of intelligent wheelchairs are studied, and the following work is mainly accomplished:
(1) In this paper, a novel intelligent wheelchair system is developed, which innovatively integrates multimodal human–computer interaction technology and autonomous navigation technology. A multi-modal intelligent human–computer interaction framework including occupant–wheelchair and guardian–wheelchair, as well as a “human-in-the-loop” intelligent wheelchair autonomous navigation framework are proposed.
(2) A novel multi-modal human–computer interaction framework based on the principle of functional substitution is proposed to solve the problem of missing human functions in the “human-in-the-loop” system. In terms of occupant–wheelchair interaction, quick control is realized by installing a handle on the traditional wheelchair; different gestures and voice commands are recognized by gesture and voice recognition sensors, so that occupants with hand disabilities can control the wheelchair conveniently. The occupant’s head gesture is recognized and mapped onto the two-dimensional plane to control the wheelchair movement; and in terms of guardian-wheelchair interaction, remote human–computer interaction is realized by means of a cloud server and a cell phone APP.
(3) A “human-in-the-loop” autonomous navigation framework for smart wheelchairs is proposed. The technology innovatively integrates speech technology and indoor navigation technology. By integrating sensor modules such as LiDAR and inertial measurement unit (IMU), the system realizes accurate mapping of the indoor environment and real-time sensing of movement status. Based on the environment map and path planning algorithm, the system outputs precise speed commands. In addition, combined with the embedded processor and voice recognition technology, the wheelchair is able to realize accurate autonomous control and fixed-point navigation based on voice commands, providing a more intelligent and flexible travel solution.
The remainder of the paper is organized as follows.
Section 2 gives the general design of the intelligent wheelchair. The third part investigates the human–computer interaction techniques of the wheelchair, including wheelchair–occupant and wheelchair–guardian human–computer interaction.
Section 4 explores the indoor navigation techniques of the intelligent wheelchair.
Section 5 describes wheelchair experiments.
Section 6 summarizes the thesis.
3. Intelligent Wheelchair Multi-Mode Human–Computer Interaction Technology
In the process of using traditional wheelchairs, there are the following problems in human–computer interaction: (1) handle controllers are designed primarily for people without physical impairments, and for users with missing hands or inability to hold the handle have to rely on others to realize autonomous travel; (2) in traditional wheelchair travel, the family members of the occupants are unable to obtain the travel information of the wheelchair, and are also unable to deal with the emergencies of the wheelchair occupants promptly, and need to realize remote monitoring and control of the wheelchair.
The human–computer interaction of the intelligent wheelchair developed in this paper can be divided into two aspects. (1) The first is two-way human–computer interaction between the occupant and the wheelchair, in which the occupant can view the status information of the wheelchair through the display screen, and control the movement of the wheelchair through gestures, voice or head posture. (2) The second is bidirectional human–computer interaction between the guardian and the wheelchair, in which the family members of the occupant can remotely view the wheelchair travel information with the help of the cell phone APP and remotely control the wheelchair.
Figure 3 shows this study’s architecture design of the multimodal human–computer interaction.
3.1. Wheelchair Mobility Control Based on Gesture Recognition
Traditional wheelchairs require a person to push them to move manually. A motorized rocker controls an electric wheelchair to move forward and backward and steer the wheelchair. This is not friendly to people with muscular atrophy and arm weakness, so we designed a gesture recognition-based wheelchair control to control the movement of the wheelchair by detecting changes in the occupant’s gestures.
To consider recognition accuracy and cost, we chose the ATK-PAJ7620 gesture module, which supports recognizing 4 gesture types: forward, backward, left turn, and right turn. When the chip works, it emits infrared signals by driving infrared LEDs; the raw feature data collected by the information extraction array on the sensor array is stored in the register, the raw data is recognized by the gesture recognition array, and the results are stored. The recognition results are output using the IIC bus.
When using the PAJ7620 module, it is necessary to drive the sensor through three steps, wake-up, initialization, and recognition, and read and write to the bank register area to obtain the gesture information. The movement direction of the wheelchair is defined according to different gestures, as follows: (1) gesture forward, the wheelchair moves forward; (2) gesture left, the wheelchair turns left; (3) gesture right, the wheelchair turns right; and (4) gesture back, the wheelchair stops moving. In addition, when the embedded processor Stm32 detects a change in gesture posture, it sends motor rotation direction and speed control commands to the motor driver via the 485 bus, which, in turn, controls the motion of the wheelchair. The wheelchair motion control command communication protocol is shown in
Table 2. The function bits represent different control methods, where 0x11 is touch screen control, 0x12 is APP control, 0x13 is gesture control, 0x14 is head posture control, and 0x15 is navigation control. Among them, the polarity represents the rotation direction of the left and right wheels: 0x01: left wheel forward rotation, right wheel forward rotation; 0x02: left wheel forward rotation, right wheel reverse rotation; 0x03: left wheel reverse rotation, right wheel forward rotation; and 0x04: left wheel reverse rotation, right wheel reverse rotation. Different control modes are distinguished by setting the priority, while manual and automatic control modes are distinguished by setting the flag bit. The two control modes, gesture and head posture, are suitable for priority control, where gesture control is set to a higher priority. Specifically, when both gesture and head posture control commands are received, the system will prioritize the execution of the gesture control command.
3.2. Smart Wheelchair Voice Recognition Control
The traditional electric wheelchair is still a challenge for people with upper limb disability, quadriplegia, or even more severe physical disabilities, to achieve autonomous mobility. In this study, the audio signal of the occupant is captured using the KeDaXunFei M260C ring array and R818 noise reduction board. The module performs pre-processing, such as filtering and endpoint detection on the captured audio, and then performs feature extraction on the processed signal. The extracted features are used for training or recognition; the training features are used to build a template library. In contrast, the feature recognition is compared with the features in the template library, and the recognition result is outputted according to the degree of match. In this study, the voice control function is turned on by setting the voice wake-up word, and “forward,” “stop”, “turn left”, “turn right” and “autonomous” are set. The speech recognition and control process are shown in
Figure 4.
3.3. Smart Wheelchair Head Posture Control
In this study, a simple vision-based head motion control scheme is implemented, and head posture control can help patients with physical disabilities to realize autonomous travel. A 3D rigid body has two types of motions relative to the camera: translational and rotational. These include , , and axes translational motions and three rotational motions, namely, roll, pitch, and yaw. The estimation of the occupant’s head pose solves for these six parameters.
Suppose a point
in the world coordinate system is known. Assume that the rotation matrix and translation vector are known to be
and
, respectively, to obtain the position of P in the camera coordinate system (
):
From the camera coordinate system to the pixel coordinate system can be calculated by Equation (2):
where
and
are the focal lengths in the
and
-axis directions, respectively, (
) is the optical center, and
is the scale factor.
The relationship between the pixel coordinate system and the world coordinate system can be calculated by Equation (3):
Solve by DLT (Direct Linear Transform) with least squares to find the Euler angles from the rotation matrix.
In this study, the wheelchair motion is controlled by the Yaw and Pitch parameters of the head pose. The Yaw and Pitch of the occupant’s head are mapped to a two-dimensional plane, as shown in
Figure 5. There are five zones, namely, forward zone, stop zone, left turn zone, correct turn zone, and keep current state zone. Within a specific range of parameters, the wheelchair is controlled by the occupant’s head posture to move forward, stop, turn left, and turn right, respectively. When the head is in the middle of a specific range, the current state of the wheelchair is maintained to move forward or stop.
3.4. Remote Control of Smart Wheelchair Based on Cloud Server
With the development of cloud computing and 5G technology, remote monitoring and wheelchair control is also possible. In this study, we designed a wheelchair remote video and control scheme based on the Ali cloud server, using a lightweight video streaming server MJPEG-Streamer to realize the video transmission; the embedded processor uses a 5G module to transmit the data in real time to the Ali cloud server, and the APP on the cell phone is connected to the cloud platform to receive the video data from the cloud platform in real time. At the same time, combined with the cloud platform to realize the remote control of the wheelchair, the specific control scheme is as follows. The occupant clicks on the control button in the cell phone APP, the APP sends control commands to the server, the cloud platform receives the commands and forwards them to the intelligent wheelchair, and the embedded platform of the smart wheelchair analyzes the control commands, and then controls the motor drive module to realize the movement control of the wheelchair. The wheelchair remote monitoring and control program is shown in
Figure 6.
6. Conclusions
To improve the control performance of the traditional wheelchair so that it can be controlled by different types of occupants and complete autonomous travel indoors while the guardian can remotely view and control the wheelchair, we have studied the multi-mode human–computer interaction and assisted driving technology, explored the human–computer interaction technology combining multiple control modes, and explored the possibility of the autonomous navigation of the intelligent wheelchair indoors, with the main functions as follows:
(1) Multi-mode human–computer interaction: In this paper, we design and implement a multi-mode human–computer interaction mode to realize the gesture control, voice control, and head posture control of the wheelchair; the three control modes can be better adapted to different people, and learn the wheelchair’s forward, left turn, right turn and stop functions. The guardian can also view the video data and control the wheelchair remotely with the help of an APP. In addition, the delay time of the remote control is about 150 ms, and the accuracy rate of gesture, voice, and head posture control is over 95%.
(2) Autonomous navigation of smart wheelchair: this paper researches and realizes the indoor navigation technology of the smart wheelchair, including constructing a 2D environment map, global and local path planning, data input, and marking the start and end points. Users mark the endpoint on the map, and the intelligent wheelchair can autonomously navigate to the target point. This dramatically facilitates wheelchair users’ traveling. The accuracy of the autonomous navigation endpoint is less than 10 cm.
The intelligent wheelchair described in this paper has the advantages of a high degree of intelligence, multiple operation modes, and high safety performance. However, after our experiments, there are still the following shortcomings: (1) more sensors are installed on the wheelchair, the wiring is more chaotic, and most of the electrical components are not waterproof, which needs to be solved in the future; (2) when the wheelchair is controlled by the occupant, it is still necessary to increase the number of sensors for detecting obstacles, further enhancing the safety performance of wheelchair travel. In future work, we will consider the implementation of smart wheelchair autonomous navigation target point setting on the guardian APP side.