Next Article in Journal
Design and Simulation of Microbolometer with Dual Cavity for High Figure of Merits
Next Article in Special Issue
A Methodology for Enhancing SSVEP Features Using Adaptive Filtering Based on the Spatial Distribution of EEG Signals
Previous Article in Journal
A Two-Stage Screw Detection Framework for Automatic Disassembly Using a Reflection Feature Regression Model
Previous Article in Special Issue
Localized Surface Hydrophilicity Tailoring of Polyimide Film for Flexible Electronics Manufacturing Using an Atmospheric Pressure Ar/H2O Microplasma Jet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MEMS Devices-Based Hand Gesture Recognition via Wearable Computing

1
School of Intelligence and Electronic Engineering, Dalian Neusoft University of Information, Dalian 116023, China
2
Key Laboratory of Intelligent Control and Optimization for Industrial Equipment of Ministry of Education, Dalian University of Technology, Dalian 116024, China
3
College of Aeronautical Engineering, Taizhou University, Taizhou 318000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Micromachines 2023, 14(5), 947; https://doi.org/10.3390/mi14050947
Submission received: 6 March 2023 / Revised: 19 April 2023 / Accepted: 24 April 2023 / Published: 27 April 2023
(This article belongs to the Special Issue Wearable and Implantable Bio-MEMS Devices and Applications)

Abstract

:
Gesture recognition has found widespread applications in various fields, such as virtual reality, medical diagnosis, and robot interaction. The existing mainstream gesture-recognition methods are primarily divided into two categories: inertial-sensor-based and camera-vision-based methods. However, optical detection still has limitations such as reflection and occlusion. In this paper, we investigate static and dynamic gesture-recognition methods based on miniature inertial sensors. Hand-gesture data are obtained through a data glove and preprocessed using Butterworth low-pass filtering and normalization algorithms. Magnetometer correction is performed using ellipsoidal fitting methods. An auxiliary segmentation algorithm is employed to segment the gesture data, and a gesture dataset is constructed. For static gesture recognition, we focus on four machine learning algorithms, namely support vector machine (SVM), backpropagation neural network (BP), decision tree (DT), and random forest (RF). We evaluate the model prediction performance through cross-validation comparison. For dynamic gesture recognition, we investigate the recognition of 10 dynamic gestures using Hidden Markov Models (HMM) and Attention-Biased Mechanisms for Bidirectional Long- and Short-Term Memory Neural Network Models (Attention-BiLSTM). We analyze the differences in accuracy for complex dynamic gesture recognition with different feature datasets and compare them with the prediction results of the traditional long- and short-term memory neural network model (LSTM). Experimental results demonstrate that the random forest algorithm achieves the highest recognition accuracy and shortest recognition time for static gestures. Moreover, the addition of the attention mechanism significantly improves the recognition accuracy of the LSTM model for dynamic gestures, with a prediction accuracy of 98.3%, based on the original six-axis dataset.

1. Introduction

Gesture recognition is a topic in computer science and language technology [1]. As an extremely efficient non-verbal interaction method, gesture interaction will provide strong technical support for excellence in emerging fields such as smart medical devices, assistive devices for the disabled, smart homes, and smart military operations [2,3]. Most of the current major research work on gesture recognition is focused on machine-vision-based recognition methods, which can pose many limitations in practical applications. Inertial sensor-based gesture recognition is mainly focused on the improvement of recognition algorithms. This limits the application of gesture recognition algorithms in practical products [4,5]. The inertial sensor-based recognition method has an important role in helping to improve the accuracy of gesture recognition. Qiu et al. [6] introduced the devices and key applications of common wearable sensors, and discussed further research directions. It can be seen that wearable device technology could be the main research direction in gesture recognition research.
Gesture recognition technology can be divided into static gesture recognition technology and dynamic gesture recognition technology according to whether it can recognize gestures based on time series [4]. At present, there are two main ways to collect gesture data, which are non-contact, based on machine vision sensors, and contact, based on data gloves [7]. The mainstream data-acquisition gloves are divided into three categories, as shown in Figure 1. Vision-based action recognition has been widely used and generally consists of four steps: gesture detection and segmentation, gesture tracking, feature extraction, and gesture classification [8]. Kinect [9] is a depth vision sensor released by Microsoft in 2010. Based on its built-in algorithm, it can automatically identify and track the dynamic skeletal structure of the human body and apply it to the hand to research human gestures. Researchers have two methods of gesture recognition using Kinect: (1) recognition based on the dynamic skeleton of the human body [10]; (2) recognition based on spatial depth sensing [5]. In the first approach, Ren et al. [11] obtained the skeleton data of 25 joint points of the human body by Kinect, obtained their coordinates in 3D space in real-time, and investigated the importance of each joint bone in dynamic gesture expression. However, they based this on the visual sensor to process the information of all parts of the whole body, and the gesture data were not expressed in enough detail. In the second approach, Wang et al. [12] obtained a higher accuracy rate by studying people’s gesture habits and the depth data from the Kinect depth sensor to control the 3D-printed robot. However, the results indicated that the recognition speed still needs to be improved. The leap motion sensor [13] is a somatosensory controller released by Leap in 2013. Unlike Kinect, it mainly performs skeletal motion tracking of the hand. Li et al. [14] generated finger-motion data with the help of Leap Motion Controller, which was used to calculate the angle of finger joints. This is a typical application of vision-based recognition; in addition to this, the research on inertial sensor-based recognition also has to be further improved.
The research on gesture recognition based on data gloves generally uses inertial sensors, myoelectric sensors, pressure sensors, and bending sensors on data gloves to obtain various gesture signals during hand movement. Alemayoh et al. [15] used inertial force sensors to capture motion data and then trained four neural networks using deep learning methods. The results showed that the vision transformer (ViT) network performed the best, with 99.05% recognition accuracy. Lin et al. [16] designed a data glove with multi-channel data transmission based on hand poses and emotion recognition to achieve simultaneous control of a robotic hand and a virtual hand. Zhao et al. [17] designed a motion-capture device based on a human sensor network with 15 sensor nodes and used the gradient descent method to fuse sensor data to improve the localization accuracy of the motion capture system. Liu et al. [18] proposed a novel gesture recognition device, which consists of a data glove with bending sensors and inertial sensors and a data arm ring with myoelectric sensors, to build a gesture recognition device. However, the sensors need to be used with electrode patches, which are very inconvenient to wear and replace; this is an area in urgent need of improvement. Fu et al. [19] proposed a gesture-recognition method based on a data glove and back propagation neural network. Only the gesture data of numbers 0–10 were used in the experiments, which lacked the recognition of dynamic gestures. Gałka J. et al. [20] introduced a construction of an accelerometer glove and its application in sign language gesture recognition. The basic data of inertial motion sensors and the design of gesture acquisition systems are also introduced. The solution presents the results of gesture recognition, selects a specific set of sign language gestures, and uses a description method based on the hidden Markov model (HMM) and parallel HMM methods. Using parallel hidden Markov models for sensor fusion modelling reduces the error rate by more than 60 % while maintaining 99.75 % recognition accuracy. Qiu et al. [21,22] used inertial sensors and data fusion algorithms to calculate the joint angles of kayakers, four machine learning algorithms were used to investigate the effect of different data combinations on phase classification, and extended Kalman filtering methods were used to fuse the sensor information all of which show good classification accuracy. Tai et al. [23] studied the continuous recognition of six types of gestures using smartphones combined with long- and short-term memory neural networks (LSTM), but the gesture actions used were too simple. The LSTM algorithm in [24] can be combined with convolutional neural networks for VGR-based gesture recognition; although these algorithms have been found to be effective, they still recognize only a single continuous gesture, and the problem of multi-class dynamic gesture recognition remains to be solved. Yuan et al. [25] design a wearable device with two arm loops and a data glove with integrated flexible sensors to capture fine arm and joint movements, and introduced an LSTM model with fused feature vectors as input to verify that the contextual information of gestures can be integrated in the gesture-recognition task and achieve excellent recognition results. Fan et al. [26] proposed a two-stage multi-headed attention human interaction action recognition model based on inertial measurement unit, which can accurately recognize seven interaction actions with an average recognition accuracy of 98.73%.
In order to recognize two types of gestures based on inertial sensors in indoor scenes, we propose recognition and analysis algorithms based on machine learning and deep learning. Applying traditional machine learning algorithms to static gesture recognition, we propose a bidirectional long- and short-term memory neural network model (attention-BiLSTM) based on the attention mechanism for the recognition study of 10 dynamic sign languages. The raw data are collected through a homemade data-collection glove and the gesture information is predicted using different machine learning algorithms. The aim was to improve the accuracy of gesture recognition and reduce the time of gesture recognition by building a gesture model, thus expanding the application scenarios of gesture recognition. The important contributions of this paper are as follows.
  • The raw data were filtered using a Butterworth low-pass filter, the magnetometer data were corrected using an ellipsoidal fitting method, and the dataset was constructed using a gesture-assisted segmentation algorithm.
  • We used four machine learning algorithms to identify static gesture data and evaluate the prediction effect by cross-validation.
  • We constructed a hidden Markov model and an attention-based mechanism neural network model to design recognition methods for dynamic gestures.
This paper is structured as follows. Section 2 describes the hardware and data-acquisition information of the system. The research methodology is described in Section 3. The Section 4 shows the algorithm design of gesture recognition. Section 5 explains the results of this study. Finally, Section 6 shows the discussion and conclusion.

2. Systematic Data Collection and Participants

2.1. System Setup

The gesture data acquired by the inertial sensors were processed by gesture segmentation, filtering and gesture fusion algorithms. The flowchart is shown in Figure 2a. A home-made data acquisition system was the main source of gesture data. The composition of the data acquisition system is shown in Figure 2. This is mainly composed of a pair of data gloves, a WiFi transceiver node and a personal computer (PC) host computer, which can complete the collection and storage of gesture data.
Each glove contains 16 inertial nodes, and the data collected by the inertial nodes can be sent to the PC host computer through the wireless module in real-time. The hardware part of the glove includes 15 inertial nodes and 1 sink node, as shown in Figure 3a. The aggregation node adopts STM32F407VGT6 Microcontroller Unit(MCU) as the main controller and is equipped with an ESP8266 serial peripheral interface bus (SPI) interface WiFi module produced by Espressif. The sampling frequency of the node was set to 100 Hz and the collected data can meet the needs of the gesture recognition algorithm. Because it uses wireless transmission, too high a sampling rate may increase the packet loss rate during transmission. The data of each sensor node are collected and processed by the sink node, and the PC receives them synchronously through WiFi. The positions of 15 inertial nodes correspond to 15 finger bones, which can detect the data of each finger. The convergence node is located on the back of the palm, which can detect the data of the palm. The glove material currently used has a certain degree of elasticity, and can increase the size of the internal volume within a certain range after the installation of sensors. Common hand sizes can basically meet the wearing requirements. The wearing effect is shown in Figure 3b.
The inertial node consists of MPU9250 9-axis sensors: a 3-axis accelerometer, a 3-axis gyroscope, and a 3-axis magnetometer. Taking node 1 of the dynamic one-hand gesture “Sorry” as an example, the obtained nine-axis raw data are shown in Figure 4b. All gesture data were obtained from the author and classmates wearing data gloves and completing the specified gestures. Prior to data acquisition, we had to perform initial calibration, i.e., calibration of the subject’s hand position by specific movements, due to the variation in glove-wearing position between participants, as well as the effects of the system’s duty cycle and the external environment. When collecting gestures, the person being collected faces due north, and the initial gesture posture is naturally drooping. When the system starts acquisition, the upper-computer data information can be observed, and after 3 s the acquisition system completes the coordinate system calibration to facilitate the conversion of gesture data. During the subsequent acquisition actions, the collector can face in any direction. The gesture collection is shown in Figure 4a.
In this paper, we used different recognition classification algorithms for static gesture and dynamic gesture recognition. Static gestures are hand patterns that are fixed at a certain moment in time, and each frame captured by the data glove can be used as a set of gesture samples. A large amount of information about dynamic features can be removed when performing feature extraction, and then traditional machine learning algorithms are used for gesture recognition. Specifically, these include: support vector machines, back-propagation neural networks, decision trees and random forest model algorithms. The dynamic gesture features are obviously more complex than static gestures, with the sign language actions changing all the time and the execution time varying between sign languages. Therefore, we used HMM- and attention-BiLSTM-based bi-directional long- and short-term memory neural network models for the recognition study of 10 dynamic sign languages, and the latter served to validate dynamic sign language recognition on deep learning methods.

2.2. Participant and Gesture Acquisition Actions

Four students recruited from the school participated in the preliminary study. Their average weight was 70.5 ± 2.3 kg and their average height was 1.74 ± 0.58 m. The participants consisted of three male students and one female student, and each student sampled 50 sets of gesture data, for a total of 200 samples per gesture. All participants recorded their height and weight, and received adequate information, and particpants’ consent was obtained.
A total of 20 static-letter gestures and 10 dynamic one-handed gestures were selected for this study. Twenty static-letter gestures were as follows: “A”, “C”, “D”, “E”, “F”, “G”, “H”, “K”, “L”, “M”, “N”, “O”, “Q”, “R”, “S”, “U”, “V”, “W ”, “X”, “Y”. Ten dynamic one-handed gestures were as follows: “Sorry”, “Angry”, “Sad”, “You”, “Hello”, “Effort”, “They”, “Me”, “Thanks”, “Goodbye”. All sign standards are referenced in the “Sign Language Dictionary | SpreadTheSign” page. Chinese sign language standards are used.

3. Methods

3.1. Definition of Coordinate System and Conversion Relationship

The gesture recognition system will involve conversion between common coordinate systems and multiple coordinate systems. The measurement values of the sensor nodes can be converted to the geographical coordinate system by means of the conversion relationships between different coordinate systems. In this system, three coordinate systems are included and each coordinate system is based on the standard right-handed 3D Cartesian coordinate system. Detailed information is as follows.
  • Sensor coordinate system (SCS): typically, the origin of the sensor coordinate system is the sensor center, and the pointing of the three axes is based on the three-axis gyroscope in the sensor.
  • Navigation coordinate system (NCS): the origin is the center of gravity of the hand when standing, and the three axes point to the northeast and the ground direction. It should be noted that north here refers to the north in the geomagnetic sense.
  • Body coordinate system(BCS): according to the spatial posture of the palm and each finger segment, the center of mass of the hand is used as the origin of coordinates.
In the initial state, the data are collected facing north so that the sensor coordinate system and the body coordinate system coincide, which facilitates gesture posture calculation and spatial coordinate conversion [17].
At present, the automatic heading reference system (AHRS) attitude calculation algorithm is the main means of aircraft attitude calculation [27]. Since the spatial description of the gesture node has similar characteristics to the spatial description of the aircraft, Euler angles and quaternions are also used to describe the gesture attitude. Quaternions are defined as follows, where i, j, k are the standard orthonormal basis for unit vector representation in 3D space.
q = q 0 + q 1 i + q 2 j + q 3 k
In the actual gesture data acquisition process, the motion data output from inertial nodes, including acceleration, angular velocity, and magnetic field strength, correspond to SCS. We need to convert the sensor output to gesture pose under NCS. This can be achieved by matrix changes between coordinate systems. Let C B n serve as the pose transformation matrix from the sensor coordinate system to the navigation coordinate system, and C B n can be expressed as a quaternion.
C B n = 1 2 q 2 2 + q 3 2 2 ( q 1 q 2 + q 0 q 3 ) 2 ( q 1 q 3 q 0 q 2 ) 2 ( q 1 q 2 q 0 q 3 ) 1 2 q 1 2 + q 3 2 2 ( q 2 q 3 + q 0 q 1 ) 2 ( q 1 q 3 + q 0 q 2 ) 2 ( q 2 q 3 q 0 q 1 ) 1 2 q 1 2 + q 2 2
The relationship between the attitude angle and the quaternion can be expressed as follows:
θ = arcsin 2 q 0 q 2 q 1 q 3 γ = arctan q 0 q 3 + q 1 q 2 1 2 q 2 2 + q 3 2 ψ = arctan q 0 q 1 + q 2 q 3 1 2 q 1 2 + q 2 2
where γ , θ , ψ represent the cross roll angle, pitch angle and heading angle, respectively.
The inertial measurement unit (IMU) sensor needs to be filtered due to the high level of noise. The that remain data after removing the noise are used for attitude-solving. In this paper the raw data are processed using Butterworth filter, which is a signal processing filter also known as the maximum flat filter. Its filtering effect is shown in Figure 5.

3.2. Calibration of Magnetometer Based on Ellipsoidal Fitting Method

At the beginning of the gesture-recognition process, the magnetometer needs to be calibrated due to the presence of hard iron distortion and soft iron distortion in the surrounding environment. The main sources of hard iron distortion are permanent magnets and magnetized metals. To eliminate the ferromagnetic interference, the ellipsoidal fitting method is used to eliminate the ferromagnetic interference. Soft iron is relatively small and negligible [28]. We used the eight-character calibration method and the single-axis rotation method for magnetometer calibration. That is, after the gesture data acquisition system works, it is rotated around the figure of eight three times and then rotated along each single axis three times. The ellipsoidal fitting results are shown in Figure 6. As can be seen from the figure, before the magnetometer calibration, the measured values are disturbed by the surrounding environment, the fitted ellipsoidal spherical center is not located at the coordinate origin and the sphere is tilted. After the calibration, the center of the fitted sphere is located at the origin and the position is accurate.

3.3. Gesture Dataset Segmentation

The segmentation process of the gesture dataset is as follows. The average of every 50 sets of static gesture data is used as one sample. The sampling frequency is set to 100 Hz, i.e., the average of every 0.5 s of data is used as one gesture sample. Since the task is dealing with time-series-based gesture classification, the sliding window method is used for segmentation, and a total of 200 sets of samples are obtained for each gesture. The length of the frontal sequence of the sliding window is defined, and the valid gesture data are intercepted sequentially starting from the starting point. The detailed data segmentation algorithm is shown in the Algorithm 1. The dynamic gesture raw data consist of two parts, transition gestures and valid gestures, as shown in Figure 7.
Algorithm 1: Assisted Segmentation Algorithm
Input: Continuous gesture data: D = d 1 , d 2 , d 3 , . . . , d n ;
Output: Split start point list S, split end point list O.
1
Transition state mean: m = i = 1 100 d i 100 ;
2
Define the length l of the sliding window sequence, the current window value A 0 , the window value A 1 after sliding 5 times, the window value A 2 after sliding 10 times, E 1 = A 0 A 1 , E 2 = A 1 A 2 ;
3
Traverse continuous gesture data D;
4
if  E 1 > 10 & & A 0 m < 12 & & A 2 m < 12  then Read the current position and add it to the starting point list S; (
5
if  E 2 > 10 & & A 0 m < 12 & & A 2 m < 12  then Read the current position and add it to the end point list O; (
6
End;

4. Design of Gesture Recognition Algorithm

4.1. Static Gesture Recognition Method Based on Machine Learning Algorithm

This section introduces the implementation of the four machine learning algorithm models for gesture recognition. The specific flowchart is shown in Figure 8.
Static gesture refers to the static spatial pose of the hand, and most sign languages convey meaning using the static pose of the hand in the air. For example, thumbs up means “good”. In this process, it is not necessary to reflect the change in the hand posture with time and space; only the static posture of the hand is discussed. The original data collected by the data glove are the data of the 16-node three-axis accelerometer, three-axis gyroscope and three-axis magnetometer. The original data structure is as follows:
R a w D a t a = a x i , a y i , a z i , g x i , g y i , g z i , m x i , m y i , m z i , i 1 , 16
After filtering and attitude fusion algorithm processing, each group of gesture data are composed of Euler angle data of 16 nodes. The data structure is as follows:
S t a t i c G e s t u r e = y a w i , r o l l i , p i t c h i , i 1 , 16
  • Support Vector Machines
    SVM [29] is a binary classifier based on supervised learning first proposed by Corinna Cortes and Vapnik et al. in 1995, whose decision boundary is the maximum margin hyperplane solved for the learned samples.
    The grid search method is an optimization method of parameter selection and cross-validation by specifying a selection list of parameters to be optimized, evaluating the model for all parameter combinations, and finally obtaining the optimal parameters in the list. This method is used in the experimental parameter optimization. The structure of the dataset used is as follows: Training set: T r a i n D a t a S e t = D P e o p l e 1 i , D P e o p l e 2 i , i 1 , 70 . Test set: T e s t D a t a S e t = D P e o p l e 1 j , D P e o p l e 2 j , j 71 , 100 . The training set part is used for algorithm model training, and the test set data are used to test the algorithm recognition accuracy of the generated model.
  • Back-Propagation Neural Network
    Neural network (NN) [30] is a mathematical model or computational model that simulates the structure of biological nerve cells to receive stimulation and generate output signals, and simulates the excitation function, as shown by Michael Houston et al.
  • Decision Tree algorithm
    Decision tree (DT) is a common classification algorithm based on supervised learning in machine learning. Its advantages are its simple structure, logic in line with human intuition, and fast processing speed for large amounts of data. In this paper, we use “information entropy” and “Gini index” to classify the attributes for model training. The CART algorithm and ID3 algorithm have overfitting problems, and the generalization ability of the model can be improved by discarding the over-divided attributes.
  • Random forest algorithm
    Random forest (RF) is an ensemble algorithm consisting of multiple decision trees. The random forest is composed of several decision trees, and the Gini index is better than the information entropy in dividing conditions with judgmental attributes, so the Gini index is also used to construct the random forest to prevent model overfitting. The maximum depth of the tree is defined as 6 and the number is 20.

4.2. Dynamic Gesture Recognition Based on Hidden Markov Model

Dynamic gestures are action sequences composed of gesture transformations in space based on time sequences. Richer and more complex gesture information can be expressed. Different datasets express the gesture features in different ways, and the raw data contain all the gesture information, in addition to valid data, as well as noise and interference. The Eulerian angle dataset better describes the gesture changes in gesture motion. Quaternions can solve the shortcomings of Euler angles in gesture description. Each of the three data structures is shown below:
R aw D a t a t = a x i t , a y i t , a z i t , g x i t , g y i t , g z i t , m x i t , m y i t , m z i t , i 1 , 16
D y n a m i c G e s t u r e E t = y a w i t , r o l l i t , p i t c h i t , i 1 , 16 D y n a m i c G e s t u r e Q t = q 1 i t , q 2 i t , q 3 i t , q 4 i t , i 1 , 16
where DynamicGestureE represents Euler angular data and DynamicGestureQ represents quaternion data. Taking the one-handed dynamic gesture “Sorry” as an example, the preprocessed quaternion and Euler angle convergence node (node1) data are shown in Figure 9a,b, respectively.
Hidden Markov model (HMM) [31] is a statistical model created in the 1970s. Hidden Markov model is a kind of Markov chain (MC) [32], which is the simplest dynamic Bayesian network. HMM is effective in solving time-series-based data-recognition problems, and the dynamic-gesture recognition problem is a strong time-series-related recognition problem. In the gesture-recognition task, the feature data of each node are a visible observation sequence, and the meaning of the gesture transition is an unknown state sequence. HMM can generate gesture transition-state sequences based on the observed sequence of gestures and generate recognition models.
Hidden Markov models can be used in prediction problems: from a determined sequence of observations, a sequence of states is obtained computationally. For example, in the gesture recognition problem, the state is the meaning of the gesture and the observation is the sequence of gesture actions. The algorithm is generated as follows:
  • We generate the first gesture meaning given the initial state probability vector π ;
  • According to the previous gesture meaning, the next gesture meaning is randomly generated using the state transfer probability matrix A;
  • After generating the sequence of gesture meanings, the observation sequence of the corresponding position is generated using the observation probability matrix B according to each gesture meaning.
In the gesture-recognition problem, we can train the model by studying the recognition problem. The HMM training process is shown in Figure 10.
First, the training set of gesture data is used as input, and divided into gesture segments that need to be recognized. Each gesture segment corresponds to a gesture semantics, and the gesture semantics are marked and trained. Each individual semantic needs to train a corresponding hidden Markov model, train the model parameters with the highest matching degree through the specified hidden Markov model structure, and finally save all the models. This process can be solved using the Baum–Welch algorithm.
The best hidden Markov model corresponding to each gesture category is obtained by solving the training problem, and the classification problem of gesture recognition can be solved by solving the evaluation problem. The evaluation problem uses a new observation sequence O = o 1 , o 2 , , o n . The probability P O ; λ of producing this observation sequence O must be found. The process of dynamic gesture recognition is shown in Figure 11.
Then, we use the forward–backward algorithm to solve the unknown gesture-classification problem. The gesture category corresponding to the model with the highest matching degree is selected as the prediction result.

4.3. Design of Gesture-Recognition Algorithm Based on Deep Learning

Long short term memory neural network (LSTM) is a recurrent neural network model proposed by Hochreiter and Schmidhuber in 1997 [33], whose single-neuron structure is shown in Figure 12.
However, such recurrent neural network models can only perform derivation and learn information before that moment in a unidirectional way, and are unsuitable for more refined recognition, such as textual information and time sequences. We introduce a bi-directional long short term memory neural network (Bi-LSTM), which can learn the bi-directional information of sequences. The Bi-LSTM model adopts the traditional encoding–decoding method, and sequence samples are input to Bi-LSTM and edited into a fixed-length vector representation regardless of their length. Although Bi-LSTM can memorize the forward and backward data information, in practical applications, due to the differences in sequence lengths of different samples, it may appear that certain key factors are overlooked when training the model, resulting in model-recognition performance degradation. In response to its characteristics, an attention mechanism is introduced, which essentially imitates a person observing something important in a scene and focusing his or her attention on that part [34]. Introducing the attention mechanism can break the problem of fixed vector length in the encoding process of Bi-LSTM, and provide the corresponding weights according to the characteristics of the sequence to show the key information more clearly, which can improve the model training efficiency and help the model in accurate recognition. The structure diagram is shown in Figure 13, where y i is the hidden layer vector output by Bi-LSTM at each moment as key, and y n output at the last moment as query, which is calculated as follows:
S t = α ( y t , y n )
a t = exp ( S t ) t = 1 n exp ( S t )
c = t = 1 n a t y y
where S t is the similarity score between y t and y n at each moment calculated by the learning function α ; then, it is normalized by the s o f t m a x function to obtain the weight a t of y t at each moment, and finally the vector c is calculated by Equation (10).
Considering that the sign language data consist of time series of acceleration, angular velocity, and pose quaternions, and the dataset is small, Bi-LSTM can better learn features from the contextual information of sign language time series. Finally, we designed the Bi-LSTM, which introduces the attention mechanism (attention-BiLSTM), and its overall structure is shown in Figure 14. Firstly, the collected sign language acceleration, gyroscope and pose quaternion data are passed into the input layer, and initial feature learning is performed by the Bi-LSTM layer for sign language information; then, the data are input to the lower Bi-LSTM layer through the Dropout layer for the second stage of feature learning, after which the similarity is calculated by passing the dropout layer into the attention layer. Finally, the softmax layer normalizes these, further calculates the weight information of each moment, and sends the result to the fully connected layer. The output layer gives the sign language recognition result.

5. Results

We performed static and dynamic gesture-recognition tasks to verify the recognition performance of the machine learning and deep learning algorithms used, and provide detailed recognition results for the two types of gestures under different algorithms.

5.1. Static Gesture Recognition Using Machine Learning Algorithms

1.
Support Vector Machines
  • When building a support vector machine model, the kernel functions are first selected and the accuracy of all kernel functions is tested. The multiclassification SVM model built using the one vs. rest (OvR) strategy. The kernel function test results are shown in Table 1.
2.
BP Neural Network
  • Due to the low complexity of static gesture data and to prevent overfitting, a neural network structure with 2 hidden layers and 20 neurons per layer was built. The prediction accuracy of different activation functions was tested by the grid optimization algorithm, as shown in Table 2.
3.
Decision Tree algorithm
  • Generally, the most important part in deciding the superiority of decision tree classification is the judgment algorithm of attribute division. Information entropy and the Gini index are used for model training, and the accuracy is shown in Table 3.
4.
Random Forest algorithm
  • We define the maximum depth of the tree as 6 and the number as 20. The random forest and decision tree are cross-validated 10 times, and the validation effect is shown in Figure 15.
After the model is built and trained, to evaluate the fit of the model, data outside the training set needs to be put into the model for evaluation, i.e., the test set. In order to better test the model, the idea of cross-validation is used to evaluate the model. We use 5-fold cross-validation, that is, the training set is divided into 5 equal parts, 5 unrelated subsets are obtained, and each subset is taken out in turn as the test set. The remaining 4 subsets are used as the training set, and cross-validation is repeated 5 times to obtain 5 test results. The final prediction results can be obtained by averaging the results, and the cross-validation results of each combination are shown in Figure 16. The final results of each model prediction are shown in Table 4.
From the table, it can be seen that the Random Forest algorithm has the highest accuracy rate and the shortest training time for static gestures, while the SVM algorithm has the lowest accuracy rate and the longest training time. This finding is also consistent with the cross-validation results.

5.2. Dynamic Gesture Recognition Based on Hidden Malcove Model

Although the hidden Markov model can be parameterized by the Baum–Welch algorithm, the priority needs to be determining the number of states for each hidden Markov model. Different dynamic gestures have different complexities, and choosing a suitable number of states according to a specific complexity can not only minimize the training time, but also improve the degree of model fitting and prevent the overfitting phenomenon that occurs with too many nodes. We build the model with different number of nodes, train it with the training set, and then test the model output with random samples from the test set. The results are shown in Figure 17, and the model fits and differentiates well when the number of nodes is 13.
We trained and evaluated the models on the original dataset and quaternion dataset, respectively. The experimental results are shown in Table 5. It can be observed that the prediction model based on the original six-axis data achieves the highest recognition accuracy. However, as the raw data contain all features of dynamic gesture data, the training and testing times of the model are long. Therefore, it is necessary to extract features from the original data, remove invalid features, or reduce the dimensionality of the data. The confusion matrix obtained for different datasets is shown in Figure 18.

5.3. Dynamic Gesture Recognition Based on Deep Learning Methods

As the deep learning model may cause underfitting or overfitting problems due to the small training dataset or the existence of sample imbalances, we can configure the following parameters for this experiment. In the attention-BiLSTM model, we set the learning rate to 0.001, the training sample batch size to 256, and the number of iterations to 50. The training process, testing process, and loss function of the final model are shown in Figure 19. The training process of the model and the testing process of the loss function finally converge. We used 6-axis data and quaternion data for the recognition of 10 dynamic gestures, respectively, with 98.3% and 94.6% recognition accuracy when fed into a pre-prepared dataset that was not involved in training and validation. The confusion matrix of the model is shown in Figure 20. In comparison with the hidden Markov model, the bidirectional long- and short-term memory neural-network model based on the attention mechanism achieves better recognition results. It is worth noting that we also conducted recognition tasks for 10 dynamic gestures using the LSTM model, and the results are presented in Table 6. The addition of attention led to a significant improvement in gesture-recognition accuracy, which can be attributed to our effective utilization of both pre- and post-moment information in the gesture sequence.

6. Discussion and Conclusions

Gesture-recognition technology can be applied to many scenarios, such as virtual reality, robot control, and remote operation. The main sensors currently used regarding gesture recognition are IMU, video-based optical capture, and surface electromyography sensors. The main problems are the inconvenience of wearing and the vulnerability to environmental interference. Some studies focus on the structural design of the data glove and ignore the influence of the recognition algorithm on the recognition accuracy.
In this paper, an inertial sensor-based gesture data acquisition system is used, with the goal of constructing a gesture-recognition model based on the collected static and dynamic gesture datasets. Traditional machine learning algorithms can perform gesture recognition classificatio. We evaluated the prediction effectiveness of four algorithms. For static gestures, the model prediction performance was evaluated by cross-validation comparison, and we obtained the conclusion that the random forest algorithm has the highest recognition accuracy and the shortest recognition time. For dynamic gestures, we used HMM and a deep-learning-based attention-BiLSTM model and, according to the results, the latter achieved a higher recognition accuracy. The model can integrate the time series information of sign language acceleration, angular velocity, and hand posture to predict the sign language category, introduce a dropout layer to avoid model overfitting, and use the Adam optimization algorithm to accelerate the model convergence speed.
However, this does not indicate that deep learning methods are superior to traditional machine learning algorithms in the field of gesture recognition. This is because we can obtain a more comprehensive understanding of the data and the underlying algorithm of the model compared to the black-box structure of the deep model. Finally, in the field of practical engineering, traditional machine learning methods often require much less computational cost than deep learning methods. In the direction of gesture recognition based on wearable devices, it is often necessary to consider portability, power consumption, cost, comfort, etc. In the case of the compatible consideration of these factors, it is difficult to add the computational units required for deep learning, so it is difficult to determine the performance of deep learning models. In contrast, traditional machine learning models are fast to train, simple to deploy, and the required engineering costs are concentrated on data processing and feature optimization in the pre-model, thus allowing for faster update iterations in hardware products and the ability to try different model approaches in a short period of time. These aspects are not attainable by deep learning at this stage.
In addition, participants believed that prolonged wear would also cause hand discomfort. In this case, there is an urgent need for more comfortable gesture-recognition monitoring solutions or the use of fewer miniature inertial sensor nodes with guaranteed recognition performance. There were no strict criteria for the gesture data collected in this study, and the participants had no experience with gesture learning. These factors are worth considering in the future. In the future, we will consider designing more lightweight, miniature wearable device modules that can be integrated into existing electronics, such as watches and rings, to create a more comprehensive gesture-capture interaction system.

Author Contributions

H.W. and B.R. drafted the manuscript and performed the experiment and analysis; X.M. were responsible for the design of the hardware platform, data analysis and interpretation of the results; Q.G. and M.H. proposed the idea and were responsible for equipment configuration, L.L. and S.Q. performed data visualization and manuscript revision. All authors have read and agreed to the published version of the manuscript.

Funding

This work was jointly supported by the National Natural Science Foundation of China under Grant No. 62272081, 61803072 and U21A20477, and Natural Science Foundation of Liaoning Province, China under Grant 2021-MS-111, and in part by the Fundamental Research Funds for the Central Universities, China under Grant No. DUT22YG128, and Special Fund for High-level Researchers supported by Taizhou University.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

We have placed some of the gesture data used in the repository below. https://github.com/DUTRB/gestureData.

Acknowledgments

The authors would like to express their thanks to all funding bodies.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHRSAttitude and heading reference system
BCSBody coordinate system
BPBack-propagation neural network
Bi-LSTMBidirectional long- and short-term memory neural network
CNNConvolutional neural network
DTDecision tree
HMMHidden Markov model
HOGHitogram of gradient orientation
HD-SEMGHigh-density surface electromyography
IMUInertial measurement unit
LSTMLong- and short-term memory neural network
MCUMicrocontroller unit
MEMSMicro-electro-mechanical System
MCMarkov chain
NCSNavigation coordinate system
NNNeural networks
OvROne vs. rest
PCPersonal computer
RFRandom forest
RNNRecurrent neural networks
ROIRegion of interest
SPISerial peripheral interface bus
SCSSensor coordinate system unit
sEMGSurface electromyography
SVMSupport vector machine
ViTVision transformer

References

  1. Chen, Z.; Chen, J. Design and implementation of FPGA-based gesture recognition system. Wirel. Internet Technol. 2020, 17, 3. [Google Scholar]
  2. Del Rio Guerra, M.S.; Martin-Gutierrez, J. Evaluation of Full-Body Gestures Performed by Individuals with Down Syndrome: Proposal for Designing User Interfaces for All Based on Kinect Sensor. Sensors 2020, 20, 3930. [Google Scholar] [CrossRef] [PubMed]
  3. Siddiqui, U.A.; Ullah, F.; Iqbal, A.; Khan, A.; Ullah, R.; Paracha, S.; Shahzad, H.; Kwak, K.S. Wearable-Sensors-Based Platform for Gesture Recognition of Autism Spectrum Disorder Children Using Machine Learning Algorithms. Sensors 2021, 21, 3319. [Google Scholar] [CrossRef] [PubMed]
  4. Ye, S. Research on Hand Gesture Recognition Based on Multi-MEMS Inertial Sensors. Master’s Thesis, Nanjing University of Aeronautics and Astronautics, Nanjing, China, 2020. [Google Scholar]
  5. Lin, Q. The Research of Hand Detection and Tracking Using Kinect. Master’s Thesis, Nanjing University of Posts and Telecommunications, Nanjing, China, 2020. [Google Scholar]
  6. Qiu, S.; Zhao, H.; Jiang, N.; Wang, Z.; Liu, L.; An, Y.; Zhao, H.; Miao, X.; Liu, R.; Fortino, G. Multi-sensor information fusion based on machine learning for real applications in human activity recognition: State-of-the-art and research challenges. Inf. Fusion 2022, 80, 241–265. [Google Scholar] [CrossRef]
  7. Chang, W.D.; Matsuoka, A.; Kim, K.T.; Shin, J. Recognition of Uni-Stroke Characters with Hand Movements in 3D Space Using Convolutional Neural Networks. Sensors 2022, 22, 6113. [Google Scholar] [CrossRef]
  8. Ren, H.; Zhu, Y.; Xu, G.; Lin, X.; Zhang, X. Vision-Based Recognition of Hand Gestures:A Survey. Acta Electron. Sin. 2000, 28, 118–121. [Google Scholar]
  9. Wang, S. Kinect-based Gesture Recognition and Robot Control Technology Research. Master’s Thesis, Beijing Jiaotong University, Beijing, China, 2014. [Google Scholar]
  10. Mazhar, O.; Ramdani, S.; Navarro, B.; Passama, R.; Cherubini, A. Towards Real-Time Physical Human-Robot Interaction Using Skeleton Information and Hand Gestures. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–6. [Google Scholar] [CrossRef]
  11. Ren, C.; Shen, J.; Wang, L.; Cai, X. Gesture recognition based on Kinect skeleton data. Comput. Eng. Des. 2019, 40, 1440–1444,1450. [Google Scholar]
  12. Wang, J. Kinect-Based Gesture Recognition and Human-Machine Interaction. Master’s Thesis, Nanjing University of Posts and Telecommunications, Nanjing, China, 2018. [Google Scholar]
  13. Huang, J.; Jing, H. Gesture Control Research Based on Leap Motion. Comput. Syst. Appl. 2015, 24, 259–263. [Google Scholar]
  14. Li, X.; Wan, K.; Wen, R.; Hu, Y. Development of finger motion reconstruction system based on leap motion controller. In Proceedings of the 2018 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA), Ottawa, ON, Canada, 12–13 June 2018; pp. 1–5. [Google Scholar] [CrossRef]
  15. Alemayoh, T.T.; Shintani, M.; Lee, J.H.; Okamoto, S. Deep-Learning-Based Character Recognition from Handwriting Motion Data Captured Using IMU and Force Sensors. Sensors 2022, 22, 7840. [Google Scholar] [CrossRef]
  16. Lin, W.; Li, C.; Zhang, Y. Interactive Application of Data Glove Based on Emotion Recognition and Judgment System. Sensors 2022, 22, 6327. [Google Scholar] [CrossRef]
  17. Qiu, S.; Zhao, H.; Jiang, N.; Wu, D.; Song, G.; Zhao, H.; Wang, Z. Sensor network oriented human motion capture via wearable intelligent system. Int. J. Intell. Syst. 2022, 37, 1646–1673. [Google Scholar] [CrossRef]
  18. Liu, L. Multi-sensor Gesture Design and Recognition Method Based on Wearable Devices. Master’s Thesis, Xidian University, Xi’an, China, 2019. [Google Scholar]
  19. Fu, Q.; Fu, J.; Guo, J.; Guo, S.; Li, X. Gesture Recognition based on BP Neural Network and Data Glove. In Proceedings of the 2020 IEEE International Conference on Mechatronics and Automation (ICMA), Beijing, China, 13–16 October 2020; pp. 1918–1922. [Google Scholar]
  20. Gaka, J.; Msior, M.; Zaborski, M.; Barczewska, K. Inertial Motion Sensing Glove for Sign Language Gesture Acquisition and Recognition. IEEE Sens. J. 2016, 16, 6310–6316. [Google Scholar] [CrossRef]
  21. Qiu, S.; Hao, Z.; Wang, Z.; Liu, L.; Liu, J.; Zhao, H.; Fortino, G. Sensor Combination Selection Strategy for Kayak Cycle Phase Segmentation Based on Body Sensor Networks. IEEE Internet Things J. 2021, 9, 4190–4201. [Google Scholar] [CrossRef]
  22. Liu, L.; Liu, J.; Qiu, S.; Wang, Z.; Zhao, H.; Masood, H.; Wang, Y. Kinematics analysis of arms in synchronized canoeing with wearable inertial measurement unit. IEEE Sens. J. 2023, 23, 4983–4993. [Google Scholar] [CrossRef]
  23. Tai, T.M.; Jhang, Y.J.; Liao, Z.W.; Teng, K.C.; Hwang, W.J. Sensor-Based Continuous Hand Gesture Recognition by Long Short-Term Memory. IEEE Sens. Lett. 2018, 2, 6000704. [Google Scholar] [CrossRef]
  24. Shin, S.; Sung, W. Dynamic hand gesture recognition for wearable devices with low complexity recurrent neural networks. In Proceedings of the 2016 IEEE International Symposium on Circuits and Systems (ISCAS), Montreal, QC, Canada, 22–25 May 2016; pp. 2274–2277. [Google Scholar] [CrossRef]
  25. Yuan, G.; Liu, X.; Yan, Q.; Qiao, S.; Wang, Z.; Yuan, L. Hand Gesture Recognition Using Deep Feature Fusion Network Based on Wearable Sensors. IEEE Sens. J. 2021, 21, 539–547. [Google Scholar] [CrossRef]
  26. Qiu, S.; Fan, T.; Jiang, J.; Wang, Z.; Wang, Y.; Xu, J.; Sun, T.; Jiang, N. A novel two-level interactive action recognition model based on inertial data fusion. Inf. Sci. 2023, 633, 264–279. [Google Scholar] [CrossRef]
  27. Geiger, W.; Bartholomeyczik, J.; Breng, U.; Gutmann, W.; Hafen, M.; Handrich, E.; Huber, M.; Jackle, A.; Kempfer, U.; Kopmann, H.; et al. MEMS IMU for ahrs applications. In Proceedings of the 2008 IEEE/ION Position, Location and Navigation Symposium, Monterey, CA, USA, 5–8 May 2008; pp. 225–231. [Google Scholar]
  28. Namchol, C.; Zhao, H.; Qiu, S.; Yongguk, S. Study of calibration of low-cost MEMS magnetometer. J. Dalian Univ. Technol. 2018, 58, 105–110. [Google Scholar]
  29. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B 2011, 42, 513–529. [Google Scholar] [CrossRef]
  30. Jiao, L.; Yang, S.; Liu, F.; Wang, S.; Feng, Z. Seventy Years Beyond Neural Networks: Retrospect and Prospect. Chin. J. Comput. 2016, 39, 1697–1717. [Google Scholar]
  31. Yamato, J.; Ohya, J.; Ishii, K. Recognizing human action in time-sequential images using hidden Markov model. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR, Champaign, IL, USA, 15–18 June 1992; Volume 92, pp. 379–385. [Google Scholar]
  32. Liu, Y.; Li, J.; Jia, X. Reliability of k-out-of-n: G system in supply chain based on Markov chain. In Proceedings of the 2008 IEEE International Conference on Service Operations and Logistics, and Informatics, Beijing, China, 12–15 October 2008; Volume 1, pp. 1390–1393. [Google Scholar]
  33. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  34. Mnih, V.; Heess, N.; Graves, A.; Kavukcuoglu, K. Recurrent Models of Visual Attention. In NeurIPS Proceedings, Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; MIT Press: Cambridge, MA, USA, 2014; Volume 2, pp. 2204–2212. [Google Scholar]
Figure 1. Mainstream data glove types. (a) Optical gloves (b) Fiber optic gloves (c) Inertia gloves.
Figure 1. Mainstream data glove types. (a) Optical gloves (b) Fiber optic gloves (c) Inertia gloves.
Micromachines 14 00947 g001
Figure 2. System configuration (a) gesture data collection and processing flow. (b) Gesture data acquisition system based on inertial sensor.
Figure 2. System configuration (a) gesture data collection and processing flow. (b) Gesture data acquisition system based on inertial sensor.
Micromachines 14 00947 g002
Figure 3. Physical hardware (a) sink and inertial nodes (b) wear display effect.
Figure 3. Physical hardware (a) sink and inertial nodes (b) wear display effect.
Micromachines 14 00947 g003
Figure 4. Raw data collection. (a) subjects wear data gloves for gesture data-collection experiments; (b) raw data collected by inertial nodes.
Figure 4. Raw data collection. (a) subjects wear data gloves for gesture data-collection experiments; (b) raw data collected by inertial nodes.
Micromachines 14 00947 g004
Figure 5. Raw data filtering effect.
Figure 5. Raw data filtering effect.
Micromachines 14 00947 g005
Figure 6. The calibration outcome of magnetometer. (a) before fitting. (b) after fitting.
Figure 6. The calibration outcome of magnetometer. (a) before fitting. (b) after fitting.
Micromachines 14 00947 g006
Figure 7. Dynamic gesture data segmentation.
Figure 7. Dynamic gesture data segmentation.
Micromachines 14 00947 g007
Figure 8. Static gesture recognition flowchart.
Figure 8. Static gesture recognition flowchart.
Micromachines 14 00947 g008
Figure 9. Dynamic gesture “Sorry” data graph. (a) Quaternion data; (b) Euler angle data.
Figure 9. Dynamic gesture “Sorry” data graph. (a) Quaternion data; (b) Euler angle data.
Micromachines 14 00947 g009
Figure 10. HMMs training flow chart.
Figure 10. HMMs training flow chart.
Micromachines 14 00947 g010
Figure 11. HMMs identification flow chart.
Figure 11. HMMs identification flow chart.
Micromachines 14 00947 g011
Figure 12. LSTM structure diagram.
Figure 12. LSTM structure diagram.
Micromachines 14 00947 g012
Figure 13. Structure of attentional mechanisms.
Figure 13. Structure of attentional mechanisms.
Micromachines 14 00947 g013
Figure 14. Attention-BiLSTM model structure.
Figure 14. Attention-BiLSTM model structure.
Micromachines 14 00947 g014
Figure 15. Random Forest vs. Decision Tree Cross-Validation.
Figure 15. Random Forest vs. Decision Tree Cross-Validation.
Micromachines 14 00947 g015
Figure 16. Cross-validation results.
Figure 16. Cross-validation results.
Micromachines 14 00947 g016
Figure 17. The number of nodes and the degree of model fit.
Figure 17. The number of nodes and the degree of model fit.
Micromachines 14 00947 g017
Figure 18. Confusion matrix results based on HMM. (a) Confusion matrix for gesture recognition based on six-axis data. (b) Effect of gesture recognition based on quaternions.
Figure 18. Confusion matrix results based on HMM. (a) Confusion matrix for gesture recognition based on six-axis data. (b) Effect of gesture recognition based on quaternions.
Micromachines 14 00947 g018
Figure 19. Model training process.
Figure 19. Model training process.
Micromachines 14 00947 g019
Figure 20. Confusion matrix results based on Attention-BiLSTM. (a) 6-axis data. (b) Quaternion data.
Figure 20. Confusion matrix results based on Attention-BiLSTM. (a) 6-axis data. (b) Quaternion data.
Micromachines 14 00947 g020
Table 1. Kernel function accuracy.
Table 1. Kernel function accuracy.
NameExperssionPrediction Accuracy
Linear Kernel K x , z = x T z 98.12%
polynomial kernel K x , z = ( x T z ) d 96.96%
Gaussian kernel K x , z = exp x z 2 2 σ 2 98.57%
Sigmoid Kernel K x , z = tanh β x T z + θ 28.93%
Table 2. Common activation functions and prediction accuracy.
Table 2. Common activation functions and prediction accuracy.
Activation Function NameExpressionPrediction Accuracy
Identity f x = x 98.83%
Logistic f x = 1 1 + e x 84.67%
Tanh f x = tanh x 92.58%
Relu f x = max 0 , x 95.92%
Table 3. Common discriminant algorithms and prediction results.
Table 3. Common discriminant algorithms and prediction results.
Discriminant AlgorithmPrediction Accuracy
Entropy73.08%
Gini index88.42%
Table 4. Model accuracy rate and training time.
Table 4. Model accuracy rate and training time.
Model NameAccuracy RateTraining Time (s)
SVM92.4%1.616
BP Neural Network96.7%0.249
Decision Trees95.9%0.011
Random Forest98.3%0.064
Table 5. Recognition accuracy and training time based on HMM.
Table 5. Recognition accuracy and training time based on HMM.
Data CharacteristicsPrediction AccuracyTraining Time(s)
Six-axis data95.6%518
Quaternions82.1%239
Table 6. Comparison of gesture recognition results based on attention-BiLSTM and LSTM models.
Table 6. Comparison of gesture recognition results based on attention-BiLSTM and LSTM models.
Data CharacteristicsAttention-BiLSTMLSTM
Six-axis data98.3%83.8%
Quaternions94.6%87.8%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Ru, B.; Miao, X.; Gao, Q.; Habib, M.; Liu, L.; Qiu, S. MEMS Devices-Based Hand Gesture Recognition via Wearable Computing. Micromachines 2023, 14, 947. https://doi.org/10.3390/mi14050947

AMA Style

Wang H, Ru B, Miao X, Gao Q, Habib M, Liu L, Qiu S. MEMS Devices-Based Hand Gesture Recognition via Wearable Computing. Micromachines. 2023; 14(5):947. https://doi.org/10.3390/mi14050947

Chicago/Turabian Style

Wang, Huihui, Bo Ru, Xin Miao, Qin Gao, Masood Habib, Long Liu, and Sen Qiu. 2023. "MEMS Devices-Based Hand Gesture Recognition via Wearable Computing" Micromachines 14, no. 5: 947. https://doi.org/10.3390/mi14050947

APA Style

Wang, H., Ru, B., Miao, X., Gao, Q., Habib, M., Liu, L., & Qiu, S. (2023). MEMS Devices-Based Hand Gesture Recognition via Wearable Computing. Micromachines, 14(5), 947. https://doi.org/10.3390/mi14050947

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop