Next Article in Journal
Performance Evaluation of UOWC Systems from an Empirical Channel Model Approach for Air Bubble-Induced Scattering
Next Article in Special Issue
Validity and Reliability of Wearable Sensors for Continuous Postoperative Vital Signs Monitoring in Patients Recovering from Trauma Surgery
Previous Article in Journal
Development and Validation of a New Type of Displacement-Based Miniatured Laser Vibrometers
Previous Article in Special Issue
The Overlay, a New Solution for Volume Variations in the Residual Limb for Individuals with a Transtibial Amputation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Hand Gesture Recognition with Surface Electromyogram and Machine Learning

by
Mujeeb Rahman Kanhira Kadavath
*,
Mohamed Nasor
and
Ahmed Imran
College of Engineering and Information Technology, Ajman University, Ajman P.O. Box 346, United Arab Emirates
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5231; https://doi.org/10.3390/s24165231
Submission received: 18 June 2024 / Revised: 30 July 2024 / Accepted: 8 August 2024 / Published: 13 August 2024
(This article belongs to the Special Issue Advanced Wearable Sensors for Medical Applications)

Abstract

:
This study delves into decoding hand gestures using surface electromyography (EMG) signals collected via a precision Myo-armband sensor, leveraging machine learning algorithms. The research entails rigorous data preprocessing to extract features and labels from raw EMG data. Following partitioning into training and testing sets, four traditional machine learning models are scrutinized for their efficacy in classifying finger movements across seven distinct gestures. The analysis includes meticulous parameter optimization and five-fold cross-validation to evaluate model performance. Among the models assessed, the Random Forest emerges as the top performer, consistently delivering superior precision, recall, and F1-score values across gesture classes, with ROC-AUC scores surpassing 99%. These findings underscore the Random Forest model as the optimal classifier for our EMG dataset, promising significant advancements in healthcare rehabilitation engineering and enhancing human–computer interaction technologies.

1. Introduction

Electromyography (EMG) is a fascinating field of study that explores the electrical activity of muscles and nerves in the body. Using EMG, one can record the electrical impulses generated by muscle fibers, which reveal important details about the condition and operation of the muscular system. It can offer a dependable method of detecting and classifying muscle activity associated with various hand movements [1,2]. EMG is a diverse field with origins in engineering and neuroscience, presenting a variety of potential for research and innovation. For example, EMG can be used to control prosthetic limbs. By detecting the electrical signals generated by residual muscle activity in an amputee’s stump, EMG can be used to manage a prosthetic limb’s motion [3]. This technique, known as myoelectric control, has shown promising results in clinical trials and is becoming increasingly popular in the field of prosthetics [4]. One widely used method for acquiring EMG signals involves placing EMG electrodes around specific muscle sites and using an EMG amplifier module for signal acquisition. This method is non-invasive, ensuring no harm to the body [5,6]. Figure 1 illustrates a single-channel EMG recording setup. It entails attaching two main electrodes (also called active electrodes) and a common electrode (also known as a reference electrode) to the subject’s biceps muscle. Figure 2 displays the matching single-channel EMG signal, where each burst represents a voluntary muscle contraction.
The dexterity of our finger movements relies on the intricate coordination of various muscles. The flexor digitorum profundus and flexor digitorum superficialis muscles facilitate the flexing of our fingers, while the extensor digitorum communis and extensor indicis muscles are responsible for straightening them [7,8]. This coordination is illustrated in Figure 3.
Through the analysis of EMG data, which captures the electrical impulses sent by these muscles, we can unravel the complex choreography that forms the basis of every movement of the fingers. Pinpointing a single muscle maestro for each finger movement remains a challenge with EMG. However, the electrical signals captured by EMG electrodes, strategically placed around the forearm, reveal characteristic shifts that can tell us which fingers are playing a role in the motion [9]. Beyond deciphering finger movements, EMG unlocks a world of possibilities. For those with prosthetics, EMG from remaining muscles translates into natural control of grip strength and finger movements [10]. Assistive technology benefits too, with EMG empowering individuals to operate wheelchairs, robotic arms, and communication tools [11]. Human–computer interaction receives a makeover with EMG-controlled cursors, games, and even gesture-based commands [12]. Rehabilitation sees EMG used to monitor muscle activity during exercises, optimizing treatment plans [12,13]. Virtual reality receives a touch upgrade, as EMG allows for more realistic manipulation of virtual objects [14]. The potential extends further, with EMG facilitating gesture recognition for smart home control, menu navigation, and even playing musical instruments through specific hand movements [15].
Recognizing gestures from arm EMG signals necessitates advanced signal processing and analysis techniques, such as pattern recognition and feature extraction. Machine learning algorithms are particularly adept at classifying signals or images based on the extracted features. By harnessing the robust capabilities of ML, we can train algorithms to identify patterns within EMG data. This training enables the algorithms to accurately predict the class of unknown EMG signals, thereby effectively recognizing gestures [14].
We adopted a wearable EMG sensor module called the Myo armband to record the electrical activity of the forearm muscles. The Myo armband operates by detecting and measuring the electrical activity generated by muscle contractions. As shown in Figure 4, it consists of eight electrodes strategically placed around the circumference of the forearm to capture the electrical signals produced by the muscles. When a muscle contracts, it generates an electrical signal known as an electromyogram (EMG). Figure 5 demonstrates the mounting of the Myo armband sensor on the subject’s arm for experimentation. The electrodes detect these signals by measuring the voltage differences on the skin surface caused by the underlying muscle activity.
The raw EMG signals are typically very weak, so they need to be amplified. The Myo armband contains built-in amplifiers to boost the signal strength, making it suitable for further processing. The amplified EMG signals are then filtered to remove noise and other artifacts, ensuring that the signals are clean and accurately represent muscle activity. These filtered analog signals are converted into digital signals using an analog-to-digital converter (ADC), necessary for processing by digital systems and algorithms.
Key quantities related to muscle activity include amplitude, which reflects the level of muscle activation; frequency, which provides information about the type of muscle fibers being activated and the nature of the muscle contraction; duration, which corresponds to the length of time the muscle is active during a contraction; and timing, indicating when a muscle starts and stops contracting. Additionally, analyzing changes in frequency and amplitude over time can assess muscle fatigue. These quantities are crucial for understanding muscle behavior.
The Myo armband provides eight channels of raw EMG data, which are kept separate and never mixed. We divided the raw EMG signals into 6822 segments, each 150 samples long. From these segments, ten features were extracted and combined to create the data required for building the model. The Myo armband connects wirelessly to devices using Bluetooth Low Energy (BLE) to transmit acquired data, ensuring efficient use of its battery life.
The user-friendly design of the Myo armband, lauded for its simplicity, lightness, and comfort, has made it a popular choice for researchers exploring sign language recognition in various applications like gaming and virtual reality [15]. One such study by Abreu et al. focused on Brazilian Sign Language (LIBRAS) recognition using a Myo armband and an SVM classifier [16]. Their dataset encompassed 20 static postures representing letters from the LIBRAS alphabet. However, the model achieved higher accuracy with simpler letter shapes (L: 77%, R: 91%, S: 46%, W: 95%) compared to more complex ones (F: 8%, M: 8%, P: 8%, T: 4%, U: 5%). This highlights the challenge of differentiating intricate hand formations based on EMG signals. Additionally, letters like A and S, or G and Q, despite being visually distinct, can cause misclassification due to similar signal patterns or applying pressure to the same finger. The study underscores the importance of precise gesture replication during training, emphasizing location and exerted force for the model to learn accurate representations of each sign and achieve better real-time recognition.
Mani et al. conducted a study to assess the feasibility of using the wearable Myo armband device for exploring patterns and relationships in finger movements during musical instrument playing using unsupervised learning [17]. They employed the Myo armband to aid in learning the electric keyboard, tracking finger movements and detecting muscle signals to identify played musical themes based on finger presses. Electromyography (EMG) signals from all eight channels of the Myo armband were recorded. Eleven features were extracted and processed using Linear Discriminant Analysis (LDA) for dimension reduction. Subsequently, the data were input into a Random Forest (RF) classifier. The study included EMG data collection in both single-user and multi-user scenarios, covering six distinct targets. Notably, electrodes placed on the anterior side of the device proved highly sensitive and beneficial for classifying finger movements. Findings showed that for multiple-finger chord playing on the keyboard, using two electrodes of the EMG signal achieved 95.83% accuracy. Additionally, for classifying individual musical notes, the Myo armband with its four electrodes achieved close to 75% accuracy.
Anil et al. conducted a research study utilizing machine learning with EMG signals captured by the Myo armband sensor module [18]. They applied discrete wavelet transform (DWT) to process these dynamic EMG signals, focusing on recognizing five specific hand gestures: Rest, Fist, Wavein, Waveout, and Fingerspread. Their approach achieved an impressive accuracy of 83% using a dataset of 335 samples, with 80% used for training and the remaining 20% for testing. Opting for SVM due to its effectiveness with smaller datasets, the authors employed SVM’s hyperplane to classify and separate data points. The study underscored the efficacy of DWT in effectively processing dynamic EMG signals. Overall, their research demonstrated robust accuracy in identifying various gestures across the dataset.
In another study focused on classifying EMG signals captured from the Myo armband, researchers aimed to discern specific finger activities like thumb, index, ring, and little finger movements, as well as resting states. Unlike traditional methods, Srinivasan et al., 2018 [19], explored deep learning techniques, specifically convolutional neural networks (CNNs). They implemented a model incorporating Inception Net v1 using surface electrodes. Training with 80% of the data and testing with the remaining 20%, their deep learning approach achieved an accuracy of 72.5% for classifying five finger movements, marking a significant 16.5% improvement over the commonly used SVM method. Subsequent advancements included expanding their dataset to further enhance classification capabilities.
The literature highlights challenges in using the Myo armband for gesture recognition, including dataset size, feature selection, and machine learning algorithms. Future studies should focus on increasing dataset sizes, incorporating dynamic gestures, exploring advanced machine learning algorithms, and ensuring thorough user calibration. The primary goal of this project was to identify the best-performing ML classifier for seven distinct hand gestures using EMG signals recorded by the Myo armband.

2. Materials and Methods

This study involved utilizing the Myo armband to capture EMG data associated with seven distinct hand gestures. The Myo armband was comfortably worn on the participant’s arm to capture the EMG signals, as illustrated in Figure 6. The selected hand gestures for analysis encompassed the index finger, middle finger, ring finger, pinky finger, thumb, resting gesture, and victory gesture, as visualized in Figure 7. EMG data were collected while participants performed these gestures in a predefined sequence, facilitated by a visual cue for each gesture. In this dataset, each finger movement requires nearly 1 s and is reflected as a sudden change in the EMG amplitude. Creating the dataset for this project took 18.7 min per subject, resulting in a total sample length of 223,793 at a sampling rate of 200 Hz. Finger motions typically appear near the center of each segment. We segmented the EMG data according to the hand gestures, with each segment initially consisting of 200 samples. However, the ends of each peak are considered static and do not play a significant role, so we retained 150 samples per segment. The resulting dataset contains 6822 segments, encompassing 7 classes of hand gestures. The entire trial lasted 18.7 min. For detailed information on each hand motion, including their frequencies and percentage distributions, refer to Table 1. The Myo armband collects eight-channel EMG data from the participant at a sample rate of 200 Hz.
A segment represents EMG data from one of the channels, consisting of 150 sample points, which corresponds to a specific finger motion. For each finger motion, we can have eight segments, each originating from one of the eight channels.
The unprocessed EMG raw data corresponding to the index finger, captured over a randomly chosen segment of 200 samples (for the interval 600 to 800), are shown in Figure 8a,b, which shows the raw EMG data recorded by electrode 1, corresponding to the seven hand gestures.
The architecture of the proposed methodology is presented in Figure 9. The input to the model is the raw EMG data, which then undergo a series of preprocessing steps. These steps include the application of band-pass filters (band width: 20 to 100 Hz) to remove noise and isolate relevant signal frequencies, normalization to standardize the data, rectification to convert all values to positive, and envelope extraction to capture the amplitude variations of the rectified signal. Additionally, the EMG data are segmented into seven classes based on the labels provided in the dataset.
Following these preprocessing and segmentation steps, the data are prepared for feature extraction and subsequent analysis within the model. In the subsequent stage, the algorithm extracts a diverse set of features from each segment of the EMG data. Our experiments covered statistical, time-domain, and frequency-domain features, leading us to identify the most effective ones, as described in the following section [20,21,22].
(a)
Minimum (min): The smallest value in the signal over a given window.
For a given EMG signal x(t) over a specific timeframe [t1,t2], the minimum value of the signal over the given window [t1,t2] is defined as:
M i n x = m i n t t 1 , t 2 x t          
(b)
Maximum (max): The largest value in the signal over a given window.
These two features together define the overall range of muscle activation within a specific timeframe.
M a x x = m a x t t 1 , t 2 x t              
(c)
Standard Deviation (SD): Measures the amount of variation or dispersion of a set of values. Captures how spread out the EMG signal is from its mean value. A high SD indicates high variability, suggesting a mix of muscle activation levels.
S D x = 1 N 1 i = 1 N x ( t i μ ) 2    
where N is the number of samples in the timeframe and μ is the mean of the signal over the window:
μ x = 1 N i = 1 N x ( t i )  
(d)
Zero Crossings (ZC): Counts the number of times the signal crosses zero, reflecting the frequency of muscle fiber recruitment and relaxation cycles. To accurately capture all zero crossings, we analyze the EMG data before the rectification stage
Z C x = i = 1 N 1   [ ( x ( t i ) .   x t i + 1 < 0 ) ]
where 1 is the indicator function, which equals 1 when the condition inside is true and 0 otherwise.
(e)
Root Mean Square (RMS): The square root of the mean of the squares of the values. It represents the overall power of the signal, reflecting the average muscle activity level.
R M S x = 1 N i = 1 N x ( t i ) 2  
(f)
Amplitude Change (AAC): Measures the change in amplitude between consecutive samples. It is useful for identifying rapid changes in muscle activity.
A A C x = 1 N 1 i = 1 N x t i + 1 x t i        
(g)
Amplitude of First Burst (AFB): Identifies the peak value within a specific time window after the signal crosses a threshold, indicating the initial burst of muscle activation.
The first crossing time ( t c ) of the threshold T is given by:
t c = min t ϵ t 1 , t 2     x t   T    
Amplitude of max burst within the window [ t c ,   t c + t b ] can be estimated by using
A F B x = m a x t t c ,     t c + t b     { x t }
(h)
Mean Absolute Value (MAV): The average of the absolute values of the signal over a specified time window. Similar to RMS, it reflects the average magnitude of the signal, regardless of positive or negative values.
M A V x = 1 N i = 1 N x ( t i )      
(i)
Length of Waveform (Len): Duration or length of the signal.
(j)
Willison Amplitude (WAMP): Counts the number of times the EMG signal crosses a specific threshold within a particular time window. It provides a rough estimate of the dominant frequency of the EMG signal, potentially reflecting muscle fiber recruitment patterns.
i = 1 N 1   [ ( x t i + 1 x t i T ) ]
where 1 is the indicator function, which equals 1 when the condition inside is true and 0 otherwise.
In the next step, we employed feature engineering to enhance the performance of our machine learning models by selecting, transforming, and creating features. This process involved various techniques to make raw data more suitable for predictive modeling, thereby improving model accuracy and efficiency. Key techniques included identifying and selecting the most relevant features and transforming features through normalization (scaling features to a specific range) and standardization (scaling features to have zero mean and unit variance). Additionally, we applied encoding methods such as one-hot encoding, label encoding, and target encoding to convert categorical variables into numerical formats.
Initially, we built the model using all ten features from each of the eight electrodes (80 features in total). We applied z-score normalization to these features and used one-hot encoding for the seven target columns. To assess feature redundancy, we calculated the correlation coefficients between the features and found that 30 features exhibited correlations of 90% or higher. To streamline the feature set, we used Principal Component Analysis (PCA), which showed that the first 49 principal components retained 99.9% of the information, as illustrated in the cumulative PCA plot in Figure 10. Consequently, we trained the model using these 49 components, achieving nearly the same (slightly less) performance scores as in the previous stage. This comprehensive feature engineering process optimized our dataset for the subsequent machine learning tasks.
We included four widely recognized machine learning algorithms to construct our proposed hand gesture classifier: Logistic Regression (LR), Support Vector Machine (SVM), K-Nearest Neighbors (kNN), and Random Forest (RF).
LR is a linear model that predicts the probability of a binary outcome using a logistic function. It is known for its simplicity, interpretability, and efficiency in linearly separable datasets [23].
SVM is a powerful classification technique that finds an optimal hyperplane to separate data into different classes. It is effective in high-dimensional spaces and can handle non-linear relationships using kernel functions. SVM with a linear kernel finds an optimal hyperplane to separate data into different classes in a linearly separable feature space. It is efficient and works well when the classes are linearly separable. SVM with a Radial Basis function (RBF kernel) can handle non-linear relationships by mapping input data into a higher-dimensional space. It computes similarity to nearby points and is effective in capturing complex decision boundaries [24,25].
kNN is a non-parametric method that classifies new instances based on their similarity to training examples in the feature space. It is intuitive and robust but can be computationally expensive with large datasets [26].
RF is an ensemble learning method that constructs multiple decision trees during training and outputs the mode of the classes as the prediction (classification) or the average prediction (regression) of the individual trees. It is known for its robustness against overfitting and ability to handle complex interactions in data [27].
The metrics used to assess the performance of the candidate algorithms were precision, recall, F1-score, accuracy, and area under the ROC curve (AUC).
Precision is a metric that measures the accuracy of the positive predictions made by a machine learning model. It is defined as the ratio of true positive predictions to the total number of positive predictions. Precision is particularly important in situations where the cost of false positives is high [28].
P r e c i s i o n = T r u e   P o s i t i v e s   T o t a l   n u m b e r   o f   p o s i t i v e s
Recall, also known as sensitivity or true positive rate, measures the ability of a model to identify all relevant instances. It is the ratio of true positive predictions to the total number of actual positive instances. Recall is crucial when the cost of false negatives is high, such as in medical diagnoses [28].
R e c a l l = T r u e   P o s i t i v e s   T r u e   P o s i t i v e s + F a l s e   n e g a t i v e s
The F1-score is the harmonic mean of precision and recall, providing a single metric that balances both concerns. It is useful when you need a balance between precision and recall and when dealing with imbalanced datasets. The F1-score ranges from 0 to 1, with 1 being the best possible score [28].
F 1 = 2 × P r e c i s i o n × R e c a l l   T P r e c i s i o n + R e c a l l
Accuracy is the ratio of correctly predicted instances (both true positives and true negatives) to the total number of instances. While accuracy is a straightforward and widely used metric, it can be misleading in the case of imbalanced datasets where one class significantly outnumbers the other [28,29].
A c c u r a c y = T r u e   p o s i t i v e s + T r u e   N e g a t i v e s     T o t a l   t e s t   s a m p l e s  
The AUC is a performance measurement for classification problems at various threshold settings. The ROC curve is a plot of the true positive rate (TPR) against the false positive rate (FPR). The AUC represents the degree or measure of separability; it tells how much the model is capable of distinguishing between classes. An AUC of 1 indicates a perfect model, while an AUC of 0.5 suggests no discriminative power [30,31].
We did not use a dedicated test dataset to evaluate our model. Instead, we employed five-fold cross-validation, a widely recognized technique in EMG signal processing and machine learning analysis, to test our model. In this approach, the dataset is partitioned into five equally sized folds. During each iteration, one fold is used as the test set while the remaining four folds are combined to form the training set. This process is repeated five times, with each fold serving as the test set exactly once. By averaging the performance metrics across these five iterations, we obtain a robust measure of the model’s generalizability and performance. This technique ensures that every data point is used for both training and testing, providing a comprehensive and reliable evaluation of the model [32,33].

3. Results

This section presents the cross-validated performance scores of the four different models. The evaluation metrics include precision, recall, F1-score, accuracy, and AUC for the seven-class hand gesture test data. The results provide a comprehensive overview of each model’s ability to accurately classify the gestures, considering both overall performance and per-class performance. Table 2 shows the percentage scores of the five metrics for the linear regression model, while Table 3 and Table 4 present the scores for the SVM with linear and RBF kernels, respectively.
Table 5 and Table 6 display the scores for the Random Forest and kNN classifiers, respectively. Table 7 presents the overall scores for the four models, where SVM-1 and SVM-2 refer to the two versions of the SVM models using linear and RBF kernel functions, respectively.

4. Discussion

The primary objective of this research was to identify the most effective multi-class machine learning classifier for recognizing hand gestures using surface EMG signals obtained from the Myo armband module. The Myo armband is a user-friendly wearable device that captures EMG signals from eight specific locations on the subject’s arm. We designed an EMG data acquisition setup and selected seven distinct hand gestures for this study, collecting a total of 6822 hand gestures across these classes. Although we aimed to balance the dataset, classes 6 and 7 remained slightly unbalanced compared to the other classes. Through experimentation, we identified ten key features from the EMG segments to construct a feature vector for training four distinct ML models. We employed five-fold cross-validation to validate our model scores and calculated precision, recall, F1-score, accuracy, and AUC for each class, as well as the overall scores for each model.
The results demonstrated that the Random Forest classifier achieved over 91% across all metrics for all classes, consistently exceeding 95% in overall scores. Specifically, the precision, recall, F1-score, accuracy, and AUC scores for the Random Forest model were 95.58%, 96.60%, 95.57%, 95.38%, and 99.72%, respectively. This highlights the model’s robustness and reliability in gesture recognition tasks. Interestingly, the SVM model with an RBF kernel outperformed its linear kernel counterpart, underscoring the importance of kernel choice in SVM performance. The kNN model performed slightly below the Random Forest model. Conversely, the logistic regression model exhibited the lowest effectiveness among the classifiers evaluated.
The EMG data utilized in this project were obtained from a single subject, suggesting that even higher accuracy might be achievable with a larger and more diverse dataset. Addressing the unbalanced class distribution could further enhance model performance. However, inter-subject variability in EMG characteristics poses a challenge for generalizing across different individuals due to several factors. Physiological differences, such as variations in muscle structure, skin thickness, and the placement of muscles and nerves, significantly affect the EMG signals collected from different individuals. Signal variability is another issue, as EMG signals are inherently noisy and can be influenced by electrode placement, skin impedance, and movement artifacts, which can differ from person to person. Additionally, each individual may adapt and perform gestures slightly differently, leading to differences in EMG signal patterns even for the same gesture. Muscle fatigue also plays a role, as it can alter EMG signal characteristics, with the rate and manner of fatigue varying among individuals. These factors make it challenging to create a model that generalizes well across different subjects without significant customization or adaptation. Therefore, while a model trained on data from one individual may perform well for that individual, its performance may degrade when applied to a different person without additional training or calibration. This is why inter-subject variability is a key consideration in EMG-based gesture recognition studies.
As indicated by previous research, machine learning classifiers for bio-signals like EMG, EOG, and EEG are generally more suitable for intra-subject cases (within the same individual) rather than inter-subject cases (across different individuals) for several reasons. EMG signals vary significantly between different individuals due to factors such as muscle physiology, skin impedance, electrode placement, and unique movement patterns. This variability makes it challenging to develop a generalized model that performs well across multiple subjects. Moreover, EMG signals from the same individual tend to be more consistent over time, which allows for more reliable and stable model performance. Collecting sufficient and representative training data from multiple subjects to build a robust inter-subject model is often more difficult and resource-intensive compared to collecting data from a single subject. Intra-subject models can be fine-tuned to the specific signal characteristics of an individual, resulting in better performance and higher accuracy. This personalization helps in capturing the unique features of an individual’s EMG signals. Therefore, in this research, we employed intra-subject EMG classifiers to achieve higher accuracy and reliability.
A notable limitation of this study is the restricted number of hand gestures analyzed; only seven gestures were included. Future work will aim to incorporate a broader range of hand gestures, including those that are distinctly different from the current set.
Recognizing hand gestures has the potential to revolutionize various fields such as prosthetics for amputees, robotics, advanced human–computer interaction, rehabilitation, and assistive technologies for individuals with motor impairments. The advancements in this area could lead to significant improvements in the quality of life and functional independence for many individuals.

5. Conclusions

This study focused on identifying the optimal ML classifier for recognizing hand gestures using EMG signals from the Myo armband module. Among the models evaluated—Random Forest, SVM with linear and RBF kernels, logistic regression, and k-Nearest Neighbors—the Random Forest classifier demonstrated robust performance with mean scores consistently exceeding 95% across precision, recall, F1-score, accuracy, and AUC metrics. The SVM model with an RBF kernel outperformed its linear counterpart, highlighting kernel choice’s crucial role in classification accuracy. Conversely, logistic regression showed the least effectiveness. Challenges like inter-subject variability in EMG characteristics complicate model generalization across individuals, emphasizing the need for customized approaches. Future research should expand datasets, address class imbalance, and explore broader gesture sets to advance EMG-based gesture recognition for applications in prosthetics, robotics, and rehabilitation.

Author Contributions

Conceptualization, methodology, coding, and documentation, M.R.K.K.; writing—review and editing, validation, formal analysis, resources, and project administration, M.N. and A.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ajman University under grant number 2023-IRG-ENIT-32.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Raw data used in this study are available upon request to the corresponding author.

Acknowledgments

We would like to express our gratitude to Ajman University, Ajman, United Arab Emirates particularly the Department of Biomedical Engineering, for their facilities and resources that made this research possible.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Michell, A. Understanding EMG; OUP: Oxford, UK, 2013. [Google Scholar]
  2. Konrad, P. A practical introduction to kinesiological electromyography 1. In The ABC of EMG; Noraxon Inc.: Scottsdale, AZ, USA, 2005; pp. 30–35. [Google Scholar]
  3. Sudarsan, S.; IEEE Student Member; Sekaran, E.C. Design and development of EMG controlled prosthetics limb. Procedia Eng. 2012, 38, 3547–3551. [Google Scholar] [CrossRef]
  4. Kamavuako, E.N.; Scheme, E.J.; Englehart, K.B. Combined surface and intramuscular EMG for improved real-time myoelectric control performance. Biomed. Signal Process. Control 2014, 10, 102–107. [Google Scholar] [CrossRef]
  5. Ahmad, I.; Ansari, F.; Dey, U.K. A review of EMG recording technique. Int. J. Eng. Sci. Technol. 2012, 4, 530–539. [Google Scholar]
  6. De Luca, G. Fundamental Concepts in EMG Signal Acquisition; Copyright Delsys Inc.: Natick, MA, USA, 2003. [Google Scholar]
  7. Kenyon, K.; Kenyon, J. The Physiotherapist’s Pocketbook: Essential Facts at Your Fingertips; Elsevier Health Sciences: Amsterdam, The Netherlands, 2009. [Google Scholar]
  8. Jones, L.A.; Lederman, S.J. Human Hand Function; Oxford University Press: Oxford, UK, 2006. [Google Scholar]
  9. Zehr, E.P. Inventing Iron Man: The Possibility of a Human Machine; JHU Press: Baltimore, MD, USA, 2011. [Google Scholar]
  10. Hoozemans, M.J.; Van Dieen, J.H. Prediction of handgrip forces using surface EMG of forearm muscles. J. Electromyogr. Kinesiol. 2005, 15, 358–366. [Google Scholar] [CrossRef] [PubMed]
  11. Vogel, J.; Hagengruber, A.; Iskandar, M.; Quere, G.; Leipscher, U.; Bustamante, S.; Dietrich, A.; Höppner, H.; Leidner, D.; Albu-Schäffer, A. EDAN: An EMG-controlled daily assistant to help people with physical disabilities. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021. [Google Scholar]
  12. Dalal, A. Design of Wireless EMG and Its Use in the Classification of Finger Movements. Master’s Thesis, San Diego State University, San Diego, CA, USA, 2012. [Google Scholar]
  13. Ghassemi, M.; Triandafilou, K.; Barry, A.; Stoykov, M.E.; Roth, E.; Mussa-Ivaldi, F.A.; Kamper, D.G.; Ranganathan, R. Development of an EMG-controlled serious game for rehabilitation. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 283–292. [Google Scholar] [CrossRef] [PubMed]
  14. Blana, D.; Kyriacou, T.; Lambrecht, J.M.; Chadwick, E.K. Feasibility of using combined EMG and kinematic signals for prosthesis control: A simulation study using a virtual reality environment. J. Electromyogr. Kinesiol. 2016, 29, 21–27. [Google Scholar] [CrossRef] [PubMed]
  15. Qi, J.; Jiang, G.; Li, G.; Sun, Y.; Tao, B. Intelligent human-computer interaction based on surface EMG gesture recognition. IEEE Access 2019, 7, 61378–61387. [Google Scholar] [CrossRef]
  16. Rawat, S.; Vats, S.; Kumar, P. Evaluating and exploring the MYO ARMBAND. In Proceedings of the 2016 International Conference System Modeling & Advancement in Research Trends (SMART), Moradabad, India, 25–27 November 2016. [Google Scholar]
  17. Abreu, J.G.; Teixeira, J.M.; Figueiredo, L.S.; Teichrieb, V. Evaluating sign language recognition using the myo armband. In Proceedings of the 2016 XVIII Symposium on Virtual and Augmented Reality (SVR), Gramado, Brazil, 21–24 June 2016; pp. 64–70. [Google Scholar]
  18. Mani, S.; Rao, M. Feasibility Study of using MyoBand for Learning Electronic Keyboard. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 10196–10202. [Google Scholar]
  19. Anil, N.; Sreeletha, S.H. EMG based gesture recognition using machine learning. In Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 14–15 June 2018; pp. 1560–1564. [Google Scholar]
  20. Srinivasan, V.B.; Islam, M.; Zhang, W.; Ren, H. Finger movement classification from myoelectric signals using convolutional neural networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), Kuala Lumpur, Malaysia, 12–15 December 2018; pp. 1070–1075. [Google Scholar]
  21. Inam, S.; Al Harmain, S.; Shafique, S.; Afzal, M.; Rabail, A.; Amin, F.; Waqar, M. A brief review of strategies used for EMG signal classification. In Proceedings of the 2021 International Conference on Artificial Intelligence (ICAI), Islamabad, Pakistan, 5–7 April 2021. [Google Scholar]
  22. Yousefi, J.; Hamilton-Wright, A. Characterizing EMG data using machine-learning tools. Comput. Biol. Med. 2014, 51, 1–13. [Google Scholar] [CrossRef] [PubMed]
  23. Turlapaty, A.C.; Gokaraju, B. Feature analysis for classification of physical actions using surface EMG data. IEEE Sens. J. 2019, 19, 12196–12204. [Google Scholar] [CrossRef]
  24. Schein, A.I.; Ungar, L.H. Active learning for logistic regression: An evaluation. Mach. Learn. 2007, 68, 235–265. [Google Scholar] [CrossRef]
  25. Mujeeb Rahman, K.K.; Nasor, M.; Imran, A. Automatic screening of diabetic retinopathy using fundus images and machine learning algorithms. Diagnostics 2022, 12, 2262. [Google Scholar] [CrossRef] [PubMed]
  26. Ben-Hur, A.; Ong, C.S.; Sonnenburg, S.; Schölkopf, B.; Rätsch, G. Support vector machines and kernels for computational biology. PLoS Comput. Biol. 2008, 4, e1000173. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for knn classification. ACM Trans. Intell. Syst. Technol. 2017, 8, 1–19. [Google Scholar] [CrossRef]
  28. Petkovic, D.; Altman, R.; Wong, M.; Vigil, A. Improving the explainability of Random Forest classifier–user centered approach. Pac Symp. Biocomput. 2018, 23, 204–215. [Google Scholar] [PubMed]
  29. Zhou, J.; Gandomi, A.H.; Chen, F.; Holzinger, A. Evaluating the quality of machine learning explanations: A survey on methods and metrics. Electronics 2021, 10, 593. [Google Scholar] [CrossRef]
  30. Vujović, Ž. Classification model evaluation metrics. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 599–606. [Google Scholar] [CrossRef]
  31. Rainio, O.; Teuho, J.; Klén, R. Evaluation metrics and statistical tests for machine learning. Sci. Rep. 2024, 14, 6086. [Google Scholar] [CrossRef] [PubMed]
  32. Anguita, D.; Ghelardoni, L.; Ghio, A.; Oneto, L.; Ridella, S. The’K’in K-fold Cross Validation. In Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 25–27 April 2012; Volume 102. [Google Scholar]
  33. Jung, Y. Multiple predicting K-fold cross-validation for model selection. J. Nonparametr. Stat. 2018, 30, 197–215. [Google Scholar] [CrossRef]
Figure 1. Single-channel EMG recorded using surface electrodes.
Figure 1. Single-channel EMG recorded using surface electrodes.
Sensors 24 05231 g001
Figure 2. Single-channel EMG signal recorded from the biceps.
Figure 2. Single-channel EMG signal recorded from the biceps.
Sensors 24 05231 g002
Figure 3. Flexor and extensor muscles responsible for hand gestures.
Figure 3. Flexor and extensor muscles responsible for hand gestures.
Sensors 24 05231 g003
Figure 4. Myo armband sensor module.
Figure 4. Myo armband sensor module.
Sensors 24 05231 g004
Figure 5. Mounting of Myo armband.
Figure 5. Mounting of Myo armband.
Sensors 24 05231 g005
Figure 6. Myo armband fitted on the arm for data acquisition.
Figure 6. Myo armband fitted on the arm for data acquisition.
Sensors 24 05231 g006
Figure 7. Seven hand gestures used in the study.
Figure 7. Seven hand gestures used in the study.
Sensors 24 05231 g007
Figure 8. (a) Eight-channel (Electrode-1 to Electrode-8) raw EMG data used for the study corresponds to the index finger gesture. (b) Raw EMG data recorded by electrode 1 corresponds to the seven hand gestures.
Figure 8. (a) Eight-channel (Electrode-1 to Electrode-8) raw EMG data used for the study corresponds to the index finger gesture. (b) Raw EMG data recorded by electrode 1 corresponds to the seven hand gestures.
Sensors 24 05231 g008
Figure 9. Architecture of the proposed methodology.
Figure 9. Architecture of the proposed methodology.
Sensors 24 05231 g009
Figure 10. Plot of cumulative explained variance by PCA.
Figure 10. Plot of cumulative explained variance by PCA.
Sensors 24 05231 g010
Table 1. Hand gestures and their respective counts.
Table 1. Hand gestures and their respective counts.
Activity/
Class ID
Finger/GestureNumber of Segments (Counts)Percentage
(%)
1Index finger extension103915.23
2Middle finger extension101714.91
3Ring finger extension104515.32
4Little finger extension 108415.89
5Thumbs up 101614.89
6Victory72610.64
7Relaxed hand89513.12
Table 2. Cross-validated scores of the logistic regression model.
Table 2. Cross-validated scores of the logistic regression model.
Metric/ClassClass 1Class 2Class 3 Class 4Class 5Class 6Class 7
Precision75.4185.9995.9485.8282.096.4796.23
Recall79.2183.2895.9885.9778.8397.9396.64
F1-score 77.2484.6095.9485.8980.3697.1996.43
Accuracy79.2183.2895.9885.9778.8397.9396.64
AUC96.0697.6799.7398.1896.7499.9999.90
Table 3. Cross-validated scores of the SVM model with a linear kernel.
Table 3. Cross-validated scores of the SVM model with a linear kernel.
Metric/Class Class 1Class 2Class 3 Class 4Class 5Class 6Class 7
Precision75.0387.1095.6788.6487.8698.0796.61
Recall84.2185.5496.4588.2878.8398.2195.30
F1-score 79.3386.3096.0588.4483.0998.1395.95
Accuracy84.2185.5496.4588.2878.8398.2195.30
AUC94.1395.9899.4896.3195.1799.9699.66
Table 4. Cross-validated scores of the SVM model with an ‘rbf’ kernel.
Table 4. Cross-validated scores of the SVM model with an ‘rbf’ kernel.
Metric/ClassClass 1Class 2Class 3 Class 4Class 5Class 6Class 7
Precision83.9090.6696.7790.6090.5196.8798.20
Recall89.4187.8097.5193.1783.5697.5297.87
F1-score 86.5489.2097.1391.8686.8997.1998.03
Accuracy89.4187.8097.5193.1783.5397.5297.87
AUC97.1497.5799.6398.4896.5799.9399.78
Table 5. Cross-validated scores of the Random Forest classifier.
Table 5. Cross-validated scores of the Random Forest classifier.
Metric/ClassClass 1Class 2Class 3 Class 4Class 5Class 6Class 7
Precision91.6093.7098.0694.4195.5998.3791.33
Recall93.8492.9197.0396.1291.6310097.65
F1-score 92.6893.2797.5495.2493.5699.1897.49
Accuracy93.8492.9197.0396.1291.6310097.65
AUC99.5099.5599.8299.7399.5399.9999.94
Table 6. Cross-validated scores of the kNN classifier.
Table 6. Cross-validated scores of the kNN classifier.
Metric/ClassClass 1Class 2Class 3 Class 4Class 5Class 6Class 7
Precision88.6190.8797.7094.4294.7198.2398.86
Recall93.4590.5597.7094.9289.7697.3196.87
F1-score 90.9690.8697.7094.6692.1698.7697.85
Accuracy93.4590.8597.7094.9289.7999.3196.87
AUC98.6698.5599.6399.3798.1499.9699.85
Table 7. Cross-validated mean scores of the four models.
Table 7. Cross-validated mean scores of the four models.
Metric/ModelLRSVM-1SVM-2RFkNN
Precision88.2789.8692.595.5894.77
Recall88.2689.5192.4195.694.69
F1-score 88.2489.6192.4195.5794.71
Accuracy87.6889.0792.1195.3894.47
AUC98.3797.2498.4499.7299.17
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kadavath, M.R.K.; Nasor, M.; Imran, A. Enhanced Hand Gesture Recognition with Surface Electromyogram and Machine Learning. Sensors 2024, 24, 5231. https://doi.org/10.3390/s24165231

AMA Style

Kadavath MRK, Nasor M, Imran A. Enhanced Hand Gesture Recognition with Surface Electromyogram and Machine Learning. Sensors. 2024; 24(16):5231. https://doi.org/10.3390/s24165231

Chicago/Turabian Style

Kadavath, Mujeeb Rahman Kanhira, Mohamed Nasor, and Ahmed Imran. 2024. "Enhanced Hand Gesture Recognition with Surface Electromyogram and Machine Learning" Sensors 24, no. 16: 5231. https://doi.org/10.3390/s24165231

APA Style

Kadavath, M. R. K., Nasor, M., & Imran, A. (2024). Enhanced Hand Gesture Recognition with Surface Electromyogram and Machine Learning. Sensors, 24(16), 5231. https://doi.org/10.3390/s24165231

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop