Next Article in Journal
Some Insights Regarding Symmetry Relevance in Biomedicine
Next Article in Special Issue
Distributed Network of Adaptive and Self-Reconfigurable Active Vision Systems
Previous Article in Journal
Postface for Applied Designs in Chemical Structures with High Symmetry
Previous Article in Special Issue
Introducing Urdu Digits Dataset with Demonstration of an Efficient and Robust Noisy Decoder-Based Pseudo Example Generator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network

1
University Institute of Information Technology, Pir Mehr Ali Shah Arid Agriculture University, Rawalpindi 46300, Pakistan
2
Industrial Engineering Department, College of Engineering, King Saud University, Riyadh 11421, Saudi Arabia
3
Department of Statistics and Operations Research, College of Science, King Saud University, Riyadh 11451, Saudi Arabia
4
School of Information Technology, Deakin University, Burwood, VIC 3128, Australia
5
Space and Upper Atmosphere Research Commission, Islamabad 44000, Pakistan
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(10), 2045; https://doi.org/10.3390/sym14102045
Submission received: 25 August 2022 / Revised: 15 September 2022 / Accepted: 26 September 2022 / Published: 30 September 2022
(This article belongs to the Special Issue Computer Vision, Pattern Recognition, Machine Learning, and Symmetry)

Abstract

:
This paper investigated the effects of variant lighting conditions on the recognition process. A framework is proposed to improve the performance of gesture recognition under variant illumination using the luminosity method. To prove the concept, a workable testbed has been developed in the laboratory by using a Microsoft Kinect sensor to capture the depth images for the purpose of acquiring diverse resolution data. For this, a case study was formulated to achieve an improved accuracy rate in gesture recognition under diverse illuminated conditions. For data preparation, American Sign Language (ASL) was used to create a dataset of all twenty-six signs, evaluated in real-time under diverse lighting conditions. The proposed method uses a set of symmetric patterns as a feature set in order to identify human hands and recognize gestures extracted through hand perimeter feature-extraction methods. A Scale-Invariant Feature Transform (SIFT) is used in the identification of significant key points of ASL-based images with their relevant features. Finally, an Artificial Neural Network (ANN) trained on symmetric patterns under different lighting environments was used to classify hand gestures utilizing selected features for validation. The experimental results showed that the proposed system performed well in diverse lighting effects with multiple pixel sizes. A total aggregate 97.3% recognition accuracy rate is achieved across 26 alphabet datasets with only a 2.7% error rate, which shows the overall efficiency of the ANN architecture in terms of processing time.

1. Introduction

A hand gesture is simply defined as “the known movement of a single or both hands”. A Gesture usually conveys a precise single message. For example, waving a hand is to say ‘Good Bye’, and waving both hands shows excitement [1]. For humans, the eyes capture the hand movement, and the brain processes the image to recognize a gestural movement. In this process, the combination of sharp vision and right visibility produces a perfect picture in the brain. This results in accurate human gesture recognition. Similarly, machines require complete and clear input information for accurate vision-based gesture recognition. In addition, an efficient embedded system is required to extract meaningful information from the environment.
In a vision-based gestural system, expressive input is required to recognize gestures with human-like accuracy. The input types used for vision-based gesture recognition technology are static and dynamic gestures. A static gesture image contains the hand in a static state. This means that the posture of the hand, its position, and shape remain the same. On the other hand, a dynamic gesture comprises a sequence of still photos that have been acquired frame by frame. Dynamic gestures can match the characteristics of actual hand movement by having different images express the gestural movement from start to finish. The accuracy of the dynamic gestural recognition system relies on an accurate static gesture recognition model [2]. This is because factors that affect the static gesture recognition process ultimately affect dynamic gesture recognition.
Image noise is one such factor that reduces the accuracy of the recognition system. Noise is defined as the undesired information that comes along with the input image. An example of this is the variation of brightness level within an image. This illumination variation occurs either because of the image sensor or the image surroundings. The change in the number of lumens (lm) emitted by the lighting device is referred to as brightness variation. More lumens mean brighter light, fewer lumens mean dim light, and in between is ambient or general lighting. Usually, rooms and kitchens are set to general lighting, whereas bathrooms have brighter lighting and bedrooms have dim lighting. Real-time recognition becomes complex when the user is moving around the house in different lighting conditions where the brightness level varies unexpectedly. Hence, image enhancement is necessary for accurate recognition.
Many researchers have proposed different image enhancement techniques to improve the feature extraction and overall gestural recognition process. Some of the applications, such as sign language recognition, need a high level of precision and reliability in detecting the hand and recognizing gestures. Therefore, distinguishing characteristics need to be identified for this purpose. However, the vast majority of efforts have only looked at a single lighting scenario to recognize the gesture pattern, so there is still a lot of room for improvement in the area of gesture recognition under varying lighting scenarios. In this study, symmetric patterns and a related luminosity-based filter are considered for use in gesture recognition. The main contributions of this paper are as follows:
  • Firstly, we proposed a symmetry-pattern-based gesture recognition framework that works well in diverse illumination lighting effects.
  • Secondly, the dataset is created based on 26 American Sign Language (ASL) hand gesture images under diverse illumination conditions. Then, an efficient method for gesture feature extraction is used that is based on luminosity-based grey-scale image conversion and perimeter feature extraction.
  • Thirdly, segmentation and identifying the significant points to enhance the number of Scale-Invariant Feature Transform (SIFT) key points and minimized the time taken for key point localization within features.
  • Then, the gesture recognition process is validated by different Artificial Neural Network (ANN) architectures to enhance the recognition accuracy rate and avoid any uncertainty management in decision-making.
  • Finally, a comparison has been performed between our work and other available researchers’ published work in a similar domain to show the efficiency of our proposed framework process.
The organization of the paper is as follows in different sections: Section 2 literature review. Section 3 briefly explains the proposed hand gesture recognition framework with mathematical modelling. Section 4 demonstrated the testbed environment and results. Finally, Section 5 discussed the conclusion and possible future work.

2. Literature Review

This section reviews the most relevant state-of-the-art literature on gesture detection in a variant illuminated background environment, considering its application in ASL. Researchers [3,4,5,6] have proposed different techniques to address the illumination variation problem in vision-based hand gesture recognition systems. For sign language recognition covering languages of their origin, that may affect image recognition. In a study [7], the approach of pattern recognition for surface electromyography (sEMG) signals of nine different finger movements is described. The authors in [8] proposed log-spiral codes of symmetric patterns in the unique method that was developed to identify human hands and understand motions from video streams using long spiral codes. In a recent study [9], the authors present a symmetric CNN called HDANet. This CNN is built on the self-attention mechanism of the Transformer and makes use of symmetric convolution in order to capture the relationships of image information in two dimensions, specifically spatial and channel. In research [10], using a generative adversarial network to capture the implicit relationship between glyphs from Oracle Bone Characters and modern Chinese characters was the basis of a research project that proposed a method for image translation from Oracle Bone Characters to modern Chinese characters. This method could translate images from Oracle Bone Characters to modern Chinese characters. Another research [11] presents a collaborative surgical robot system for percutaneous treatment directed by hand gestures and supplemented by an AR-based surgical field. The use of hand gestures to instruct the surgical robot improved needle insertion accuracy in experiments. Whereas [12] proposed a depth-based palm biometrics system. The technology splits the user’s palm and retrieves finger dimensions from the depth picture. In addition, studies [13,14,15] present a thorough review of hand gesture techniques to eliminate the effect of illumination variations. Some of the most common classifiers proposed by the researchers include k-NN, presented in [12], the SVM classifier discussed in [16], and the tree-based random forest classifier elaborated in [17].
The authors in [18] presented a novel recognition algorithm based on a double-channel convolutional neural network, which separates the varying illumination from the gesture. The study [19], recognized sign language gestures in seven categories using visibility, shape, and orientation of the hand features. At the preprocessing step, they applied skin detection using colour properties in the HSV domain to form a uniform linear binary pattern. The multiclass Support Vector Machine classifier classified the images from the dataset of 3414 signs corresponding to 37 Pakistan Sign Language alphabets with good categorization results. Chen et al. [20] proposed an event-based system that uses a biologically inspired neuromorphic vision sensor, an encoding process to identify objects, and a flexible system to classify hand movements. According to [21], the fitness function for gamma correction preserves the brightness and details of the image of both brighter and low-contrast images. Particle Swarm Optimization can be applied to make the gamma correction adaptive by calculating the optimal gamma values.
As shown by [22], a method for compensating for the poor ambient illumination in the scene is by balancing it against incident illumination. Demonstrated by [23], details about imaging hardware, the collection procedure, the organization of the database, several potential uses of the database, and how to obtain the database. The study collected a database of over 40,000 images of 68 people. Each person is captured in 13 different poses, under 43 different illumination conditions, and with 4 different expressions. In another study [24], a large-scale dataset was collected with various illumination variations to evaluate the performance of the Remote Photo Plethysmography (RPG) algorithm. The study also proposed a low-light enhancement solution for remote heart rate estimation under low-light conditions. In a recent study [25], fine perceptive generative adversarial networks (FP-GANs) are proposed to construct super-resolution (SR) MR images from low-resolution equivalents. FP-GANs use a divide-and-conquer strategy to process low-and high-frequency MR image components individually and in tandem. In another study [26] mild cognitive impairment and Alzheimer’s disease are assessed using a tensorizing GAN with high-order pooling. The proposed model can benefit from brain structure by tensorizing a three-player cooperative gaming framework. By introducing high-order pooling into the classifier, the suggested model can employ second-order MRI statistics (MRI). The study [27] proposed state-of-the-art XAI algorithms for EMG hand gesture classification to understand the outcome of machine learning models with respect to physiological processes to recognise hand movements by mapping and merging synergic muscle activity.
The systematic review shown in Table 1 that highlights the potential related research focused on diverse illumination variation factors is given by.

3. Materials and Methods

This section proposed a framework of sensor-based sign language gesture recognition which consists of the following main phases, i.e., acquiring the image, image preprocessing, feature extraction, and classification, as shown in Figure 1.
The recognition process starts with the acquisition of depth images using the Kinect sensor. Each pixel in a depth image reflects the distance between the image plane and an RGB image object. Following that, hand shapes are precisely segmented in order to locate and track the hands’ shapes, similar to the human’s communication approach. To develop the datasets, ASL-based gesture images are stored in a database and converted into PNG format. The next step is the pre-processing of the acquired images to transform the captured image into a uniform level of brightness. For that, the system performs the luminosity method based on the grey-scale conversion of the input image. Grayscale conversion reduces complexity and is much easier to work with a variety of tasks such as image segmentation problems. Greyscale conversion is carried out through the weighted method, also called the luminosity method. The main reason for proposing the luminosity method is to equalize the weights of red, green, and blue according to their wavelengths. The luminosity method is a better version of the average method. As discussed in [35], luminosity-based greyscale conversion can be calculated as follows:
Grayscale = 0.299R + 0.587G + 0.114B
where R, G, and B represent red, green, and blue colours, respectively.
The next step is to extract the appropriate features and their selection. The selection of the number of features is a critical step because more features consume additional space and computational time. Fewer features affect the accuracy. For that, the SIFT method is proposed to select and extract the appropriate features from acquired data. The proposed method extracts four significant features (perimeter, hand size, centre of hand and finger distance) from a given input image.
To define the shape of the hand and calculate the perimeter value, the perimeter feature extraction (PFE) technique is used to detect the edges and boundaries of the human hand by counting the pixels having values of 1 and 0 for neighboring pixels while skipping the grey shades. The shape is calculated by finding the projection of the hand that provides the size of the hand. For that, vertical and horizontal values are calculated by adding up all the values of rows and columns as follows:
v i   ( c ) = c = 0 n 1 P i ( r , c )
h i   ( r ) = r = 0 n 1 P i ( r , c )
where v and h represent the vertical and horizontal positions, and r and c represent the row and column, respectively, of pixel P in an image. In Equation (2), the letter “n” stands for the maximum “height” value, which is the vertical side of a hand. In Equation (3), the letter “n” stands for the maximum “width” value, which is the horizontal side of a hand.
The hand size feature is useful to recognize the change in the size of a hand because hand size differs from person to person. The size component of a hand represents the hand size at a specific time. To calculate the hand size, we define the function, Mi(r, c) as in Equations (4) and (5).
M i   ( r , c ) = { 1   i f   P ( r , c ) = i t h   object   number 0 otherwise   }
A i = r = 0 n 1   c = 1 n 1 M i ( r , c )
where r and c represent the row and column, respectively, of pixel P in an image. Then, the area Ai is measured in pixels and indicates the relative size of the hand.
The centre of the hand feature recognizes the orientation and position of the hand. Finding the centre of the object is important for detecting any change in hand shape, palm position, and movement of the hand or fingers. We can define the centre of hand by the pair ( r i , c i ) for rows and columns, respectively, measured as follows:
r i = 1 A i r = 0 n 1   c = 1 n 1   r   M i ( r , c )
c i = 1 A i r = 0 n 1   c = 1 n 1 c   M i ( r , c )
where both r and c represent the row and column, respectively, of pixel P in an image, and Mi(r, c) is a size function.
Finally, the fingers position is calculated by finding the distance between two open or closed fingers. The finger’s distance feature helps define the gesture. It can be carried out by counting the continuous pixel having a value of 0 until a neighboring pixel with a value of 1. The significant points can be calculated as follows:
S p = ( avg   x , avg   y )
S p = ( 1 5 p = s x p = x + 5 ,   1 5 p = s y p = y + 5 + b )
where Sp is a significant point plotted on the plane averaging 5-pixel values along the x-axis and y-axis, respectively, sx is the starting point along the x-axis averaging the next five values, sy is the starting point along the y-axis averaging the next five-pixel values. Additionally, b is a bias that gives the neural network an extra parameter to tune by initializing non-zero random values.
After that, extracted features are combined into the form of a feature vector set F s , the data is displayed in the form of these symmetric patterns. To measure the boundary, size, and orientation of the hand for a particular gesture defined as
F s = [ P e r i m e t e r , H a n d   S i z e ,   C e n t e r   o f   H a n d ,   F i n g e r   P o s i t i o n ]
where
P e r i m e t e r   F 1 = [ a 1 , a 2 , a 3 , , a n ]
H a n d   S i z e   F 2 = [ b 1 , b 2 , b 3 , , b n ]
C e n t e r   o f   H a n d   F 3 = [ x 1 , x 2 , x 3 , , x n ]
F i n g e r   P o s i t i o n   F 4 = [ y 1 , y 2 , y 3 , , y n ]
where the letter “n” denotes the total number of extracted ‘feature points’ for each defined feature (F1, F2, F3, and F4) for a hand gesture. The symmetric feature set enables the description of symmetric patterns such as the perimeter, center, finger position, and the size of the hand. Where ‘perimeter features’ outline the physical shape of the hand, ‘size feature’ extracts features for comparison of scale variation of the hand, ‘center of hand’ handles the orientation of the hand, and ‘finger position’ feature measures the distance between two fingers. The feature set created in matrix form shows all the features combined in Equation (15).
X = [   F 1   F 2   F 3   F 4 ] [ a 1   a 2   a 3 a n b 1   b 2   b 3 b n x 1   x 2   x 3 x n y 1   y 2   y 3   y n ] [ 0   0   0   1   1   0   .   .   .   0 0   0   1   1   0   0   .   .   .   1 0   0   1   1   1   0   .   .   .   0 0   1   1   1   0   0   .   .   .   1 ]
where each pixel corresponds to a matrix element value. “0” and “1” represent pixel values, with 1 indicating the presence of a feature point and 0 indicating its absence. Finally, the SIFT technique is employed, which was proposed by Lowe, D.G. [36] and later the anatomical structure was discussed by the authors in their work [37]. The SIFT technique extracts features and identifies the significant key points of ASL-based images with their relevant features in diverse illumination conditions. The SIFT algorithm is formulated as follows in Algorithm 1.
Algorithm 1: SIFT Keypoints Generation
1:   Gaussian scale-space computation
2:    Input: i image
3:    Output: s scale-space
4:   Difference of Gaussians (DoG)
5:    Input: s scale-space
6:    Output: d DoG
7:   Finding keypoints (extrema of DoG)
8:    Input: d DoG
9:    Output: {(rd, cd, αd)} list of discrete extrema (position and scale)
10:   Keypoints localization to sub-pixel precision
11:    Input: d DoG and {(rd, cd, αd)} discrete extrema
12:    Output: {(r, c, α)} extreme points
13:   Filter unstable extrema
14:    Input: d DoG and {(r, c, α)}
15:    Output: {(r, c, α)} filtered keypoints
16:   Filter poorly localized keypoints on edges
17:    Input: d DoG and {(r, c, α)}
18:    Output: {(r, c, α)} filtered keypoints
19:   Assign a reference orientation to each keypoint
20:    Input: (∂mv, ∂nv) scale-space gradient and {(r, c, α)} list of keypoints
21:    Output: {(x, y, α, θ)} list of oriented keypoints
22:   SIFT Feature descriptor generator
23:    Input: (∂mv, ∂nv) scale-space gradient and {(x, y, α, θ)} list of keypoints
24:    Output: {(r, c, 1: α, θ, f)} list of described keypoints

4. Experimentation and Results

To prove the concept, a workable testbed has been developed in the laboratory by using the Microsoft Kinect sensor to capture the images and convert them into depth images for acquiring the diverse resolution data, as shown in Figure 2.
For the development of a case study, a total of three subjects were considered to acquire the data in diverse illuminated conditions and save it into PNG format according to Equation (1), as shown in Figure 3.
The subsequent step is to convert the ASL images into grayscale for the segmentation to separate the hand object from its background, as shown in Figure 4. The next stage is to calculate the SIFT points from the segmented hand objects to identify the feature points as mentioned in Equation (15). A feature descriptor method is used to process the significant image points from identified feature points to convert them into significant vector points. For that, the Matlab tool is used to extract the required feature point values and convert them into significant points using the SIFT algorithm from 26 letters of the alphabets.
After calculating the SP using SIFT, the average processing time of every step is calculated under different lighting conditions and resolution rates as shown in the following Table 2. As we can observe from Table 2, the reasonable processing time is at a 1024 × 768-resolution rate in ambient light conditions, which is a good resolution rate for analysis. It is noticed that higher resolution rates consume more processing time. Therefore, we will consider the 1024 × 768 resolution rate at the next stage.
After converting the 26 alphabets into grayscale, segmentation and significant point calculation of A–Z, four letters are considered to show the efficiency of the proposed framework. Letters “P”, “A”, “I”, and “R” are used for segmentation, SIFT, and measure the significant points of the hand gesture as shown in the following Figure 5.
Referring to Equation (15), and Figure 6, we considered the mean ( x ¯ ), standard deviation (σ), variance (µ) and average deviation (AD) values against each feature of selected SP and converted them into the compact form of featured SP, respectively. The details are shown in Table 3.
Selection of applicable gesture features is a key task as inputs of ANN for training to measure the accuracy from acquired data and it may indistinct the network structure. However, the right feature selection improves the efficiency of the ANN network, and training time may also be reduced by adopting the right ANN hidden layer architecture according to the input and output. In classification and data processing, ANN learning accommodates a variety of conditions better than any other classification technique [38]. In this paper, we have chosen four different features as inputs ( F 1 ,   F 2 , F 3   a n d   F 4 ) for each layer of neurons. Each network consists of one hidden layer that contains multiple neurons according to the inputs. The number of hidden layer neurons has a reliable impact on the performance of the ANN model. Therefore, the selection of a number of the hidden layer neurons depends on ANN accuracy in primary trials. For the target output, a vector of classes to recognize the hand gesture is written as follows:
  • [1; 0; 0]: Hand Natural Position;
  • [0; 1; 0]: Hand Gestural Position;
  • [0; 0; 1]: Hand Unknown Position.
A multi-layer Feedforward Neural Network (FFNN) method is used in this paper for the recognition of hand gestures from datasets. The proposed architecture of ANN for a single hand gesture is presented in Figure 7.
Based on the same ANN training process, we used a similar classification architecture for the recognition of multiple hand gestures as objects. Every hand gesture feature’s vector class was used as input data and classified through a similar network architecture. The output layer in Figure 6 presents the current state of the gesture. It contains a total of four NN input nodes, and the hidden layer activation function ( l o g s i g ) is employed for every proposed output. All object feature values in Table 3 were stored in the Mat extension file and assigned these values with each hand gesture and divided all features into sub-features. The training goal is set at 0.01 target value and the Back Propagation (BP) learning method is adopted for training. Figure 7 presents the inside architecture of each neural network for each hand object.
After initializing the ANN model for non-linear system modelling, specific ANN data must be assessed targeting nodes precedents. The hidden layer neurons and transfer function are set up to compute the training objective. Then, the layer weight is set for output. Table 4 shows the brief explanation and ANN layer setup information.
For the selection of suitable hidden layer neurons architecture, three types of ANN architecture (the [4 × 4 × 3], [4 × 14 × 3] and [4 × 24 × 3]) were tested in this paper for training purposes, as shown in Figure 8. To alter the weights of the hidden layer until the targeted output was achieved at reasonable epoch numbers with less error rate, as shown in Table 5.
Table 4. Description of the implemented ANN.
Table 4. Description of the implemented ANN.
NN StepsArtificial Neural Network Structure for Performance Matrices
Network ModeFFNN
Learning PatternBack Propagation
Training Goal0.001
Input dataFour inputs of 1D ANN matrix where all data were placed in each image’s class for recognition process index
No. of neurons in hidden layerDiverse N architectures are used with different values of neurons inside hidden layer. For example, [4 × 4 × 3], [4 × 14 × 3] and [4 × 24 × 3] (see Figure 9).
Vector of classes for the target outputsMathematical matrices refer to the classified vector classes with value 0 or 1.
It can be observed from Table 5 that the selected architecture [4 × 14 × 3] has presented better mean squared error (MSE) performance with reasonable epoch numbers and error rate than other ANN architectures. The next process is to measure the validation of acquired features in Table 2. Figure 9, Figure 10, Figure 11 and Figure 12 show the training performance graph of the ANN architecture [4 × 14 × 3], which attained a good and considerable performance result during ANN testing.
After measuring and testing the performance graph of the ANN architecture [4 × 14 × 3], the next stage of validation is to calculate the accuracy by calculating the Confusion Matrix (CM). To construct the CM, the four features input ( F 1 ,   F 2 , F 3   a n d   F 4 ) values are inserted into ANN architecture in Figure 7 by adjusting the height of the hidden layer. The combined confusion matrices of all features against each character of the PAIR word are shown in Figure 13, Figure 14, Figure 15 and Figure 16.
In Figure 13, Figure 14, Figure 15 and Figure 16 above, each corner cell shows the accurately tested pattern cases of gestures through the proposed ANN architecture [4 × 14 × 3] to decide the recognition of the right-hand gesture. In the confusion matrices graph, the confusion grid holds the features processed training data between the target and output classes, consisting of three procedural stages: preparing, testing, and training of the gesture recognition and individually measuring the performance of the ANN architecture.
To perform these procedural stages, four horizontal target and vertical output classes were defined to demonstrate the precise data validation testing process to reflect all possible targeted sample values of feature sets. The green cells in the CM grid graphs show those data groups that are accurately classified and have completed a successful training process. Each grey corner cell in horizontal position shows those data groups of targeted classes that are accurately classified and complete the testing phase in the training process. The red cell presents those data sets that are wrongly classified or might not be properly validated in the testing phase. Finally, the blue cell presents the overall percentage of correctly classified gesture test cases from datasets. From confusion matrix diagrams, we can easily observe that each class has been tested under 1200 test cases and show percentages in green cells to observe the targeted class output parentage with error rates which are accurately classified during the testing phase with less than 1% wrongly classified in all trained datasets. Overall, a maximum 97.4% accurate rate of the word “PAIR” was achieved in the blue cell with only a 2.6% error rate, which shows the overall efficiency of the ANN architecture in terms of processing time.
After measuring, testing, and calculating the accuracy of four selected features, the next step is to measure the accuracy of the whole dataset (A–Z). For that, the same ANN architecture [4 × 14 × 3] is utilized with the previous configuration settings that were used in the case study. Figure 17 shows the training performance graph of ANN architecture [4 × 14 × 3], NN testing produced an excellent and significant performance result. The combined confusion matrix of all chosen features against each character (A–Z) word is shown in Figure 17. From Figure 18, we can certainly perceive that each class has been tested under 1400 test cases from trained datasets to observe the accurately classified target output class and only less than 1% are wrongly classified in all trained datasets. Overall, a maximum of 97.3% accuracy rate of all 26 alphabets was achieved in the blue cell with only a 2.7% error rate, which shows the overall efficiency of the adopted NN architecture [4 × 14 × 3] in terms of processing time.
Finally, a comparison has been performed between our work and other researcher’s published works in a similar domain to show the efficiency of our proposed framework process. Table 6 compares the performance of our and other works based on gesture image datasets and recognition methods in terms of the number of gestures, frame resolution, response time, recognition approaches, and accuracy rate with an error rate under various illuminated conditions. The combination of preprocessing process, PFE, segmentations, significant point extraction, and utilising multiple ANN architectures for classification to reduce the error rate and achieve the high accuracy rate in gesture recognition is the reason for achieving the high accuracy rate compared with others.

5. Conclusions

This paper presented an efficient framework to improve the performance of gesture recognition under variant illumination using the luminosity method. Symmetric patterns and a related luminosity-based filter are considered for use in gesture recognition. The proposed framework consists of four main phases, i.e., acquiring the image, image pre-processing, feature extraction, and classification. To develop the datasets, a workable testbed has been developed in the laboratory by using two Microsoft Kinect sensors to capture the depth images for the purpose of acquiring diverse resolution data. ASL-based gesture images are stored in the database and converted into PNG format. The next step is the pre-processing of the acquired images to transform the captured image into a uniform level of brightness. For that, the system performs the luminosity method based on the grey-scale conversion of the input image. Grayscale conversion reduces complexity and is much easier to work with a variety of tasks such as image segmentation problems. Greyscale conversion was carried out through the weighted method. The next step is to extract the appropriate features and make their selection. It is a critical step because more appropriate features consume additional space and computational time. For that, the SIFT method is proposed to select and extract the appropriate features from acquired data. The proposed method extracts four significant features (perimeter, hand size, centre of hand, finger distance) from a given input image. After that, extracted features are combined into the form of a feature vector set to measure the boundary, size, and orientation of the hand for a particular gesture. Then, SIFT algorithm is to identify the significant key points of ASL-based images with their relevant features in diverse illuminated conditions. A feature descriptor method is used to calculate the significant image points from identified feature points and convert them into significant vector points. From the results, we can observe that the reasonable processing time is at a 1024 × 768 resolution rate, which is a good resolution rate for analysis. It is also noticed that higher resolution rates consume higher processing time and are shown in tabular form. Finally, we have chosen four different features as inputs ( F 1 ,   F 2 , F 3   a n d   F 4 ) for each layer of neurons. Each network consists of one hidden layer that contains multiple neurons with NN architectures ([4 × 4 × 3], [4 × 14 × 3], and [4 × 24 × 3]), respectively. After training, it is identified that the architecture [4 × 14 × 3] presented a better mean squared error (MSE) performance with reasonable epoch numbers and error rate among other NN architectures. From confusion matrix diagrams, we can easily observe that each class has been tested under 1200 test cases and show percentages in green cells to observe the targeted class output parentage with error rates which are accurately classified during the testing phase with less than 1% wrongly classified in all trained datasets. Overall, a maximum 97.4% accurate rate of the word “PAIR” was achieved with only a 2.6% error rate, which shows the overall efficiency of the NN architecture.
Future development would be extended toward the incorporation of complex depth images with more gesture angles in other languages (regional sign language) by using multiple high-resolution camera modules. Further complexity in recognition can be created to select more features at multiple angles and lighting conditions in outdoor environments.

Author Contributions

Conceptualization, M.H., M.Z., S.A. (Shafiq Ahmad) and S.A. (Saud Altaf); methodology, M.H., S.A. (Saud Altaf) and M.Z.; software, M.H.; validation, M.H., S.A. (Shafiq Ahmad) and S.I.; formal analysis, M.H., M.Z. and S.A. (Saud Altaf); investigation, M.H.; resources, S.A. (Saud Altaf); data curation, S.A. (Saud Altaf), M.Z. and S.A. (Shafiq Ahmad); writing—original draft preparation, M.H., S.H. and S.I.; writing—review and editing, S.A. (Saud Altaf); visualization, M.H. and S.A. (Saud Altaf), M.Z.; supervision, S.H.; project administration, M.Z.; funding acquisition, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by King Saud University, grant number RSP-2021/387, and the APC was funded by RSP-2021/387.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors extend their appreciation to King Saud University for funding this work through Researchers Supporting Project number (RSP-2021/387), King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Mahmud, H.; Hasan, M.K.; al Tariq, A.; Kabir, M.H.; Mottalib, M.A. Recognition of Symbolic Gestures Using Depth Information. Adv. Hum.-Comput. Interact. 2018, 2018, 1069823. [Google Scholar] [CrossRef]
  2. Sagayam, K.M.; Andrushia, A.D.; Ghosh, A.; Deperlioglu, O.; Elngar, A.A. Recognition of Hand Gesture Image Using Deep Convolutional Neural Network. Int. J. Image Graph. 2022, 22, 2140008. [Google Scholar] [CrossRef]
  3. Khan, N.S.; Abid, A.; Abid, K. A Novel Natural Language Processing (NLP)–Based Machine Translation Model for English to Pakistan Sign Language Translation. Cogn. Comput. 2020, 12, 748–765. [Google Scholar] [CrossRef]
  4. Rezende, T.M.; Almeida, S.G.M.; Guimarães, F.G. Development and validation of a Brazilian sign language database for human gesture recognition. Neural Comput. Appl. 2021, 33, 10449–10467. [Google Scholar] [CrossRef]
  5. Van, Q.P.; Binh, N.T. Vietnamese Sign Language Recognition using Dynamic Object Extraction and Deep Learning. In Proceedings of the 2020 IEEE Eighth International Conference on Communications and Electronics, Phu Quoc Island, Vietnam, 13–15 January 2021; pp. 402–407. [Google Scholar]
  6. Mustafa, M. A study on Arabic sign language recognition for differently abled using advanced machine learning classifiers. J. Ambient Intell. Humaniz. Comput. 2021, 12, 4101–4115. [Google Scholar] [CrossRef]
  7. Moh, A.; Wahyu, C.; Ariyanto, M.; Mol, M.; Setiawan, J.D.; Glowacz, D. Pattern Recognition of Single-Channel sEMG Signal Using PCA and ANN Method to Classify Nine. Symmetry 2020, 12, 541. [Google Scholar]
  8. Nemati, H.; Fan, Y.; Alonso-fernandez, F. Hand Detection and Gesture Recognition Using Symmetric Patterns. Stud. Comput. Intell. 2016, 642, 365–375. [Google Scholar]
  9. Zhang, Q.; Feng, L.; Liang, H.; Yang, Y. Hybrid Domain Attention Network for Efficient Super-Resolution. Symmetry 2022, 14, 697. [Google Scholar] [CrossRef]
  10. Gao, F.; Zhang, J.; Liu, Y.; Han, Y. Image Translation for Oracle Bone Character Interpretation. Symmetry 2022, 14, 743. [Google Scholar] [CrossRef]
  11. Karbasi, M.; Zabidi, A.; Yassin, I.M.; Waqas, A.; Bhatti, Z.; Alam, S. Malaysian Sign Language Dataset for Automatic Sign Language Recognition System. J. Fundam. Appl. Sci. 2017, 9, 459–474. [Google Scholar] [CrossRef]
  12. Pariwat, T.; Seresangtakul, P. Multi-Stroke Thai Finger-Spelling Sign Language Recognition System with Deep Learning. Symmetry 2021, 13, 262. [Google Scholar] [CrossRef]
  13. Wen, R.; Tay, W.L.; Nguyen, B.P.; Chng, C.B.; Chui, C.K. Hand gesture guided robot-assisted surgery based on a direct augmented reality interface. Comput. Methods Programs Biomed. 2014, 116, 68–80. [Google Scholar] [CrossRef] [PubMed]
  14. Nguyen, B.P.; Tay, W.L.; Chui, C.K. Robust Biometric Recognition from Palm Depth Images for Gloved Hands. IEEE Trans. Hum.-Mach. Syst. 2015, 45, 799–804. [Google Scholar] [CrossRef]
  15. Oudah, M.; Al-Naji, A.; Chahl, J. Hand Gesture Recognition Based on Computer Vision: A Review of Techniques. J. Imaging 2020, 6, 73. [Google Scholar] [CrossRef] [PubMed]
  16. Sagayam, K.M.; Hemanth, D.J. A probabilistic model for state sequence analysis in hidden Markov model for hand gesture recognition. Comput. Intell. 2019, 35, 59–81. [Google Scholar] [CrossRef]
  17. Chang, C.C.; Lin, C.J. LIBSVM: A Library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  18. Wu, X.Y. A hand gesture recognition algorithm based on DC-CNN. Multimed. Tools Appl. 2020, 79, 9193–9205. [Google Scholar] [CrossRef]
  19. Saqlain Shah, S.M.; Abbas Naqvi, H.; Khan, J.I.; Ramzan, M.; Zulqarnain; Khan, H.U. Shape based Pakistan sign language categorization using statistical features and support vector machines. IEEE Access 2018, 6, 59242–59252. [Google Scholar] [CrossRef]
  20. Chen, G.; Xu, Z.; Li, Z.; Tang, H.; Qu, S.; Ren, K.; Knoll, A. A Novel Illumination-Robust Hand Gesture Recognition System with Event-Based Neuromorphic Vision Sensor. IEEE Trans. Autom. Sci. Eng. 2021, 18, 508–520. [Google Scholar] [CrossRef]
  21. Mahmood, A.; Khan, S.A.; Hussain, S.; Almaghayreh, E.M. An Adaptive Image Contrast Enhancement Technique for Low-Contrast Images. IEEE Access 2019, 7, 161584–161593. [Google Scholar] [CrossRef]
  22. Al Delail, B.; Bhaskar, H.; Zemerly, M.J.; Werghi, N. Balancing Incident and Ambient Light for Illumination Compensation in Video Applications. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 1762–1766. [Google Scholar] [CrossRef]
  23. Yu, W.; Lei, B.; Ng, M.K.; Cheung, A.C.; Shen, Y.; Wang, S. Tensorizing GAN With High-Order Pooling for Alzheimer’s Disease Assessment. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 4945–4959. [Google Scholar] [CrossRef] [PubMed]
  24. You, S.; Lei, B.; Wang, S.; Chui, C.K.; Cheung, A.C.; Liu, Y.; Gan, M.; Wu, G.; Shen, Y. Fine Perceptive GANs for Brain MR Image Super-Resolution in Wavelet Domain. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef] [PubMed]
  25. Gozzi, N.; Malandri, L.; Mercorio, F.; Pedrocchi, A. XAI for myo-controlled prosthesis: Explaining EMG data for hand gesture classification. Knowl.-Based Syst. 2022, 240, 108053. [Google Scholar] [CrossRef]
  26. Wang, Y.; Ren, A.; Zhou, M.; Wang, W.; Yang, X. A novel detection and recognition method for continuous hand gesture using fmcw radar. IEEE Access 2020, 8, 167264–167275. [Google Scholar] [CrossRef]
  27. Xi, L.; Chen, W.; Zhao, C.; Wu, X.; Wang, J. Image Enhancement for Remote Photoplethysmography in a Low-Light Environment. In Proceedings of the 15th IEEE International Conference on Automatic Face and Gesture Recognition, Buenos Aires, Argentina, 16–20 November 2020; pp. 761–764. [Google Scholar] [CrossRef]
  28. Li, G.; Tang, H.; Sun, Y.; Kong, J.; Jiang, G.; Jiang, D.; Tao, B.; Xu, S.; Liu, H. Hand gesture recognition based on convolution neural network. Clust. Comput. 2019, 22, 2719–2729. [Google Scholar] [CrossRef]
  29. Wang, J.; Liu, T.; Wang, X. Human hand gesture recognition with convolutional neural networks for K-12 double-teachers instruction mode classroom. Infrared Phys. Technol. 2020, 111, 103464. [Google Scholar] [CrossRef]
  30. Al-Hammadi, M.; Muhammad, G.; Abdul, W.; Alsulaiman, M.; Bencherif, M.A.; Mekhtiche, M.A. Hand Gesture Recognition for Sign Language Using 3DCNN. IEEE Access 2020, 8, 79491–79509. [Google Scholar] [CrossRef]
  31. Mujahid, A.; Awan, M.J.; Yasin, A.; Mohammed, M.A.; Damaševičius, R.; Maskeliūnas, R.; Abdulkareem, K.H. Real-time hand gesture recognition based on deep learning YOLOv3 model. Appl. Sci. 2021, 11, 4164. [Google Scholar] [CrossRef]
  32. Al-Hammadi, M.; Muhammad, G.; Abdul, W.; Alsulaiman, M.; Bencherif, M.A.; Alrayes, T.S.; Mathkour, H.; Mekhtiche, M.A. Deep learning-based approach for sign language gesture recognition with efficient hand gesture representation. IEEE Access 2020, 8, 192527–192542. [Google Scholar] [CrossRef]
  33. Tan, Y.S.; Lim, K.M.; Lee, C.P. Hand gesture recognition via enhanced densely connected convolutional neural network. Expert Syst. Appl. 2021, 175, 114797. [Google Scholar] [CrossRef]
  34. Pinto, R.F.; Borges, C.D.B.; Almeida, A.M.A.; Paula, I.C. Static Hand Gesture Recognition Based on Convolutional Neural Networks. J. Electr. Comput. Eng. 2019, 2019, 4167890. [Google Scholar] [CrossRef]
  35. Pin, X.; Chunrong, Z.; Gang, H.; Yu, Z. Object Intelligent Detection and Implementation Based on Neural Network and Deep Learning. In Proceedings of the 2020 International Conference on Computer Engineering and Application ICCEA 2020, Guangzhou, China, 18–20 March 2020; pp. 333–338. [Google Scholar] [CrossRef]
  36. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  37. Rey-Otero, I.; Delbracio, M. Anatomy of the SIFT method. Image Process. Line 2014, 4, 370–396. [Google Scholar] [CrossRef]
  38. Gopal, P.; Gesta, A.; Mohebbi, A. A Systematic Study on Electromyography-Based Hand Gesture Recognition for Assistive Robots Using Deep Learning and Machine Learning Models. Sensors 2022, 22, 3650. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Proposed framework of sensor-based sign language gesture recognition.
Figure 1. Proposed framework of sensor-based sign language gesture recognition.
Symmetry 14 02045 g001
Figure 2. Data acquisition setup for development of case study.
Figure 2. Data acquisition setup for development of case study.
Symmetry 14 02045 g002
Figure 3. Data acquisition of ASL Images under (a) bright light, (b) ambient light, (c) dark light.
Figure 3. Data acquisition of ASL Images under (a) bright light, (b) ambient light, (c) dark light.
Symmetry 14 02045 g003
Figure 4. ASL conversion into grey scale and segmentation.
Figure 4. ASL conversion into grey scale and segmentation.
Symmetry 14 02045 g004
Figure 5. Case study feature and significant point’s calculation using SIFT.
Figure 5. Case study feature and significant point’s calculation using SIFT.
Symmetry 14 02045 g005
Figure 6. Proposed ANN classification architecture for the gesture.
Figure 6. Proposed ANN classification architecture for the gesture.
Symmetry 14 02045 g006
Figure 7. The internal design of NN architecture.
Figure 7. The internal design of NN architecture.
Symmetry 14 02045 g007
Figure 8. Overview of the different ANN architectures chosen: (a) [4 × 4 × 3]; (b) [4 × 14 × 3]; (c) [4 × 24 × 3].
Figure 8. Overview of the different ANN architectures chosen: (a) [4 × 4 × 3]; (b) [4 × 14 × 3]; (c) [4 × 24 × 3].
Symmetry 14 02045 g008aSymmetry 14 02045 g008b
Figure 9. Performance graphs using [4 × 14 × 3] neural network architecture for gesture P.
Figure 9. Performance graphs using [4 × 14 × 3] neural network architecture for gesture P.
Symmetry 14 02045 g009
Figure 10. Performance graphs using [4 × 14 × 3] neural network architecture for gesture A.
Figure 10. Performance graphs using [4 × 14 × 3] neural network architecture for gesture A.
Symmetry 14 02045 g010
Figure 11. Performance graphs using [4 × 14 × 3] neural network architecture for gesture I.
Figure 11. Performance graphs using [4 × 14 × 3] neural network architecture for gesture I.
Symmetry 14 02045 g011
Figure 12. Performance graphs using [4 × 14 × 3] neural network architecture for gesture R.
Figure 12. Performance graphs using [4 × 14 × 3] neural network architecture for gesture R.
Symmetry 14 02045 g012
Figure 13. Confusion matrices for letter P on architecture [4 × 14 × 3].
Figure 13. Confusion matrices for letter P on architecture [4 × 14 × 3].
Symmetry 14 02045 g013
Figure 14. Confusion matrices for letter A on architecture [4 × 14 × 3].
Figure 14. Confusion matrices for letter A on architecture [4 × 14 × 3].
Symmetry 14 02045 g014
Figure 15. Confusion matrices for letter I on architecture [4 × 14 × 3].
Figure 15. Confusion matrices for letter I on architecture [4 × 14 × 3].
Symmetry 14 02045 g015
Figure 16. Confusion matrices for letter R on architecture [4 × 14 × 3].
Figure 16. Confusion matrices for letter R on architecture [4 × 14 × 3].
Symmetry 14 02045 g016
Figure 17. Performance graphs using [4 × 14 × 3] neural network architecture for the gestures A−Z.
Figure 17. Performance graphs using [4 × 14 × 3] neural network architecture for the gestures A−Z.
Symmetry 14 02045 g017
Figure 18. Confusion matrices for letters A–Z on the architecture [4 × 14 × 3].
Figure 18. Confusion matrices for letters A–Z on the architecture [4 × 14 × 3].
Symmetry 14 02045 g018
Table 1. Comparison of various existing techniques under different illuminated conditions.
Table 1. Comparison of various existing techniques under different illuminated conditions.
Paper# of GesturesTechniqueLighting ChangesBackground
[18]10DC-CNN2-VariantRedundant
[28]8CNNN/ACluttered
[29]10ANN2-VariantColourful
[30]83D-CNNvariantOcclusion
[31]24Darknet2-VariantN/A
[32]40D Learn2-VariantCluttered
[33]24DC-CNNDissimilarNoise
[34]24ANNArtificial Cluttered
Table 2. Average processing time under variant lightening conditions at different resolution.
Table 2. Average processing time under variant lightening conditions at different resolution.
ResolutionAverage Processing Time (Sec)
Bright LightAmbient LightDark Light
260 × 1752.382.242.34
320 × 2402.502.202.46
640 × 4802.582.242.25
800 × 6002.212.182.41
1024 × 7682.282.212.35
2048 × 15402.282.362.32
4160 × 31202.392.502.84
Table 3. Feature values calculation.
Table 3. Feature values calculation.
FeaturesPAIR
F1 x ¯ 348.57335.12278.37433.92
µ18744.811651.613358.4928356.31
σ136.91107.94115.57168.39
AD108.6476.0191.82125.29
F2 x ¯ 362.96317.5432.61402.57
µ17503.45888.7510446.195307.95
σ132.3076.73102.2072.85
AD103.2270.7576.4343.01
F3 x ¯ 456.12392.47457.19413.45
µ3845.169542.871481.615662.65
σ62.0097.6838.4975.25
AD47.5679.8830.4763.90
F4 x ¯ 494.23527.55519.97540.97
µ5157.584899.463547.316015.84
σ71.8169.9959.5577.56
AD61.9257.3648.7867.11
Table 5. Different ANN architecture for classification performance.
Table 5. Different ANN architecture for classification performance.
ArchSampleMSENo. of EpochAccuracyClassification Error
[4 × 4 × 3]F17.56 × 10−27093.66.4
F27.22 × 10−26292.77.3
F36.56 × 10−27293.16.9
F47.22 × 10−29893.96.1
[4 × 14 × 3]F18.96 × 10−211496.73.3
F28.75 × 10−212296.83.2
F37.5 × 10−213097.42.6
F49.28 × 10−212597.12.9
[4 × 24 × 3]F17.65 × 10−237293.76.3
F26.22 × 10−230492.87.2
F37.90 × 10−238590.99.1
F47.56 × 10−237490.49.6
Table 6. Performance comparison with other gesture recognition research.
Table 6. Performance comparison with other gesture recognition research.
Paper# of GesturesTest ImageFrame ResolutionRecognition Time (sec)Technique Accuracy (%)Error Rate (%)
[18]10400128 × 1280.4DC-CNN94.85.2
[25]8195320 × 2400.09–0.11CNN93.96.1
[26]10600512 × 4240.133ANN95.64.4
[27]8220112 × 1120.033D-CNN95.84.2
[28]24300416 × 4160.0666Darknet96.73.3
[29]4090112 × 112N/AD Learn96.23.8
[30]2466320 × 240N/ADC-CNN94.55.5
[31]24135400 × 4000.19ANN95.74.3
Our work268001024 × 7680.013ANN97.42.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Haroon, M.; Altaf, S.; Ahmad, S.; Zaindin, M.; Huda, S.; Iqbal, S. Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network. Symmetry 2022, 14, 2045. https://doi.org/10.3390/sym14102045

AMA Style

Haroon M, Altaf S, Ahmad S, Zaindin M, Huda S, Iqbal S. Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network. Symmetry. 2022; 14(10):2045. https://doi.org/10.3390/sym14102045

Chicago/Turabian Style

Haroon, Muhammad, Saud Altaf, Shafiq Ahmad, Mazen Zaindin, Shamsul Huda, and Sofia Iqbal. 2022. "Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network" Symmetry 14, no. 10: 2045. https://doi.org/10.3390/sym14102045

APA Style

Haroon, M., Altaf, S., Ahmad, S., Zaindin, M., Huda, S., & Iqbal, S. (2022). Hand Gesture Recognition with Symmetric Pattern under Diverse Illuminated Conditions Using Artificial Neural Network. Symmetry, 14(10), 2045. https://doi.org/10.3390/sym14102045

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop