Next Article in Journal
An Outlook on Physical and Virtual Sensors for a Socially Interactive Internet
Next Article in Special Issue
Lightweight Visual Odometry for Autonomous Mobile Robots
Previous Article in Journal
General Signal Model for Multiple-Input Multiple-Output GMTI Radar
Previous Article in Special Issue
Fast Visual Odometry for a Low-Cost Underwater Embedded Stereo System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Handshape Recognition Using Skeletal Data

by
Tomasz Kapuscinski
* and
Patryk Organisciak
Department of Computer and Control Engineering, Rzeszow University of Technology, 35-959 Rzeszow, Poland
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(8), 2577; https://doi.org/10.3390/s18082577
Submission received: 10 July 2018 / Revised: 2 August 2018 / Accepted: 5 August 2018 / Published: 6 August 2018
(This article belongs to the Special Issue Visual Sensors)

Abstract

:
In this paper, a method of handshapes recognition based on skeletal data is described. A new feature vector is proposed. It encodes the relative differences between vectors associated with the pointing directions of the particular fingers and the palm normal. Different classifiers are tested on the demanding dataset, containing 48 handshapes performed 500 times by five users. Two different sensor configurations and significant variation in the hand rotation are considered. The late fusion at the decision level of individual models, as well as a comparative study carried out on a publicly available dataset, are also included.

1. Introduction

Handshapes are the basis of so-called finger alphabets that are used by deaf people to express words for which there are no separate signs in sign languages. The same handshapes, shown for various positions and orientations of the hand, are also important components of dynamic signs occurring in sign languages. Moreover, in the case of the so-called minimal pairs, the shape of the hand is the only distinguishing feature. Therefore, building a complete system for automatic recognition of the manual part of sign language is not possible without solving the problem of recognizing static handshapes.
The problem is challenging. Handshapes occurring in finger alphabets are complicated. A projection, that takes place during the image formation in a camera, results in significant loss of information. Individual fingers overlap each other or remain completely covered. In addition, some handshapes are very similar. Moreover, a movement trajectory is not available and therefore a detailed analysis of the shape is required. In the case of typical cameras, including stereo cameras, a big challenge is a dependence on variable backgrounds and lighting conditions. Individual differences in showing particular shapes by different users need to be considered as well. Therefore, systems developed in a controlled and sterile laboratory environment do not always work in demanding real-world conditions.
Currently, there are imaging devices on the market which operate both in the visible and near-infrared and provide accurate and reliable 3D information in the form of point clouds. These clouds can be further processed to extract skeletal information. An example of such a device is the popular Kinect controller, which, along with the included software, provides the skeletal data for the entire body of the observed person. There are similar solutions, with smaller observation area and higher resolution, for obtaining skeletal data for the observed hand. Examples of such devices are some time-of-flight cameras or a Leap Motion controller (LMC). These are in early stages of development but technological progress in this area is fast. For example, the first version of the Leap Motion Software Development Kit (SDK) was able to track only visible parts of the hand, but the version 2 uses some prediction algorithms, and the individual joints of each finger are tracked even when the controller cannot see them. It is expected that sooner or later, newer solutions will emerge. Therefore, it is reasonable to undertake research on the handshape recognition based on skeletal data.
Despite a number of works in this field, the problem remains unresolved. Current works are either dedicated to one device only or deal with a few simple static shapes or dynamic gestures, for which the great support is the distinctive role of the motion trajectory.
In this paper, a method of handshapes recognition, based on skeletal data is described. The proposed feature vector encodes the relative differences between vectors associated with the pointing directions of the fingers and the palm normal. Different classifiers are tested on the demanding dataset, containing 48 handshapes performed by five users. Each shape is repeated 500 times by each user. Two different sensor configurations and significant variation in the hand rotation are considered. The late fusion at the decision level of individual models, as well as comparative study carried out on a publicly available dataset, are also included.
The remainder of this paper is organized as follows. The recent works are characterized in Section 2, Section 3 describes the method, Section 4 discusses the experiment results, and Section 5 summarizes the paper. Appendix A contains the full versions of the tables with the results of leave-one-person-out cross-validation.

2. Recent Works

The suitability of the skeletal data, obtained from the LMC, for Australian Sign Language (Auslan) recognition has been explored in [1]. Testing showed that despite the problems with accurate tracking of fingers, especially when the hand is perpendicular to the controller, there is a potential for the use of the skeletal data, after some further improvement of the provided API.
An extensive evaluation of the quality of skeletal data, obtained from the LMC, was also tested in [2]. Static and dynamic measurements were performed using a high-precision motion tracking system. For static objects, the 3D position estimation with the standard deviation less than 0.5 mm was reported. A spatial dependency of the controller’s precision was also tested. In [1,2] the early version of the provided software was used. Recently, the stability of tracking has been significantly improved.
In [3], the skeletal data was used to recognize a subset of 10 letters from American Manual Alphabet. Handshapes were presented 10 times by 14 users. The feature vector was based on the positions and orientations of the fingers measured by the LMC. The multi-class support vector machine (SVM) classifier was used. The recognition accuracy was 80.86%. When the feature vector was augmented by features calculated from the depth map obtained with the Kinect sensor, the recognition accuracy increased to 91.28%.
In [4], the 26 letters of the English alphabet in American Sign Language (ASL) performed by two users were recognized using the features derived from the skeletal data. The recognition rate was 72.78% for the k-nearest neighbor (kNN) classifier and 79.83% for SVM.
Twenty-eight signs corresponding to the Arabic alphabet, performed 100 times by one person were recognized using 12 selected attributes of the hand skeletal data [5]. For the Naive Bayes (NB) classifier, the recognition rate was 98.3% and for the Multilayer Perceptron (MP) 99.1%.
In [6], the 50 dynamic gestures from Arabic Sign Language (ArSL), performed by two persons, were recognized using the feature vector composed of positions of fingers and distances between them and multi-layer perceptron neural network. The recognition accuracy was 88%.
A real-time multi-sensor system for ASL recognition was presented in [7]. The skeletal data, collected from Leap Motion sensors, was fused using multiple sensors data fusion and the classification was performed using hidden Markov models (HMM). The 10 gestures, corresponding to the digits from 0 to 9, were performed by eight subjects. The recognition accuracy was 93.14%.
In [8], the 24 letters from ASL were recognized using the feature vector that consists of the normal vector of the palm, coordinates of fingertips and finger bones, the arm direction vector, and the fingertip direction vector. These features were derived from the skeletal data provided by LMC. The decision tree (DT) and genetic algorithm (GA) were used as the classifier. The recognition accuracy was 82.71%.
Five simple handshapes were used to control a robotic wheelchair in [9]. Skeletal data was acquired by LMC. Feature vector consisted of the palm roll, pitch and yaw angles, and the palm normal direction vector. Block Sparse Representation (BSR) based classification was applied. According to the authors, the method yields accurate results but no detailed information about experiments and obtained recognition accuracy are given.
In [10], 10 handshapes corresponding to the digits in Indian Sign Language were recognized. The feature vector consisted of the distances between the consecutive fingertips and palm center and the distances between the fingertips. The features were derived from skeletal data acquired by LMC. Multi-Layer Perceptron (MP) neural network with back propagation algorithm was used. Each shape was presented by four subjects. The recognition accuracy of 100% is reported in the paper.
In [11], 28 letters of the Arabic Sign Language were recognized using the body and hand skeletal data acquired by Kinect sensor and LMC. One thousand four hundred samples were recorded by 20 subjects. One hundred and three features for each letter were reduced to 36 using the Principal Component Analysis algorithm. For the SVM classifier, the recognition accuracy of 86% is reported.
In [12], 25 dynamic gestures from Indian Sign Language were recognized using a multi-sensor fusion framework. Data was acquired using jointly calibrated Kinect sensor and LMC. Each word was repeated eight times by 10 subjects. Different data fusion schemes were tested and the best recognition accuracy of 90.80% was reported for the Coupled Hidden Markov Models (CHMM).
Twenty-eight handshapes corresponding to the letters of the Arabic alphabet were recognized using skeletal data from LMC and RGB image from Kinect sensor [13]. Gestures were performed at least two times by four users. Twenty-two of 28 letters were recognized with 100% accuracy.
In [14], Rule Based-Backpropagation Genetic Algorithm Neural Network (RB-BGANN) was used to recognize 26 handshapes corresponding to the alphabet in Sign System of Indonesian Language. Thirty-four features, related to the fingertips positions and orientations, taken from the hand skeletal data acquired by LMC, were used. Each gesture was performed five times. The recognition accuracy was 93.8%.
The skeletal data provided by the hand tracking devices LMC and Intel RealSense was used for recognizing 20 of the 26 letters from ASL [15]. The SVM classifier was used. The developed system was evaluated with over 50 individuals, and the recognition accuracy for particular letters was in the range of 60–100%.
In [16], a method to recognize static sign language gestures, corresponding to 26 American alphabet letters and 10 digits, performed by 10 users, was presented. The skeletal data acquired by LMC was used. Two variants of the feature vector were considered: (i) the distances between fingertips and the center of the palm, and (ii) the distances between the adjacent fingertips. The nearest neighbor classifier with four different similarity measures (Euclidean, Cosine, Jaccard, and Dice) was used. The obtained recognition accuracy varied from 70–100% for letters and 95–100% for digits.
Forty-four letters of Thai Sign Language were recognized using the skeletal data acquired by LMC and the decision trees [17]. The recognition accuracy of 72.83% was reported, but the authors do not indicate how many people performed gestures.
In [18], the skeletal data, acquired from two Leap Motion controllers, was used to recognize 28 letters from Arabic Sign Language. Handshapes were presented 10 times by one user. For the data fusion at features level and Linear Discriminant Analysis (LDA) classifier, the average accuracy was about 97.7%, while for classifier level fusion using Dempster-Shafer theory of evidence—97.1%.
Ten static gestures performed 20 times by 13 individuals were recognized using the new feature called Fingertips Tip Distance, derived from LMC skeletal data, and Histogram of Oriented Gradients (HOG), calculated from undistorted, raw sensor images [19]. After dimension reduction, based on Principal Component Analysis (PCA), and feature weighted fusion, the multiclass SVM classifier was used. Several variants of feature fusion were explored. The best recognition accuracy was 99.42%.
In [20], 28 isolated manual signs and 28 finger-spelling words, performed four times by 10 users, were recognized. The proposed feature vector consisted of fingertip positions and orientations derived from the skeletal data obtained with LMC. The SVM classifier was used to differentiate between manual and finger spelling sequences and the Bidirectional Long Short-Term Memory (BLSTM) recurrent neural networks were used for manual sign and fingerspelled letters recognition. The obtained recognition accuracy was 63.57%.
Eight handshapes, that can be used to make orders in a bar, were recognized in [21]. Each shape was presented three times by 20 participants. The feature vector consisted of normalized distances between the tips of the fingers and the center of the palm and was calculated from row skeletal data provided by LMC. Three classification methods: kNN, MP and Multinomial Logistic Regression (MLR) were considered. The best recognition accuracy of 95% was obtained for kNN classifier.
In [22], fingertip distances, fingertip inter-distances, and hand direction, derived from skeletal data acquired by LMC as well as the RGB-D data provided by Kinect sensor were used for sign language recognition in a multimodal system. Ten handshapes, performed 10 times by 14 users were recognized using data-level, feature-level, and decision-level multimodal fusion techniques. The best recognition accuracy of 97.00% was achieved for the proposed decision level fusion scheme.
The current works are summarized in Table 1.

3. Proposed Method

3.1. Hand Skeletal Data

The skeletal hand model considered in this paper is shown in Figure 1.
It consists of bones visualized in the form of straight line sections and connections between them (joints) depicted as numbered balls. There are four kinds of bones in this model: (i) four metacarpals (between joints P 5 P 6 , P 10 P 11 , P 15 P 16 , P 20 P 21 ), (ii) five proximal phalanges ( P 1 P 2 , P 6 P 7 , P 11 P 12 , P 16 P 17 , P 21 P 22 ), (iii) five intermediate phalanges ( P 2 P 3 , P 7 P 8 , P 12 P 13 , P 17 P 18 , P 22 P 23 ), and (iv) five distal phalanges ( P 3 P 4 , P 8 P 9 , P 13 P 14 , P 18 P 19 , P 23 P 24 ).
In a contactless way, such a model can be acquired directly using LMC released in 2012 [23] or Intel RealSense device released in 2015 [24] and embedded in some laptop models. A simplified version of the model, sufficient to determine the feature vector proposed in this paper, can be also obtained using Softkinetic DepthSense 325 camera along with the Close Interaction Library [25]. LMC has been recently evaluated [26], but there are no many publications about RealSense due to its recent release. It is expected that in the near future these devices and the supplied software will be further improved to allow for reliable skeletal hand tracking.

3.2. Feature Vector

The proposed feature vector encodes the relative differences between vectors associated with the pointing directions of the fingers and the palm normal. Let P c be the center of the palm, n c normal to the palm at point P c , P i the end of the i-th finger, and n i the vector pointed by that finger (Figure 2).
The relative position of vectors n c and n i can be unambiguously described giving four values determined from the Formulas (1)–(4) [27]:
α i = a c o s ( v i · n i )
ϕ i = a c o s u · d i | d i |
Θ i = a t a n w i · n i u · n i
d i = P i - P c
where the vectors u, v i , and w i define the so-called Darboux frame [28]:
u = n c
v i = d i | d i | × u
w i = u × v i
and · indicates the scalar and ×—vector products. Since the d i vectors depend on the size of the hand, they have been omitted. The feature vector consists of 15 values calculated for individual fingers using the Formulas (1)–(3):
V = [ α 1 , ϕ 1 , Θ 1 , α 2 , ϕ 2 , Θ 2 , α 3 , ϕ 3 , Θ 3 , α 4 , ϕ 4 , Θ 4 , α 5 , ϕ 5 , Θ 5 ]
In the case of LMC, the palm center, the palm normal and the pointing directions of the fingers are returned along with the skeletal data. For other devices, they can be derived from the skeletal data using the Formulas (9)–(11) (see Figure 1):
P c = 1 10 j J P j
n c = P 15 P 16 × P 5 P 6
n i = P 3 + 5 ( i - 1 ) P 4 + 5 ( i - 1 )
where J = { 1 , 2 , 5 , 6 , 10 , 11 , 15 , 16 , 20 , 21 } .

3.3. Classification

The following classification methods have been tested: decision trees (DT) [29], linear and quadratic discriminants (LD and QD) [30,31], support vector machines with linear, quadratic, cubic and Gaussian kernel function (SVM Lin/Quad/Cub/Gauss) [32,33,34], different version of k-nearest neighbor classifiers (1 NN, 10 NN, 100 NN, 10 NN Cos, 10 NN W) [35,36], different ensemble classifiers, that meld results from many weak learners into one model (Ens Boost/Bag/RUS/SubD/kNN) [37,38,39,40] and fast approximate nearest neighbors with randomized kd-trees (FLANN) [41]. A detailed list of tested classifiers with their initial parameters is provided in Table 2.

4. Experiments

4.1. Datasets

Two datasets were considered.

4.1.1. Dataset 1: Authors’ Own Dataset

Forty-eight static handshapes, occurring in Polish Finger Alphabet (PFA) and Polish Sign Language (PSL) were considered (Figure 3) [25].
The gestures were recorded in two configurations: (i) LMC lies horizontally on the table (configuration user-sensor); (ii) the sensor is attached to the monitor and directed towards the signer (configuration user-user). In the configuration (i), two variants were additionally considered: (a) gestures are made with fixed hand orientation (like in PFA); (b) spatial hand orientation changes in a wide range (like in PSL). In the configuration (i) variant (a) five people, designated hereinafter A, B, C, D, and E, participated in the recordings. In other cases, the gestures of person A were recorded. Gestures were shown by each person 500 times. During the data collection, visual feedback was provided, and when an abnormal or incomplete skeleton was observed, the process was repeated to ensure that 500 correct shapes were registered for each class. Incorrect data was observed for approximately 5% of frames. It was also noticed that the device works better when the whole hand with very visible fingers is presented first and then slowly changes to the desired shape.

4.1.2. Dataset 2: Microsoft Kinect and Leap Motion Dataset

In order to evaluate the method for more users and to make a comparative analysis, the database provided in the work [3] was used. The database contains the recordings of 10 letters from ASL, performed 10 times by 14 people and acquired by jointly calibrated LMC and depth sensor.

4.2. Results

The results of 10-fold cross-validation for the dataset 1 are shown in Table 3, Table 4 and Table 5.
For LMC lying on the table (configuration (i)) the best recognition rates (≥99.5%) were for SVM, kNN, Ens Bag, Ens Sub kNN and FLANN, wherein the results obtained under large variation in hand’s rotation (variant (b)) were only slightly worse. For configuration (ii), the results are better. This configuration seems to be more natural for a user accustomed to showing gestures to another person.
However, the results of the leave-one-subject-out cross-validation experiment, shown in Table 6 and Table A1, are much worse for all considered classification methods. The best recognition rates (≥50.0%) were for: LD, SVM Lin, Ens Bag.
The performances of the individual gestures are strongly dependent on the user, and the training set consisting of four people is not sufficiently representative to correctly classify the gestures of the fifth, unknown person.
The results obtained for dataset 2, and shown in Table 7, Table 8 and Table A2, confirm that when the training set consists of more users, the discrepancy between 10-fold cross-validation and leave-one-subject-out cross-validation is significantly lower. However, it should be mentioned that in this case, the number of recognized classes is much smaller.
For the dataset 2 and 10-fold CV, the best results (≥88.0%) were for SVM Gauss, kNN 1, kNN W, Ens Bag and Ens Sub kNN, whereas for leave-one-subject-out cross-validation the best results (≥88.0%) were for kNN1, kNN W, Ens Sub kNN.
Because for the most demanding case (Table 6) the best results were obtained for SVM Lin and Ens Bag—the parameters of these two classifiers were further analyzed (see Table 9 and Table 10).
The SVM classifier is by nature binary. It classifies instances into one of the two classes. However, it can be turned into a multinomial classifier by two different strategies: one-vs-one and one-vs-all. In one-vs-one, a single classifier for each pair of classes is trained. The decision is made by applying all trained classifiers to an unseen sample and a voting scheme. The class that has been recognized most times is selected. In one-vs-all, a single classifier per class is trained. The samples of that class are positive samples, and all other samples are negatives. The decision is made by applying all trained classifiers to an unseen sample and selecting the one with the highest confidence score.
In SVM Lin classifier, the change of the multiclass method from one-vs-one to one-vs-all leads to decrease in the recognition accuracy. For Ens Bag classifier, the recognition accuracy increases with the number of learners, but the response time increases as well (see Figure 4).
The experiment was stopped when the response time reached 100 ms, i.e., the value at which the typical user will notice the delay [42].
In Table 8 the best results were obtained for 1 NN, 10 NN W, and Ens Sub kNN. The FLANN version of the kNN classifier turned out to be the fastest one. Therefore, a further analysis of kNN classifiers has been carried out.
In Table 11, the nearest neighbor classifier 1 NN with brute-force search in the dataset was compared with the FLANN version with a different number of the randomized trees.
As should be expected, the results obtained for the exact version are slightly better than for the classifier, which finds the approximate nearest neighbor. However, if we compare the processing times, the FLANN version is over 400 times faster. Therefore, this classifier is a particularly attractive choice in practical applications.
An experiment was also carried out to check whether the late fusion of classifiers, at the decision level of individual models, leads to improved recognition accuracy. A simple method was used, in which every classifier votes for a given class. According to [43], simple unweighted majority voting is usually the best voting scheme. All possible combinations of classifiers were tested. The best result of leave-one-subject-out cross-validation on dataset 1, 56.7%, was obtained when the outputs of the classifiers LD, QD, SVM Lin, Ens Boost, Ens Bag were fused. The result is better than the best result obtained for a single classifier by 4.4%. However, the fusion of classifiers leads to a decrease in the individual classes recognition. The voting deteriorates the prediction in classes F, I, Xm, Yk.

4.3. Computational Efficiency

The average response times of the individual classifiers are shown in Table 12.
Together with the average time needed for data acquisition and feature vectors calculation, which is equal to 6 ms, they do not exceed 100 ms, so the typical user will not notice the time delay between presentation of the given gesture and the predicted response of the system [42]. However, all experiments were carried out on a fairly powerful workstation, equipped with a 2.71 GHz processor, 32 GB of RAM and a fast SSD. For less-efficient systems, e.g., mobile or embedded devices, the preferred choice is FLANN or DT. Moreover, in the case of FLANN classifier, the randomized trees can be searched in parallel.

4.4. Comparative Analysis

According to the authors’ knowledge, the only database of static hand skeletal data available on the Internet for which comparative analysis can be carried out is Dataset 2 [3]. Table 13 compares the recognition accuracy obtained for this database.
The first row quotes the results obtained for LMC, without additional data from the Kinect sensor. The proposed feature vector allows obtaining better results even with the same classifier (SVM).

5. Conclusions

Handshape recognition based on its skeleton becomes an important research problem because there are more and more new devices on the market that enable the acquisition of such data. In this paper:
  • A feature vector was proposed, which describes the relative differences between the pointing directions of individual fingers and the hand normal vector.
  • A demanding dataset containing 48 hand shapes, shown 500 times by five persons in two different sensor placement, has been prepared and made available [44].
  • The registered data has been used to perform classification. Seventeen known and popular classification methods have been tested.
  • For classifiers SVM Lin and Ens Bag, given the best recognition accuracies, an analysis of the impact of their parameters on the obtained results was carried out.
  • It was found that the weaker result for leave-one-person-out validation may be caused by individual character of performances of individual gestures, a difficult dataset, containing as many as 48 classes, among which there are very similar shapes, and imperfections of the LMC, which in the case of individual fingers occlusions tries predict their position and spatial orientation. It is worth mentioning that other works on static handshape recognition, cited in the literature, concern a smaller number of simpler gestures.
  • The proposed feature vector allows obtaining better results.
  • It was determined experimentally that although late fusion improves the results, it causes the deterioration of recognition efficiency in some classes, which in some applications may be undesirable.
To recognize complicated handshapes occurring in the sign languages, a feature vector invariant to translation, rotation, and scale, which is sensitive to the subtle differences in shape, is needed. The proposed feature vector is inspired by research on local point cloud descriptors [27]. Angular features, describing the mutual position of two vectors normal to the cloud surface, are used there to form a representation of the local geometry. Such a descriptor is sensitive to subtle differences in shape [45]. In our proposition, the fingertips and the palm center are treated as a point cloud, and the finger directions and the palm’s normal are used instead of the surface normals. It is also worth noting that the proposed feature vector is invariant to position, orientation, and scale. This is not always the case in the literature, where the features depending on the hand size or orientation are used. This invariance is particularly important in the case of sign language, where unlike in the finger alphabet, the hand’s position and orientation are not fixed. An interesting proposition of an invariant feature vector was proposed in [3] and enhanced in [19]. In Section 4.4, it was compared to our proposal.
Analysis of the confusion matrices obtained for the dataset 1 shows that the most commonly confused shapes are: B-Bm, C-100, N-Nw, S-F, T-O, Z-Xm, Tm-100, Bz-Cm and 4z-Cm. In fact, these are very similar shapes (see Figure 3). In adverse lighting conditions, when they are viewed from some distance or from the side, they can be confused even by a person. When sequences of handshapes, corresponding to fingerspelled words, are recognized, disambiguation can be achieved by using the temporal context. However, this is not always possible because often fingerspelling is used to convey difficult names, foreign words or proper names. If the similar shapes are discarded from the dataset 1, leave-one-subject-out cross-validation gives recognition efficiencies of about 80%.
The proposed system is fast and requires no special background or specific lighting. One of the reasons for the weaker results of leave-one-person-out validation is the imperfection of a sensor, that does not cope well with fingers occlusions. Therefore, as part of further work, the processing of point clouds registered with two calibrated sensors is considered in order to obtain more accurate and reliable skeletal data. Further work will also include recognition of letter sequences and integration of the presented solution with the sign language recognition system.

Author Contributions

T.K. conceived and designed the experiments; P.O. prepared the data; T.K. and P.O. performed the experiments and analyzed the data; T.K. wrote the paper.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Results of Leave-One-Person-Out Cross-Validation

Table A1. Leave-one-subject-out cross-validation results for dataset 1, configuration (i), variant (1).
Table A1. Leave-one-subject-out cross-validation results for dataset 1, configuration (i), variant (1).
TrainingTestingDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
B, C, D, EA37.451.332.146.141.038.84.043.143.3
A, C, D, EB31.237.629.633.032.931.05.632.331.7
A, B, D, EC47.954.741.657.857.555.421.860.661.2
A, B, C, ED41.653.851.158.654.051.718.752.351.3
A, B, C, DE48.957.158.166.160.157.419.854.555.5
Avg41.450.942.552.349.146.914.048.648.6
TrainingTesting100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
B, C, D, EA45.044.343.238.739.650.442.227.640.9
A, C, D, EB32.632.731.837.640.531.335.916.433.8
A, B, D, EC60.455.761.646.659.051.060.435.356.5
A, B, C, ED48.951.751.345.256.144.852.030.150.8
A, B, C, DE57.755.355.448.558.756.559.033.956.1
Avg48.947.948.743.350.846.849.928.747.6
Table A2. Leave-one-subject-out cross-validation results for the dataset 2.
Table A2. Leave-one-subject-out cross-validation results for the dataset 2.
TrainingTestingDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
2–14185.076.077.077.077.071.079.095.074.0
1, 3–14289.087.091.090.090.091.091.089.090.0
1–2, 4–14381.088.091.091.084.077.083.082.084.0
1–3, 5–14489.092.097.096.095.087.097.094.094.0
1–4, 6–14590.090.092.091.091.087.092.086.092.0
1–5, 7–14691.092.093.092.092.089.093.093.092.0
1–6, 8–14787.085.088.088.086.080.087.084.089.0
1–7, 9–14886.090.092.092.087.090.093.091.092.0
1–8, 10–14986.084.087.087.084.078.087.088.086.0
1–9, 11–141084.077.089.089.089.086.081.086.085.0
1–10, 12–141179.072.074.080.080.075.073.076.074.0
1–11, 13–141290.094.0100.0100.0100.096.0100.095.097.0
1–12, 141385.076.077.077.077.073.079.095.074.0
1–131485.076.077.077.077.072.079.095.074.0
Avg86.284.287.587.686.482.386.789.285.5
TrainingTesting100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
2–14173.076.095.077.080.076.095.074.095.0
1, 3–14289.090.091.091.089.087.090.089.089.0
1–2, 4–14388.084.082.084.083.088.082.090.082.0
1–3, 5–14489.094.095.097.097.092.094.092.094.0
1–4, 6–14590.091.089.092.092.091.086.092.080.0
1–5, 7–14688.092.093.093.093.092.093.093.087.0
1–6, 8–14782.089.087.088.086.088.083.088.078.0
1–7, 9–14888.091.090.092.090.091.091.091.091.0
1–8, 10–14986.085.086.087.088.084.088.087.088.0
1–9, 11–141075.084.084.085.089.076.086.089.086.0
1-10, 12-141174.074.078.078.079.072.075.079.069.0
1–11, 13–141285.096.094.0100.0100.089.097.098.095.0
1–12, 141373.076.095.077.082.076.095.074.070.0
1–131473.076.095.077.080.076.095.074.091.0
Avg82.485.689.687.087.784.189.386.485.4

References

  1. Potter, L.E.; Araullo, J.; Carter, L. The Leap Motion Controller: A View on Sign Language. In Proceedings of the 25th Australian Computer-Human Interaction Conference: Augmentation, Application, Innovation, Collaboration, Adelaide, Australia, 25–29 November 2013; ACM: New York, NY, USA, 2013; pp. 175–178. [Google Scholar]
  2. Guna, J.; Jakus, G.; Pogačnik, M.; Tomažič, S.; Sodnik, J. An Analysis of the Precision and Reliability of the Leap Motion Sensor and Its Suitability for Static and Dynamic Tracking. Sensors 2014, 14, 3702–3720. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Marin, G.; Dominio, F.; Zanuttigh, P. Hand gesture recognition with leap motion and kinect devices. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 1565–1569. [Google Scholar]
  4. Chuan, C.H.; Regina, E.; Guardino, C. American Sign Language Recognition Using Leap Motion Sensor. In Proceedings of the 2014 13th International Conference on Machine Learning and Applications, Detroit, MI, USA, 3–6 December 2014; pp. 541–544. [Google Scholar]
  5. Mohandes, M.; Aliyu, S.; Deriche, M. Arabic sign language recognition using the leap motion controller. In Proceedings of the 2014 IEEE 23rd International Symposium on Industrial Electronics (ISIE), Istanbul, Turkey, 1–4 June 2014; pp. 960–965. [Google Scholar]
  6. Elons, A.S.; Ahmed, M.; Shedid, H.; Tolba, M.F. Arabic sign language recognition using leap motion sensor. In Proceedings of the 2014 9th International Conference on Computer Engineering Systems (ICCES), Cairo, Egypt, 22–23 December 2014; pp. 368–373. [Google Scholar]
  7. Fok, K.Y.; Ganganath, N.; Cheng, C.T.; Tse, C.K. A Real-Time ASL Recognition System Using Leap Motion Sensors. In Proceedings of the 2015 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery, Xi’an, China, 17–19 September 2015; pp. 411–414. [Google Scholar]
  8. Funasaka, M.; Ishikawa, Y.; Takata, M.; Joe, K. Sign Language Recognition using Leap Motion Controller. In Proceedings of the 2015 International Conference on Parallel and Distributed Processing Techniques and Applications (PDPTA’15), Las Vegas, NV, USA, 27–30 July 2015; pp. 263–269. [Google Scholar]
  9. Boyali, A.; Hashimoto, N.; Matsumato, O. Hand Posture Control of a Robotic Wheelchair Using a Leap Motion Sensor and Block Sparse Representation based Classification. In Proceedings of the Third International Conference on Smart Systems, Devices and Technologies, Paris, France, 20–24 July 2014; pp. 20–25. [Google Scholar]
  10. Naglot, D.; Kulkarni, M. ANN based Indian Sign Language numerals recognition using the leap motion controller. In Proceedings of the 2016 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, 26–27 August 2016; Volume 2, pp. 1–6. [Google Scholar]
  11. Almasre, M.A.; Al-Nuaim, H. Recognizing Arabic Sign Language gestures using depth sensors and a KSVM classifier. In Proceedings of the 2016 8th Computer Science and Electronic Engineering (CEEC), Colchester, UK, 28–30 September 2016; pp. 146–151. [Google Scholar]
  12. Kumar, P.; Gauba, H.; Roy, P.P.; Dogra, D.P. Coupled HMM-based multi-sensor data fusion for sign language recognition. Pattern Recognit. Lett. 2017, 86, 1–8. [Google Scholar] [CrossRef]
  13. Miada, A.; Almasre, H.A.N. A Real-Time Letter Recognition Model for Arabic Sign Language Using Kinect and Leap Motion Controller v2. Int. J. Adv. Eng. Manag. Sci. 2016, 2, 514–523. [Google Scholar]
  14. Nurul Khotimah, W.; Andika Saputra, R.; Suciati, N.; Rahman Hariadi, R. Alphabet Sign Language Recognition Using Leap Motion Technology and Rule Based Backpropagation-Genetic Algorithm Nueral Network (RBBPGANN). JUTI J. Ilm. Teknol. Inf. 2017, 15, 95–103. [Google Scholar]
  15. Quesada, L.; López, G.; Guerrero, L. Automatic recognition of the American sign language fingerspelling alphabet to assist people living with speech or hearing impairments. J. Ambient Intell. Humaniz. Comput. 2017, 8, 625–635. [Google Scholar] [CrossRef]
  16. Auti, A.; Amolic, R.; Bharne, S.; Raina, A.; Gaikwad, D.P. Sign-Talk: Hand Gesture Recognition System. Int. J. Comput. Appl. 2017, 160, 13–16. [Google Scholar] [CrossRef]
  17. Tumsri, J.; Kimpan, W. Thai Sign Language Translation Using Leap Motion Controller. In Proceedings of the International Multi Conference of Engineers and Computer Scientists 2017, Hong Kong, China, 15–17 March 2017; Volume I, pp. 46–51. [Google Scholar]
  18. Mohandes, M.; Aliyu, S.; Deriche, M. Prototype Arabic Sign language recognition using multi-sensor data fusion of two leap motion controllers. In Proceedings of the 2015 IEEE 12th International Multi-Conference on Systems, Signals Devices (SSD15), Mahdia, Tunisia, 16–19 March 2015; pp. 1–6. [Google Scholar]
  19. Du, Y.; Liu, S.; Feng, L.; Chen, M.; Wu, J. Hand Gesture Recognition with Leap Motion. CoRR. 2017. Available online: http://xxx.lanl.gov/abs/1711.04293 (accessed on 9 July 2018).
  20. Kumar, P.; Saini, R.; Behera, S.K.; Dogra, D.P.; Roy, P.P. Real-time recognition of sign language gestures and air-writing using leap motion. In Proceedings of the 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, Japan, 8–12 May 2017; pp. 157–160. [Google Scholar]
  21. Toghiani-Rizi, B.; Lind, C.; Svensson, M.; Windmark, M. Static Gesture Recognition using Leap Motion. arXiv, 2017; arXiv:1705.05884. [Google Scholar]
  22. Ferreira, P.M.; Cardoso, J.S.; Rebelo, A. Multimodal Learning for Sign Language Recognition. In Pattern Recognition and Image Analysis; Alexandre, L.A., Salvador Sánchez, J., Rodrigues, J.M.F., Eds.; Springer: Cham, Switzerland, 2017; pp. 313–321. [Google Scholar]
  23. Leap Motion. Available online: https://www.leapmotion.com/ (accessed on 9 March 2018).
  24. Intel RealSense Technology. Available online: https://www.intel.com/content/www/us/en/architecture-and-technology/realsense-overview.html (accessed on 9 March 2018).
  25. Close Interaction LIbrary. Available online: https://www.sony-depthsensing.com/products/CIlib (accessed on 9 March 2018).
  26. Weichert, F.; Bachmann, D.; Rudak, B.; Fisseler, D. Analysis of the Accuracy and Robustness of the Leap Motion Controller. Sensors 2013, 13, 6380–6393. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Rusu, R.B.; Marton, Z.C.; Blodow, N.; Beetz, M. Learning informative point classes for the acquisition of object model maps. In Proceedings of the 2008 10th International Conference on Control, Automation, Robotics and Vision, Hanoi, Vietnam, 17–20 December 2008; pp. 643–650. [Google Scholar]
  28. Spivak, M. A Comprehensive Introduction to Differential Geometry, 3rd ed.; Publish or Perish: Houston, TX, USA, 1999; Volume III. [Google Scholar]
  29. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  30. Rayens, W.S. Discriminant Analysis and Statistical Pattern Recognition. Technometrics 1993, 35, 324–326. [Google Scholar] [CrossRef]
  31. Rencher, A. Methods of Multivariate Analysis; Wiley Series in Probability and Statistics; Wiley: Hoboken, NJ, USA, 2003. [Google Scholar]
  32. Press, W.H.; Teukolsky, S.A.; Vetterling, W.T.; Flannery, B.P. Numerical Recipes 3rd Edition: The Art of Scientific Computing, 3rd ed.; Cambridge University Press: New York, NY, USA, 2007. [Google Scholar]
  33. Boser, B.E.; Guyon, I.M.; Vapnik, V.N. A Training Algorithm for Optimal Margin Classifiers. In Proceedings of the Fifth Annual Workshop on Computational Learning Theory, Pittsburgh, PA, USA, 27–29 July 1992; ACM: New York, NY, USA, 1992; pp. 144–152. [Google Scholar]
  34. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef] [Green Version]
  35. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef] [Green Version]
  36. Dudani, S.A. The Distance-Weighted k-Nearest-Neighbor Rule. IEEE Trans. Syst. Man Cybern. 1976, SMC-6, 325–327. [Google Scholar] [CrossRef]
  37. Freund, Y.; Schapire, R.E. A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  38. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  39. Seiffert, C.; Khoshgoftaar, T.M.; Hulse, J.V.; Napolitano, A. RUSBoost: Improving classification performance when training data is skewed. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  40. Ho, T.K. The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 832–844. [Google Scholar]
  41. Muja, M.; Lowe, D.G. Scalable Nearest Neighbor Algorithms for High Dimensional Data. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2227–2240. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Card, S.K. The Model Human Processor: A Model for Making Engineering Calculations of Human Performance. Proc. Hum. Factors Soc. Annu. Meet. 1981, 25, 301–305. [Google Scholar] [CrossRef]
  43. Moreno-Seco, F.; Iñesta, J.M.; de León, P.J.P.; Micó, L. Comparison of Classifier Fusion Methods for Classification in Pattern Recognition Tasks. In Structural, Syntactic, and Statistical Pattern Recognition; Yeung, D.Y., Kwok, J.T., Fred, A., Roli, F., de Ridder, D., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 705–713. [Google Scholar]
  44. Dataset. Available online: vision.kia.prz.edu.pl (accessed on 9 July 2018).
  45. Rusu, R.B.; Bradski, G.; Thibaux, R.; Hsu, J. Fast 3D recognition and pose using the Viewpoint Feature Histogram. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 2155–2162. [Google Scholar]
Figure 1. Hand skeletal model.
Figure 1. Hand skeletal model.
Sensors 18 02577 g001
Figure 2. Feature vector construction.
Figure 2. Feature vector construction.
Sensors 18 02577 g002
Figure 3. Static handshapes, occurring in Polish Finger Alphabet and Polish Sign Language.
Figure 3. Static handshapes, occurring in Polish Finger Alphabet and Polish Sign Language.
Sensors 18 02577 g003
Figure 4. Response time for Ens Bag classifier for different number of learners.
Figure 4. Response time for Ens Bag classifier for different number of learners.
Sensors 18 02577 g004
Table 1. Recent works on handshape recognition using skeletal data.
Table 1. Recent works on handshape recognition using skeletal data.
WorkSignSignUsersDeviceMethodAccuracyData
TypeVocabulary[%]Available
[3]static10 letters ASL14LMCSVM80.86Yes
LMC + Kinect91.28
[4]static26 letters ASL2LMCkNN72.78No
SVM79.83
[5]static28 letters ArSL1LMCNB98.3No
MP99.1
[6]dynamic50 sign ArSL2LMCMP88No
[7]static10 digits ASL82 LMsHMM93.14No
[8]static24 letters ASL1LMCDT + GA82.71No
[9]static5 simple shapes?LMCBSR?No
[10]static10 digits ISL4LMCMP100No
[11]static28 letters ArSL20LMC + KinectSVM86No
[12]dynamic25 sign ISL10LMC + KinectCHMM90.80Yes
[13]static28 letters ArSL4LMC + KinectkNN100No
[14]static26 letters SIBI1LMCRB-BGANN + GA93.8No
[15]static20 letters ASL50LMC, RealSenseSVM60–100No
[16]static26 letters ASL10LMCkNN70–100No
10 digits ASL95–100
[17]static44 letters ThSL?LMCDT72.83No
[18]static28 letters ArSL12 LMsLDA97.7No
[19]static10 hand shapes13LMCSVM99.42No
[20]dynamic28 signs and10LMCSVM + BLSTM63.57No
28 words ISL
[21]static8 hand shapes20LMCkNN95Yes
[22]static10 hand shapes14LMC + Kinect 97No
Table 2. Tested classifiers and their parameters.
Table 2. Tested classifiers and their parameters.
ClassifierParameterValue
DTMaximum number of splits100
Split criterionGini’s diversity index
LDCovariance structureFull
QDCovariance structureFull
SVM Lin/Quad/Cub/GaussKernel functionLinear/Quadratic/Cubic/Gaussian
Box constraint level1
Multiclass methodOne-vs-one
1 NN/10 NN/100 NNNumber of neighbors1 /10/100
Distance metricEuclidean
10 NN CosNumber of neighbors10
Distance metricCosine
10 NN WNumber of neighbors10
Distance metricEuclidean
Distance weightSquared inverse
Ens Boost/Bag/RUSEnsemble methodBoosted/bagged/random subspace trees
Learner typeDecision tree
Number of learners30
Ens Sub D/Sub kNNEnsemble methodSubspace
Learner typeDiscriminant/1 NN
Number of learners30
Subspace dimension8
FLANNNumber of neighbors1
Number of trees8
Number of times the trees128
should be recursively traversed
Table 3. 10-fold cross-validation results for dataset 1, configuration (i), variant (a).
Table 3. 10-fold cross-validation results for dataset 1, configuration (i), variant (a).
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]81.272.899.799.5100.0100.0100.0100.099.9
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]99.299.9100.064.2100.069.9100.039.9100.0
Table 4. 10-fold cross-validation results for dataset 1, configuration (i), variant (b).
Table 4. 10-fold cross-validation results for dataset 1, configuration (i), variant (b).
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]83.178.797.296.799.499.799.199.898.5
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]88.998.599.567.199.777.399.835.899.7
Table 5. 10-fold cross-validation results for dataset 1, configuration (ii).
Table 5. 10-fold cross-validation results for dataset 1, configuration (ii).
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]100.0100.0100.0100.0100.0100.0100.0100.0100.0
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]100.0100.0100.0100.0100.0100.0100.043.8100.0
Table 6. Leave-one-subject-out cross-validation results for dataset 1, configuration (i), variant (1).
Table 6. Leave-one-subject-out cross-validation results for dataset 1, configuration (i), variant (1).
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]41.450.942.552.349.146.914.048.648.6
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]48.947.948.743.350.846.849.928.747.6
Table 7. 10-fold cross-validation results for dataset 2.
Table 7. 10-fold cross-validation results for dataset 2.
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]87.684.586.487.686.284.188.488.688.0
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]82.687.989.187.888.985.188.586.985.9
Table 8. Leave-one-subject-out cross-validation results for the dataset 2.
Table 8. Leave-one-subject-out cross-validation results for the dataset 2.
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Recognition rate [%]86.284.287.587.686.482.386.789.285.5
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Recognition rate [%]82.485.689.687.087.784.189.386.485.4
Table 9. Support Vector Machines classifier with linear kernel function performance when the multiclass method was changed from one-vs-one to one-vs-all.
Table 9. Support Vector Machines classifier with linear kernel function performance when the multiclass method was changed from one-vs-one to one-vs-all.
TrainingTestingSVM Lin
B, C, D, EA36.8
A, C, D, EB21.1
A, B, D, EC47.8
A, B, C, ED52.0
A, B, C, DE46.6
Avg40.8
Table 10. Ens Bag performance for a different number of learners.
Table 10. Ens Bag performance for a different number of learners.
TrainingTesting102030405010020040080010002000
B, C, D, EA44.647.439.646.845.346.846.946.145.445.946.3
A, C, D, EB40.140.340.538.038.637.038.839.240.041.438.7
A, B, D, EC53.256.659.056.155.853.958.658.357.457.758.0
A, B, C, ED54.854.656.154.858.557.957.057.457.958.157.9
A, B, C, DE58.760.058.760.658.663.460.561.262.563.061.7
Avg50.451.850.851.351.351.852.452.452.653.252.5
Table 11. 1 NN vs. FLANN with a different number of trees (given in parenthesis).
Table 11. 1 NN vs. FLANN with a different number of trees (given in parenthesis).
TrainingTesting1 NNFLANN (1)FLANN (2)FLANN (4)FLANN (8)FLANN (16)FLANN (32)
B, C, D, EA43.138.041.239.540.441.140.4
A, C, D, EB32.331.130.933.633.633.033.6
A, B, D, EC60.655.856.557.456.157.056.4
A, B, C, ED52.352.151.351.351.151.551.2
A, B, C, DE54.555.156.656.555.755.555.6
Avg48.646.447.347.647.447.647.4
Table 12. Average response times of the individual classifiers.
Table 12. Average response times of the individual classifiers.
ClassifierDTLDQDSVMSVMSVMSVM1 NN10 NN
LinQuadCubGauss
Response time [ms]0.070.460.4226.7331.9830.1264.8724.1526.64
Classifier100 NN10 NN10 NNEnsEnsEnsEnsEnsFLANN
CosWBoostBagSub DSub kNNRUS
Response time [ms]29.2422.5023.143.222.959.347.654.140.06
Table 13. 10-fold cross validation results of different methods obtained for the Dataset 2.
Table 13. 10-fold cross validation results of different methods obtained for the Dataset 2.
LpReferenceFeaturesMethodRecognition Rate
1[3]Fingertips distances, angles and elevationsMulticlass SVM80.9%
2[19]Fingertips Tip distanceMulticlass SVM81.1%
3This paperAs described in Section 3.2SVM Lin87.6%
4This paperAs described in Section 3.210NN W89.1%

Share and Cite

MDPI and ACS Style

Kapuscinski, T.; Organisciak, P. Handshape Recognition Using Skeletal Data. Sensors 2018, 18, 2577. https://doi.org/10.3390/s18082577

AMA Style

Kapuscinski T, Organisciak P. Handshape Recognition Using Skeletal Data. Sensors. 2018; 18(8):2577. https://doi.org/10.3390/s18082577

Chicago/Turabian Style

Kapuscinski, Tomasz, and Patryk Organisciak. 2018. "Handshape Recognition Using Skeletal Data" Sensors 18, no. 8: 2577. https://doi.org/10.3390/s18082577

APA Style

Kapuscinski, T., & Organisciak, P. (2018). Handshape Recognition Using Skeletal Data. Sensors, 18(8), 2577. https://doi.org/10.3390/s18082577

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop