Next Article in Journal / Special Issue
Robust and Accurate Vision-Based Pose Estimation Algorithm Based on Four Coplanar Feature Points
Previous Article in Journal
A Cooperative Traffic Control of Vehicle–Intersection (CTCVI) for the Reduction of Traffic Delays and Fuel Consumption
Previous Article in Special Issue
Detecting Target Objects by Natural Language Instructions Using an RGB-D Camera
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
Beijing Key Laboratory of Computational Intelligence and Intelligent System, Beijing 100124, China
3
Department of Computing Science, University of Alberta, Edmonton, AB T6G2E8, Canada
*
Author to whom correspondence should be addressed.
Sensors 2016, 16(12), 2171; https://doi.org/10.3390/s16122171
Submission received: 16 October 2016 / Revised: 1 December 2016 / Accepted: 7 December 2016 / Published: 17 December 2016
(This article belongs to the Special Issue Video Analysis and Tracking Using State-of-the-Art Sensors)

Abstract

:
Noise and constant empirical motion constraints affect the extraction of distinctive spatiotemporal features from one or a few samples per gesture class. To tackle these problems, an adaptive local spatiotemporal feature (ALSTF) using fused RGB-D data is proposed. First, motion regions of interest (MRoIs) are adaptively extracted using grayscale and depth velocity variance information to greatly reduce the impact of noise. Then, corners are used as keypoints if their depth, and velocities of grayscale and of depth meet several adaptive local constraints in each MRoI. With further filtering of noise, an accurate and sufficient number of keypoints is obtained within the desired moving body parts (MBPs). Finally, four kinds of multiple descriptors are calculated and combined in extended gradient and motion spaces to represent the appearance and motion features of gestures. The experimental results on the ChaLearn gesture, CAD-60 and MSRDailyActivity3D datasets demonstrate that the proposed feature achieves higher performance compared with published state-of-the-art approaches under the one-shot learning setting and comparable accuracy under the leave-one-out cross validation.

1. Introduction

Vision-based gesture recognition is a critical interface for non-intrusive human-robot interaction (HRI) systems, for which many natural and convenient recognition methods have been proposed [1,2,3]. In a traditional HRI system, a user usually needs to memorize a predefined gesture language or a set of instructions before interaction. Although users can learn new gesture motions, they may not enjoy learning the predefined gesture language or instructions. In particular, when a user plans to define and use arbitrary interaction gestures, existing HRI systems require labour-intensive and time-consuming procedures to collect and label all the training samples. This is because most of the traditional methods based on supervised learning require many training samples and long training times. All of these factors render a typical HRI system unsuitable for adding user-defined gestures easily and hence, motivate the current research of one-shot learning gesture recognition in solving the above mentioned problems.
One-shot learning is a concept cognition and learning method that models after a human’s innate learning ability. For example, humans can learn and generalize a new concept from one or very few samples [4]. Unlike tradition methods, one-shot learning gesture recognition is a small sample size learning problem, which means that every gesture has only one or very few training samples [5,6,7,8,9]. When learning a new gesture, the user needs to perform the gesture only once or a few times without the need to collect a large number of training samples. Meanwhile, simple classification methods, which do not need a lot of time for offline learning, can satisfy well the identification requirements. One-shot learning gesture recognition not only greatly improves the ease-of-use of HRI, but also makes the robot learn and recognize interaction gestures with the cognition mechanism conforming to the expectation of a human user.
One important challenge of one-shot learning gesture recognition is to accurately extract distinctive features from a small number of training samples. These features should represent well the between-class differences and within-class similarities of gestures [1,5]. Some recently published approaches use the color or depth information to detect spatiotemporal interest points (STIPs), i.e., keypoints, from which to build the feature descriptors. The spatiotemporal feature approaches (keypoints + feature descriptors) do not need any preprocessing steps, such as human detection, segmentation or skeleton extraction. Hence, they are highly flexible for applying to different environments. In fact, a user can perform gestures in 3D space. If the features are extracted from a single channel (color or depth), the loss of information in the other channel will greatly reduce the representative and discriminative ability of gesture features, so the study of RGB-D data-based spatiotemporal feature approaches has received much attention from researchers.
Hernandez-Vela et al. [10,11] detected keypoints using the Harris 3D detector [12,13] on RGB and depth volumes. Then, Histogram of Oriented Gradients (HOG) [14], Histogram of Optical Flow (HOF) [15], HOG/HOF [15] and Viewpoint Feature Histogram Camera Roll Histogram (VFHCRH) feature descriptors are calculated for each keypoint. The 3D motion scale invariant feature transform (3D MoSIFT) [16], 3D enhanced MoSIFT (3D EMoSIFT) [6] and 3D sparse MoSIFT (3D SMoSIFT) [1] spatiotemporal feature approaches, which are extensions of the MoSIFT method [17] using RGB-D data, can be referred to as 3D MoSIFT-based methods. All the 3D MoSIFT-based methods detect the initial interest points in whole frames. Then, these interest points of 3D MoSIFT whose grayscale velocities satisfy certain motion constraints are treated as keypoints. 3D EMoSIFT increases the depth-dependent constraints on the basis of 3D MoSIFT. 3D SMoSIFT extracts keypoints from the initial interest points by applying the grayscale or depth absolute velocity constraints, which are constant and determined by trial-and-error. Finally, the 3D MoSIFT-based methods use fused RGB-D data to construct SIFT-like feature descriptors. The Mixed Features Around Sparse Keypoints (MFSK) method [9] uses a strategy similar to that of 3D SMoSIFT to detect keypoints. The difference between them is that the MFSK method [9] detects Speeded Up Robust Features (SURF) corners [18] as initial interest points while 3D SMoSIFT uses Shi-Tomasi corners [19]. Then 3D SMoSIFT, HOG, HOF and motion boundary histogram (MBH) feature descriptors [20] are calculated from the RGB-D data. These one-shot learning gesture recognition approaches have achieved good performance using the 2011/2012 ChaLearn One-shot-learning Gesture Dataset (CGD) [21].
Besides the motion caused by the moving body parts (MBPs), motion may also exist in the background and the remaining body parts (non-MBPs). These motions may be caused by illumination changes, moving shadows, inaccurate or lost depth values (black block in a depth frame) or clothes movements (moving together with the MBPs), all of which can be collectively referred to as noise. The above mentioned spatiotemporal feature approaches directly detect keypoints in the whole frame, according to some predefined empirical motion constraints, which can incorrectly label some noise points with larger motion as keypoints. Because the speeds of different MBPs are not the same, constant motion constraints cannot guarantee correct detection of accurate and a sufficient number of keypoints in every MBP. If the constraint thresholds were set too high, features with relatively small movements would be ignored. On the contrary, if the constraint thresholds were set too low, many noise points would be detected. Different users or even the same user performing the same gesture at different times may exhibit different speeds and also the actual hardware execution environment of HRI varies. Hence, constant empirical motion constraints will not work in all situations.
To address the above problems, an adaptive local spatiotemporal feature (ALSTF) which combines both the appearance and the motion information is proposed. First, variances of grayscale and depth optical flows are used to adaptively extract motion regions of interest (MRoIs) which mainly contain MBPs, and to remove regions of the background and of the remaining body parts which contain a lot of noise. Then, Harris-Affine corners [22] are used as the initial interest points in MRoIs. The initial interest points with depth values and velocities of grayscale and depth that satisfy new local adaptively determined motion constraints in each MRoI, are regarded as the keypoints. Finally, multi-feature descriptors, computed in the new extended gradient and motion spaces, are combined to represent gestures. The major contributions of our new method include:
  • A new adaptive MRoIs extraction approach is proposed to reduce the effect of noise on the accuracy of extracting spatiotemporal features.
  • In each MRoI, new local depth and motion constraints are adaptively determined to detect keypoints. By this, not only the influence of noise can be reduced, but also the accuracy of detecting keypoints can be improved and more features can be extracted in the MBPs with large speed differences.
  • 3D SMoSIFT, HOG, HOF and MBH feature descriptors are calculated in the new extended gradient space and motion space, and employed to generate more RGB–D appearance and motion features of gestures.
  • Compared with other spatiotemporal feature approaches and with the published one-shot learning gesture recognition approaches, the proposed method achieves a better recognition performance.
The rest of this paper is organized as follows: Section 2 briefly reviews related works on spatiotemporal feature approaches. Section 3, Section 4 and Section 5 describe details of the proposed method. The adaptive MRoIs extraction approach is presented in Section 3.1; Section 3.2 shows how to adaptively detect keypoints in each MRoI. The improved feature descriptor calculation process is illustrated in Section 4. Section 5 introduces a resolution strategy for a special case wherein the human body is maintained in a relatively static condition during one stage of a gesture. Then the experimental results are shown in Section 6, where the evaluations and comparisons with state-of-the-art algorithms are discussed. Finally, Section 7 concludes this paper and includes some discussions on future work.

2. Related Works

Spatiotemporal feature approaches belong to the single-layered human gesture recognition approach. They consider a gesture as a set of appropriate features extracted from a 3D space-time volume and recognize the gesture from an unknown video sequence by categorizing it into one of the known classes [23]. In the following, some spatiotemporal features used in state-of-the-art techniques on gesture recognition and one-shot learning gesture recognition tasks are described.
Laptev et al. [12,13] proposed the Spatiotemporal Interest Point (STIP) detector, i.e., Harris 3D corner detector, to detect interest points in the spatiotemporal domain. The Harris 3D detector is a spatiotemporal extension of the Harris corner detector [24]. First, a spatiotemporal second-moment matrix μ and a normalized Laplace operator n o r m 2 L of each grayscale pixel are computed. The grayscale points with local positive maxima of the corner function H = det ( μ )     k trace 3 ( μ ) and extrema of the operator n o r m 2 L are regarded as STIPs. Then, local, spatiotemporal and scale-invariant N-jet descriptors are used to represent and classify events of interest.
The Cuboid [25] feature uses a new response function for the Harris 3D detector to detect STIPs. Its response function is composed of a 2D Gaussian smoothing kernel for the spatial dimensions and a quadrature pair of 1D Gabor filters for the temporal dimension. The locations of interest points are given by the local maxima of the response function. Once detected, the cuboid around each interest point, which contains the pixel appearance values of the interest point’s neighborhoods, is extracted [23]. By testing various transformations to be applied to cuboids to extract the final local features, the flattened vector of brightness gradients with the best performance is chosen as the descriptor, which is essentially a generalization of the Principal Components Analysis-Scale Invariant Feature Transform (PCA-SIFT) descriptor [23,25].
Lu et al. [26] employed spatiotemporal filtering and noise suppression to construct the response function in the spatiotemporal and scale domain from depth videos. Depth STIPs (DSTIPs) are selected at the local maxima of the response function. The Depth Cuboid similarity feature (DCSF) descriptor is used to encode the spatiotemporal appearances of the 3D cuboid around the DSTIP, based on self-similarity.
The Hessian detector [27] is an extension of the Hessian saliency measure [28,29] for blob detection in the spatiotemporal domain. The localization and scale selection of interest points are directly implemented and selected, without iteration, by measuring saliency with the determinant of the 3D Hessian matrix. To describe interest points, an extended version of the SURF descriptor is proposed. Descriptors are calculated in both spatial and temporal domains simultaneously. For rotation invariance, the dominant orientation is used in the spatiotemporal domain. Meanwhile, all the Haar-wavelets are extended over the full length of the temporal scale of the interest point [27].
Wang et al. [30] proposed representing human actions using dense trajectories and motion boundary descriptors. First, feature points are densely sampled on a grid space and the Shi-Tomasi detector [19] is used to remove points in the image areas that have no structure. Then, each feature point is tracked over L frames to form a trajectory. Finally, a trajectory shape descriptor is employed to encode local motion patterns of actions. Besides, a concatenated feature descriptor of HOG, HOF and MBH is used to represent the appearance and motion information.
Chen et al. [17] proposed the MoSIFT method, based on RGB and optical flow information. Since it is derived from SIFT [31], it is scale and rotation invariant. First, the Gaussian pyramids and the Difference of Gaussian (DoG) pyramid are constructed for two consecutive grayscale frames. The local extrema of the DoG pyramid are selected as the initial interest points. Then, the optical flow pyramids corresponding to the grayscale pyramids are calculated, and for each interest point to be a keypoint, its grayscale optical flow must satisfy certain empirical motion constraint thresholds. Finally, the MoSIFT feature descriptors of the keypoints are computed using the grayscale Gaussian pyramid and the optical flow pyramids.
Hernandez-Vela et al. [10,11] used the Harris 3D detector to detect keypoints S R G B in the RGB volumes and keypoints S D in the depth volumes. Then, the HOG, HOF and HOG/HOF feature descriptors for S R G B and the VFHCRH descriptors for S D are calculated to represent gestures. This approach is an extension of the Harris 3D detector using RGB and depth data.
3D MoSIFT [16], 3D EMoSIFT [6] and 3D SMoSIFT [1] are derived from MoSIFT using RGB-D data. 3D MoSIFT and 3D EMoSIFT adopt a similar strategy to detect initial interest points. 3D SMoSIFT just detects the Shi-Tomasi corners in grayscale scale space as initial interest points to speed up processing. To extract keypoints from the initial interest points, different 3D MoSIFT-based methods have their own individual strategies. In particular, 3D MoSIFT takes the grayscale motion constraints which are the same as that used in MoSIFT. On the basis of MoSIFT and 3D MoSIFT methods, 3D EMoSIFT includes a depth-dependent constraint to remove some noise points. 3D SMoSIFT adopts grayscale or depth motion constraints to extract keypoints. The difference between 3D MoSIFT and 3D SMoSIFT is that 3D SMoSIFT uses the magnitude of the velocity while 3D MoSIFT uses the magnitudes of the horizontal and vertical velocity components. After keypoints are detected, 3D MoSIFT-based methods construct a 3D gradient space and a 3D motion space to calculate SIFT-like feature descriptors in these spaces. They adopt the same 3D gradient space which are constructed using the grayscale and depth spatial information. The temporal variations of the grayscale optical flow and of the depth information are employed for constructing a 3D motion space for 3D MoSIFT and 3D EMoSIFT. 3D SMoSIFT improves the 3D motion space by simultaneously using the grayscale and the depth optical flow simultaneously.
The MFSK [9] spatiotemporal features are derived from 3D SMoSIFT. The difference between the two is that, for detecting the initial interest points, MFSK uses the SURF detector while 3D SMoSIFT employs the Shi-Tomasi corner detector; on the basis of 3D SMoSIFT feature descriptors, MFSK includes the HOG, HOF, and MBH feature descriptors, which can well represent the appearance and motion features of gestures.
It can be inferred that the above feature-based approaches detect keypoints from whole frames in the spatiotemporal domain. Some approaches rely on grayscale or depth cues to detect keypoints, which include keypoints due to noise. Although other approaches add motion information to filter out noise, constant global empirical motion constraints are not very good to adapt to various possible scenarios. Therefore, the adaptive local spatiotemporal feature is worth studying.

3. Keypoint Detection

Firstly, to reduce the effect of noise, MRoIs are adaptively extracted on the basis of variance information of grayscale and depth optical flows. Then, a sufficient number of accurate keypoints are detected using adaptive local motion and depth constraints in each MRoI.
To introduce the proposed approach more intuitively, two pairs of consecutive frames are used to illustrate every stage, as shown in Figure 1. The grayscale frames G t , G t + 1 (converted from RGB frames) and the depth frames D t , D t + 1 are captured by Kinect at time t and t + 1 , respectively. The frame resolution is N × M , where M is the number of rows, and N the number of columns.

3.1. Adaptive MRoIs Extraction

The Farneback algorithm [32] is adopted to obtain dense grayscale (depth) optical flow of G t , G t + 1 ( D t , D t + 1 ). The grayscale (depth) optical flow consists of the horizontal velocity V G t , r ( V D t , r ) and the vertical velocity V G t , c ( V D t , c ), as illustrated in Figure 2. Then, variance vectors S G t , r = { s G t , r , 1 , , s G t , r , i , , s G t , r , M } , S D t , r = { s D t , r , 1 , , s D t , r , i , , s D t , r , M } , S G t , c = { s G t , c , 1 , , s G t , c , j , , s G t , c , N } and S D t , c = { s D t , c , 1 , , s D t , c , j , , s D t , c , N } are calculated, respectively. s G t , r , i and s D t , r , i ( s G t , c , j and s D t , c , j ) are the magnitude variances of the ith row (jth column) of V G t , r and V D t , r ( V G t , c and V D t , c ), respectively.
When s G t , r , i or s D t , r , i ( s G t , c , j or s D t , c , j ) is larger, the corresponding ith row (jth column) of G t or D t has a faster horizontal (vertical) motion. If a row (column) interval of G t or D t contains large horizontal (vertical) motions, all the s G t , r , i or s D t , r , i ( s G t , c , j or s D t , c , j ) of this interval are larger. These motions are not caused by gesture alone, but noise as well. For example, there are relatively large s G t , c , j in intervals A and B (Figure 3b), which are caused by the obvious vertical motions of shadow in regions A1, A2 and B1, B2 (Figure 1a,b), respectively. The corresponding grayscale vertical velocities are shown in regions A3 and B3 of Figure 2b. The larger depth horizontal velocities in region C3 (Figure 2c) are caused by the loss of depth values in regions C1 and C2 (Figure 1c,d), and shown as row interval C in Figure 3c, in which all the s D t , r , i are larger. If keypoints are detected within the whole frame according to constant empirical motion constraints, as the 3D MoSIFT-based and MFSK methods, numerous noise points will be falsely detected as keypoints in the larger motion regions.
Usually, the motion of MBPs would cause changes in both grayscale and depth. But the depth of regions with moving shadow or illumination change remain nearly unchanged; the grayscale of regions with larger motion caused by depth value loss are mostly stable. Although clothes movements would cause changes in grayscale and depth, the changes of these regions are generally small. So, the grayscale and depth velocity magnitude variances are combined to extract MRoIs by eliminating noise to the extent possible.
After normalizing S G t , r and S D t , r , the weighted combined vector S G , D t , r of the S G t , r and S D t , r is given by:
S G , D t , r = { s G , D t , r , i | s G , D t , r , i = k 1 exp ( s G t , r , i ) + ( 1 k 1 )   exp ( s D t , r , i ) ,   1 i M } ,
where k 1 = s D t , r , i / ( s G t , r , i + s D t , r , i ) is the weight of exp ( s G t , r , i ) , then:
s G , D t , r , i = s D t , r , i exp ( s G t , r , i ) + s G t , r , i exp ( s D t , r , i ) s G t , r , i + s D t , r , i .
If the ith row of G t or D t contains a larger horizontal noise motion, then one of s G t , r , i or s D t , r , i is much larger than the other. To reduce the influence of noise, a small weighted value s G , D t , r , i can be obtained by Equations (1) and (2) where a smaller weight is assigned to the larger magnitude variance and vice versa. But when s G t , r , i and s D t , r , i are both larger or smaller, the combined s G , D t , r , i varies similarly. The weighted combined vector S G , D t , c of the normalized S G t , c and S D t , c is calculated according to Equation (3):
S G , D t , c = { s G , D t , c , j | s G , D t , c , j = s D t , c , j exp ( s G t , c , j ) + s G t , c , i exp ( s D t , c , j ) s G t , c , j + s D t , c , j ,   1 j N } .
The weighted combination can well suppress the larger variances caused by noise, but preserve the large variances intervals including MBPs, as shown in Figure 3e,f. The S G , D t , r and S G , D t , c are normalized.
Accordingly, the ith row or the jth column of G t and D t can be treated as a part of the horizontal or vertical MRoI, while s G , D t , r , i or s G , D t , c , j satisfies constraints (4) or (5):
s G , D t , r , i α 1 , 1 i M ,
s G , D t , c , j α 2 , 1 j N ,
where α 1 and α 2 are adaptively determined from S G , D t , r and S G , D t , c , respectively. α 1 can be obtained using the following steps: (1) Calculate the binarization threshold t h 1 of S G , D t , r according to the Otsu algorithm [33], and the elements of S G , D t , r with values less than t h 1 constitute a set H 1 ; (2) Calculate the threshold t h 2 of H 1 also according to the Otsu algorithm, then α 1 = t h 2 . α 2 is calculated according to the same method.
The intersections of the horizontal and the vertical MRoIs are the final MRoIs. As indicated in Figure 4, the extracted MRoIs contain complete MBPs, while the background and the non-MBPs are removed effectively, so the next step is focused on detecting keypoints in each MRoI.

3.2. Adaptive Keypoint Detection

For one-shot learning gesture recognition, the limited number of training samples requires that the detected keypoints are affine invariant, besides being scale, rotation and illumination invariant. The Harris-Affine corners are chosen as the candidate keypoints. Using the Harris-Affine corners can potentially increase the number of keypoints, which can improve the quality of the representation of the hand appearance and of its shape.
To detect Harris-Affine corners, a Gaussian scale space is constructed first. The grayscale and depth Gaussian scale spaces P G t , P D t are constructed using Equation (6) at time t [22]:
P G t , l = g ( k l 1 σ ) G t , 1 l L , P D t , l = g ( k l 1 σ ) D t , 1 l L ,
where L is the number of levels of scale space; P G t , l and P D t , l are the l t h level of P G t and P D t , respectively; k is a scale factor; σ is an initial smoothing parameter of the Gaussian function g ( ) ; and is the convolution operator. At time t + 1 , the Gaussian scale spaces P G t + 1 , P D t + 1 for G t + 1 and D t + 1 are also constructed using Equation (6). P G t , P D t , P G t + 1 and P D t + 1 together form the final Gaussian scale space. Although the method of extracting MRoIs from G t and D t has been described in the section above, the extracted MRoIs are also applicable to all the levels of scale space. Harris-Affine corners are detected as the initial interest points in the m th MRoI of different levels of P G t , l and P D t , l ( 2 l L 1 ). Assume that N m initial interest points are detected. Because MRoIs may still include parts of the background and the torso, some noise points could be falsely detected, which need further filtering. On the one hand, the noise points in the background can be removed based on the adaptive depth constraint. On the other hand, the corners in the torso region need to be screened on the basis of grayscale and depth motion information to obtain accurate keypoints.
The greyscale and depth optical flows of an initial interest point p ( x l , y l ) are calculated from P G t , l , P G t + 1 , l and P D t , l , P D t + 1 , l using the Lucas-Kanada algorithm [34]. If the depth and the optical flow of point p satisfy the adaptive constraints, Equation (7), p is deemed as a keypoint:
{ d t p θ | v G t , p | = ( v G t , r , p ) 2 + ( v G t , c , p ) 2 β 1 | v D t , p | = ( v D t , r , p ) 2 + ( v D t , c , p ) 2 β 2 ,
where v G t , r , p and v G t , c , p are the horizontal and vertical velocities of point p , respectively, of the grayscale image; v D t , r , p and v D t , c , p are the horizontal and vertical velocities of point p , respectively, of the depth image; | v G t , p | and | v D t , p | are the velocity magnitudes of point p , respectively, of the grayscale and depth image; d t p is the depth value of point p ; θ is the local depth constraint threshold adaptively determined in the mth MRoI; β 1 and β 2 , are, respectively, the local grayscale and depth motion constraint thresholds adaptively determined in the mth MRoI.
The purpose of determining θ is to adaptively search for the optimal threshold, which can accurately distinguish between large and small depth values. The purpose of determining β 1 and β 2 is to adaptively search for the optimal thresholds, which can accurately distinguish between high and low velocities. These processes are actually binarization, so θ , β 1 and β 2 can be determined using the Otsu algorithm in the mth MRoI.
The extracted MRoIs, which have different depth and speed, can obtain their own local depth and motion constraints. For each MRoI, by its own depth and motion constraints, the initial interest points detected in the background and the torso regions are filtered well and in the MBP regions are retained as keypoints. Figure 5 shows the keypoints detected by 3D EMoSIFFT, 3D SMoSIFT4, 3D SMoSIFT2, MFSK and the proposed method. All keypoints detected at different levels of grayscale or depth scale space are mapped into the original grayscale image G t or depth frame D t . Red points denote the keypoints detected from MBPs. Falsely detected keypoints are marked as green asterisks. In the second, third and fourth columns, a large number of noise points are falsely detected in the regions where shadow moving, depth value loss and clothes movements exist. That is because 3D SMoSIFT4, 3D SMoSIFT2 and MFSK detect keypoints from the whole grayscale images or depth frames according to empirical grayscale or depth motion constraints which are specified by constant values. On the contrary, only few noise points are detected in the last column. In the first column, there are also a small amount of noise points. That come out of 3D EMoSIFT using the grayscale optical flow and depth-dependent constraints simultaneously to detect keypoints. Although there are fewer noise points, the keypoints are sparse and nonuniform in the two hands and arms in Figure 5a. When compared with the proposed method, it is observed that the keypoints of 3D EMoSIFT, 3D SMoSIFT2/4 and MFSK methods are less dense than that of the proposed method, particularly for 3D EMoSIFT. Dense keypoints are essential to achieving better feature descriptors to represent gestures.

4. Feature Descriptor

The single or combined application of 3D SMoSIFT, HOG, HOF and MBH feature descriptors achieves excellent performance in one-shot learning gesture recognition, which has been demonstrated by some state-of-the-art approaches [1,9,10,11,35,36,37], and is also widely used for human activity recognition [15,30,38]. In this paper, 3D SMoSIFT, HOG, HOF and MBH feature descriptors are concatenated to represent gestures. Contrast with the 3D SMoSIFT and MFSK methods, the gradient space and motion space are extended to calculate these descriptors, which are useful for obtaining more representative RGB-D appearance and motion features of gestures.
Suppose that keypoint p is detected from the lth level of P G t or P D t (i.e., P G t , l or P D t , l ). To calculate the feature descriptors, eight local patches Γ 1 Γ 8 ( 32 × 32 ) around p are extracted. Γ 1 and Γ 2 are extracted from P G t , l and P D t , l , respectively. Γ 3 and Γ 4 are extracted from P G t + 1 , l and P D t + 1 , l , respectively. Γ 5 , Γ 6 , Γ 7 and Γ 8 in turn contain the grayscale horizontal, depth horizontal, grayscale vertical, and depth vertical velocity information. The Lucas-Kanada algorithm is adopted to calculate these grayscale and depth velocities from two pairs of local patches Γ 1 Γ 2 and Γ 3 Γ 4 .
As indicated in Figure 6a, 3D SMoSIFT constructs the 3D gradient space with three 2D gradient planes ( x y , x z 1 , y z 2 plane) based on Γ 1 Γ 2 . For a neighborhood ( 32   ×   32 ) point p 1 ( i , j ) of the keypoint p , its horizontal gradient I x ( i , j ) and vertical gradient I y ( i , j ) are calculated from Γ 1 . D z 1 x ( i , j ) and D z 2 y ( i , j ) are the horizontal and vertical gradients of p 1 calculated from Γ 2 . I x ( i , j ) , I y ( i , j ) , D z 1 x ( i , j ) and D z 2 y ( i , j ) are given as [1]:
I x ( i , j ) = Γ 1 ( i , j + 1 ) Γ 1 ( i , j ) , I y ( i , j ) = Γ 1 ( i + 1 , j ) Γ 1 ( i , j ) , D z 1 x ( i , j ) = Γ 2 ( i , j + 1 ) Γ 2 ( i , j ) , D z 2 y ( i , j ) = Γ 2 ( i + 1 , j ) Γ 2 ( i , j ) .
I x ( i , j ) and I y ( i , j ) form plane x y ; I x ( i , j ) and D z 1 x ( i , j ) form plane x z 1 ; I y ( i , j ) and D z 2 y ( i , j ) form plane y z 2 .
Three pairs of gradient magnitude and orientation of p 1 can be calculated in planes x y , x z 1 and y z 2 , respectively. For all the neighborhood ( 32 × 32 ) points of keypoint p , we calculate their three pairs of gradient magnitude and orientation by the same method. Then, we generate three new patches Γ x y , Γ x z 1 and Γ y z 2 . The size of Γ x y , Γ x z 1 and Γ y z 2 are 32 × 32 . For each point with coordinate ( i , j ) in the patch Γ x y , Γ x z 1 or Γ y z 2 , it has two values: the gradient magnitude and orientation which are calculated in the plane x y , x z 1 or y z 2 . Finally, a SIFT-like feature descriptor with 128 dimensions is calculated in each new patch. So the dimensionality of the feature descriptor vector, calculated in the 3D gradient space is 384 ( 128   ×   3 ). On this foundation, we add a new plane z 1 z 2 , which is formed by D z 1 x ( i , j ) and D z 2 y ( i , j ) . Using the above strategy, a new patch Γ z 1 z 2 is generated, as shown in Figure 6c. A SIFT-like feature descriptor, calculated from Γ z 1 z 2 , also has 128 dimensions. Therefore, a 512-dimension ( 128 × 4 ) feature descriptor vector of keypoint p can be obtained from the extended gradient space with four planes ( x y , x z 1 , y z 2 and z 1 z 2 ).
Figure 6b is the 3D motion space with three 2D gradient planes ( x y , x z 1 , y z 2 plane) based on Γ 5 Γ 8 . V x ( i , j ) , V y ( i , j ) , V z 1 x ( i , j ) and V z 2 y ( i , j ) are given by Γ 5 ( i , j ) , Γ 7 ( i , j ) , Γ 6 ( i , j ) and Γ 8 ( i , j ) of neighborhood point p 1 ( i , j ) . V x ( i , j ) and V y ( i , j ) form plane x y ; V x ( i , j ) and V z 1 x ( i , j ) form plane x z 1 ; V y ( i , j ) and V z 2 y ( i , j ) form plane y z 2 . 3D SMoSIFT uses the same method as the feature descriptor calculated in the 3D gradient space to get a 384-dimension descriptor vector in the 3D motion space. The same strategy continues to be adopted to extend the 3D motion space. A new plane z 1 z 2 is formed by V z 1 x ( i , j ) and V z 2 y ( i , j ) , as shown in Figure 6d; so, the extended motion space includes four planes ( x y , x z 1 , y z 2 and z 1 z 2 ) as well. Thus, a feature descriptor vector with 512 ( 128 × 4 ) dimensions for keypoint p can be calculated in the extended motion space. Finally, these two feature descriptor vectors of keypoint p are concatenated into a vector with 1024 ( 512 + 512 ) dimensions. The extended 3D SMoSIFT feature descriptor contains more depth appearance and motion information to represent gesture.
The MFSK method calculates the HOG and HOF feature descriptors in Γ 1 and Γ 2 [9]. The local patch is divided into γ × γ cells. The gradient orientations are quantized into η bins [9]. The MFSK method calculates the grayscale and depth HOG descriptors in the x y and z 1 z 2 planes of the extended gradient space. They can also be calculated in the x z 1 and y z 2 planes. The HOG feature descriptor vector, calculated in each plane, has γ × γ × η dimensions; thus, a HOG feature descriptor with 4 × γ × γ × η dimensions is obtained. Similarly, a HOF feature descriptor can be computed from the extended motion space. The HOF descriptor vector also has 4 × γ × γ × η dimensions.
Optical flow includes horizontal and vertical velocity components; so, there are two classes of MBH feature descriptors, i.e., MBHx and MBHy [20]. The MFSK method calculates the MBHx feature descriptors in patches Γ 5 Γ 6 and the MBHy feature descriptors in patches Γ 7 Γ 8 . Different from the HOF feature descriptor, MBHx and MBHy represent the gradient of optical flow; so, they are calculated in the extended gradient space which has four planes ( x y , x z 1 , y z 2 and z 1 z 2 ). For the MBHx feature descriptor, the extended gradient space is constructed in Γ 5 and Γ 6 . I x ( i , j ) , I y ( i , j ) , D z 1 x ( i , j ) and D z 2 y ( i , j ) are calculated by Equation (8) where Γ 1 and Γ 2 are replaced by Γ 5 and Γ 6 , respectively. Then, the MBHx descriptors are separately computed in the four planes, according to the calculation strategy of the original MBHx descriptor; so, each MBHx feature descriptor has 4 × γ × γ × η dimensions, whereas the one calculated by the MFSK method has only 2 × γ × γ × η dimensions. This is because the MFSK method calculates the MBHx descriptor in the x y and z 1 z 2 planes only. Similarly, the MBHy feature descriptor with 4 × γ × γ × η dimensions also can be calculated in Γ 7 and Γ 8 by the same method as the one used for the MBHy descriptor. The extended gradient space is also constructed by Equation (8) where Γ 7 and Γ 8 replace Γ 1 and Γ 2 , respectively.

5. Special Case

Sometimes, when the motion speed of the user is very small in several frames, the user can be considered to be in a relatively static state at that moment. Relatively static frames do not contain useful motion features, and the appearance features of the human body cannot be extracted using the proposed method. When the motion speed of the user becomes larger, the last relatively static frame turns into the first motion frame. From this frame onwards, following the proposed method, not only the motion features of MBPs, but also the same appearance features as those of the relatively static frames can be extracted, so relatively static frames can be ignored. That does not result in loss of the motion and appearance features of gesture; instead, it improves the processing efficiency.
If G t is static relative to G t + 1 , the grayscale information of the two consecutive images does not differ much. Hence, the grayscale correlation coefficient ρ t , t + 1 of G t and G t + 1 is larger than a predetermined threshold τ , so relatively static frames can be determined using the criterion ρ t , t + 1 τ .

6. Experimental Results and Discussion

The proposed method are compared with the current state-of-the-art approaches using the 2011/2012 ChaLearn One-shot-learning Gesture Dataset (CGD) [21], the Cornell Activity Dataset-60(CAD-60) [39] and the MSRDailyActivity3D Dataset [40]. There is only one training sample per gesture category. Each training or test sample includes two videos, i.e., a grayscale video (converted from RGB video) and a depth video, both of which are captured by Kinect simultaneously. L = 5 , k = 1.4 and σ = 1.6 are used to construct the scale space [22,31].
To determine the appropriate dimension sizes for calculating feature descriptors, the effects of using different settings of γ and η on feature extraction are analyzed. The MLD scores are calculated with different values of γ [ 1 , 2 , 3 , 4 ] and η [ 2 , 4 , 8 , 16 ] on the development batch of the CGD (devel 01–devel 20). As shown in Table 1, the performance of the ALSTF features is relatively stable, and the best performance is 0.1240 when γ = 3 and η = 8 . The MLD score 0.1263 of γ = 2 and η = 8 is the second best. Although its recognition performance is slightly worse, its descriptor (HOG + HOF + MBHx + MBHy) dimension 512 is less than half of that of the best one. Then we compare the computational efficiency of calculating the feature descriptors (HOG + HOF + MBHx + MBHy) on the two settings of γ and η . 47 pairs of videos (3926 grayscale and 3926 depth frames) are selected from the development batch to form a test set. Experiments are performed on a PC with C++ programs, CPU Intel® Core™ i7-4700MQ @ 2.4 GHZ and RAM 8 GB. The average computation time of γ = 3 and η = 8 is about 119.4 ms/f. It is obviously slower than the second place setting, which is about 64.7 ms/f. Taking into account both the computational efficiency and the comparison with the MFSK approach using γ = 2 and η = 8 , we adopt the trade-off with γ and η set to 2 and 8, respectively [9]. Consequently, the dimension of the HOG, HOF, MBHx or MBHy descriptors is 128 ( 4 × 2 × 2 × 8 ).
To determine the relatively static frames, based on trial-and-error, τ = 0.995 . The Bag of Word (BoW) model is adopted to represent gestures, and the parameter setting is the same as the one used in the literature [16]. The Nearest Neighbour (NN) classifier is used for classification and recognition.

6.1. Experiments on CGD Dataset

In the following experiments, four evaluation batches of CGD are used: development batch (devel 01–devel 20), validation batch (valid 01–valid 20), final batch (round1: final 01–final 20) and final batch (round 2: final 21–final 40). Each evaluation batch includes 20 sub-batches, and each sub-batch has 47 pairs of RGB and depth videos (10 fps, 320 × 240 ), which were captured by Kinect. Each pair of videos contains one to five gestures, therefore, there are 10 training gestures (corresponding to 10 classes of gestures) and 90 test gestures in every sub-batch. Every gesture has one training sample. For the sake of comparison, the Mean Levenshtein Distance (MLD) score [41], which was used by the challenge organizers, is adopted to evaluate the recognition performance. The recognition accuracy increases as the MLD score decreases, and vice versa.

6.1.1. Evaluation of Keypoint Detection Approaches

We calculate the 3D MoSIFT, 3D EMoSIFT, 3D SMoSIFT and MFSK feature descriptors of keypoints detected by the proposed method, and compare them respectively with the 3D MoSIFT, 3D EMoSIFT, 3D SMoSIFT and MFSK spatiotemporal feature approaches on the validation and final (round 2) batches.
Table 2 shows that using the proposed keypoints, combined with each of the four feature descriptors, gets higher recognition accuracy than that of the corresponding spatiotemporal feature approach. The MLD scores decrease, on average, by 0.0313 and 0.0206 on these two batches. This is because, on the one hand, the proposed method minimizes the noise effect on the detection of accurate keypoints by adaptively extracting MRoIs, and on the other hand, the keypoints are detected by determining the adaptive local depth and motion constraints in each MRoI. Because of these reasons, the MBPs with speed differences can be covered by more spatiotemporal keypoints.

6.1.2. Evaluation of Feature Descriptors

To evaluate the performance of the extended feature descriptors, the MFSK method is used in detecting the keypoints in the development batch (devel 01–devel 20) and in calculating the original (the same as [9]) and the extended 3D SMoSIFT, HOGHOF and MBH descriptors. It can be seen from Table 3 that, as compared with the original descriptors, the MLD scores 0.194, 0.187, and 0.181 of the extended descriptors decreased by 0.014, 0.011 and 0.007, respectively. Compared with the 3D SMoSIFT + HOGHOF + MBH descriptor, the accuracy of the extended 3D SMoSIFT + HOGHOF + MBH descriptor increases by 0.016. It is an improvement of about 10.3%. This is because the extended feature descriptors contain richer RGB-D appearance and motion information, which can represent gestures more fully.

6.1.3. Comparison with Other Spatiotemporal Feature Approaches

This experiment further demonstrates the recognition performance of the proposed spatiotemporal feature on the final batch (round 2: final 21–final 40). It is compared with Cuboid, Harris 3D, Dense Trajectory, 3D MoSIFT, 3D EMoSIFT, 3D SMoSIFT and MFSK feature methods, and the results are shown in Table 4. Our method achieves the lowest score 0.0737, implying that its recognition accuracy is higher than that of the other state-of-the-art spatiotemporal feature approaches. In Table 4, RGB denotes that feature descriptors are calculated with the RGB data; RGB-D denotes that the RGB and depth data are simultaneously used to calculate the feature descriptors. The 3D MoSIFT-based, MFSK and the proposed features are originally designed to use the RGB-D data, so these features are not extracted from the RGB videos. The MLD scores of Cuboid, Harris 3D, 3D MoSIFT and 3D EMoSIFT features are derived from [6]. The MLD scores of the 3D SMoSIFT feature are derived from [1]. The MLD scores of Dense Trajectory and MFSK features are derived from [9].
For the Cuboid, Harris 3D and Dense Trajectory approaches, the descriptors calculated from the RGB-D data achieve higher recognition accuracy than that calculated from the RGB data alone. It is noteworthy that all the 3D MoSIFT-based methods, based on the RGB-D data, obtain the desired results. This, in effect, means that the use of RGB-D double channel data enhances the representation and discrimination capability of spatiotemporal feature.
The MLD scores of the proposed method, and also of the 3D MoSIFT-based and MFSK methods, are significantly lower than that of Cuboid or the Harris 3D feature approach. This is because the 3D MoSIFT-based, MFSK and the proposed method fuse the RGB and depth information well to represent the appearance and motion pattern of MBPs. It is difficult to extract distinctive appearance and motion pattern from only one RGB training sample for the Cuboid and Harris 3D approaches. When the RGB-D data is used, the recognition accuracy of the Cuboid and Harris 3D approaches improves, but they simply calculate the feature descriptors only in the RGB and depth channel and do not fuse the RGB-D information well. Therefore, their MLD scores are still relatively high. The Dense Trajectory feature approach, using only the RGB data, can get a respectable accuracy of 0.1470 which is comparable with that of the 3D MoSIFT and 3D EMoSIFT methods. This is closely related to the dense spatiotemporal features extracted by the Dense Trajectory approach. The dense features can overcome the effect of some noise, besides well representing the appearance and motion of the human body. However, the effect of noise still exists; so, even after using the RGB and depth data simultaneously, the MLD score 0.1365 of the Dense Trajectory approach is still higher than that of the 3D SMoSIFT, MFSK and the proposed method.
Additionally, the proposed approach should be categorized into the multi-feature-based gesture recognition approach. It is unfair to compare it with the single feature-based ones (i.e., Cuboid, Harris 3D, Dense Trajectory, 3D MoSIFT, 3D EMoSIFT, 3D SMoSIFT). So the MLD score of each adaptive feature approach (i.e., the extended 3D SMoSIFT, HOGHOF or MBH) is calculated on the final batch (round 2), as shown in the last three lines of the Table 4. The MLD scores of the extended 3D SMoSIFT, HOGHOF and feature approaches are 0.0833, 0.0953 and 0.1027, respectively, which are lower than the single feature-based approaches. That means the recognition accuracy of each adaptive feature approach is also superior to the single feature-based ones.

6.1.4. Comparison with Other One-Shot Learning Gesture Recognition Approaches

The proposed method is compared using the validation batch (valid 01–valid 20) and the final batch (round 1: final 01–final 20) with five state-of-the-art one-shot learning gesture recognition approaches: mcHMM + LDA, HOG/HOF + DTW, 3D MoSIFT, HMM-based, motion signature analysis. The five teams corresponding to the above five approaches have made outstanding achievements in the ChaLearn Challenge. The experimental results are shown in Table 5.
The recognition accuracy of the motion signature analysis approach is far higher than that of other approaches. The team Alfine is ranked first in the two rounds of the challenge. Unfortunately, the details of their approach are not made public. Although there is a large gap in the MLD scores between our method and the motion signature analysis approach, the MLD scores 0.1069, 0.1156 achieved by our method are lower than those obtained by the other published approaches. The accuracies of the five state-of-the-art approaches are derived from [42].
mcHMM + LDA, HOG/HOF + DTW and HMM-based approaches employ HOG and HOF feature descriptors to represent gestures. Their keypoints are detected from the whole frame, and hence some noise points are falsely detected. The larger motion caused by noise affects the accuracy of the HOF descriptors too seriously to represent gesture motion. In addition, the HOG and HOF feature descriptors are not scale and rotation invariant. Using them alone to represent gesture features is unsuitable for one-shot leaning gesture recognition. So, their MLD scores 0.2084, 0.1702 are higher than those of the proposed method.

6.1.5. MLD Score Analysis for Sub-Batches

In all of the above experiments, the MLD scores are computed at the batch level. To further verify the performance, the MLD scores of every sub-batch of development batch (devel 01–devel 20) are calculated using the 3D EMoSIFT, 3D SMoSIFT, MFSK methods and our method. The results are shown in Figure 7. On most of the sub-batches, the above methods achieve good recognition results. The MLD scores of the four approaches, on average, are 0.1943, 0.1881, 0.1552 and 0.1263. On devel 04, devel 09 and devel 13, only one or two gestures are falsely detected by our method. Especially on devel 17, all the test gestures are correctly recognized.
The MLD scores of devel 03 and devel 19 obtained by these four feature-based methods are higher. To find the reasons, the confusion matrices (Figure 8) are calculated using our method. Devel 03 includes eight training gestures and 92 test gestures. Devel 19 includes nine training gestures and 91 test gestures. It reveals that the gestures, which can easily be falsely recognized, have similar trajectories in the same sub-batch.
Their main differences are in the hand shapes and appearances, such as those shown in Figure 9. Feature descriptors calculated in the extended motion space contain the motion magnitude and orientation information of keypoints. We can consider that these feature descriptors represent the trajectory features of gestures. Since these four methods detect sparse keypoints, feature descriptors calculated in the gradient space could not well represent the shape and appearance of the hand when its region is large.
The overall recognition accuracy 0.1263 of the proposed method is better than that of the 3D EMoSIFT, 3D SMoSIFT and MFSK feature methods. This is because of relatively dense and accurate keypoints together with the corresponding extended feature descriptors, which are helpful to represent the motion and appearance features of gestures.

6.1.6. Comparison of Computational Efficiency

We compare the computational efficiency of the proposed approach and related techniques. Forty seven pairs of videos (3926 grayscale and 3926 depth frames) are selected from the development batch to form a test set. The computational efficiency of 3D EMoSIFT, 3D SMoSIFT, MFSK and our methods are quantitatively compared using the test data. Experiments are performed on a PC with C++ programs, CPU Intel® Core™ i7-4700MQ @ 2.4 GHZ and RAM 8 GB. As shown in Table 6, the computational efficiency of MFSK and our methods is higher than that of the 3D EMoSIFT method, but not as good as that of the 3D SMoSIFT method. Compared with the 3D SMoSIFT method, the addition of HOG, HOF and MBH feature descriptors leads to a decrease in computational cost of the MFSK method. Our method calculates the dense optical flow to extract MRoIs before detecting keypoints, and extends the dimension sizes of the feature descriptors, so it costs more time than the MFSK method. The unoptimized code also affects the computing efficiency of our method. If the code is optimized and a higher performance computer is used, our method may meet the requirement for real-time applications.

6.2. Evaluation on CAD-60 Dataset

CAD-60 includes RGB and aligned depth videos (30 fps, 640 × 480 ). It was captured using a Microsoft Kinect sensor in five different environments: office, kitchen, bedroom, bathroom, and living room. Three to four common activities are identified for each location, giving a total of twelve unique activities and several random activities [39]. The activities were performed by two males and two females. To test the proposed feature, we experimented with two test settings: leave-one-out cross validation setting and one-shot learning setting.

6.2.1. Experiments in Leave-One-Out Cross Validation Setting

For leave-one-out cross validation setting, the model was trained on three of the four people, and tested on the fourth [39]. As can be seen in Table 7, the results of our method are comparable with those of the state-of-the-art. Our method obtains 89.7% precision, 86.1% recall and 87.86% F1 score. Although the results, on the whole, are not the best, they are comparable with other approaches. With the approaches described in [43,44,45], it is easier to locate the human’s moving parts and extract accurate features using skeleton information, so they achieve a higher accuracy than our method. Although the Kinect can effectively provide skeleton information, there are some cases where the skeleton cannot be extracted correctly and hence, cannot be extracted. Because our method does not need any preprocessing, it can still work in those cases.

6.2.2. Experiment in One-Shot Learning Setting

In the other one-shot learning setting, only one sample per activity is used for training and the rest for testing. Further, random activities are ignored and only the rest of the twelve unique activities are used to evaluate the proposed feature.
The proposed feature is compared with the MFSK feature, as shown in Figure 10. The MFSK feature obtains the better accuracy compared with other published approaches on CGD Dataset. The proposed feature can obtain 60.2% precision, 59.1% recall and 58.51% F1 score, which are higher than the results of the MFSK feature: 57%, 57.6% and 53.8% [9].

6.3. Experiments on MSRDailyActivity3D Dataset

This dataset, which consists of 16 daily activities in the living room, was captured by a Kinect sensor. There are ten subjects and each subject performs one activity in two different poses: “sitting” and “standing”. So, the total number of the activity samples is 320 [40]. This dataset is particularly challenging, because many of the daily activities involve human-object interaction.

6.3.1. Experiment in Leave-One-Out Cross Validation Setting

Table 8 shows the recognition accuracies of different state-of-the-art spatiotemporal feature approaches in leave-one-out cross validation setting. Our method achieves a recognition accuracy of 96.8%. This result is better than other approaches, considering the difficulties in this dataset.

6.3.2. Experiment in One-Shot Learning Setting

In one-shot learning setting, two samples (standing and sitting) of each activity are randomly selected as training samples. The recognition accuracy of 3D SMoSIFT, MFSK and the proposed methods are compared. The results are shown in Figure 11. Compared with the other two approaches, our method has significantly improved the recognition performance, achieving 44.5% accuracy. The accuracies have increased by 7.2% and 3.3%, respectively.

7. Conclusions

A distinctive adaptive local spatiotemporal feature has been developed to represent appearance and motion information of gesture for one-shot learning gesture recognition using RGB-D data. Adaptive extraction of MRoIs by utilizing the variance of grayscale and depth optical flows can minimize the effect of noise. MRoIs mainly include MBPs where sparse spatiotemporal features will be extracted. In each MRoI, Harris-Affine corners that satisfy adaptive local depth constraint, motion of grayscale and depth constraints are treated as keypoints. These keypoints, which are affine, scale, rotation and illumination-invariant, are accurately and sufficiently distributed in MBPs. The 3D SMoSIFT, HOG, HOF and MBH feature descriptors are calculated in the extended grayscale and motion spaces to represent the rich appearance and motion features of gesture.
The experimental results on the CGD dataset show that the proposed feature can outperform other state-of-the-art spatiotemporal feature-based approaches and the published one-shot learning approaches. It also obtains respectable results on the CAD-60 and MSRDailyActivity3D datasets under both leave-one-out cross validation setting and one-shot learning setting.
Future work, based on the results of this paper, will be focused on accurately detecting the hand region to obtain more hand shape and appearance features. This will be beneficial for improving the performance in recognizing gestures which have similar trajectories but different hand shapes. We need to study motion estimation in dynamic 3D scene [51,52], which is to enable the robot to recognize gestures in motion. We also plan to design a real-time HRI system for practical application.

Acknowledgments

This work is supported by National Natural Science Foundation of China under Grant (Grant No. 61375086) and Key Project of S & T Plan of Beijing Municipal Commission of Education (Grant No. KZ201610005010). Yang is supported by NSERC.

Author Contributions

Jia Lin contributed to the study of the approach, the performance the experiments and the manuscript writing; Xiaogang Ruan and Naigong Yu provided significant suggestions on the study of the approach and the design of the experiments; Yee-Hong Yang contributed to the study of the approach and the manuscript writing; Jia Lin and Xiaogang Ruan analyzed the results.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wan, J.; Ruan, Q.; Li, W.; An, G.; Zhao, R. 3D SMoSIFT: Three-dimensional Sparse Motion Scale Invariant Feature Transform for Activity Recognition from RGB-D Videos. J. Electron. Imaging 2014, 23. [Google Scholar] [CrossRef]
  2. Sempere, A.D.; Serna-Leon, A.; Gil, P.; Puente, S.; Torres, F. Control and Guidance of Low-Cost Robots via Gesture Perception for Monitoring Activities in the Home. Sensors 2015, 15, 31268–31292. [Google Scholar] [CrossRef] [PubMed]
  3. Santos, D.G.; Fernandes, B.J.T.; Bezerra, B.L.D. HAGR-D: A Novel Approach for Gesture Recognition with Depth Map. Sensors 2015, 15, 28646–28664. [Google Scholar] [CrossRef] [PubMed]
  4. Lake, B.M.; Salakhutdinov, R.; Tenenbaum, J.B. Human-level Concept Learning through Probabilistic Program Induction. Science 2015, 350, 1332–1338. [Google Scholar] [CrossRef] [PubMed]
  5. Krishnan, R.; Sarkar, S. Conditional Distance Based Matching for One-shot Gesture Recognition. Pattern Recognit. 2015, 48, 1302–1314. [Google Scholar] [CrossRef]
  6. Wan, J.; Ruan, Q.; Li, W.; Deng, S. One-shot Learning Gesture Recognition from RGB-D Data using Bag of Features. J. Mach. Learn. Res. 2013, 14, 2549–2582. [Google Scholar]
  7. Belgacem, S.; Chatelain, C.; Paquet, T. Adaptive Biometric Systems-A Hybrid CRF/HMM for One-Shot Gesture Learning. In Adaptive Biometric Systems, 1st ed.; Rattani, A., Roli, F., Granger, E., Eds.; Springer: Cham, Switzerland, 2015; pp. 51–72. [Google Scholar]
  8. Jiang, F.; Zhang, S.; Wu, S.; Gao, Y.; Zhao, D. Multi-layered Gesture Recognition with Kinect. J. Mach. Learn. Res. 2015, 16, 227–254. [Google Scholar]
  9. Wan, J.; Guo, G.; Li, S. Explore Efficient Local Features from RGB-D Data for One-Shot Learning Gesture Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 1626–1639. [Google Scholar] [CrossRef] [PubMed]
  10. Hernandez-Vela, A.; Bautista, M.A.; Perez-Sala, X.; Baro, X.; Pujol, O.; Angulo, C.; Escalera, S. BoVDW: Bag-of-visual-and-depth-words for gesture recognition. In Proceedings of the IEEE International Conference on Pattern Recognition, Tsukuba, Japan, 11–15 November 2012; pp. 1–6.
  11. Hernandez-Vela, A.; Bautista, M.A.; Perez-Sala, X.; Ponce-Lopez, V.; Escalera, S.; Baro, X.; Pujol, O.; Angulo, C. Probability-based Dynamic Time Warping and Bag-Of-Visual-And-Depth-Words for Human Gesture Recognition in RGB-D. Pattern Recognit. Lett. 2014, 50, 112–121. [Google Scholar] [CrossRef]
  12. Laptev, I.; Lindeberg, T. Space-time Interest Points. In Proceedings of the IEEE International Conference on Computer Vision, Beijing, China, 13–16 October 2003; pp. 432–439.
  13. Laptev, I. On Space-time Interest Points. Int. J. Comput. Vis. 2005, 64, 107–123. [Google Scholar] [CrossRef]
  14. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–26 June 2005; pp. 886–893.
  15. Laptev, I.; Marszałek, M.; Schmid, C.; Rozenfeld, B. Learning Realistic Human Actions from Movies. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 24–26 June 2008; pp. 1–8.
  16. Ming, Y.; Ruan, Q.; Hauptmann, A.G. Activity Recognition from RGB-D Camera with 3D Local Spatio-Temporal Features. In Proceedings of the International Conference on Multimedia and Expo, Melbourne, Australia, 9–13 July 2012; pp. 344–349.
  17. Chen, M.; Hauptmann, A. MoSIFT: Recognition Human Actions in Surveillance Videos; Research Report; Carnegie Mellon University: Pittsburgh, PA, USA, 2009. [Google Scholar]
  18. Bay, H.; Tuytelaars, T.; Gool, L.V. SURF: Speeded Up Robust Features. Comput. Vis. Image Underst. 2006, 110, 404–417. [Google Scholar] [CrossRef]
  19. Shi, J.; Tomasi, C. Good Features to Track. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 593–600.
  20. Dalal, N.; Triggs, B.; Schmid, C. Human Detection Using Oriented Histograms of Flow and Appearance. In Proceedings of the European Conference on Computer Vision, Graz, Austria, 7–13 May 2006; pp. 428–441.
  21. ChaLearn. ChaLearn Gesture Dataset. 2011. Available online: http://gesture.chalearn.org/data (accessed on 9 August 2015).
  22. Mikolajczyk, K.; Schmid, C. Scale & Affine Invariant Interest Point Detectors. Int. J. Comput. Vis. 2004, 60, 63–86. [Google Scholar] [CrossRef]
  23. Aggarwal, J.K.; Ryoo, M.S. Human Activity Analysis: A Review. ACM Comput. Surv. 2011, 43. [Google Scholar] [CrossRef]
  24. Harris, C.; Stephens, M. A Combined Corner and Edge Detector. In Proceedings of the Alvey vision Conference, Manchester, UK, 31 August–2 September 1988; pp. 147–151.
  25. Dollar, P.; Rabaud, V.; Cottrell, G.; Belongie, S. Behavior Recognition via Sparse Spatiotemporal Features. In Proceedings of the Joint IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, Beijing, China, 15–16 October 2005; pp. 65–72.
  26. Lu, X.; Aggarwal, J.K. Spatio-Temporal Depth Cuboid Similarity Feature for Activity Recognition using Depth Camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 25–27 June 2013; pp. 2834–2841.
  27. Willems, G.; Tuytelarrs, T.; Van Gool, L. An Efficient Dense and Scale-invariant Spatio-temporal Interest Point Detector. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; pp. 650–663.
  28. Schuldt, C.; Laptev, I.; Caputo, B. Recognizing Human Actions: A Local SVM Approach. In Proceedings of the International Conference on Pattern Recognition, Cambridge, UK, 23–26 August 2004; pp. 32–36.
  29. Laptev, I.; Lindeberg, T. Local Descriptors for Spatio-Temporal Recognition. In Proceedings of the International Conference on Spatial Coherence for Visual Motion Analysis, Prague, Czech Republic, 15 May 2004; pp. 91–103.
  30. Wang, H.; Klaser, A.; Schmid, C.; Liu, C. Dense Trajectories and Motion Boundary Descriptors for Action Recognition. Int. J. Comput. Vis. 2013, 103, 60–79. [Google Scholar] [CrossRef]
  31. Lowe, D.G. Distinctive Image Features from Scale-invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  32. Farneback, G. Two-frame Motion Estimation Based on Polynomial Expansion. In Proceedings of the 13th Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2003; pp. 363–370.
  33. Otsu, N. A Threshold Selection Method from Gray-level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  34. Lucas, B.D.; Kanade, T. An Iterative Image Registration Technique with an Application to Stereo Vision. In Proceedings of the International Joint Conference on Artificial Intelligence, Vancouver, BC, Canada, 24–28 August 1981; pp. 674–679.
  35. Konecny, J.; Hagara, M. One-shot-learning Gesture Recognition Using HOG-HOF Features. J. Mach. Learn. Res. 2014, 15, 2513–2532. [Google Scholar]
  36. Malgireddy, M.R.; Inwogu, I.; Govindaraju, V. A Temporal Bayesian Model for Classifying, Detecting and Localizing Activities in Video Sequences. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 43–48.
  37. Malgireddy, M.R.; Nwogu, I.; Govindaraju, V. Language-motivated Approaches to Action Recognition. J. Mach. Learn. Res. 2013, 14, 2189–2212. [Google Scholar]
  38. Wang, H.; Ullah, M.M.; Klaser, A.; Laptev, I.; Schmid, C. Evaluation of Local Spatio-temporal Features for Action Recognition. In Proceedings of the British Machine Vision Conference, London, UK, 7–10 September 2009; pp. 124.1–124.11.
  39. Sung, J.; Ponce, C.; Selman, B.; Saxena, A. Unstructured Human Activity Detection from RGBD Images. In Proceedings of the IEEE Conference on Robotics and Automation, Saint Paul, MN, USA, 14–18 May 2012; pp. 842–849.
  40. Wang, J.; Liu, Z.; Wu, Y.; Yuan, J. Mining Actionlet Ensemble for Action Recognition with Depth Cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 1290–1297.
  41. Guyon, I.; Athitsos, V.; Jangyodsuk, P.; Escalante, H.J. The ChaLearn Gesture Dataset (CGD 2011). Mach. Vis. Appl. 2014, 25, 1929–1951. [Google Scholar] [CrossRef]
  42. Guyon, I.; Athitsos, V.; Jangyodsuk, P.; Escalante, H.J.; Hamner, B. Results and Analysis of the ChaLearn Gesture Challenge 2012. In Proceedings of the International Workshop on Depth Image Analysis, Tsukuba, Japan, 11 November 2012; pp. 186–204.
  43. Zhu, Y.; Chen, W.; Guo, G. Evaluating Spatiotemporal Interest Point Features for Depth-based Action Recognition. Image Vis. Comput. 2014, 32, 453–464. [Google Scholar] [CrossRef]
  44. Parisi, C.; Weber, C.; Wermter, S. Self-organizing Neural Integration of Pose-motion Features for Human Action Recognition. Front. Neurorobot. 2015, 9. [Google Scholar] [CrossRef] [PubMed]
  45. Faria, D.R.; Premebida, C.; Nunes, U. A Probabilistic Approach for Human Everyday Activities Recognition Using Body Motion from RGB-D Images. In Proceedings of the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 732–737.
  46. Ni, B.; Pei, Y.; Moulin, P.; Yan, S. Multilevel Depth and Image Fusion for Human Activity Detection. IEEE Trans. Cybern. 2013, 43, 1383–1394. [Google Scholar] [CrossRef] [PubMed]
  47. Gupta, R.; Chia, Y.-S.A.; Rajan, D. Human Activities Recognition Using Depth Images. In Proceedings of the 21st ACM International Conference on Multimedia, Barcelona, Catalunya, Spain, 21–25 October 2013; pp. 283–292.
  48. Zhang, C.; Tian, Y. RGB-D Camera-based Daily Living Activity Analysis. In Proceedings of the 4th Asia-Pacific Signal & Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; pp. 1–6.
  49. Oreifej, O.; Liu, Z. Hon4d: Histogram of Oriented 4D Normal for Activity Recognition from Depth sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 25–27 June 2013; pp. 716–723.
  50. Liu, L.; Shao, L. Learning Discriminative Representations from RGB-D Video Data. In Proceedings of the International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013; pp. 1493–1500.
  51. He, H.; Tan, J. Ambient Motion Estimation in Dynamic Scenes using Wearable Visual-inertial Sensors. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014; pp. 1002–1007.
  52. Herbst, E.; Xiaofeng, R.; Dieter, F. RGB-D Flow: Dense 3-D Motion Estimation using Color and Depth. In Proceedings of the IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; pp. 2276–2282.
Figure 1. Two pairs of consecutive gesture frames captured at time t and t + 1 . (a) Grayscale frame G t ; (b) Grayscale frame G t + 1 ; (c) Depth frame D t ; (d) Depth frame D t + 1 . Noise exists in regions A1, A2, B1, B2, C1 and C2.
Figure 1. Two pairs of consecutive gesture frames captured at time t and t + 1 . (a) Grayscale frame G t ; (b) Grayscale frame G t + 1 ; (c) Depth frame D t ; (d) Depth frame D t + 1 . Noise exists in regions A1, A2, B1, B2, C1 and C2.
Sensors 16 02171 g001
Figure 2. Grayscale and depth optical flow corresponding to Figure 1. (a) Grayscale horizontal velocity V G t , r ; (b) Grayscale vertical velocity V G t , c ; (c) Depth horizontal velocity V D t , r ; (d) Depth vertical velocity V D t , c .
Figure 2. Grayscale and depth optical flow corresponding to Figure 1. (a) Grayscale horizontal velocity V G t , r ; (b) Grayscale vertical velocity V G t , c ; (c) Depth horizontal velocity V D t , r ; (d) Depth vertical velocity V D t , c .
Sensors 16 02171 g002
Figure 3. Row and column variance vectors of grayscale and depth velocity magnitude. (a) Row variance vector S G t , r of grayscale horizontal velocity magnitude; (b) Column variance vector S G t , c of grayscale vertical velocity magnitude; (c) Row variance vector S D t , r of depth horizontal velocity magnitude; (d) Column variance vector S D t , c of depth vertical velocity magnitude; (e) Weighted combined row variance vector S G , D t , r of horizontal velocity magnitude; (f) Weighted combined column variance vector S G , D t , c of vertical velocity magnitude. S G t , r , S G t , c , S D t , r and S D t , c are normalized.
Figure 3. Row and column variance vectors of grayscale and depth velocity magnitude. (a) Row variance vector S G t , r of grayscale horizontal velocity magnitude; (b) Column variance vector S G t , c of grayscale vertical velocity magnitude; (c) Row variance vector S D t , r of depth horizontal velocity magnitude; (d) Column variance vector S D t , c of depth vertical velocity magnitude; (e) Weighted combined row variance vector S G , D t , r of horizontal velocity magnitude; (f) Weighted combined column variance vector S G , D t , c of vertical velocity magnitude. S G t , r , S G t , c , S D t , r and S D t , c are normalized.
Sensors 16 02171 g003
Figure 4. Extracted horizontal, vertical and final MRoIs. (ac) The MRoIs of two moving hands frame; (df) The MRoIs of single moving hands frame. (a,d) The horizontal MRoIs; (b,e) The vertical MRoIs; (c,f) The final MRoIs.
Figure 4. Extracted horizontal, vertical and final MRoIs. (ac) The MRoIs of two moving hands frame; (df) The MRoIs of single moving hands frame. (a,d) The horizontal MRoIs; (b,e) The vertical MRoIs; (c,f) The final MRoIs.
Sensors 16 02171 g004
Figure 5. Keypoints detected by different spatiotemporal feature approaches. (a,f) The keypoints detected by 3D EMoSIFT; (b,g) The keypoints detected by 3D SMoSIFT4; (c,h) The keypoints detected by 3D SMoSIFT2; (d,i) The keypoints detected by MFSK; (e,j) The keypoints detected by the proposed approaches. Red points denote the keypoints detected from MBPs. Falsely detected keypoints are marked as green asterisks.
Figure 5. Keypoints detected by different spatiotemporal feature approaches. (a,f) The keypoints detected by 3D EMoSIFT; (b,g) The keypoints detected by 3D SMoSIFT4; (c,h) The keypoints detected by 3D SMoSIFT2; (d,i) The keypoints detected by MFSK; (e,j) The keypoints detected by the proposed approaches. Red points denote the keypoints detected from MBPs. Falsely detected keypoints are marked as green asterisks.
Sensors 16 02171 g005
Figure 6. Extended gradient space and motion space. (a) 3D gradient space with x y , x z 1 and y z 2 planes; (b) 3D motion space with x y , x z 1 and y z 2 plane; (c) z 1 z 2 plane made of D z 1 x ( i , j ) and D z 2 y ( i , j ) ; (d) z 1 z 2 plane made of V z 1 x ( i , j ) and V z 2 y ( i , j ) . (a,c) form the extended gradient space; (b,d) form the extended motion space.
Figure 6. Extended gradient space and motion space. (a) 3D gradient space with x y , x z 1 and y z 2 planes; (b) 3D motion space with x y , x z 1 and y z 2 plane; (c) z 1 z 2 plane made of D z 1 x ( i , j ) and D z 2 y ( i , j ) ; (d) z 1 z 2 plane made of V z 1 x ( i , j ) and V z 2 y ( i , j ) . (a,c) form the extended gradient space; (b,d) form the extended motion space.
Sensors 16 02171 g006
Figure 7. The results evaluated on sub-batches devel 01–devel 20 for 3D EMoSIFT, 3D SMoSIFT, MFSK and the proposed approaches.
Figure 7. The results evaluated on sub-batches devel 01–devel 20 for 3D EMoSIFT, 3D SMoSIFT, MFSK and the proposed approaches.
Sensors 16 02171 g007
Figure 8. Confusion matrix of devel 03 and devel 19. (a) Confusion matrix of devel 03; (b) Confusion matrix of devel 19.
Figure 8. Confusion matrix of devel 03 and devel 19. (a) Confusion matrix of devel 03; (b) Confusion matrix of devel 19.
Sensors 16 02171 g008
Figure 9. (ac) Three gestures that have similar trajectories (color curves) but different hand shapes and appearances. Each column shows one stage of these gestures from the start frame to the last frame.
Figure 9. (ac) Three gestures that have similar trajectories (color curves) but different hand shapes and appearances. Each column shows one stage of these gestures from the start frame to the last frame.
Sensors 16 02171 g009
Figure 10. Performance comparison between MFSK and the proposed feature on the CAD-60 dataset in one-shot learning setting. Precision, recall and F1 score are used as accuracy measure.
Figure 10. Performance comparison between MFSK and the proposed feature on the CAD-60 dataset in one-shot learning setting. Precision, recall and F1 score are used as accuracy measure.
Sensors 16 02171 g010
Figure 11. The comparison of 3D SMoSIFT, MFSK and the proposed spatiotemporal features on the MSR DailyActivity3D dataset in one-shot learning setting.
Figure 11. The comparison of 3D SMoSIFT, MFSK and the proposed spatiotemporal features on the MSR DailyActivity3D dataset in one-shot learning setting.
Sensors 16 02171 g011
Table 1. The analysis of using some different settings of γ and η to calculate feature descriptors. The values in the brackets are feature descriptor (HOG + HOF + MBHx + MBHy) dimensions.
Table 1. The analysis of using some different settings of γ and η to calculate feature descriptors. The values in the brackets are feature descriptor (HOG + HOF + MBHx + MBHy) dimensions.
η24816
γ
10.1564 (32)0.1484 (64)0.1394 (128)0.1420 (256)
20.1382 (128)0.1439 (256)0.1263 (512)0.1343 (1024)
30.1393 (288)0.1384 (576)0.1240 (1152)0.1413 (2304)
40.1337 (512)0.1305 (1024)0.1459 (2048)0.1486 (4096)
Table 2. The comparison of keypoint detection methods on the validation batch (valid 01–valid 20) and final batch (round 2: final 21–final 40).
Table 2. The comparison of keypoint detection methods on the validation batch (valid 01–valid 20) and final batch (round 2: final 21–final 40).
ApproachMLD Score
Valid 01–Valid 20Final 01–Final 20
3D MoSIFT keypoint + 3D MoSIFT descriptor [42]0.18240.1448
3D EMoSIFT keypoint + 3D EMoSIFT descriptor [6]0.15950.1331
3D SMoSIFT keypoint + 3D SMoSIFT descriptor [1]0.17400.1130
MFSK keypoint + MFSK descriptor [9]0.12700.0925
Proposed keypoint + 3D MoSIFT descriptor0.14120.1221
Proposed keypoint + 3D EMoSIFT descriptor0.13450.1127
Proposed keypoint + 3D SMoSIFT descriptor0.12940.0879
Proposed keypoint + MFSK descriptor0.11250.0783
Table 3. The performance validation of the extended 3D SMoSIFT, HOGHOF and MBH feature descriptors (devel 01–devel 20).
Table 3. The performance validation of the extended 3D SMoSIFT, HOGHOF and MBH feature descriptors (devel 01–devel 20).
DescriptorMLD Score (Devel 01–20)
MBH [9]0.208
HOGHOF [9]0.198
3D SMoSIFT [9]0.188
3D SMoSIFT + HOGHOF + MBH0.155
Extended MBH0.194
Extended HOGHOF0.187
Extended 3D SMoSIFT0.181
Extended 3D SMoSIFT + HOGHOF + MBH0.139
Table 4. The performance comparison of different spatiotemporal feature approaches on the final batch (round 2: final 21–final 40).
Table 4. The performance comparison of different spatiotemporal feature approaches on the final batch (round 2: final 21–final 40).
Spatiotemporal FeatureMLD Score (Final 21–Final 40)
RGBRGB-D
Cuboid0.31390.2806
Harris 3D + HOG0.23460.2268
Harris 3D + HOF0.29060.2712
Harris 3D + HOGHOF0.18860.1819
Dense Trajectory0.14700.1365
3D MoSIFT-0.1448
3D EMoSIFT-0.1331
3D SMoSIFT-0.1130
MFSK-0.0925
Proposed feature (ALSTF)-0.0737
Proposed keypoint + Extended MBH 0.1027
Proposed keypoint + Extended HOGHOF 0.0953
Proposed keypoint + Extended 3D SMoSIFT 0.0833
Table 5. The performance comparison with 5 state-of-the-art one-shot learning gesture recognition approaches on the valid and final (round 1: final 01–final 20) batches.
Table 5. The performance comparison with 5 state-of-the-art one-shot learning gesture recognition approaches on the valid and final (round 1: final 01–final 20) batches.
Approach/TeamMLD Score
Vaild 01–Valid 20Final 01–Final 20
mcHMM + LDA/Immortals0.24880.1847
HOGHOF + DTW/Turtle Tamers0.20840.1702
3D MoSIFT/Joewan0.18240.1680
HMM-based/Pennect0.17970.1652
Motion signature analysis/Alfine0.09950.0734
Proposed feature (ALSTF)0.10690.1156
Table 6. The computational efficiency of 3D EMoSIFT, 3D SMoSIFT, MFSK and the proposed feature.
Table 6. The computational efficiency of 3D EMoSIFT, 3D SMoSIFT, MFSK and the proposed feature.
ApproachAverage Time (ms/f)
3D EMoSIFT646.4
3D SMoSIFT45.3
MFSK119.1
Proposed feature (ALSTF)163.7
Table 7. The performance comparison of several state-of-the-art approaches on the CAD-60 dataset under leave-one-out cross validation setting.
Table 7. The performance comparison of several state-of-the-art approaches on the CAD-60 dataset under leave-one-out cross validation setting.
ApproachAccuracy (%)
PrecisionRecallF1 Score
Hierarchical MEMM [39]67.955.561.08
3D SMoSIFT [1]74.865.870.01
Ni et al., 2013 [46]75.969.572.56
Guptal et al., 2013 [47]78.175.476.73
Zhang and Tian 2012 [48]86.084.084.99
MFSK 2016 [9]87.183.885.42
Zhu et al., 2014 [43]93.284.688.69
Parisi et al., 2015 [44]91.990.291.47
DBMM [45]91.191.991.50
Proposed feature (ALSTF)89.786.187.86
Table 8. The comparison of different spatiotemporal features on the MSR DailyActivity3D dataset in leave-one-out cross validation setting.
Table 8. The comparison of different spatiotemporal features on the MSR DailyActivity3D dataset in leave-one-out cross validation setting.
Spatiotemporal FeatureAccuracy (%)
3D MoSIFT75.9
Hon4d [49]80.0
RGGP [50]85.6
Fourier temporal pyramid feature + Actionlet ensemble model [40]85.8
3D EMoSIFT86.0
3D SMoSIFT93.2
MFSK95.7
Proposed feature (ALSTF)96.8

Share and Cite

MDPI and ACS Style

Lin, J.; Ruan, X.; Yu, N.; Yang, Y.-H. Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition. Sensors 2016, 16, 2171. https://doi.org/10.3390/s16122171

AMA Style

Lin J, Ruan X, Yu N, Yang Y-H. Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition. Sensors. 2016; 16(12):2171. https://doi.org/10.3390/s16122171

Chicago/Turabian Style

Lin, Jia, Xiaogang Ruan, Naigong Yu, and Yee-Hong Yang. 2016. "Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition" Sensors 16, no. 12: 2171. https://doi.org/10.3390/s16122171

APA Style

Lin, J., Ruan, X., Yu, N., & Yang, Y. -H. (2016). Adaptive Local Spatiotemporal Features from RGB-D Data for One-Shot Learning Gesture Recognition. Sensors, 16(12), 2171. https://doi.org/10.3390/s16122171

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop