Next Article in Journal
The Influence of the sEMG Amplitude Estimation Technique on the EMG–Force Relationship
Previous Article in Journal
Security and Privacy Analysis of Youth-Oriented Connected Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Unified Local–Global Feature Extraction Network for Human Gait Recognition Using Smartphone Sensors

National Institute of Technology Rourkela, Rourkela 769008, India
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(11), 3968; https://doi.org/10.3390/s22113968
Submission received: 13 April 2022 / Revised: 12 May 2022 / Accepted: 16 May 2022 / Published: 24 May 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Smartphone-based gait recognition has been considered a unique and promising technique for biometric-based identification. It is integrated with multiple sensors to collect inertial data while a person walks. However, captured data may be affected by several covariate factors due to variations of gait sequences such as holding loads, wearing types, shoe types, etc. Recent gait recognition approaches either work on global or local features, causing failure to handle these covariate-based features. To address these issues, a novel weighted multi-scale CNN (WMsCNN) architecture is designed to extract local to global features for boosting recognition accuracy. Specifically, a weight update sub-network (Ws) is proposed to increase or reduce the weights of features concerning their contribution to the final classification task. Thus, the sensitivity of these features toward the covariate factors decreases using the weight updated technique. Later, these features are fed to a fusion module used to produce global features for the overall classification. Extensive experiments have been conducted on four different benchmark datasets, and the demonstrated results of the proposed model are superior to other state-of-the-art deep learning approaches.

1. Introduction

Human gait is a biometric attribute that is useful and attracting attention in different fields such as surveillance, biomedical engineering, clinical analysis, etc. Commonly, gait analysis is essential in a clinical investigations such as fall detection [1], rehabilitation [2,3], physical therapy [4], etc., for the well-being of a patient suffering from underlying diseases such as strokes, Parkinson’s, or progressive supranuclear palsy (PSP). Current studies focus on the recent development of human gait rehabilitation therapy based on the state of the brain by employing the brain–computer interface (BCI) system [5,6,7]. BCI systems are capable of decoding the cognitive state of a patient to provide feedback to an external device such as a wheelchair, robotic prostheses/orthoses, or muscle simulator by acquiring brain signals from electroencephalographic (EEG), as discussed in these papers [6,8,9]. In [10], the authors utilized EEG-based brain signals for distinguishing between a healthy person and a patient by measuring the level of attention of a person toward his gait. Furthermore, to measure the attention level, numerous methods have been developed such as the continuous performance test (CPT) and the test of variable attention (T.O.V.A.) referred in [11]. Apart from that, the eye-movement tracking technique [4] is adopted among the PSP patient to improve temporal aspects of the gait of the patient by estimating the eye-movement parameters through a GP3 eye-tracker [12,13].
Although human gait is very familiar in the era of clinical analysis, the current paper exploits this attribute in individual recognition. Generally, gait recognition models are commonly implemented either through vision-based methods, which utilize the video and image data [14,15,16,17,18,19] or through inertial-based devices such as wearable sensors/floor sensors/smartphones’ sensors to capture signals of human movement [20,21,22,23] to infer gait identity. Although the vision-based method has been extensively studied and can achieve a high recognition rate, its application is limited due to the high acquisition cost and difficulty in the deployment of cameras in a real-life environment. On the other hand, inertia-sensor-based technology such as smart devices with built-in sensors, wearable sensors, and smartphones are on excess demand due to its low cost, convenience carrying, and good real-time performance [23,24,25]. Today, smartphones are featured with many inertial sensors such as an accelerometer and gyroscope to capture the speed and direction of a moving person [26,27,28]. Therefore, it is beneficial to track the person in survieillance. Currently, many research studies [24,29,30,31] have been completed in this area, which motivates us to utilize smartphone sensor data for gait recognition.
In this article, an ideal approach is proposed to effectively handle covariate-based gait signals by utilizing multi-scale CNN concepts to get deep spatial features using down-sampled signals referred to as local features. However, the key difference between traditional multi-scale CNN and our proposed approach is to predict features at different scales to obtain discriminant features. To accomplish this task, a branch network called a weight update subnetwork (Ws) is coupled to each CNN to highlight the relevance feature vectors and specify more weights by using the fisher discriminant criterion [32]. The down-sampled signal from low scale to high scale indicates elusive variations between gait poses due to the effect of the covariates. Therefore, a fusion module is implemented to generate the effect of dependencies between low-scale samples to high-scale samples. Eventually, all these weighted features are flattened into a 1D array for producing a single feature vector. In the end, a softmax layer followed by a fully connected network (FCN) is employed to process the feature vectors for final classification.
The main contributions of this article are briefly outlined as follows:
  • Inspired by the multi-scale approach, the proposed model leverages multi-scale convolutional neural networks [33], a fusion network, and a weight update sub-network, and it combines them in an end-to-end manner to address the covariate issues.
  • In particular, it aims to highlight relevant local features in each scale with respect to label-based gait patterns by incorporating weight update sub-networks (Ws). Furthermore, global features are extracted with the help of a fusion network. The significance of discriminative local and global features is to handle intra-class variations and inter-class variations, respectively.
  • The proposed framework has been gone through extensive empirical evaluations using four benchmark gait-based inertial datasets: OU-ISIR, whuGAIT, Gait-mob-ACC, and IDNet, and the results are compared with many state-of-the-art gait recognition models such as IdNet [23], CNN [34], LSTM [30], DeepConv [35], CNN+LSTM [24], and the proposed model outperforms others.
The remainder of this paper is organized as follows. Literature related to the proposed method is discussed in Section 2. The framework of the proposed model and its corresponding architecture is described in Section 3. The experimental setup and results are presented in Section 4 and discussed in Section 5. Section 6 provides the conclusion.

2. Related Work

2.1. Sensor-Based Gait Identification

Recently, sensors-based gait analysis has become a rapidly growing research platform [21,24,30,36,37,38,39,40]. In early research, Nickel et al. [41] captured accelerometer data through smartphones, where cepstral coefficients are extracted from the data to consider as a feature set, and support vector machine (SVM) has been used for training these features. In 2012, Juefei-Xu et al. [42] developed a step-independent gait identification model from a continuous tracking of smartphone-based acceleration and gyroscope data. Furthermore, several studies have been proposed for handling multi-modal sensor data in gait identification using fusion-based techniques [43], a Gaussian mixture model (GMM-UBM) [42], and CNN methodologies [44].

2.2. Deep Learning Approaches on Gait Analysis

In the last few years, several deep learning models have been proposed for gait-based identification [24,30,39,45,46]. For example, convolutional neural networks (CNN) are widely used in many existing gait recognition methods [23,24,47]. IDNet [23] incorporates both a CNN-based deep learning approach and machine learning tools such as SVM to process inertial signals captured from smartphones for gait authentication. Here, the CNN network has been adopted as a universal feature extractor and SVM for gait classification. Another related work of deep learning is multi-scale analysis, which has achieved a series of progress in the field of detection, classification, and identification. So far, a multi-scale strategy has been widely used in deep learning for gait-based recognition [48,49,50] where it explores spatial features at multiple scales and learns more details about different gait regions to extract local features. However, it fails to find dependencies among the spatial features as well as overall gait variations. Gait recognition methods based on global representations deal with gait data as a whole and do not pay attention to local gait details; some examples include GaitNet-1 [51] and GaitNet2 [52], but these methods are sensitive to the covariate factors. To address the above issues, for the first time, in this context, a novel model (WMsCNN-Local-Global) has been proposed to extract more comprehensive features, which contains both local and global information of inertia signals acquired from smartphones.

3. System Overview

The proposed framework is comprised of five parts: acquisition of inertial gait data, segmentation of gait cycle, deep feature extraction, training, and classification. The schematic diagram of the proposed framework is shown in Figure 1. The acquisition of inertial gait data is done through an accelerometer and gyroscope sensor, which are useful for tracking a person’s movement along the X, Y, and Z directions, denoted as A x , A y , A z , and G x , G y , G z respectively. All the sensor data are normalized using L 2 norm to avoid uncertain movements of smartphones such as shifting of smartphones from left to right or up to down positions. Furthermore, the gait cycle segmentation task is carried out using the acceleration data along the X, Y, and Z directions. The paper adopts U-net [24] to perform this task. All the gait cycles are randomly split into a gallery (train) and probe (test) sequences. To obtain deep features from the multi-scale technique, samples are further down-sampled into different time scales and processed through several convolutional layers, which are treated as an independent feature set. A novelty of the proposed method lies in the localization of the important feature map and assigning weights to the feature vector for training and classification. To perform this task, a weight update subnetwork (Ws) is designed to connect each CNN architecture. Later, all the locally weighted features are fused to get dependence among them to utilize for overall gait variations. Eventually, all the fused features are flattened and fed to the fully connected layer for classification.

3.1. Proposed Approach

The objective of the paper is to estimate the importance of feature vectors with respect to their label prediction and ignore other features that may misguide a classifier. In other words, different weights can be assigned to the local feature vectors from different scales by giving more weight to the representative features and less weight to others. In order to accomplish this, a multi-scale signal is reconstructed from a single scale by down-sampling and further processed through a stack of CNN structures to get deep features at different time scales. The detailed design of the proposed model is shown in Figure 2.
Multi-scale signal reconstruction: The inertial data acquired from the accelerometer and gyroscope are simultaneously considered inputs. It can be expressed as x t = [ A x , A y , A z , G x , G y , G z ] at time step t along the X, Y, and Z-axis. Combining all the time steps can be represented as a gait cycle X = [ x 1 , x 2 , , x N ] , where N is the number of steps to be considered in each gait cycle. Assume each gait cycle ’X’ is down-sampled at a time scale ’ τ ’ is expressed below.
x t τ = 1 τ k = ( t 1 ) τ + 1 t τ X k , 1 t N τ ,
where x t τ is a down-sampled signal computed by taking an average of consecutive data points t of the input signal X k at time index k. The whole expression of the multi-scale signal is denoted as x τ = { x 1 τ , , x t τ , , x N / τ τ } .
So far, the effectiveness of the convolutional neural network has been proven as a good feature extractor in the field of motion data, image analysis, speech signal processing, etc. [53]. Thus, we are motivated to incorporate CNN architecture into each scaled signal to obtain significant features. Each scaled sub-sample x t τ is fed to the four convolutional layers of the CNN network, which is followed by a pooling layer. The output of the layer is expressed as below.
x t l , τ = R e L U ( m = 1 M W m l 1 , t x t + m 1 l 1 , τ + b t l ) ,
where x t l , τ denotes the output layer, l ( 1 , 2 , 3 , 4 ) ; * denotes the convolutional operator; M is the kernel size; b t l is the bias term at layer l; W is the weight of the l t h feature map; ReLU denotes an activation function; m and l denote the index of the kernel and convolutional layer, respectively. Later, by applying a pooling layer, the output local feature is given as
f t l , τ = m a x ( x t l 1 , τ ( n ) ) ,   n [ ( j 1 ) w , t w ] ,
where f t l , τ is the output of the maximum value among the ( l 1 ) t h layer obtained from samples x t l 1 , τ ( n ) , n represents the n t h output neurons at the j t h position of local features, and w is the width size of the pooling layer.
Weight update sub-network (Ws): The proposed sub-network aims to explore a novel spatial adaptive weighting technique using the Fisher-based discrimination [54] among the feature vectors with respect to their labels. The main idea is to map the classifier weights to each feature vector to perform localized classification. Subsequently, weights are assigned to each feature vector depending on its contribution to its label data. To accomplish this task, a sub-network is inserted between the last CNN layer and a classifier. A global average pooling layer and a soft-max layer are the part of the sub-network that finds localized features for each class label. The architecture of the weight update sub-network (Ws) is shown in Figure 3.
Suppose in multi-scale signal analysis, F τ , k represents the output feature map of CNN for each scale of unit k after passing through a global average pooling layer (GAP), which is specified as below.
F τ , k = 1 h i f i , k ,
where f i , k R c , i = 1 , 2 , , h is the local feature vectors at unit k. The localized classification is performed using the dot product between the feature vectors and the weights of the classifier, as described in (5).
f i ^ = i k w k c f i ,
where f i ^ R N is the localized classification score at class c, w k c is the class-specific weight vector assigned to local features, and ’i’ is the location of each feature. Subsequently, weights are updated by projecting the localized classification scores from high-dimensional space to low-dimensional space, based on their intra (within) and inter (between) classes distributions.
Let the localized scores f ^ i be projected from the N dimensional space to N dimensional space for separating two different classes. Then, the weight λ i is computed by considering an N eigenvector corresponding to the maximum eigenvalue given below.
^ w = i f ^ c i ( f ^ m i ) ( f ^ m i )
^ b = ( m i m ) ( m i m )
λ i = m a x N ( e i g ( ^ w 1 ^ b ) ) ,
where ^ w , ^ b are the within-class matrix and between-class matrix, which are computed in (6) and (7), respectively. m ¯ i and m are the mean of the local and global class, respectively.

3.2. Fusion Network

All the locally weighted features from low-scaled gait variations to large-scaled variations are fused to obtain linear dependency among them. As it is a linear combination of discriminative features from small gait sequences to large gait sequences, the resultant feature set is named the global feature set. It is computed as follows.
F ^ g l o b a l = W τ 1 F ^ τ 1 + W τ 2 F ^ τ 2 + + W τ s F ^ τ s ,
where the fusion weights W τ 1 , W τ 2 , , W τ s are the adaptive parameters learned from the training sets. Subsequently, the global feature F ^ g l o b a l is fed to a fully connected layer (FC) and a softmax layer. The output expressions of both the layers are presented below:
F ^ g l o b a l l = b l + F ^ g l o b a l l 1 · w l
o = S o f t m a x ( F ^ g l o b a l W 0 + b 0 )

3.3. Training and Classification

The training of the proposed model is performed in an end-to-end manner, learning combined with multiple weight update sub-networks (Ws) and overall networks in a single unified fashion using a backpropagation algorithm. To do so, Ws sub-networks are trained independently from fewer scales to more scales to obtain local optimization under the supervision of label-based gait sample patterns. The total classification loss of the local features is observed below.
L l o c a l = i = 1 i = s α i L i ( F ^ i , y ) ,
where s represents the total number of sub-networks in the local module, α i is the weight parameter of the each sub-network, and y is the label of gait patterns at different conditions. Then, the overall training is computed at the final layer to obtain global optimization, and the gradients are propagated backwards layer-by-layer to update the weights. The overall loss of the proposed framework (WMsCNN-Local-Global) can be represented by
L o v e r a l l = α L L o c a l ( F i , y ) + β L g l o b a l ( F g l o b a l , y ) ,
where α and β are both weight updated parameters. Each loss function is defined in terms of cross-entropy loss.
The network is iteratively trained through several epochs to update the model using the training set. Furthermore, the training set is split up into distinct batches B , and each batch B has B segments. In each epoch, the training set is shuffled and computes a set of output vectors O based on its loss function. Let each vector o i O be the estimated prediction score for each label. o ^ i O is the actual score for label i. Then, the cross-entropy loss-based classification problem can be formulated as below.
L B ( o , o ^ ) = 1 B i = 1 K o ^ i l n o i + ( ( 1 o ^ i ) l n ( 1 o i ) ) ,
where L B the cross-entropy loss function used to update network’s internal parameters through back-propagation [55]. When all the batches have been used to train the network, one training epoch is completed; then, the process is repeated with a new epoch until it meets a stopping condition as referred in Section 4.1. It is observed from (14) that a large difference between o i and o ^ i results in a high value of entropy loss. Basically, the training network adopted this concept for optimization.

4. Experimental Setup and Result Analysis

The experiment is conducted by integrating weight update sub-networks (Ws) into various CNN architectures. All the experiments are implemented using the Keras API and Caffe framework. The proposed network evaluates different challenging datasets having covariate conditions and compares them with several state-of-the-art deep learning approaches, such as CNN, LSTM, CNN+LSTM, IdNet, and Deepconv modules. A brief description of the datasets is given in Table 1.

4.1. Different Sensor-Based Gait Dataset

whuGAIT datasets [24]: Here, 118 subjects are taken into consideration in the data collection, out of which 20 subjects have a large number of data, where each holds thousands of samples. The rest of the subjects contain a smaller amount of data, each holding hundreds of samples. Furthermore, each data sample contains a three-axis accelerometer and gyroscope data. Here, all the data are sampled at 50 Hz. The dataset is organized into eight subsets from Dataset #1 to Dataset #8. In this paper, Dataset #1 and Dataset #2 are used for classification, while the rest, Dataset #5–#6 and #7–#8, used for gait authentication and gait data extraction, respectively.
IdNet dataset [23]: It has 50 subjects and collects data from both a tri-axial gyroscope and accelerometer embedded in a smartphone. The sampling rate of the sensor data is 100 Hz. These data include two such variations, such as people wearing different shoe types and different clothes at a different time of gait data acquisition.
OU-ISIR dataset [21]: So far, it is the largest population dataset in terms of capturing inertial-sensor-based gait data. Two types of devices such as 3IMUZ sensors and Motorola ME860 are used to capture the sensor data. The first one captures both accelerometer and gyroscope data, while the second one collects triaxial accelerometer data. Each sensor works at 100 Hz. The experiments are performed on two different sets of users on the basis of two different conditions. One experiment is conducted for evaluation in the presence of a large set of the population around 744 subjects; another one is conducted on 408 subjects in the presence of two different ground surfaces, i.e., sloppy surface and plain surface.
Gait-mob-ACC-dataset [22]: It is the most challenging dataset that incorporates eight types of covariates along with speed variations. There are three sets of data such as Dataset #1, Dataset #2, and Dataset #3, which are captured from an accelerometer and kinetic sensor simultaneously. Here, inertial data from accelerometers are only included in the experiments. Among the three datasets, Dataset #1 contains 10 subjects, and each subject contain 100 samples. Out of 100 samples, half of the samples are collected from the fast walk and another half are collected from the normal walk. Dataset #2 has 50 subjects, with ten data samples for each subject. Dataset #3 has 50 subjects and 48 data samples from each subject. In particular, each subject requests to walk in eight different conditions, i.e., freestyle walking, hand in a pocket (left or right or both hands), holding a book either right or left hand, carrying loads either right or left hand.

4.2. Network Architecture

The proposed network has been built in an end-to-end fashion such that a gait sample is accepted from one end; then, it passes through sub-networks, which are tied together, and produces its identity at the other end. Each sub-network is connected with CNN, having four optimum numbers of convolutional layers in the order of a kernel regularization layer ( L 2 ) , a ReLU activated layer followed by a max-pooling layer of size 2, and a dropout layer. Each layer has a filter attached, and the maximum depth is set as 32, 32, 64, and 128 in the order of four layers. An Adam optimizer is compiled with a learning rate of 0.001. The dropout layer is recognized to be the best option to reduce overfitting. Here, dropout is set at a rate of 0.5 after convolutional layers and 0.8 after the fusion layer to force other weights to neutralize. This leads to higher accuracy and a better understanding of the data. The weights of the convolutional layers and fully connected layers are initialized using the Kaiming initializer. The weighting factors α and β are manually tuned and set to 0.99 and 0.87, and a batch size of 32 is used for all experiments. The number of epochs for training is 200. The early stopping condition is set if no improvement is taking place after 50 consecutive epochs. The detailed parameters of the proposed single scale CNN network are given in Table 2. For multi-scale analysis, each input signal has a fixed dimension of 200 samples of length.
The experiments and the results are discussed on the following points:
  • Experiments on the effect of using the proposed weight update sub-network (Ws) into various CNN architectures.
  • Performance of the proposed methods in handling gait data collected under different covariate conditions.
  • Evaluation of the proposed method for identification and authentication.

4.3. Experiments on the Effect of Using the Proposed Weight Updated Sub-Networks (Ws) into Various CNN Architectures

The proposed Ws layer is integrated into various CNN backbones such as AlexNet [56], VGG14 [57], VGG16 [57], and ResNet50 [58], and we compare their performance for handling sensor-based gait signals in multi-scale analysis. To do so, all the fully connected layers are removed from each of the CNN backbones and replaced with Ws layers followed by a fully connected softmax layer. For example, in AlexNet, the layers after conv5 have been replaced with Ws. In both the architecture of VGG14 and VGG16, its single and triple FCN layers are replaced with Ws, respectively. In ResNet, the proposed layer is connected after the max-pooling layer to perform the task. From Table 3, it is observed that by employing Ws, the identification rate improves to 1–3% in each model. This is because each sub-network guides the extraction of more correlated features by focusing on the semantically relevant class-specific samples and ignoring the uncorrelated patterns. The performance of the proposed model is the best among all other models irrespective of different covariate conditions. Regarding the architectures, we find that ResNet50 performs comparably to the proposed model. Meanwhile, both VGG-14 and VGG-16 have similar performance in the identification rate, but VGG-16 shows a quite significant improvement in identification rate of 0.5 % to 1.5 % on the Gait-mob-ACC dataset. Furthermore, we observe that the performance of each model slowly declines as the size of sub-network s varies from 4 to 5. The best performance is recorded at the ensemble of 4 sub-networks.

4.4. Performance Evaluation of the Proposed Network under Different Covariate Conditions

The paper analyzes the performance of the proposed model on the most challenging dataset, i.e., Gait-mob-ACC [22], which contains possible co-variate factors ties in our daily life.
To evaluate the proposed model on the above covariate conditions, the Gait-mob-ACC dataset is divided into five sub-datasets and named as Gait-normal, Gait-fast, Gait-mixed, Gait-fast-Covar, and Gait-normal-Covar, each having an equal number of subjects. Some experiments have been conducted with varying batch sizes, steps, and training samples to obtain the highest performance of the proposed model. The comparative results are shown in Figure 4a–c. It is observed from Figure 4a that the model obtains the best performance on different covariates by varying the batch size B . Increasing B from 16 to 32, the accuracy gradually improves from 94 % to 94.6 % in the normal walk, it improves around 0.45 % more in both fast and mixed walks, and it improves 0.36 % more in covariate conditions. However, when B reached more than 64, the accuracy is degraded to more than 0.07 % . This is because after a certain increase of batch size, the overlapping may take the place of two gait cycles over two different persons, which shows erroneous results.
Another important setting for improving the performance is considering the number of walking cycles of a given model. The accuracy will increase with increasing the number of steps. It is shown from Figure 4 that at normal walking speed, the accuracy increased at a rate of 0.01–0.05%, whereas for fast walking, the rate of accuracy increase is about 0.1–0.3%. So, a higher step always gives better performance; however, higher steps for a person also entail a longer acquisition time, which we would rather avoid. Therefore, we restrict the number of steps N s = 2 in all the experiments, as it provides a good trade-off between accuracy and complexity across evaluations.
From Figure 4c, it is observed that the model obtains good recognition accuracy in all five cases, e.g., over 0.95 in a normal gait speed, 0.92 during fast walking, and 0.90 during mixed walking using 30 % of training. Moreover, the model shows almost equal accuracy under normal and fast pace.

4.5. Identification and Authentication of Gait Based Bio-Metric System

The whole dataset is divided into two sub-datasets: a training and a test set. Both the training and the testing sets are made disjoint from each other. The experiment shows its performance in terms of identification and verification process. In the identification process, a identification rate (IR) has been used for rank-based classification. For the verification process, receiver operating characteristics (ROC) curves are obtained by plotting pairs of verification rates and false acceptance rates at various threshold values.

4.5.1. Experimental Results on Identification

All the datasets for gait identification are processed through a common experimental set up. Here, each dataset is split into a training set considered as a gallery set, and the remaining is used for testing as the probe set. The distance scores between the whole gallery set are compared to each other to obtain the smallest score as IR. Table 4 demonstrates the Rank1 IR as compared to the other state-of-the-art methods on different benchmark datasets.
For the whuGait dataset, Dataset #2 achieves better performance as compared to Dataset #1 with an IR of more than 96 % . This is because there are more samples per subject in Dataset #2 than in Dataset #1. It is also observed that both standalone networks CNN and LSTM perform approximately 0.3 % better than parallely connected CNN and LSTM. One possible reason is that the parallel network may face over-fitting problems. Furthermore, it is noticeable that the performance of the CNN network is better than that of the LSTM network. C N N f i x + L S T M and L S T M f i x + C N N are both complementary networks of each other. Both are designed with parallel connection by fixing the parameter of one network and updating the other network. These two networks achieve an IR of approximately 93 % and 92 % on Dataset #1 and Dataset #2, respectively. The proposed network outperforms the other two networks such as IdNet and DeepconvLSTM with IR values of more than 2.34 % and 2.05 % , respectively. This is because the WMsCNN-Local model is a single-scale CNN architecture attached with a CWs sub-network that gives a competitive performance for its discriminative local feature analysis, whereas the multi-scale approach is incorporated with the proposed network modeled as (WMsCNN-Local-Global), which gives the best performance of around 99.96 % .
In the IDNet dataset, all the collected gait samples are free style walking. Therefore, the IR values of all networks are quite high. The proposed approach achieves 99.96 % IR.
In the OU-ISIR dataset, the LSTM network achieves better performance than the C N N + L S T M network in the presence of variation of gait sequences. For both Dataset #1 and # 2 , the proposed network obtains more than 73 % IR. The result signifies that the proposed network can effectively handle variations of gait sequences better than other approaches.
In the Gait-mob-ACC database, six different covariates are incorporated in Dataset #3. It is the most challenging dataset, having speed variations from normal walk to fast walk. The last column of the table gives a detailed comparison. Deepconv competes with the proposed approach with a performance of less than 2 % , but it achieves better results than other approaches. Our multi-scale approach can effectively handle complex features generated from covarite conditions such as both hands in the pocket, carrying loads, etc.

4.5.2. Experiments on Authentication

The authentication task is performed by transforming the multi-class identification problem into a binary classification problem, which is based on the hypothesis of either positive acceptance or false acceptance. The authentication performance is evaluated by the metric of the average receiver operating characteristic (ROC) curve. It is created by plotting the true acceptance rate (TAR) against the false acceptance rate (FAR) at varying threshold settings. In the ROC curve, the value of FAR is set as 0.001 % as the standard FAR for bio-metric authentication. The TAR and FAR are defined as
T A R = T r u e P o s i t i v e T r u e P o s i t i v e + F a l s e P o s i t i v e
F A R = F a l s e P o s i t i v e F a l s e P o s i t i v e + T r u e N e g a t i v e
To evaluate the system performance, the model is incorporated into different types of state-of-the-art methods. The experiments are conducted to examine the relative behavior of the false accept rate and the verification rate under different covariate conditions using (15) and (16). The ROC curves for the proposed method and the other state-of-the-art methods are plotted in Figure 5. The model achieves a higher verification rate at very low FARs. As we find from Figure 5b, the proposed network achieves limited improvement on the OU-ISIR dataset, while it has a notable performance on the whuGait dataset, IdNet datasets, and Gait-mob-ACC datasets as in Figure 5a,b,f respectively. In Figure 5d, the proposed network produces a very competing performance with LSTM, but later, it achieves equal performance with it when FAR is around 0.001. In response to the real environment, the Gait-mob-ACC dataset is considered, having multiple covariates along with speed variations from normal to fast. The performance of the proposed network is superior to others. After that, CNN finds it better than the other three networks. It can be observed from most figures that the multi-scale network uniformly outperforms overall networks, which may simply indicate that the multi-scale features are more discriminative by describing the detailed gait subdynamics. According to the above analysis of experimental results, we conclude that the combination of discriminate local features and global features is more suitable for the gait analysis on covariate conditions.

5. Discussion

A major contribution of this work is the joint use of the discriminative local features and global features to handle covariate factors and overall gait sequence variations, respectively. From Table 4, it is observed that WMsCNN-Local achieves good recognition accuracy using the benefits of Ws. However, combining all the local to global features further improves the recognition accuracy both in the identification and authentication module. It is reasonable that the global features only focus on overall gait cycle variations and ignore the multiple pose variations due to the effect of several covariates. Similarly, only local features ignore the overall variations. From the experimental results of Figure 5a–c, it is observed that the performance of the CNN network is quite appreciable because the features of CNN have more discriminable properties than the LSTM. Therefore, the proposed model (WMsCNN-Local-Global) outperforms as it captures CNN data at different time scales and combines them for a better representation of the feature sets. In addition, it is observed that direct features of LSTM are not appropriate for discriminating complex features such as gait, resulting in lowering the accuracy. Moreover, Table 4 shows the performance of the single-scale proposed model (WMsCNN-Local) and the multi-scale proposed model (WMsCNN-Local-Global), which gives the inference that the ensemble of sub-networks improves the performance of a single network. However, Table 4 reveals that more than 4 sub-networks degrade the performance of the overall network. Furthermore, in the evaluation of results, it is noticeable that some inconsistencies are found between the performances of the identification and authentication model. The performance of authentication is a little bit lower than the performance of identification. One possible reason for this is over-fitting, since only one test is used in the authentication process.

6. Conclusions

In this paper, an improved deep learning network is designed for gait recognition using smartphones. The novelty of the proposed approach lies in the feature extraction technique, which is based on a multi-scale signal approach and it is incorporated with a weight update feature sub-network to exploit significant local features. These sub-networks of each CNN architecture assign more weights to become discriminative feature regions for better classification. The significant of the local features from each scale are combined using a fusion network to achieve global-based features. The experiment performs on four benchmark datasets with different covariate conditions. The acquired results of the proposed framework reach an accuracy of 99.96 % and 73.56 % in the normal gait and most challenging gait database, respectively. The overall performance of the proposed model is superior compared to other state-of-the-art networks.

Author Contributions

S.D. Conceptualization, Methodology, Software, Validation, Writing—original draft preparation. S.M. Supervision, Investigation, Formal analysis. U.K.S. Supervision, Investigation, Formal analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tao, W.; Liu, T.; Zheng, R.; Feng, H. Gait analysis using wearable sensors. Sensors 2012, 12, 2255–2283. [Google Scholar] [CrossRef] [PubMed]
  2. Castermans, T.; Duvinage, M.; Cheron, G.; Dutoit, T. Towards effective non-invasive brain-computer interfaces dedicated to gait rehabilitation systems. Brain Sci. 2013, 4, 1–48. [Google Scholar] [CrossRef] [PubMed]
  3. Hamid, H.; Naseer, N.; Nazeer, H.; Khan, M.J.; Khan, R.A.; Shahbaz Khan, U. Analyzing Classification Performance of fNIRS-BCI for Gait Rehabilitation Using Deep Neural Networks. Sensors 2022, 22, 1932. [Google Scholar] [CrossRef] [PubMed]
  4. Andreoni, G.; Mazzola, M.; Zambarbieri, D.; Forzoni, L.; D’Onofrio, S.; Viotti, S.; Santambrogio, G.C.; Baselli, G. Motion analysis and eye tracking technologies applied to portable ultrasound systems user interfaces evaluation. In Proceedings of the 2013 International Conference on Computer Medical Applications (ICCMA), Sousse, Tunisia, 20–22 January 2013; pp. 1–6. [Google Scholar]
  5. Katona, J.; Kovari, A. A brain–computer interface project applied in computer engineering. IEEE Trans. Educ. 2016, 59, 319–326. [Google Scholar] [CrossRef]
  6. Katona, J. A review of human–computer interaction and virtual reality research fields in cognitive InfoCommunications. Appl. Sci. 2021, 11, 2646. [Google Scholar] [CrossRef]
  7. Hong, K.S.; Khan, M.J. Hybrid brain–computer interface techniques for improved classification accuracy and increased number of commands: A review. Front. Neurorobot. 2017, 11, 35. [Google Scholar] [CrossRef] [Green Version]
  8. Katona, J.; Ujbanyi, T.; Sziladi, G.; Kovari, A. Speed control of Festo Robotino mobile robot using NeuroSky MindWave EEG headset based brain-computer interface. In Proceedings of the 2016 7th IEEE international conference on cognitive Infocommunications (CogInfoCom), Wrocław, Poland, 16–18 October 2016; pp. 251–256. [Google Scholar]
  9. Katona, J.; Ujbanyi, T.; Sziladi, G.; Kovari, A. Examine the effect of different web-based media on human brain waves. In Proceedings of the 2017 8th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Debrecen, Hungary, 11–14 September 2017; pp. 407–412. [Google Scholar]
  10. Costa, Á.; Iáñez, E.; Úbeda, A.; Hortal, E.; Del-Ama, A.J.; Gil-Agudo, A.; Azorín, J.M. Decoding the attentional demands of gait through EEG gamma band features. PLoS ONE 2016, 11, e0154136. [Google Scholar] [CrossRef]
  11. Katona, J. Examination and comparison of the EEG based Attention Test with CPT and TOVA. In Proceedings of the 2014 IEEE 15th International Symposium on Computational Intelligence and Informatics (CINTI), Budapest, Hungary, 19–21 November 2014; pp. 117–120. [Google Scholar]
  12. Katona, J. Clean and dirty code comprehension by eye-tracking based evaluation using GP3 eye tracker. Acta Polytech. Hung. 2021, 18, 79–99. [Google Scholar] [CrossRef]
  13. Katona, J. Analyse the Readability of LINQ Code using an Eye-Tracking-based Evaluation. Acta Polytech. Hung. 2021, 18, 193–215. [Google Scholar] [CrossRef]
  14. Chen, X.; Xu, J. Uncooperative gait recognition: Re-ranking based on sparse coding and multi-view hypergraph learning. Pattern Recognit. 2016, 53, 116–129. [Google Scholar] [CrossRef]
  15. Choudhury, S.D.; Tjahjadi, T. Robust view-invariant multiscale gait recognition. Pattern Recognit. 2015, 48, 798–811. [Google Scholar] [CrossRef] [Green Version]
  16. Luo, J.; Tang, J.; Tjahjadi, T.; Xiao, X. Robust arbitrary view gait recognition based on parametric 3D human body reconstruction and virtual posture synthesis. Pattern Recognit. 2016, 60, 361–377. [Google Scholar] [CrossRef] [Green Version]
  17. Das, S.; Purnananda, H.Y.; Meher, S.; Sahoo, U.K. Minimum Spanning Tree Based Clustering for Human Gait Classification. In Proceedings of the IEEE Region 10 Conference (TENCON), Kochi, India, 17–20 October 2019; pp. 982–985. [Google Scholar]
  18. Xing, X.; Wang, K.; Yan, T.; Lv, Z. Complete canonical correlation analysis with application to multi-view gait recognition. Pattern Recognit. 2016, 50, 107–117. [Google Scholar] [CrossRef]
  19. Zeng, W.; Wang, C.; Yang, F. Silhouette-based gait recognition via deterministic learning. Pattern recognit. 2014, 47, 3568–3584. [Google Scholar] [CrossRef]
  20. Sprager, S.; Juric, M.B. Inertial sensor-based gait recognition: A review. Sensors 2015, 15, 22089–22127. [Google Scholar] [CrossRef]
  21. Ngo, T.T.; Makihara, Y.; Nagahara, H.; Mukaigawa, Y.; Yagi, Y. The largest inertial sensor-based gait database and performance evaluation of gait-based personal authentication. Pattern Recognit. 2014, 47, 228–237. [Google Scholar] [CrossRef] [Green Version]
  22. Zou, Q.; Ni, L.; Wang, Q.; Li, Q.; Wang, S. Robust gait recognition by integrating inertial and RGBD sensors. IEEE Trans. Cybern. 2017, 48, 1136–1150. [Google Scholar] [CrossRef] [Green Version]
  23. Gadaleta, M.; Rossi, M. Idnet: Smartphone-based gait recognition with convolutional neural networks. Pattern Recognit. 2018, 74, 25–37. [Google Scholar] [CrossRef] [Green Version]
  24. Zou, Q.; Wang, Y.; Wang, Q.; Zhao, Y.; Li, Q. Deep Learning-Based Gait Recognition Using Smartphones in the Wild. IEEE Trans. Inf. Forensics Secur. 2020, 15, 3197–3212. [Google Scholar] [CrossRef] [Green Version]
  25. De Marsico, M.; Mecca, A.; Barra, S. Walking in a smart city: Investigating the gait stabilization effect for biometric recognition via wearable sensors. Comput. Electr. Eng. 2019, 80, 106501. [Google Scholar] [CrossRef]
  26. Bashar, S.K.; Al Fahim, A.; Chon, K.H. Smartphone Based Human Activity Recognition with Feature Selection and Dense Neural Network. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 5888–5891. [Google Scholar]
  27. Mekruksavanich, S.; Jitpattanakul, A. Smartwatch-based human activity recognition using hybrid lstm network. In Proceedings of the 2020 IEEE Sensors, Rotterdam, The Netherlands, 25–28 October 2020; pp. 1–4. [Google Scholar]
  28. Mondal, R.; Mukherjee, D.; Singh, P.K.; Bhateja, V.; Sarkar, R. A New Framework for Smartphone Sensor-Based Human Activity Recognition Using Graph Neural Network. IEEE Sens. J. 2020, 21, 11461–11468. [Google Scholar] [CrossRef]
  29. Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Cell phone-based biometric identification. In Proceedings of the 4th IEEE International Conference on Biometrics, Theory, Applications and Systems (BTAS), Washington, DC, USA, 27–29 September 2010; pp. 1–7. [Google Scholar]
  30. Watanabe, Y.; Kimura, M. Gait identification and authentication using LSTM based on 3-axis accelerations of smartphone. Procedia Comput. Sci. 2020, 176, 3873–3880. [Google Scholar] [CrossRef]
  31. Nemes, S.; Antal, M. Feature learning for accelerometer based gait recognition. In Proceedings of the 2021 IEEE 15th International Symposium on Applied Computational Intelligence and Informatics (SACI), Timisoara, Romania, 19–21 May 2021; pp. 479–484. [Google Scholar]
  32. Loog, M.; Duin, R.P.W.; Haeb-Umbach, R. Multiclass linear dimension reduction by weighted pairwise Fisher criteria. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 762–766. [Google Scholar] [CrossRef]
  33. Sermanet, P.; Eigen, D.; Zhang, X.; Mathieu, M.; Fergus, R.; LeCun, Y. Overfeat: Integrated recognition, localization and detection using convolutional networks. In Proceedings of the 2nd International Conference on Learning Representations (ICLR 2014), Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  34. Delgado-Escaño, R.; Castro, F.M.; Cózar, J.R.; Marín-Jiménez, M.J.; Guil, N. An end-to-end multi-task and fusion CNN for inertial-based gait recognition. IEEE Access 2018, 7, 1897–1908. [Google Scholar] [CrossRef]
  35. Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  36. Raccagni, C.; Gaßner, H.; Eschlboeck, S.; Boesch, S.; Krismer, F.; Seppi, K.; Poewe, W.; Eskofier, B.M.; Winkler, J.; Wenning, G.; et al. Sensor-based gait analysis in atypical parkinsonian disorders. Brain Behav. 2018, 8, e00977. [Google Scholar] [CrossRef]
  37. Wei, Z.; Qinghui, W.; Muqing, D.; Yiqi, L. A new inertial sensor-based gait recognition method via deterministic learning. In Proceedings of the 2015 34th Chinese Control Conference (CCC), Hangzhou, China, 28–30 July 2015; pp. 3908–3913. [Google Scholar]
  38. Donath, L.; Faude, O.; Lichtenstein, E.; Pagenstert, G.; Nüesch, C.; Mündermann, A. Mobile inertial sensor based gait analysis: Validity and reliability of spatiotemporal gait characteristics in healthy seniors. Gait Posture 2016, 49, 371–374. [Google Scholar] [CrossRef]
  39. Tran, L.; Choi, D. Data augmentation for inertial sensor-based gait deep neural network. IEEE Access 2020, 8, 12364–12378. [Google Scholar] [CrossRef]
  40. Khabir, K.M.; Siraj, M.S.; Ahmed, M.; Ahmed, M.U. Prediction of gender and age from inertial sensor-based gait dataset. In Proceedings of the 2019 Joint 8th International Conference on Informatics, Electronics & Vision (ICIEV) and 2019 3rd International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Spokane, WA, USA, 30 May–2 June 2019; pp. 371–376. [Google Scholar]
  41. Nickel, C.; Brandt, H.; Busch, C. Classification of acceleration data for biometric gait recognition on mobile devices. In Proceedings of the BIOSIG 2011–Proceedings of the Biometrics Special Interest Group, Darmstadt, Germany, 8–9 September 2011. [Google Scholar]
  42. Juefei-Xu, F.; Bhagavatula, C.; Jaech, A.; Prasad, U.; Savvides, M. Gait-id on the move: Pace independent human identification using cell phone accelerometer dynamics. In Proceedings of the 5th IEEE International Conference on Biometrics: Theory, Applications and Systems (BTAS), Arlington, VA, USA, 23–27 September 2012; pp. 8–15. [Google Scholar]
  43. Kumar, P.; Mukherjee, S.; Saini, R.; Kaushik, P.; Roy, P.P.; Dogra, D.P. Multimodal gait recognition with inertial sensor data and video using evolutionary algorithm. IEEE Trans. Fuzzy Syst. 2018, 27, 956–965. [Google Scholar] [CrossRef]
  44. Gong, J.; Goldman, M.D.; Lach, J. Deepmotion: A deep convolutional neural network on inertial body sensors for gait assessment in multiple sclerosis. In Proceedings of the IEEE Wireless Health (WH), Bethesda, MD, USA, 25–27 October 2016; pp. 1–8. [Google Scholar]
  45. Sena, J.; Barreto, J.; Caetano, C.; Cramer, G.; Schwartz, W.R. Human activity recognition based on smartphone and wearable sensors using multiscale DCNN ensemble. Neurocomputing 2021, 444, 226–243. [Google Scholar] [CrossRef]
  46. Yao, S.; Hu, S.; Zhao, Y.; Zhang, A.; Abdelzaher, T. Deepsense: A unified deep learning framework for time-series mobile sensing data processing. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 May 2017; pp. 351–360. [Google Scholar]
  47. Hoang, H.V.; Tran, M.T. DeepSense-inception: Gait identification from inertial sensors with inception-like architecture and recurrent network. In Proceedings of the 13th International Conference on Computational Intelligence and Security (CIS), Hong Kong, China, 15–18 December 2017; pp. 594–598. [Google Scholar]
  48. Wu, Z.; Huang, Y.; Wang, L.; Wang, X.; Tan, T. A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 209–226. [Google Scholar] [CrossRef] [PubMed]
  49. Chao, H.; He, Y.; Zhang, J.; Feng, J. Gaitset: Regarding gait as a set for cross-view gait recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8126–8133. [Google Scholar]
  50. Fu, Y.; Wei, Y.; Zhou, Y.; Shi, H.; Huang, G.; Wang, X.; Yao, Z.; Huang, T. Horizontal pyramid matching for person re-identification. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8295–8302. [Google Scholar]
  51. Zhang, Z.; Tran, L.; Yin, X.; Atoum, Y.; Liu, X.; Wan, J.; Wang, N. Gait recognition via disentangled representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4710–4719. [Google Scholar]
  52. Zhang, Z.; Tran, L.; Liu, F.; Liu, X. On learning disentangled representations for gait recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef] [PubMed]
  53. Cui, Z.; Chen, W.; Chen, Y. Multi-scale convolutional neural networks for time series classification. arXiv 2016, arXiv:1603.06995. [Google Scholar]
  54. Liu, C.; Wechsler, H. Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition. IEEE Trans. Image Process. 2002, 11, 467–476. [Google Scholar] [PubMed] [Green Version]
  55. Chen, D.S.; Jain, R.C. A robust backpropagation learning algorithm for function approximation. IEEE Trans. Neural Netw. 1994, 5, 467–479. [Google Scholar] [CrossRef] [PubMed]
  56. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  57. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
Figure 1. Overview of the proposed framework.
Figure 1. Overview of the proposed framework.
Sensors 22 03968 g001
Figure 2. Detailed design of the WMs–CNN–Local–Global model.
Figure 2. Detailed design of the WMs–CNN–Local–Global model.
Sensors 22 03968 g002
Figure 3. Architecture of a weight update sub-network (Ws) to achieve discriminative features.
Figure 3. Architecture of a weight update sub-network (Ws) to achieve discriminative features.
Sensors 22 03968 g003
Figure 4. Performance evaluation of the proposed network in terms of accuracy on the five different types of gait sequences, with the influence of varying (a) batch size B (b) number of steps N s per gait cycle, (c) amount of training data.
Figure 4. Performance evaluation of the proposed network in terms of accuracy on the five different types of gait sequences, with the influence of varying (a) batch size B (b) number of steps N s per gait cycle, (c) amount of training data.
Sensors 22 03968 g004
Figure 5. A comparative ROC curves of state-of-the-art deep learning networks: IDnet, CNN, LSTM, CNN+LSTM, and the proposed models. The performance of four benchmarks, each having different sub-datasets as referred in Table 1, is shown in (af). (a,b) refer to whuGait Dataset #1 and Dataset #2, respectively, (c) refers to the IdNet dataset, (d,e) refer to sub-dataset #1 and sub-dataset #2 of the OU-ISIR dataset, respectively, and (f) refers to the Gait-mob-ACC dataset.
Figure 5. A comparative ROC curves of state-of-the-art deep learning networks: IDnet, CNN, LSTM, CNN+LSTM, and the proposed models. The performance of four benchmarks, each having different sub-datasets as referred in Table 1, is shown in (af). (a,b) refer to whuGait Dataset #1 and Dataset #2, respectively, (c) refers to the IdNet dataset, (d,e) refer to sub-dataset #1 and sub-dataset #2 of the OU-ISIR dataset, respectively, and (f) refers to the Gait-mob-ACC dataset.
Sensors 22 03968 g005
Table 1. Details of four challenging datasets.
Table 1. Details of four challenging datasets.
Database
Name
No.Number of
Subjects
Sampling
Rate
Challenges
OU-ISIR#1
#2
745
408
100 HzA large database with fewer samples on each subject and
each subject walks on a plain and sloppy surface
whuGAIT#1
#2
118
20
50 Hz
Gait-mob-ACC#1
#2
#3
10
50
50
Variation of walking speed: normal and fast with seven different covariates:
either hand/both hand in pocket, either hand holding
book, and either hand with loadings
IDNet-50100 HzWear different shoe types and different clothes
Table 2. Detailed parameters of the proposed single-scale CNN network.
Table 2. Detailed parameters of the proposed single-scale CNN network.
Layer NameInputKernel
Size
Number of
Kernels
Feature
Map
Number of
Parameters
Conv1_1 200 × 1 × 6 9 × 1 32 192 × 1 × 32 1760
MaxPool1 192 × 1 × 32 2 × 1 / 96 × 1 × 32
Conv2_1 96 × 1 × 32 5 × 1 64 92 × 1 × 64 10,304
Conv2_2 92 × 1 × 64 5 × 1 128 46 × 1 × 128 41,088
MaxPool2 46 × 1 × 128 2 × 1 / 23 × 1 × 128
Conv3_1 23 × 1 × 128 3 × 1 128 21 × 1 × 128 49,280
Table 3. Rank-1 and Rank-5 identification rates, and verification rate (VR) of different gait datasets are reported by integrating 2/3/4/5 numbers of the Ws sub-network layers into various CNN architectures at the presence of different time scales ( τ ). Bold font indicates the best performance.
Table 3. Rank-1 and Rank-5 identification rates, and verification rate (VR) of different gait datasets are reported by integrating 2/3/4/5 numbers of the Ws sub-network layers into various CNN architectures at the presence of different time scales ( τ ). Bold font indicates the best performance.
sub-Networks (s)whuGait IDnet OU-ISIR Gait-Mob-ACC
Rank1Rank-5VR Rank-1Rank-5VR Rank-1Rank-5VR Rank-1Rank-5VR
IdId(FAR = 10 3 ) IdId(FAR = 10 3 ) IdId(FAR = 10 3 ) IdId(FAR = 10 3 )
2 ( τ = 4 , 5 ) AlexNet78.6982.940.79 81.9883.990.79 55.8961.820.43 75.8880.750.74
VGG-1483.6686.430.83 86.2490.910.86 57.7661.830.44 75.9479.040.74
VGG-1684.5687.060.83 86.3191.650.87 57.8862.950.44 76.3279.990.75
ResNet-5093.0695.830.88 92.9594.090.91 67.4770.540.49 86.2189.770.91
CWs-AlexNet80.9885.910.82 84.2493.460.81 56.9161.980.43 78.8883.180.77
CWs-VGG1485.8790.050.85 88.0693.320.88 60.7864.210.46 83.1988.060.79
CWs-VGG1686.3391.870.86 89.8794.540.89 61.4365.320.47 84.6788.420.8
CWs-ResNet5095.0497.760.91 96.1197.970.93 69.5373.890.51 89.5593.670.94
proposed95.3298.080.92 96.3498.560.93 70.3475.850.52 91.3294.430.94
3 ( τ = 3 , 4 , 5 ) AlexNet89.0192.640.81 88.8792.570.84 59.5661.840.45 80.0183.650.87
VGG1491.3294.540.82 92.0195.360.87 61.1965.920.47 83.8887.730.88
VGG1691.7895.810.83 92.4295.750.88 61.7664.200.46 84.8888.780.89
ResNet5093.5497.470.88 96.2497.960.94 68.5472.860.51 89.6593.640.92
CWs-AlexNet90.5493.040.83 90.7694.350.85 61.8265.470.47 83.7185.450.90
CWs-VGG1493.1296.130.89 94.0297.610.89 64.8968.780.48 87.2191.650.91
CWs-VGG1693.0396.430.89 94.2497.860.90 65.5669.350.49 88.4593.670.93
CWs-ResNet5096.3298.530.93 98.241000.96 72.8676.590.53 94.0198.150.96
Proposed97.3699.780.94 99.961000.96 73.3877.510.54 94.0598.640.96
4 ( τ = 2 , 3 , 4 , 5 )AlexNet87.4392.560.83 86.2190.640.83 53.5458.430.40 72.8877.670.86
VGG1487.4391.010.84 88.1292.890.86 60.1964.840.46 80.3284.430.88
VGG1688.3492.430.84 88.5193.300.86 60.4865.480.46 82.9885.130.89
ResNet5091.8995.940.88 90.8994.140.89 64.9769.990.50 86.1690.980.92
CWs-AlexNet90.3495.890.86 88.1393.030.85 55.7660.200.41 76.7180.510.87
CWs-VGG1491.0395.960.86 90.0795.460.87 64.9868.780.48 84.2388.890.92
CWs-VGG1691.8996.650.87 90.4594.550.88 65.1669.980.49 86.8991.780.94
CWs-ResNet5094.1298.270.9 94.3997.060.92 69.9773.320.52 91.1391.980.95
Proposed97.0199.750.94 99.931000.96 73.5678.840.55 94.8899.140.97
5 ( τ = 1 , 2 , 3 , 4 , 5 )AlexNet79.4583.650.76 79.6384.320.82 52.1756.990.41 77.4480.210.81
VGG1481.6984.320.78 86.3390.560.82 53.1157.420.4 79.6482.430.83
VGG1681.9484.870.79 87.4491.210.83 54.9857.960.4 81.0185.780.85
ResNet5090.1594.470.89 92.3197.090.9 62.3369.080.50 88.5692.110.90
CWs-AlexNet83.1785.540.79 82.5487.540.83 54.1258.730.42 80.3284.040.82
CWs-VGG1489.1291.350.85 89.3293.760.87 61.4164.560.45 84.3687.550.89
CWs-VGG1689.5492.980.85 91.0793.860.88 61.5965.320.46 86.1490.760.91
CWs-ResNet5093.3297.890.91 95.8997.320.91 65.7671.340.50 90.1294.320.93
Proposed96.3898.640.93 98.7699.320.94 72.1676.180.53 93.8998.560.96
Table 4. Comparison of state-of-the-art methods on different benchmark datasets in terms of Rank-1 identification rate.
Table 4. Comparison of state-of-the-art methods on different benchmark datasets in terms of Rank-1 identification rate.
MethodswhuGait
Dataset1
(118 Subjects)
whuGait
Dataset2
(20 Subjects)
IDNet
Dataset
(50 Subjects)
OU-ISIR
Dataset
(745 Subjects)
OU-ISIR
Dataset2
(408 Subjects)
Gait-Mob-ACC
Dataset3
(50 Subjects)
IdNet [23]92.91%96.78%99.58%44.29%46.20%74.75%
CNN [34]92.89%97.02%99.71%40.60%47.14%90.2%
LSTM [30]91.88%96.98%99.46%66.36%65.32%81.65%
DeepConv [35]92.25%96.80%99.24%37.33%41.32%86.23%
CNN+LSTM [24]92.51%96.82%99.61%34.28%53.96%89.22%
C N N f i x +LSTM [24]92.94%97.04%99.64%---
CNN+ L S T M f i x [24]93.52%97.33%99.75%---
WMsCNN-Local93.36%98.28%99.81%65.74%72.13%90.49%
WMsCNN-Local-Global95.75%98.98%99.96%73.56%76.42%94.71%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Das, S.; Meher, S.; Sahoo, U.K. A Unified Local–Global Feature Extraction Network for Human Gait Recognition Using Smartphone Sensors. Sensors 2022, 22, 3968. https://doi.org/10.3390/s22113968

AMA Style

Das S, Meher S, Sahoo UK. A Unified Local–Global Feature Extraction Network for Human Gait Recognition Using Smartphone Sensors. Sensors. 2022; 22(11):3968. https://doi.org/10.3390/s22113968

Chicago/Turabian Style

Das, Sonia, Sukadev Meher, and Upendra Kumar Sahoo. 2022. "A Unified Local–Global Feature Extraction Network for Human Gait Recognition Using Smartphone Sensors" Sensors 22, no. 11: 3968. https://doi.org/10.3390/s22113968

APA Style

Das, S., Meher, S., & Sahoo, U. K. (2022). A Unified Local–Global Feature Extraction Network for Human Gait Recognition Using Smartphone Sensors. Sensors, 22(11), 3968. https://doi.org/10.3390/s22113968

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop