Next Article in Journal
Investigation of Polyurethane Matrix Membranes for Salivary Nitrate ISFETs to Prevent the Drift
Next Article in Special Issue
Classification of Postprandial Glycemic Status with Application to Insulin Dosing in Type 1 Diabetes—An In Silico Proof-of-Concept
Previous Article in Journal
Microcapsule-Based Signal Amplification Method for Biomolecules
Previous Article in Special Issue
Using Wearable and Non-Invasive Sensors to Measure Swallowing Function: Detection, Verification, and Clinical Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Stage Latent Dynamics Modeling and Filtering for Characterizing Individual Walking and Running Patterns with Smartphone Sensors

1
Department of Mathematics, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Korea
2
Department of Control and Instrumentation Engineering, Korea University, 2511 Sejong-ro, Sejong-City 30019, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2019, 19(12), 2712; https://doi.org/10.3390/s19122712
Submission received: 2 May 2019 / Revised: 14 June 2019 / Accepted: 15 June 2019 / Published: 17 June 2019
(This article belongs to the Special Issue Wearable Sensors in Healthcare: Methods, Algorithms, Applications)

Abstract

:
Recently, data from built-in sensors in smartphones have been readily available, and analyzing data for various types of health information from smartphone users has become a popular health care application area. Among relevant issues in the area, one of the most prominent topics is analyzing the characteristics of human movements. In this paper, we focus on characterizing the human movements of walking and running based on a novel machine learning approach. Since walking and running are human fundamental activities, analyzing their characteristics promptly and automatically during daily smartphone use is particularly valuable. In this paper, we propose a machine learning approach, referred to as ’two-stage latent dynamics modeling and filtering’ (TS-LDMF) method, where we combine a latent space modeling stage with a nonlinear filtering stage, for characterizing individual dynamic walking and running patterns by analyzing smartphone sensor data. For the task of characterizing movements, the proposed method makes use of encoding the high-dimensional sequential data from movements into random variables in a low-dimensional latent space. The use of random variables in the latent space, often called latent variables, is particularly useful, because it is capable of conveying compressed information concerning movements and efficiently handling the uncertainty originating from high-dimensional sequential observation. Our experimental results show that the proposed use of two-stage latent dynamics modeling and filtering yields promising results for characterizing individual dynamic walking and running patterns.

1. Introduction

Recently, data from built-in sensors such as gyroscope and accelerometers in smartphones have been readily available, and analyzing data for various types of health information from smartphone users has become a popular health care application area. Among the relevant issues in the area, one of the most prominent issues is analyzing the characteristics of human movements. In this paper, we consider the problem of characterizing the human movements of walking and running by means of a novel machine learning approach. Since many health care topics are related to walking and running, a great deal of current research efforts focus on questions and problems concerning how to characterize human walking and running patterns via machine learning. In particular, various machine learning methods have successfully addressed distinguishing human activities such as walking (see e.g., [1,2,3,4]) utilizing data from wearable sensors. Sekine et al. [1] distinguished ambulatory patterns of elderly subjects walking on stairways versus walking on level ground using waist acceleration signals, utilizing wavelet coefficients. Papagiannaki et al. [2] proposed an activity recognition scheme for older people based on feature extraction from wearable sensors and machine learning methods, and considered the problem of recognizing physical activity of older people. In their works, classification was conducted by standard machine learning as well as deep learning techniques. Jiang et al. [3] applied convolutional neural networks (CNNs) to human activity recognition using activity image, and extracted optical features for six different actions. This study [3] used the activity image, which assembled time-series sensor signals of accelerometers and gyroscopes, as input to CNNs, and obtained good performance in terms of recognition accuracy and computational cost. Wang et al. [4] proposed an algorithm for detecting several human ambulatory patterns from data obtained via a triaxial accelerometer; decomposing sensor signal data into frequency scales was conducted by a discrete Fourier transform (DFT), and then classifications of the resultant features were performed by multilayer perceptron (MLP) networks. In addition, for machine learning methods for detecting falls, one may refer to papers such as [5,6,7,8,9].
In this paper, we investigate the problem concerning how to find the low-dimensional latent dynamics for walking and running with smartphone sensors, which will lead us to some intrinsic representation of the movements. For solving this problem, we propose a machine learning approach, referred to as a ’two-stage latent dynamics modeling and filtering’ (TS-LDMF) method, where we combine a latent space modeling stage with a nonlinear filtering stage. The proposed method makes use of encoding high-dimensional sequential data from human movements into random variables in a low-dimensional latent space. The use of random variables in latent space, i.e., latent variables, is particularly useful, because this is capable of carrying compressed information concerning movements and efficiently handling uncertainty deriving from high dimensional sequential observation.
In the sense that the proposed method utilizes the step of transforming high-dimensional smartphone sensor outputs into low-dimensional latent variables, the proposed method can be viewed as fulfilling nonlinear dynamic dimension reduction. Dimension reduction is one of the most fundamental issues in the area of machine learning. Among the well-known conventional linear and nonlinear dimension reduction methods, are there principal component analysis (PCA) [10] and kernel principal component analysis (KPCA) [11]. The proposed method is a type of generalized dimension reduction, which can perform nonlinear dimension reduction with the additional capacity of modeling latent dynamics and filtering for latent states in the low-dimensional latent space.
In addition, in the sense that this paper addresses multiple functions of generating joint distributions for observations and latent variables, modeling latent dynamics, and nonlinear filtering for latent states with smartphone sensor data based on TS-LDMF, the work in this paper may be viewed as a closely related extension and applications of deep generative models such as the deep Markov model [12,13]. Deep generative models are a branch of deep learning [14,15,16], and recently, they have been successfully applied to various important classes of unsupervised learning problems such as variational auto-encoders (VAE) [17,18], generative adversarial networks (GAN) [19,20], neural ordinary differential equations (neural ODE) [21,22], and deep Markov models [12,13]. One of the main advantages of deep generative models is that uncertainty information in their probabilistic models can be explicitly provided by their solutions. Because uncertainty information in their probabilistic models is valuable in addressing latent dynamics modeling, filtering, and probability density estimation, the strategies of deep generative modeling are also useful for the purpose of this paper. For the task of finding the normal latent region for walking or running, we use a modern density estimation approach based on the neural ODE [21,22]. Since the ODE-based approach can work well utilizing a relatively smaller neural network [21,22], it will be very useful to our purpose in handling high-dimensional data on smartphones. In addition, since the latent samples yielded by the training of TS-LDMF models are utilized in our density estimation step, combining the proposed TS-LDMF together with the neural ODE [21,22]-based approach is natural and seamless. Our main contributions and novelty can be summarized as follows:
  • In order to obtain low-dimensional intrinsic trajectories associated with walking and running data, we propose a novel method referred to as ’two-stage latent dynamics modeling and filtering’, which combines a latent dynamics modeling stage together with non-linear incremental filtering stage.
  • The proposed method can yield simple and intrinsic representation in latent spaces for walking and running. Providing simple and intrinsic representation in latent spaces for human movements is a great help in a variety of application fields such as the entertainment, healthcare, and medical domains.
  • Our works are based on smartphone data, which ensures easy accessibility and convenient deployment in real applications.
The remaining parts of this paper is organized as follows. In Section 2, after describing some common framework, we present the two-stage latent dynamic modeling and filtering method for the problem of characterizing individual dynamic walking and running patterns. In Section 3, the effectiveness of the proposed TS-LDMF method is demonstrated by experiments, and in Section 4, the usefulness of the proposed TS-LDMF method is discussed. Finally in Section 4, concluding remarks are given togather with issues for future work.

2. Methods

The purpose of this study is to propose a machine learning approach that can characterize low-dimensional dynamic features of individuals while walking and running with smartphone sensors. Our approach yields low-dimensional latent trajectories of human motions by processing high-dimensional raw data from smartphone sensors, as shown in Figure 1.
For characterizing the sequential data from movements in latent feature space, we propose a novel approach, termed a two-stage latent dynamics modeling and filtering (TS-LDMF) method. The two-stage latent dynamics modeling and filtering method combines a latent space modeling stage with a nonlinear filtering stage, for characterizing individual dynamic walking and running patterns by analyzing smartphone sensor data. The block diagram for its workflow is shown in Figure 2.
An established procedure for training TS-LDMF models in experiments is provided in Table 1. In the experiment, we used the built-in sensors of the iPhone. The smartphone unit includes two types of sensors, accelerometer and gyro sensor, which can obtain motion data around three orthogonal axes (x, y, z). Thus, we measured motion data consisting of six features. Furthermore, for additional information on movement intensity, we consider total magnitudes [23] of acceleration and angular velocity as well. After obtaining the motion data and additional intensity features, we have an 8-dimensional feature data set, x t , each time, t (Table 2). In Figure 3, we show the configuration of the smartphone unit in the experiments. In the following subsections, we describe details of the TS-LDMF method.

2.1. Backbone Structure of TS-LDMF

In this subsection, we describe the backbone structure of the proposed TS-LDMF, which the first and second stages of TS-LDMF utilize as a common sub-module. The backbone structure contains a transition network, an emitter network, and a probability distribution of the initial latent variable, exact meanings of which are provided in the following. In the backbone structure, we use the transition network for a process model in low-dimensional latent space, and the emitter network for a measurement model for sensors (e.g., [12,24]). For the emitter and transition networks, we use the multilayer perceptron (MLP) [25] and mixture density network [26], respectively. Under the assumption of the Markov property [12,24] in the latent dynamics, we have the following joint probability distribution for the observations, x 0 : T , and the latent variables, h 0 : T :
p θ ( x 0 : T , h 0 : T ) = p θ ( h 0 ) p θ ( x 0 | h 0 ) t = 1 T p θ ( x t | h t ) p θ ( h t | h t 1 ) ,
where p θ ( h 0 ) , p θ ( x t | h t ) , and p θ ( h t | h t 1 ) stand for the probability distribution of the initial latent variable, the conditional probability distribution for the emitter network, and the conditional probability distribution for the transition network, respectively. Note that the probabilistic model of Equation (1) is based on the key idea that the sequence of the high-dimensional sequential observation, x 0 : T , can be explained by means of the lower-dimensional sequence of the latent variables, h 0 : T , where the h 0 : T are generated via the conditional distribution of the transition network, p θ ( h t | h t 1 ) , and the x 0 : T are generated via the conditional distribution of the emitter network, p θ ( x t | h t ) . In addition, note that in Equation (1), our notation uses θ for all the parameters of the backbone structure. When the joint distribution can be factorized as in Equation (1), the true posterior inference distribution p θ ( h 0 : T | x 0 : T ) can be factorized as follows [12,27]:
p θ ( h 0 : T | x 0 : T ) = p θ ( h 0 | x 0 : T ) t = 1 T p θ ( h t | h t 1 , x t : T ) .
Motivated by the factored representation of Equation (2), the strategy of the variational approximation [27,28] usually approximate the true posterior distribution p θ ( h 0 : T | x 0 : T ) with the variational distribution of the following form [12]:
q ϕ ( h 0 : T | x 0 : T ) = q ϕ ( h 0 | x 0 : T ) t = 1 T q ϕ ( h t | h t 1 , x t : T ) ,
where ϕ stands for parameters of the variational distributions.
As mentioned, Equations (1) and (2) hold true under the assumption of the Markov property. However, since the Markov property is not guaranteed in actual experiments involving sensors, satisfying p θ ( x t , h t | h t 1 ) = p θ ( x t | h t ) p θ ( h t | h t 1 ) exactly is difficult. To alleviate this difficulty, we construct the observation and latent state vectors based on the current and past two feature data sets. More specifically, our observation y t is defined as
y t = [ x t , x t 1 , x t 2 ] ,
where x t is the feature data set of Table 2 at time t, and the corresponding latent state vector z t plays the role of { h t , h t 1 , h t 2 } . Note that as a result of this definition, there is some overlap of information among the three observation vectors, y t , y t 1 , and y t 2 . For the distributions of the transition and emitter networks, we use normal and multinomial distributions, respectively. Based on the new observation and latent state definition, the corresponding equation involving the variational distributions becomes
q ϕ ( z 0 : T | y 0 : T ) = q ϕ ( z 0 | y 0 : T ) t = 1 T q ϕ ( z t | z t 1 , y t : T ) .
In the following Section 2.2 and Section 2.3, we explain how the true posterior distribution p θ ( z 0 : T | y 0 : T ) can be adequately approximated at each stage of TS-LDMF.

2.2. First Stage of TS-LDMF for Modeling Latent Dynamics

The purpose of the first stage of TS-LDMF is to train the backbone structure with the variational approximation strategy. As mentioned, the true posterior distribution p θ ( z 0 : T | y 0 : T ) can be efficiently approximated by the variational distribution in the form of Equation (5). For the variational distributions in the right hand side of the equation, one often uses separate normal distributions for q ϕ ( z 0 | y 0 : T ) and q ϕ ( z t | z t 1 , y t : T ) , i.e.,
q ϕ ( z 0 | y 0 : T ) = N ( z 0 | μ ( y 0 : T ) , Σ ( y 0 : T ) ) ,
q ϕ ( z t | z t 1 , y t : T ) = N ( z t | μ ( z t 1 , y t : T ) , Σ ( z t 1 , y t : T ) ) , t { 1 , , T } ,
where N ( z | μ , Σ ) is the notation for the multivariate normal distribution with the mean vector μ and the covariance matrix Σ . Motivated by the observation [29] that much better training stability is obtained when the variational distribution q ϕ for z t depends exclusively on the data y t : T , we use the modification for q ϕ ( z t | z t 1 , y t : T ) in the first stage so that it be conditioned exclusively on y t : T , and modeling interactions with z t 1 work only through the transition network. Note that under this modification, both q ϕ ( z 0 | y 0 : T ) and q ϕ ( z t | z t 1 , y t : T ) can be implemented with the common form:
q ϕ ( z t | y t : T ) , t { 0 , , T } .
More specifically, our implementation for q ϕ ( z t | y t : T ) uses the hidden state h t r n n of a recurrent neural network structure running backwards in time across y t : T , and for the recurrent neural network structure, we use the gated recurrent units (GRU) [30]. We call the network used for q ϕ in the first stage the encoder network. In the training process of the first stage, we find the parameters θ and ϕ simultaneously by maximizing E L B O ( θ , ϕ ) , the variational lower bound of Equation (9) [17,18]:
log p ( y 0 : T ) E L B O ( θ , ϕ ) = E z 0 : T q ϕ ( z 0 : T | y 0 : T ) [ log p θ ( y 0 : T | z 0 : T ) ] K L ( q ϕ ( z 0 : T | y 0 : T ) p θ ( z 0 : T ) ) .
More precisely, E L B O ( θ , ϕ ) of Equation (9) can be written as follows:
t = 0 T E z t q ϕ ( z t | y t : T ) [ log p θ ( y t | z t ) ] K L ( q ϕ ( z 0 | y 0 : T ) p θ ( z 0 ) ) t = 1 T E z t 1 q ϕ ( z t 1 | y t 1 : T ) [ K L ( q ϕ ( z t | y t : T ) p θ ( z t | z t 1 ) ] .
By denoting y 0 : T 1 and y 1 : T as Y and Y ˜ , respectively, one can see that the training of the first stage can be interpreted as aiming at the following goals (Figure 4):
Y r e c o n Y , Y ˜ r e c o n Y ˜ , and Z ˜ t r a n s Z ˜ ,
where Z = z 0 : T 1 , Z ˜ = z 1 : T , and Z ˜ t r a n s is the random variables produced by means of the transition network. In Equation (11), both Y r e c o n Y and Y ˜ r e c o n Y ˜ mean that reconstructions generated by the emitter network should be close to actual observations so that the log-likelihood of observations be large, while Z ˜ t r a n s Z ˜ means that distributions of latent variables yielded by the encoder network and the transition networks should be close in the sense of Kullback-Leibler divergence [31]. Note that the interpretation shown in Figure 4 is quite general, and may be applicable in other types of variational approaches as well.

2.3. Second Stage of TS-LDMF for Estimating Latent Variables

The proposed TS-LDMF consists of two stages, i.e., the first stage for latent dynamics modeling, and the second stage for estimating latent variables via nonlinear filtering. As indicated in Figure 2, training the first stage of TS-LDMF yields a transition network, an emitter network, and an encoder network, which are for p θ ( z t | z t 1 ) , p θ ( y t | z t ) , and q ϕ ( z t | y t : T ) each time step t, respectively. Note that in the encoder network, the computation of the conditional distribution for z t involves the future sequence of observation, y t : T ; hence the encoder network obtained by training the first stage has a limitation in estimating latent states with sequential data processing. The goal of TS-LDMF is to provide a versatile way for characterizing individual dynamic walking and running patterns so that it can work sequentially and efficiently when characterizing, estimating, and predicting the latent trajectories for movements. With this type of versatility in mind, we introduce the second stage of TS-LDMF for estimating latent variables via sequential data processing. In the second stage, we use a new variational distribution q ψ , which is different from the q ϕ of the first stage, and does not rely on the future sequence of observation. We call the resultant network for q ψ the combiner network, and for its implementation, we use the multilayer perceptron [25]. Our implementation of the combiner network for z t uses the predicted state, predicted variance, and the observation y t as its inputs, which is motivated by the way that the state and covariance are updated in the correction step of linear and extended Kalman filters [32,33]. A flowchart for the filtering performed by the second stage of TS-LDMF is shown in Figure 5. Note that in the training process of the second stage, we optimize the parameters of the combiner network only, and the transition and emitter networks for p θ ( z t | z t 1 ) and p θ ( y t | z t ) remain fixed as provided by the first stage.

3. Experimental Results

In our experiments, we addressed the problem of characterizing individual human motions with smartphone sensors via the proposed two-stage latent dynamics modeling and filtering method. For these motions, we considered walking and running. We believe that the proposed method can be applicable to more types of motions, and we are planning to study its applicability in future continuing research.

3.1. Data Collection

Based on the procedure of Table 1, the experiments were conducted at the Korea University R&D Center with its WiFi networks. In experiments for the paper, we recruited 10 male and 10 female subjects to evaluate the performance of the proposed method properly. Profiles of the recruited subjects are given in Table 3.
During the entire experimental procedure, we used a single smartphone unit: the iPhone SE, one laptop computer (a MacBook Pro), and two applications: Matlab [34] and PyTorch [35] for processing sensor data, and training the TS-LDMF models, respectively. The walking and running data were sent from the smartphone to the computer over the campus WiFi network. Figure 3 shows the configuration of the smartphone in the experiments.
We set the sampling rate for data transmission from the smartphone to the PC at 30 Hz. Training was conducted with sensor data from the smartphone accelerometer and gyro sensors after minmax scaling. The minmax scaling was conducted by importing the MinMaxScaler from sklearn.preprocessing [36]. Note that the scaling part is somewhat user-dependent, because its min and max values should be chosen so that the resultant interval should cover all the subjects’ data. As shown in Figure 3, the smartphone near the left pants pocket, positioned on the side of the leg with a harness and with the screen facing outward. In the experiment, the participants walked and ran a predefined course. We thereby acquired the necessary data for simulation of the proposed method. A detailed protocol for obtaining the data for each subject is as follows:
(a)
Set the predefined course for walking and running.
(b)
Set parameter (sampling rate: 30 Hz, data collection time: 60 s) with help of MATLAB Support Package for Apple iOS Sensors.
(c)
Run Matlab Mobile on the iPhone.
(d)
Connect the iPhone to the desktop on the same Wifi network.
(e)
Position the iPhone to the predefined location and position.
(f)
Instruct the subject to walk on the predefined course.
(g)
Initiate the countdown prior to the data recording.
(h)
Let the subject begin walking before completing the countdown.
(i)
Upon completion of the data recording, have the subject stop.
(j)
Save the recorded data (angular velocity around the x, y, z-direction; acceleration along the x, y, z-direction) and conduct preprocessing for data (total magnitude of angular velocity and total magnitude of acceleration) on the desktop.
(k)
Repeat steps (d) through (j) for running.
In addition, a detailed description of the unit’s feature data set is provided in Table 2. Note that each feature data set for the configuration consists of eight dimensions. For the latent space, we chose R 2 or R 3 for convenience of visualization and easy understanding.

3.2. Experimental Results

This section explains the experimental environment and data for demonstrating the latent space based solutions for characterizing dynamic walking and running patterns.
For demonstrating the latent space based solutions, we used five sets of training data. In all these training data, a sequence of features was obtained from our twenty subjects with a frequency of 30 Hz. Additionally, the training batch size was 64. In all experiments, we constructed the observation of Equation (4) from the current and past two feature data sets; hence, the dimension of the resultant observation each time is 8 × 3 = 24 . Note that some overlap of information exists among the three consecutive observation vectors. Since the sampling rate in this paper is set at 30 Hz, there is no loss of information even with the overlapping information compared to our previous related work [9], where the sampling rate was set at 10 Hz and no overlapping information was allowed.
For splitting data into train and test sets, we used the five-fold cross-validation. More precisely, we utilized the corresponding method of the sklearn library [36], i.e., sklearn.model_selection.KFold(n_splits=5). Note that the method provides train and test indices to split data into train and test sets, and splits the data into k = 5 consecutive folds. Based on the method, each fold was used as a test set whereas the remaining k 1 = 4 folds constituted the training set. Thus, 20% of data was used as a test set while the remaining 80% of data was used for training.
Figure 6 and Figure 7 show the simulation results of the five-fold cross-validation for characterizing individual dynamic walking patterns with the proposed TS-LDMF method, and show the resulting latent trajectories in R 2 for male and female subjects, respectively. Note that in each cross-validation, we considered the setup that the training and test data include only data from one person. The subplots in the figure can be interpreted as follows: in the j-th row, which is for the j-th subject, the i-th subplot shows the latent trajectories obtained from the proposed TS-LDMF method for the i-th experiment, in which the i-th walking data set was used as the test set, and the other four walking data sets were used as the training set for estimating latent trajectories. In each individual subplot, the solid red line indicates a portion of the latent trajectories from the test data sets provided by the proposed method, while the dashed blue lines represents some portion of the latent trajectories of the training data sets. Figure 6 and Figure 7 indicate that the proposed latent space based method worked satisfactorily in characterizing dynamic walking patterns in the latent space. From the cross-validation results, obvious similarities can be seen between the latent trajectory of the test data and that of the training data. For the purpose of characterizing individual dynamic running patterns, we also conducted similar experiments. Figure 8 and Figure 9 show the results of the corresponding five-fold cross-validation for male and female subjects, respectively. As shown in the Figures, the proposed latent space based method worked well for characterizing the dynamic running patterns, and the cross-validation results of Figure 8 and Figure 9 exhibit obvious similarities between the latent trajectories of the training data and those of the test data. In addition, Figure 10, Figure 11, Figure 12 and Figure 13 show the corresponding results for the case with the three dimensional latent space.
Overall, the results of Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13 indicate that the proposed method successfully transformed high-dimensional sequences of noisy observation data from the smartphone sensors to low-dimensional latent trajectories, and the training and test data with their common characteristics in fact shared similar patterns in latent space.

3.3. Performance Comparison

For performance comparison with a conventional approach, we considered incremental principal component analysis [37]. The family of principal component analysis methods such as PCA (principal component analysis), PPCA (probabilistic principal component analysis), and IPCA (incremental principal component analysis) are all important tools for reducing dimensionality, and have often been utilized for problems involving gait and dimension reduction (e.g., [38]).
We compared the results of the proposed method to those of the incremental PCA-based method using the mean squared distance (MSE), which is defined as
M S E = 1 M k = 1 M x k x ^ k 2
where M is the number of test patterns, x k is the kth test pattern, and x ^ k is the reconstructed result for the kth test pattern. Table 4 shows the ratios of M S E I P C A / M S E T S L D M F computed for the test set of the five-fold cross-validation. Ratios being larger than one in the table shows that the proposed method performs better in terms of reconstruction capability, compared to the incremental principal component analysis.
Figure 14 shows latent trajectories obtained by the incremental principal component analysis for the first cross-validation case of the first subject. The corresponding results of Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13 indicate that our results are more smooth and easy to interpret, compared to the IPCA results of Figure 14.
Finally in order to show that the effectiveness of the proposed method does not depend on particular locations of smartphone, we also considered other choice of locations used in related works [39,40,41,42] (Figure 15). Each column of Figure 15 shows the location of smartphone together with the latent trajectories of walking and running for the first cross-validation set of the first subject. From the figure, one can see that regardless of locations, the proposed method can yield reasonable intrinsic latent trajectories for walking and running. Additionally, from the figure, one can conclude that when the location of the smartphone has more movements (e.g., foot or hand), the resultant latent trajectories tend to have more variations compared to the case with less movements (e.g., chest).

4. Discussion and Conclusions

4.1. Discussion

In this paper, we investigated the use of machine learning for characterizing dynamic walking and running patterns with smartphone sensors. The key idea behind the characterizing is that the sequence of the high-dimensional observation can be explained by means of the substantially lower-dimensional sequence of the latent variables. We believe that the key idea is reasonable because the high-dimensional observation in our experiments all originates from human motions, which are intrinsically movements in a three dimensional space. For the task of characterizing dynamic walking and running patterns in a low-dimensional latent feature space, we put forth a novel approach, referred to as two-stage latent dynamics modeling and filtering. Our approach is closely related with the deep Markov model (DMM) approach [12]. The most important difference worth noting is that the proposed method uses the second stage to estimate latent variables via filtering in the training phase. The second stage is critically important, because it ensures that the resultant networks can work in real time. Since the latent trajectories obtained by the proposed TS-LDMF method are somewhat unique for each subject, the proposed TS-LDMF method has the potential value of identifying individual dynamic walking and running patterns with smartphone sensor data. In our opinion, the capability of TS-LDMF for characterizing individual dynamic walking and running patterns can successfully be extended to other types of human motions. Furthermore, we believe that despite specific experimental environment of these experiments for verifying the proposed method, this method can be deployed in existing smartphone systems.
Once training the TS-LDMF models is completed, we can find the normal latent regions for movements based on the training results. For the task of finding the normal latent region for walking or running, we use a modern density estimation approach based on the neural ODE [21,22] in obtaining a probability density for latent samples resulting from training the TS-LDMF models. This ODE-based method [21,22] is particularly attractive because it can parameterize the derivative of the latent state by means of a relatively smaller neural network. Additionally, since the latent samples generated by the training of the two-stage latent dynamics modeling and filtering can be recycled in our density estimation step, these two modules can be combined together seamlessly. The pipeline graph for the combination is shown in Figure 16. Given any observation data, the two-stage latent dynamics modeling and filtering is capable of providing a probability density for their corresponding latent patterns. Consequently, one can obtain the support of the latent objects by thresholding the resultant probability density function. Figure 17 and Figure 18 show how relevant contours for the density of latent patterns in R 2 appeared in the experiments. For these contours for the density, we utilized matplotlib.pyplot.contour [43]. Since it can be quickly noticed if the trajectory deviates from the high-density normal region, the combination may be utilized to anomaly detection problems such as fall detection.

4.2. Conclusions

In this study, we have examined the problem of characterizing individual dynamic walking and running patterns with smartphone sensors. For the sensors, we used built-in sensors in a single smartphone unit positioned near the left pants pocket, and from this unit, acceleration, rate of turn along three perpendicular axes, and their total magnitudes were used as input features.
For characterizing movements, we proposed two-stage latent dynamics modeling and filtering, which can map noisy high-dimensional sequential observation to low-dimensional latent trajectory, and the resultant latent trajectories efficiently found intrinsic characteristics of the users’ dynamic walking and running patterns. The latent trajectories found by the proposed method showed that the low-dimensional latent trajectories associated with dynamic walking and running patterns were smooth, while the original input features from smartphone sensors were often noisy.
For the task of finding the normal latent region for walking or running, we used a modern density estimation approach based on the neural ODE, and utilized latent samples resulting from the two-stage latent dynamics modeling and filtering. Since the latent samples generated from training for two-stage latent dynamics modeling and filtering can be recycled in the task, combining the TS-LDMF and the ODE-based density estimation can be done seamlessly. Future work to be done includes further, more extensive experiments and comparison studies, which might uncover strengths and weaknesses of the proposed approach, as well as further refinements of the proposed method in various directions and with more participants. Examination of different types of neural networks and applications to other kinds of human motions are some of the topics to be covered along these lines. Issues of deploying trained networks into current smartphone systems so that they work in real-time are also reserved for future studies.

Author Contributions

J.P. conceived and designed the methodology of the paper with help of J.L. and J.K.; J.K. and J.L. performed the experiments; J.L. and J.P. wrote the computer program with help of J.K. and S.L.; J.P. wrote the paper with help of J.K., J.L., W.J. and H.K.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government(MSIT). (No. 2017R1E1A1A03070652)

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Sekine, M.; Tamura, T.; Fujimoto, T.; Fukui, Y. Classification of walking pattern using acceleration waveform in elderly people. In Proceedings of the 2000 22nd Annual International Conference of the Engineering in Medicine and Biology Society, Chicago, IL, USA, 23–28 July 2000; Volume 2, pp. 1356–1359. [Google Scholar]
  2. Papagiannaki, A.; Zacharaki, E.I.; Kalouris, G.; Kalogiannis, S.; Deltouzos, K.; Ellul, J.; Megalooikonomou, V. Recognizing Physical Activity of Older People from Wearable Sensors and Inconsistent Data. Sensors 2019, 19, 880. [Google Scholar] [CrossRef] [PubMed]
  3. Jiang, W.; Yin, Z. Human activity recognition using wearable sensors by deep convolutional neural networks. In Proceedings of the 23rd ACM international conference on Multimedia, Brisbane, Australia, 26–30 October 2015; pp. 1307–1310. [Google Scholar]
  4. Wang, N.; Ambikairajah, E.; Lovell, N.H.; Celler, B.G. Accelerometry based classification of walking patterns using time-frequency analysis. In Proceedings of the 2007 29th Annual International Conference of the Engineering in Medicine and Biology Society, Lyon, France, 23–26 August 2007; Volume 5, pp. 4899–4902. [Google Scholar]
  5. Mubashir, M.; Shao, L.; Seed, L. A survey on fall detection: Principles and approaches. Neurocomputing 2013, 100, 144–152. [Google Scholar] [CrossRef]
  6. Delahoz, Y.; Labrador, M. Survey on fall detection and fall prevention using wearable and external sensors. Sensors 2014, 14, 19806–19842. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, T.; Wang, J.; Xu, L.; Liu, P. Fall detection by wearable sensor and one-class SVM algorithm. In Intelligent Computing in Signal Processing and Pattern Recognition; Springer: Berlin/Heidelberg, Germnay, 2006. [Google Scholar]
  8. Habib, M.; Mohktar, M.; Kamaruzzaman, S.; Lim, K.; Pin, T.; Ibrahim, F. Smartphone-based solutions for fall detection and prevention: Challenges and open issues. Sensors 2014, 14, 7181–7208. [Google Scholar] [CrossRef] [PubMed]
  9. Kim, T.; Park, J.; Heo, S.; Sung, K.; Park, J. Characterizing Dynamic Walking Patterns and Detecting Falls with Wearable Sensors Using Gaussian Process Methods. Sensors 2017, 17, 1172–1185. [Google Scholar] [CrossRef] [PubMed]
  10. Jolliffe, I. Principal Component Analysis; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
  11. Schölkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2001. [Google Scholar]
  12. Krishnan, R.G.; Shalit, U.; Sontag, D. Structured inference networks for nonlinear state space models. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  13. Wu, H.; Mardt, A.; Pasquali, L.; Noe, F. Deep Generative Markov State Models. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2018; pp. 3975–3984. [Google Scholar]
  14. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  15. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  16. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. The International Conference on Learning Representations (ICLR) 2014. Available online: https://arxiv.org/pdf/1312.6114v10.pdf (accessed on 1 May 2014).
  18. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  19. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2014; pp. 2672–2680. [Google Scholar]
  20. Goodfellow, I. NIPS 2016 tutorial: Generative adversarial networks. arXiv 2016, arXiv:1701.00160. [Google Scholar]
  21. Chen, T.Q.; Rubanova, Y.; Bettencourt, J.; Duvenaud, D.K. Neural ordinary differential equations. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2018; pp. 6571–6583. [Google Scholar]
  22. Grathwohl, W.; Chen, R.T.; Bettencourt, J.; Duvenaud, D. Scalable Reversible Generative Models with Free-form Continuous Dynamics. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  23. Karantonis, D.M.; Narayanan, M.R.; Mathie, M.; Lovell, N.H.; Celler, B.G. Implementation of a real-time human movement classifier using a triaxial accelerometer for ambulatory monitoring. IEEE Trans. Inf. Technol. Biomed. 2006, 10, 156–167. [Google Scholar] [CrossRef] [PubMed]
  24. Karl, M.; Soelch, M.; Bayer, J.; van der Smagt, P. Deep variational bayes filters: Unsupervised learning of state space models from raw data. arXiv 2016, arXiv:1605.06432. [Google Scholar]
  25. Haykin, S. Neural Networks: A Comprehensive Foundation; Prentice Hall: Upper Saddle River, NJ, USA, 1994. [Google Scholar]
  26. Bishop, C.M. Mixture Density Networks; Technical Report NCRG/4288; Aston University: Birmingham, UK, 1994. [Google Scholar]
  27. Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  28. Fox, C.W.; Roberts, S.J. A tutorial on variational Bayesian inference. Artif. Intell. Rev. 2012, 38, 85–95. [Google Scholar] [CrossRef]
  29. Schmidt, F.; Hofmann, T. Deep State Space Models for Unconditional Word Generation. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Foundation, Inc.: San Diego, CA, USA, 2018; pp. 6161–6171. [Google Scholar]
  30. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  31. Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  32. Brown, R.G.; Hwang, P.Y. Introduction to Random Signals and Applied Kalman Filtering; Wiley: New York, NY, USA, 1992; Volume 3. [Google Scholar]
  33. Kim, P. Kalman Filter for Beginners: With MATLAB Examples; CreateSpace: Scotts Valley, CA, USA, 2011. [Google Scholar]
  34. MATLAB 2019a; The MathWorks, Inc.: Natick, MA, USA, 2019.
  35. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in PyTorch. In Proceedings of the NIPS 2017 Autodiff Workshop, Long Beach, CA, USA, 9 December 2017. [Google Scholar]
  36. Scikit-learn: Machine Learning in Python. Available online: http://scikit-learn.org/stable/ (accessed on 11 June 2018).
  37. Ross, D.A.; Lim, J.; Lin, R.S.; Yang, M.H. Incremental learning for robust visual tracking. Int. J. Comput. Vis. 2008, 77, 125–141. [Google Scholar] [CrossRef]
  38. Matsushima, A.; Yoshida, K.; Genno, H.; Ikeda, S.I. Principal component analysis for ataxic gait using a triaxial accelerometer. J. Neuroeng. Rehabil. 2017, 14, 37. [Google Scholar] [CrossRef] [PubMed]
  39. Zhu, Q.; Chen, Z.; Soh, Y.C. Smartphone-based human activity recognition in buildings using locality-constrained linear coding. In Proceedings of the 2015 IEEE 10th Conference on Industrial Electronics and Applications (ICIEA), Auckland, New Zealand, 15–17 June 2015; pp. 214–219. [Google Scholar]
  40. Lemoyne, R.; Mastroianni, T. Implementation of a Smartphone as a Wireless Accelerometer Platform for Quantifying Hemiplegic Gait Disparity in a Functionally Autonomous Context. J. Mech. Med. Biol. 2018, 18, 1850005. [Google Scholar] [CrossRef]
  41. del Rosario, M.; Redmond, S.; Lovell, N. Tracking the evolution of smartphone sensing for monitoring human movement. Sensors 2015, 15, 18901–18933. [Google Scholar] [CrossRef] [PubMed]
  42. Shanahan, C.J.; Boonstra, F.; Cofré Lizama, L.E.; Strik, M.; Moffat, B.A.; Khan, F.; Kilpatrick, T.J.; van der Walt, A.; Galea, M.P.; Kolbe, S.C.; et al. Technologies for advanced gait and balance assessments in people with multiple sclerosis. Front. Neurol. 2018, 8, 708. [Google Scholar] [CrossRef] [PubMed]
  43. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
Figure 1. Two-stage latent dynamics modeling and filtering (TS-LDMF) for characterizing dynamic walking and running patterns in a low-dimensional latent space.
Figure 1. Two-stage latent dynamics modeling and filtering (TS-LDMF) for characterizing dynamic walking and running patterns in a low-dimensional latent space.
Sensors 19 02712 g001
Figure 2. Block diagram for two-stage latent dynamics modeling and filtering (TS-LDMF) approach.
Figure 2. Block diagram for two-stage latent dynamics modeling and filtering (TS-LDMF) approach.
Sensors 19 02712 g002
Figure 3. Illustrations of smartphone location and position in experiments: (a) side view, (b) front view.
Figure 3. Illustrations of smartphone location and position in experiments: (a) side view, (b) front view.
Sensors 19 02712 g003
Figure 4. Interpretation for training the first stage of TS-LDMF.
Figure 4. Interpretation for training the first stage of TS-LDMF.
Sensors 19 02712 g004
Figure 5. Flowchart for the filtering performed by the second stage of TS-LDMF.
Figure 5. Flowchart for the filtering performed by the second stage of TS-LDMF.
Sensors 19 02712 g005
Figure 6. Five-fold cross validation results: Latent trajectories in R 2 for walking of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 6. Five-fold cross validation results: Latent trajectories in R 2 for walking of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g006
Figure 7. Five-fold cross validation results: Latent trajectories in R 2 for walking of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 7. Five-fold cross validation results: Latent trajectories in R 2 for walking of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g007
Figure 8. Five-fold cross validation results: Latent trajectories in R 2 for running of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 8. Five-fold cross validation results: Latent trajectories in R 2 for running of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g008
Figure 9. Five-fold cross validation results: Latent trajectories in R 2 for running of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 9. Five-fold cross validation results: Latent trajectories in R 2 for running of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g009
Figure 10. Five-fold cross validation results: Latent trajectories in R 3 for walking of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 10. Five-fold cross validation results: Latent trajectories in R 3 for walking of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g010
Figure 11. Five-fold cross validation results: Latent trajectories in R 3 for walking of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 11. Five-fold cross validation results: Latent trajectories in R 3 for walking of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g011
Figure 12. Five-fold cross validation results: Latent trajectories in R 3 for running of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 12. Five-fold cross validation results: Latent trajectories in R 3 for running of 10 male participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g012
Figure 13. Five-fold cross validation results: Latent trajectories in R 3 for running of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 13. Five-fold cross validation results: Latent trajectories in R 3 for running of 10 female participants. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g013
Figure 14. Latent trajectories obtained by the incremental principal component analysis method for the first cross-validation set of the first subject: (a) 2-dim latent trajectories for walking, (b) 2-dim latent trajectories for running, (c) 3-dim latent trajectories for walking, (d) 3-dim latent trajectories for running. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 14. Latent trajectories obtained by the incremental principal component analysis method for the first cross-validation set of the first subject: (a) 2-dim latent trajectories for walking, (b) 2-dim latent trajectories for running, (c) 3-dim latent trajectories for walking, (d) 3-dim latent trajectories for running. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g014
Figure 15. Location of smartphone together with the latent trajectories in R 2 and R 3 for walking and running for the first cross-validation of the first subject: (a) Thigh, (b) foot, (c) hand, (d) chest. Solid red lines and dashed blue lines for test data and training data, respectively.
Figure 15. Location of smartphone together with the latent trajectories in R 2 and R 3 for walking and running for the first cross-validation of the first subject: (a) Thigh, (b) foot, (c) hand, (d) chest. Solid red lines and dashed blue lines for test data and training data, respectively.
Sensors 19 02712 g015
Figure 16. Pipeline graph for the combination of TS-LDMF and density estimation for normal latent region of walking or running.
Figure 16. Pipeline graph for the combination of TS-LDMF and density estimation for normal latent region of walking or running.
Sensors 19 02712 g016
Figure 17. Density estimation results in R 2 for walking of 20 subjects.
Figure 17. Density estimation results in R 2 for walking of 20 subjects.
Sensors 19 02712 g017
Figure 18. Density estimation results in R 2 for running of 20 subjects.
Figure 18. Density estimation results in R 2 for running of 20 subjects.
Sensors 19 02712 g018
Table 1. An established procedure for training TS-LDMF models in experiments.
Table 1. An established procedure for training TS-LDMF models in experiments.
1: Obtain the training data for each category of motions (walking or running), and for each subject.
2: Obtain the test data for each category of motions (walking or running), and for each subject.
3: Train the first stage of TS-LDMF, and fix its transition and emitter networks after the training is completed.
4: Train the second stage of TS-LDMF, and fix its combiner network after the training is completed.
5: Find latent trajectories corresponding to the training and test data for each category of motions (walking or running), and for each subject.
6: Validity check: If the the obtained latent trajectories are not satisfactory, repeat the the above steps until satisfactory.
7: Report TS-LDMF results (i.e., the transition, emitter, and combiner networks), and the latent trajectories for each class of motions (walking or running) and for each subject.
Table 2. Smartphone unit’s feature data set.
Table 2. Smartphone unit’s feature data set.
NotationMeaning
ω x , ω y , ω z Angular velocities around the x , y , z -directions, respectively
ω T Square root of the sum of squares of angular velocities, ω x 2 + ω y 2 + ω z 2
A x , A y , A z Accelerations along the x , y , z -directions, respectively
A T Square root of the sum of squares of accelerations, A x 2 + A y 2 + A z 2
Table 3. Profiles of the recruited subjects.
Table 3. Profiles of the recruited subjects.
SubjectsGenderAge (yrs)Height (cm)Weight (kg)
subject 1Male3517462
subject 2Male2517580
subject 3Male2616756
subject 4Male2818584
subject 5Male5817264
subject 6Male3717070
subject 7Male4916585
subject 8Male28181100
subject 9Male3117080
subject 10Male5917267
subject 11Female2916358
subject 12Female4716758
subject 13Female5615863
subject 14Female3615347
subject 15Female2316355
subject 16Female2216048
subject 17Female2115954
subject 18Female2116548
subject 19Female2416368
subject 20Female2216152
Average33.85167.1564.95
Table 4. Comparison of average mean squared distance (MSE) ratios, M S E I P C A / M S E T S L D M F , (ICPA = incremental principal component analysis) after reconstruction.
Table 4. Comparison of average mean squared distance (MSE) ratios, M S E I P C A / M S E T S L D M F , (ICPA = incremental principal component analysis) after reconstruction.
2-dim, walk2-dim, run3-dim, walk3-dim, run
Male subjects1.011.751.211.60
Female subjects1.331.691.361.51
Average1.171.721.291.56

Share and Cite

MDPI and ACS Style

Kim, J.; Lee, J.; Jang, W.; Lee, S.; Kim, H.; Park, J. Two-Stage Latent Dynamics Modeling and Filtering for Characterizing Individual Walking and Running Patterns with Smartphone Sensors. Sensors 2019, 19, 2712. https://doi.org/10.3390/s19122712

AMA Style

Kim J, Lee J, Jang W, Lee S, Kim H, Park J. Two-Stage Latent Dynamics Modeling and Filtering for Characterizing Individual Walking and Running Patterns with Smartphone Sensors. Sensors. 2019; 19(12):2712. https://doi.org/10.3390/s19122712

Chicago/Turabian Style

Kim, Jaein, Juwon Lee, Woongjin Jang, Seri Lee, Hongjoong Kim, and Jooyoung Park. 2019. "Two-Stage Latent Dynamics Modeling and Filtering for Characterizing Individual Walking and Running Patterns with Smartphone Sensors" Sensors 19, no. 12: 2712. https://doi.org/10.3390/s19122712

APA Style

Kim, J., Lee, J., Jang, W., Lee, S., Kim, H., & Park, J. (2019). Two-Stage Latent Dynamics Modeling and Filtering for Characterizing Individual Walking and Running Patterns with Smartphone Sensors. Sensors, 19(12), 2712. https://doi.org/10.3390/s19122712

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop