Next Article in Journal
Comparison of Conventional and Variable Borehole Heat Exchangers for Use in a Desiccant Assisted Air Conditioning System
Next Article in Special Issue
Application of Lifecycle Measures for an Integrated Method of Environmental Sustainability Assessment of Radio Frequency Identification and Wireless Sensor Networks
Previous Article in Journal
npTrack: A n-Position Single Axis Solar Tracker Model for Optimized Energy Collection
Previous Article in Special Issue
Future Sensors for Smart Objects by Printing Technologies in Industry 4.0 Scenario
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Acceleration Feature Extraction of Human Body Based on Wearable Devices

1
School of Computer Science and Technology, China University of Mining and Technology, Xuzhou 221000, China
2
Library, China University of Mining and Technology, Xuzhou 221000, China
3
Engineering Research Center of Mine Digitalization of Ministry of Education, Xuzhou 221000, China
4
Department of Information Security Engineering, Soonchunhyang University, Asan 31538, Korea
5
Faculty of Engineering and Architecture, Kore University of Enna, 94100 Enna, Italy
*
Author to whom correspondence should be addressed.
Energies 2021, 14(4), 924; https://doi.org/10.3390/en14040924
Submission received: 13 January 2021 / Revised: 4 February 2021 / Accepted: 4 February 2021 / Published: 10 February 2021
(This article belongs to the Special Issue Measurement Applications in Industry 4.0)

Abstract

:
Wearable devices used for human body monitoring has broad applications in smart home, sports, security and other fields. Wearable devices provide an extremely convenient way to collect a large amount of human motion data. In this paper, the human body acceleration feature extraction method based on wearable devices is studied. Firstly, Butterworth filter is used to filter the data. Then, in order to ensure the extracted feature value more accurately, it is necessary to remove the abnormal data in the source. This paper combines Kalman filter algorithm with a genetic algorithm and use the genetic algorithm to code the parameters of the Kalman filter algorithm. We use Standard Deviation (SD), Interval of Peaks (IoP) and Difference between Adjacent Peaks and Troughs (DAPT) to analyze seven kinds of acceleration. At last, SisFall data set, which is a globally available data set for study and experiments, is used for experiments to verify the effectiveness of our method. Based on simulation results, we can conclude that our method can distinguish different activity clearly.

Graphical Abstract

1. Introduction

In recent years, wearable devices provide great convenience for the collection of human body acceleration data. With the rapid development of pervasive computing, the research of human activity recognition and behavior monitoring have attracted more and more attention [1,2]. The study of human behavior monitoring mainly includes behavior perception, data collection, behavior modeling and behavior analysis [3,4,5,6]. Currently, there are two main forms of human body monitoring:
(1)
Video based body monitoring. The behavior monitoring method based on video images is used to recognize human activities by observing the image sequence taken by the camera [7,8,9]. However, visual tools such as cameras are usually fixed and are more suitable for indoor use. There are many limitations to the use of video image-based behavior monitoring for behaviors that penetrate indoors and outdoors and in different locations. For example, it is difficult to deploy and can be easily blocked by objects. Its identification and data processing methods rely heavily on the external environment and are highly intrusive to the privacy of human activities. In addition, if a more accurate identification rate is required, the requirements for data sources are higher.
(2)
Sensor based body monitoring [10,11]. With the development of micromachines, sensors can sense more and more content with low costs. Wearable devices can be attached to the human body and move with people, thus providing continuous monitoring without interfering with the normal activities of the wearer. Considering this advantage, many researchers are more inclined to use sensors as human body data acquisition tools. Wearable devices are favored by users because of their compactness, lightness, and ability to continuously monitor human behavior data.
Wearable devices are portable devices that can be attached to the human body, which can monitor the acceleration, position, heart rate and other parameters. It does not affect the wearer’s daily activities but can provide human parameters in real time. Ling B. et al. [12] recognized 20 human behaviors with the recognition rate 84% under unsupervised condition. This is the first use of multiple sensors for behavior recognition research work. Priyanka et al. [13] build a remote heart disease monitoring system. When the relevant indicators are abnormal, the patients can be rescued in time. Chao et al. [14] used distributed sensor networks for human motion recognition and proposes a linear settlement method to process sensor network data. However, there are many sensors and large amount of data, so the real-time performance is not satisfactory. Maekawa T. and Cheng J. [15,16] used the accelerometer and gyroscope sensor of mobile phone to collect human motion data, analyzes the time domain and frequency domain of the original data, and extracts the relevant eigenvalues. Then the J48 decision-maker combined with Markov model is used to get the recognition results. The method can identify people going upstairs, downstairs, running, walking, static and other states. However, the algorithm can only recognize a certain state of a person in a period of time, it can’t count the number of actions.
Feature extraction methods have direct impacts on the recognition accuracy [17]. Human behavior recognition can be considered as the classification of time-varying data. The methods of human behavior recognition can be divided into three types [18]: template matching, statistical pattern recognition and semantic description. The method based on template matching is an early method of human behavior recognition. It is a good choice when there is not a large number of samples for training [19]. Dynamic time warping algorithm (DTW) [20] is a common technology in template matching method which is applied to gait recognition. Compared with the template matching algorithm, the recognition accuracy of statistical pattern recognition method is higher. Statistical pattern recognition method is usually used for human behavior recognition based on acceleration sensor [21]. The commonly used statistical recognition methods include decision tree, k-nearest neighbor, Bayes, SVM, neural network and hidden Markov model (HMM) [22,23].
With the developments of behavior recognition technology, classification technology also develops, and different classification technologies correspond to different recognition performance. Foody G. M. et al. [24] adopts decision tree classification algorithm. The algorithm is suitable for processing non- numerical data, but not good at continuous data type. Moreover, the self-learning ability of the algorithm is poor [25], so it needs to customize different logic rules for each people, and the amount of calculation is also relatively large. When the number of recognition types increases, the rules will be more complex, which greatly increases the error rate of model recognition [26,27,28].
Although the feature extraction method has achieved good results, there are still many problems to be solved: (1) Even if the human body is performing the same activity, the data collected by wearable devices change with different locations. Therefore, it is necessary to understand the influence of their location on the data collection, which is important for the human behavior recognition; (2) Feature extraction is critical to human behavior recognition, and the extraction of effective feature values is based on a lot of prior experience. So far, most of the feature value extraction is still done manually and subjectively by researchers, which has low efficiency and reliability; (3) The data collection, processing and other tasks require great energy consumption. At present, many algorithms are still based on multi-sensor. Therefore, it is necessary to study the acceleration feature extraction based on single sensor.
This paper mainly studies the method of extracting the characteristic value of human motion acceleration. By extracting the characteristic value of the acceleration signal collected by the sensor, the preliminary preparation for human behavior recognition is made. We combine the genetic algorithm with the Kalman filtering algorithm. By using the genetic algorithm to encode the parameters in the Kalman filter, we proposed KGA algorithm. Then, we define Standard Deviation (SD), Interval of Peaks (IoP) and Difference between Adjacent Peaks and Troughs (DAPT) to distinguishing the feature value of different motions.
The rest of the paper is organized as follows: Datasets and acceleration feature extraction method are proposed in Section 2. In Section 3, some experiments have been done to verify the analysis results. In Section 4, we make some conclusions.

2. Methods

In the process of human behavior recognition, firstly, the original data of human body acceleration is obtained from the human body sensor, and then a series of processing is carried out on the original data to generate the behavior feature vector. The processing methods include filtering, smoothing, segmentation, data feature extraction, etc. Then the classifier is used to classify the behavior feature vector. Although the data processing methods of different recognition methods are different, their basic system framework is roughly the same from the macro perspective, and they are operated on the basis of sensor signal sequence.

2.1. Datasets

The difference of human body movements, number of sensor nodes and sensor placement produces different data sets. In order to facilitate the algorithm design and analysis, researchers have created some standard data sets for this issue. In this paper, we chose the globally available SisFall dataset for study and experiments. The SisFall dataset used a gyroscope and two acceleration sensors to obtain data from a sample of 4, 505 activities from 38 subjects. As shown in Figure 1, the wearable device is fixed at the waist of the tester, because the waist is easy to wear. The three-axis acceleration sensor can measure the acceleration in three directions, the front direction of the subject is the positive direction of X axis, the gravity direction is the positive direction of Y axis and the right side of the subject is the positive direction of Z axis. The data in SisFall are listed in Table 1 [29].

2.2. Pre-Processing

The pre-processing of data is very important to the performance of the acceleration feature extraction method. The pre-processing must be simple and efficient. In order to remove the high frequency noise superimposed on the source data, we use the Butterworth filter to pre-processing the source data.
The system function G ( z ) of the Butterworth filter can be calculated by Equation (1):
G ( z ) = i = 0 M b i z i 1 i = 1 N a i z i
The cut-off frequency is calculated by Equation (2):
F p 10 0.1 R p 1 2 N F c F p 10 0.1 A s 1 2 N
where R p is the maximum allowable attenuation of the passband, A s is the minimum attenuation required for the stopband, and F p is the value after the precorrection of the passband cutoff frequency.
Then, according to the value of the filter order, the normalized system function G a n ( s ) is calculated in Equation (3):
G a n ( s ) = d 0 1 + a 1 s + a 2 s 2 + + s N = d 0 ( s s 1 ) ( s s 2 ) ( s s N )
When Ω = 0 , let | G a n ( s ) | = 1 to determine the value of d 0 , and when a 0 = 1 , let d 0 = a 0 = 1 .
At last, let s = s / F c , and convert it into a system function G a ( s ) of a low pass filter:
G a ( s ) = G a n ( s = s / F c ) = 1 1 + a 1 s F c + a 2 ( s F c ) 2 + + ( s F c ) N
The process of filtering with Butterworth filters is the process of solving Equation (5):
k = 0 N a k y ( n k ) = m = 0 M b m x ( n m )
In Equation (5), a k and b m are the system array of numerator and denominator of system function G ( z ) , sequence x ( n ) is the signal sequence before filtering, and sequence y ( n ) is the signal sequence after filtering. When a 0 = 1 ,   x ( n ) and y ( n ) have the same length. The above equation is simplified as Equation (6):
y ( n ) = m = 0 M b m x ( n m ) k = 1 N a k x ( n k )
When k < 0 , x ( k ) = 0 , y ( k ) = 0 , the value of the sequence y ( n ) can be obtained after iteration. Parameter settings of Butterworth filter is shown in Table 2 [29,30].

2.3. KGA Algorithm

As the Butterworth filter cannot remove all noise points, to further smooth the source data, we combine the genetic algorithm with the Kalman filtering algorithm. By using the genetic algorithm to encode the parameters in the Kalman filter, we proposed KGA algorithm. This method consists of three steps: initialization stage, the optimization stage and the result output stage, and its flow is shown in Figure 2.

2.3.1. Initial Stage

In the initial operation of Kalman filtering algorithm, the parameters of Kalman filtering algorithm are generated in a random manner, which allows the subsequent genetic algorithm to encode the initial values. The calculation process of the Kalman filter algorithm [31] is as follows:
First of all, we introduce a linear stochastic differential equation, as shown in Equation (7):
Y ( k ) = M Y ( k 1 ) + N U ( K ) + φ ( k )
where, Y ( k ) is the system state at time k , U ( k ) is the control quantity of system at k time, M , N is the system parameter value and φ ( k ) represents process noise. The system measurement value G ( k ) at k time can be calculated by Equation (8):
G ( k ) = I Y ( k ) + θ ( k )
where, I is the measurement system parameter, θ ( k ) is the measurement noise. Let the covariance of φ ( k ) and θ ( k ) be Q and R .
Suppose that the current system state is k , according to system model, the current system state can be predicted from the state of the system at the last moment, according to Equation (9),
Y ( k | k 1 ) = M Y ( k 1 | k 1 ) + N U ( k )
where, Y ( k | k 1 ) is the predicted result using the previous state and Y ( k 1 | k 1 ) is the optimal result of the previous state. U ( k ) is the control quantity of the current state which can be set to zero if there is no control quantity.
The covariance C corresponding to Y ( k | k 1 ) should also be updated after updating the system results, as shown in Equation (10):
C ( k | k 1 ) = M C ( k 1 | k 1 ) M + Q
where, C ( k | k 1 ) is the covariance value of Y ( k | k 1 ) , accordingly, C ( k 1 | k 1 ) is the covariance value of Y ( k 1 | k 1 ) and M is the transformation matrix of M .
Equations (9) and (10) are predicted values of the current system. The optimal estimates of the current state can be derived by combining the predicted values with the measured values of the current state:
Y ( k | k ) = Y ( k | k 1 ) + K g ( k ) ( G ( k ) I Y ( k | k 1 ) )
where, K g ( k ) is the Kalman gain, as shown in Equation (12):
K g ( k ) = C ( k | k 1 ) I [ I C ( k | k 1 ) I + R ] 1
Finally, the covariance value needs to be updated so that the Kalman filter will keep running until the end of the system process. The updated equation is shown in Equation (13):
C ( k | k 1 ) = [ E K g ( k ) I ] C ( k | k 1 )
where, E is the unit matrix.
The prediction results of initial stage are taken as the input values of fitness function in the next stage.

2.3.2. Optimization Stage

In this stage, genetic algorithm is used to optimize each parameter of Kalman filtering algorithm, each individual in the population contains all the parameters of Kalman filtering algorithm. In this paper, the fitness value is represented by the difference between the source data and the output value of initial Kalman filter. The process of genetic algorithm [14] is as follows:
(a)
Selection operation. Select the best individuals from all the generated populations or generate new populations for inheritance, then evaluate them primarily by the fitness F i of each population. Make selection by method of roulette, so that the selection probability P i of each population individual can be expressed as:
P i = F i i = 1 N F i
(b)
Crossover operation. Randomly select two population individuals for exchange and combination. Suppose the two populations are x and y respectively, and their crossing at i position can be expressed as:
{ x i = x i r a n d ( ) + y i ( 1 r a n d ( ) ) x i = y i r a n d ( ) + x i ( 1 r a n d ( ) )
where, r a n d ( ) is a random number between 0 and 1.
(c)
Mutation operation. One individual population is selected randomly from all the generated populations for mutation operation. The variation of individual x at i position can be expressed as:
If rand() > 0.5
x i = x i + ( x i x max ) ( 1 s S ) 2 r a n d ( )
If rand() = 0.5 or rand() < 0.5
x i = x i + ( x i x min ) ( 1 s S ) 2 r a n d ( )
where, x m a x and x m i n represent the upper and lower bounds of an individual separately, s and S is the current hereditary algebra and the total hereditary algebra.
The genetic algorithm calculates the new fitness value after mutation operation. If the genetic algebra reached the setting values, the optimization result of Kalman filter can be got. Otherwise, the algorithm will continue the previous loop operation.

2.4. Feature Extraction

Feature extraction is very important in human behavior monitoring. Its goal is to extract parameters that can identify human behavior. Generally speaking, two methods have been proposed to extract features from time series data: statistical method and structural method. Statistical methods use quantitative features of data to extract features, while structural methods consider the relationship between data [29,30,32].
The feature values used in human behavior monitoring methods are extracted from accelerometers, gyroscopes, pressure sensors and so on, which are widely used in human behavior monitoring methods. However, it is still a problem that how to select the characteristic value to ensure the most convenient data processing and achieve the optimal recognition effect for the specific human activity to be monitored.
This paper only considers the acceleration changes of a certain action. First of all, the acceleration amplitude was used to describe different body motion. Since accelerations are vectors, their superposition value at a certain moment can be calculated by Equation (18):
S a c c = a c c x 2 + a c c y 2 + a c c z 2
where, a c c x ,   a c c y ,   a c c z represents the acceleration value in the x axis, y axis, and z axis directions respectively. Taking D01 for example, as shown in Figure 3, the horizontal ordinate indicates time and the vertical ordinate indicates the acceleration value.
Other repetitive motion such as walking and running are all regular exercises. The graphical form of their acceleration data is similar to Figure 3. The difference lies in the intensity of exercise, stride size and stepping frequency. So, we propose Standard Deviation (SD), Interval of Peaks (IoP) and Difference between Adjacent Peaks and Troughs (DAPT) to distinguishing the feature value of different motions.

2.4.1. Standard Deviation

The degree of dispersion of the dataset can be expressed by the standard deviation. When human moves, the acceleration will increase accordingly, and the corresponding standard deviation will also have different values due to different types of motion. Therefore, using the standard deviation can clearly distinguish the static state of the human body from the motion state. We can calculate standard deviation by Equation (19):
S D = 1 N i = 1 N ( x i μ ) 2
Among them, N is the number of data, and μ is the mean value of all the data.

2.4.2. Interval of Peaks

Since the human body’s stride frequency changes with different activities, such as going up and down stairs. Therefore, the interval between two adjacent peaks is used for distinguishing the kind of motion. As shown in Equation (20):
I o P = T t T t k
where, T t k is the time when a certain wave peak appears, T t is the time when the next wave peak adjacent to the previous peak and k is the number of sampling points of adjacent peaks.

2.4.3. Difference between Adjacent Peaks and Troughs

The amplitude of human body acceleration changes with different motions. The requirement of real-time is also very high. When extracting feature points, real-time requirements are very important. Therefore, this method will synthesize the extreme points of S a c c . So, we regard the peak and trough value as critical parameters. In the actual data processing process, when a new sampled value is received, it is compared with the previous one. If the turning point doesn’t appear, it continues to observe. If there appears a turning point, then this point is the extreme point. We consider the range between adjacent peaks and troughs of S a c c t as the extreme point, which can be calculated in Equation (21):
D A P T = S a c c t S a c c t k
Among them, S a c c t is the value of a certain wave peak or trough, S a c c t k is the value of the adjacent wave trough and peak, and k is the time between adjacent peak and trough.

3. Results and Discussion

3.1. Analysis of Filtering Algorithms

In order to verify the validity of KGA, we compared KGA with the Kalman filtering algorithm. These two methods are implemented on the SisFall dataset that shown in Figure 4. The source data range from −5 g to 5 g, and the fluctuation is very large.
We first use Butterworth filter to deal with the data in Figure 3, and the result is shown in Figure 5.
As can be seen from the Figure 5, after the Butterworth filter, the high-frequency data has been removed and the low-pass filtered data become flat. The data range from −2 g to 2 g, and the data distribution is more concentrated.
If the Kalman filtering algorithm is not optimized with genetic algorithm, the result of its operation is shown in Figure 6. While the result of the GA algorithm is decoded, and the optimal parameter value obtained is added to the Kalman filtering algorithm formula, the filtering result of this algorithm is shown in Figure 7.
As can be seen from the Figure 7, after the low-pass filtered data is processed again, the data range of the two is changed from [−2 g, 2 g] to [−1 g, 0.5 g], the data distribution is more compact, and the acceleration values in the three directions are more distinct. However, the data in Figure 6 fluctuates significantly more and contains more outliers, while the data in Figure 7 is smoother.

3.2. Analysis of Running Activity

The simulation takes SisFall as dataset. Figure 8 and Figure 9 are the comparison of jogging and fast running. We randomly select an object from the data set and analyze its motion parameters. As can be seen in Figure 8 and Figure 9, the S a c c , S D , D A P T waveforms produced by jogging and fast running are very similar. Since running is a cyclical exercise, the S a c c of both of them have regular small peaks and the values are similar. Possible reasons are as follows: the difference between the speeds of jogging and fast running is not obvious enough; wearable device is not firmly worn. However, by analyzing the standard deviation of these two waveforms, it can be seen that the S D produced by fast running is significantly greater than that produced by slow running. In addition, due to that human body runs faster than jogging, correspondingly, the stepping frequency of fast running is higher than that of jogging, the I o P produced by fast running activities is less than that by jogging activities.

3.3. Analysis of Going up and down Stairs

Going up and down the stairs is a repetitive motion. Figure 10 and Figure 11 are the comparison of slow up and down stairs and fast up and down stairs. By analyzing Figure 10 and Figure 11, the S a c c generated by slow up and down stairs are significantly lower than that by the fast up and down stairs. It can be seen from the Figure 10 that the value of S D generated by slowly going up and down stairs does not exceed 0.1 g, while the S D generated by going up and down the stairs quickly is slightly larger, not exceeding 0.3 g. At the same time, the D A P T of former is also smaller than that of latter. Similar to walking activities, since the human body is slower when going up and down stairs at a slow speed, correspondingly, the stepping frequency of this activity is lower, in terms of data, the I o P is larger.

3.4. Analysis of Sit-Up Activity

This experiment takes D07 and D08 in Table 1 as examples to analyze the sit-up type activities, as shown in Figure 12. It can be seen from the Figure 13 that their S a c c have large fluctuations, but due to the different speeds, the S a c c of former has a smaller value. At the same time, there are two large fluctuations in the standard deviation of these two activities. Corresponding to the two movements of the human body, “sit down” and “get up”, since the slow activity is relatively stable. In terms of data, the S D of former is much smaller than that of latter. The difference reached 10 times the order of magnitude. This is because when the human body quickly sits down and rises quickly from a half-height chair, the motion range is larger, the action is more violent, and the corresponding acceleration is greater, and D A P T is also greater.

3.5. Analysis of Jumping Activity

Figure 14 is the analysis of jump movement. it can be found that the S a c c , S D , I o P and D A P T waveforms produced by the jumping motion are very similar to the sit-up motion. Both have large peaks and flat areas, especially the peak interval values are very similar. However, there are obvious differences in certain characteristics between the two actions of sit-up and jumping: although the two are both highly volatile activities, the S D and D A P T are different, because the former has a smoother speed and a smaller acceleration change. Correspondingly, the S D generated by sit-up activity is smaller than that by jumping activity. At the same time, the D A P T of the former is smaller, with a value around 0.5 g, while the latter produces a value around 1 g.

3.6. Comparison of Feature Values of Different Motions

In order to intuitively represent Sacc, SD, DAPT and IoP of different motions, Figure 8, Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14 are the results of feature extraction of an object in the SisFall. To further verify the effectiveness of the extracted feature values, we extract Sacc, SD, DAPT and IoP for all objects in the dataset, and their average peak values are shown in Table 3. It can be seen that the peaks of different feature values can be obviously distinguished.

3.7. Comparison of Different Methods

Based on the previously extracted Sacc, SD, DAPT and IoP, and the deep neural network classifier method proposed in literature [33], we classified D01–D11. As shown in Figure 15, we compare our method with literature [5,7]. These three methods are run on SisFall. It can be seen that the method of [5] is slightly better than our method in the identification of D07. However, the recognition effects of our method are better than the other two methods in most cases.

4. Conclusions

Wearable device is a promising method for collecting acceleration data of human body, that can be used for behavior recognition. This paper focused on investigating the feature extraction method based on wearable devices with single acceleration sensor. We associate Kalman filter algorithm with genetic algorithm for data pre-processing, which greatly reduces the amount of calculation. Then, we propose SD, IoP and DAPT as acceleration feature values that can be extracted from processed data sets. The experiments indicate that the source data filtered by KGA algorithm is more compact and the acceleration values in the three directions are more distinct. SD, DAPT and IoP are very obvious for the discrimination of seven different movements.

Author Contributions

Conceptualization, Z.H. and I.Y.; methodology, Z.H. and Q.N.; software, G.P.; validation, Z.H. and I.Y.; formal analysis, Z.H. and I.Y.; investigation, Q.N.; resources, Q.N.; data curation, G.P. and Z.H.; writing—original draft preparation, Z.H., G.P. and Q.N.; writing—review and editing, I.Y., G.P. and Q.N.; visualization, G.P. and Z.H.; supervision, I.Y.; project administration, I.Y.; funding acquisition, Z.H. and I.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by National Natural Science Foundation of China under Grant 62071470 and the Soonchunhyang University Research Fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: [http://sistemic.udea.edu.co/investigacion/proyectos/englishfalls/?lang=en].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hussain, S.; Kang, B.H.; Lee, S. A Wearable Device-Based Personalized Big Data Analysis Model. In Ubiquitous Computing and Ambient Intelligence. Personalisation and User Adapted Services; Springer: Berlin/Heidelberg, Germany, 2014; pp. 236–242. [Google Scholar]
  2. Hsu, Y.L.; Yang, S.C.; Chang, H.C.; Lai, H.-C. Human Daily and Sport Activity Recognition Using a Wearable Inertial Sensor Network. IEEE Access 2018, 6, 1–14. [Google Scholar] [CrossRef]
  3. Wilson, C.; Hargreaves, T.; Hauxwell-Baldwin, R. Smart homes and their users: A systematic analysis and key challenges. Pers. Ubiquitous Comput. 2015, 19, 463–476. [Google Scholar] [CrossRef] [Green Version]
  4. Song, F.; Zhou, Y.; Wang, Y.; Zhao, T.; You, I.; Zhang, H. Smart Collaborative Distribution for Privacy Enhancement in Moving Target Defense. Inf. Sci. 2019, 479, 593–606. [Google Scholar] [CrossRef]
  5. Chen, Y.; Shen, C. Performance Analysis of Smartphone-Sensor Behavior for Human Activity Recognition. IEEE Access 2017, 5, 3095–3110. [Google Scholar] [CrossRef]
  6. Song, F.; Zhu, M.; Zhou, Y.; You, I.; Zhang, H. Smart Collaborative Tracking for Ubiquitous Power IoT in Edge-Cloud Interplay Domain. IEEE Internet Things J. 2020, 7, 6046–6055. [Google Scholar] [CrossRef]
  7. Janidarmian, M.; Fekr, A.R.; Radecka, K.; Zilic, Z. A Comprehensive Analysis on Wearable Acceleration Sensors in Human Activity Recognition. Sensors 2017, 17, 529. [Google Scholar] [CrossRef]
  8. Song, F.; Zhou, Y.; Chang, L.; Zhang, H. Modeling Space-Terrestrial Integrated Networks with Smart Collaborative Theory. IEEE Netw. 2019, 33, 51–57. [Google Scholar] [CrossRef]
  9. Wang, A.; Chen, G.; Yang, J.; Chang, C.-Y.; Zhao, S. A Comparative Study on Human Activity Recognition Using Inertial Sensors in a Smartphone. IEEE Sens. J. 2016, 16, 4566–4578. [Google Scholar] [CrossRef]
  10. Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  11. Su, X.; Caceres, H.; Tong, H.; He, Q. Online Travel Mode Identification Using Smartphones with Battery Saving Considerations. IEEE Trans. Intell. Transp. Syst. 2016, 17, 2921–2934. [Google Scholar] [CrossRef]
  12. Ling, B.; Intille, S.S. Activity Recognition from User-Annotated Acceleration. Data Proc. Pervasive 2004, 3001, 1–17. [Google Scholar]
  13. Priyanka, K.; Tripathi, N.K.; Peerapong, K. A Real-Time Health Monitoring System for Remote Cardiac Patients Using Smartphone and Wearable Sensors. Int. J. Telemed. Appl. 2015, 2015, 1–11. [Google Scholar]
  14. Chao, C.; Shen, H. A feature evaluation method for template matching in daily activity recognition. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2013), KunMing, China, 5–8 August 2013; pp. 1–4. [Google Scholar]
  15. Maekawa, T.; Kishino, Y.; Yasushi, S. Activity recognition with hand-worn magnetic sensors. Pers. Ubiquitous Comput. 2013, 17, 1085–1094. [Google Scholar] [CrossRef]
  16. Cheng, J.; Amft, O.; Bahle, G.; Lukowicz, P. Designing Sensitive Wearable Capacitive Sensors for Activity Recognition. IEEE Sens. J. 2013, 13, 3935–3947. [Google Scholar] [CrossRef]
  17. Barshan, B.; Yüksek, M.C. Recognizing Daily and Sports Activities in Two Open Source Machine Learning Environments Using Body-Worn Sensor Units. Comput. J. 2013, 57, 1649–1667. [Google Scholar] [CrossRef]
  18. Zhang, M.; Sawchuk, A.A. A feature selection-based framework for human activity recognition using wearable multimodal sensors. In Proceedings of the 13th International Conference on Ubiquitous Computing (UbiComp 2011), Beijing, China, 17–21 September 2011; pp. 92–98. [Google Scholar]
  19. Dinov, I.D. Decision Tree Divide and Conquer Classification. In Data Science and Predictive Analytics; Springer: Berlin/Heidelberg, Germany, 2018; pp. 307–343. [Google Scholar]
  20. Aziz, O.; Russell, C.M.; Park, E.J.; Robinovitch, S.N. The effect of window size and lead time on pre-impact falls detection accuracy using support vector machine analysis of waist mounted inertial sensor data. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014; pp. 30–33. [Google Scholar]
  21. Diep, N.N.; Pham, C.; Phuong, T.M. A classifier based approach to real-time falls detection using low-cost wearable sensors. In Proceedings of the 2013 International Conference on Soft Computing and Pattern Recognition (SoCPaR), Hanoi, Vietnam, 15–18 December 2015; pp. 14–20. [Google Scholar]
  22. Jian, H.; Chen, H. A portable falls detection and alerting system based on k-NN algorithm and remote medicine. China Commun. 2015, 12, 23–31. [Google Scholar] [CrossRef]
  23. Song, F.; Ai, Z.; Zhou, Y.; You, I.; Choo, R.; Zhang, H. Smart Collaborative Automation for Receive Buffer Control in Multipath Industrial Networks. IEEE Trans. Ind. Inform. 2020, 16, 1385–1394. [Google Scholar] [CrossRef]
  24. Foody, G.M. RVM-based multi-class classification of remotely sensed data. Int. J. Remote Sens. 2008, 29, 1817–1823. [Google Scholar] [CrossRef]
  25. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. Human Activity Recognition on Smartphones Using a Multiclass Hardware-Friendly Support Vector Machine Ambient Assisted Living and Home Care. In Ambient Assisted Living and Home Care; Springer: Berlin/Heidelberg, Germany, 2012; pp. 216–223. [Google Scholar]
  26. Martín, H.; Bernardos, A.M.; Iglesias, J.; Casar, J.R. Activity logging using lightweight classification techniques in mobile devices. Pers. Ubiquitous Comput. 2013, 17, 675–695. [Google Scholar] [CrossRef]
  27. Song, F.; Ai, Z.; Zhang, H.; You, I.; Li, S. Smart Collaborative Balancing for Dependable Network Components in Cyber-Physical Systems. IEEE Trans. Ind. Inform. 2020. [Google Scholar] [CrossRef]
  28. Nukala, B.T.; Shibuya, N.; Rodriguez, A.I.; Nguyen, T.Q.; Tsay, J.; Zupancic, S.; Lie, D.Y.C. A real-time robust falls detection system using a wireless gait analysis sensor and an Artificial Neural Network. In Proceedings of the 2014 IEEE Healthcare Innovation Conference (HIC), Seattle, WA, USA, 8–10 October 2014; pp. 219–222. [Google Scholar]
  29. Angela, S.; López, J.; Vargas-Bonilla, J. SisFall: A Fall and Movement Dataset. Sensors 2017, 17, 198. [Google Scholar]
  30. Digital Signal Processing Formula Variable Program (four)-Butterworth Filter (On). Available online: https://blog.csdn.net/shengzhadon/article/details/46784509 (accessed on 7 July 2015).
  31. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Fluids Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef] [Green Version]
  32. Xianglin, W.; Jianwei, L.; Yangang, W.; Chaogang, T.; Yongyang, H. Wireless edge caching based on content similarity in dynamic environments. J. Syst. Archit. 2021, 102000. [Google Scholar] [CrossRef]
  33. Bingqian, H.; Wei, W.; Bin, Z.; Lianxin, G.; Yanbei, S. Improved deep convolutional neural network for human action recognition. Appl. Res. Comput. 2019, 36, 3111–3607. [Google Scholar]
Figure 1. Wearing of accelerometer and directions of coordinate axes.
Figure 1. Wearing of accelerometer and directions of coordinate axes.
Energies 14 00924 g001
Figure 2. Diagram of the algorithm.
Figure 2. Diagram of the algorithm.
Energies 14 00924 g002
Figure 3. Acceleration data of D01: (a) x axis acceleration; (b) y axis acceleration; (c) z axis acceleration; (d) Sacc superposition acceleration.
Figure 3. Acceleration data of D01: (a) x axis acceleration; (b) y axis acceleration; (c) z axis acceleration; (d) Sacc superposition acceleration.
Energies 14 00924 g003
Figure 4. Diagram of source data.
Figure 4. Diagram of source data.
Energies 14 00924 g004
Figure 5. Diagram of low-pass filter.
Figure 5. Diagram of low-pass filter.
Energies 14 00924 g005
Figure 6. Diagram of Kalman filter.
Figure 6. Diagram of Kalman filter.
Energies 14 00924 g006
Figure 7. Diagram of KGA.
Figure 7. Diagram of KGA.
Energies 14 00924 g007
Figure 8. Diagram of features in D03: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 8. Diagram of features in D03: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g008
Figure 9. Diagram of features in D04: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 9. Diagram of features in D04: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g009
Figure 10. Diagram of features in D05: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 10. Diagram of features in D05: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g010
Figure 11. Diagram of features in D06: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 11. Diagram of features in D06: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g011
Figure 12. Diagram of features in D07: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 12. Diagram of features in D07: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g012
Figure 13. Diagram of features in D08: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 13. Diagram of features in D08: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g013
Figure 14. Diagram of features in D11: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Figure 14. Diagram of features in D11: (a) Sacc over time; (b) SD over time; (c) DAPT over time; (d) IoP over time.
Energies 14 00924 g014
Figure 15. Comparison of different methods.
Figure 15. Comparison of different methods.
Energies 14 00924 g015
Table 1. Types of activities.
Table 1. Types of activities.
NumberTypeNumber of ExperimentsDuration (s)
D01Walk slowly1100
D02Walk quickly1100
D03Run slowly1100
D04Run quickly1100
D05Go up and down the stairs slowly525
D06Go up and down the stairs quickly525
D07Sit→get up slowly in a half-high chair512
D08Sit→get up quickly in a half-high chair512
D09Sit→get up slowly in a low chair512
D10Sit→get up quickly in a low chair512
D11Jump up512
Table 2. Parameter settings of Butterworth filter.
Table 2. Parameter settings of Butterworth filter.
ParametersValues
Sampling frequency ( H z ) 25
Cut-off frequency ( H z ) 5
Order4
Table 3. Feature values of different motions.
Table 3. Feature values of different motions.
Motion TypeSaccSDDAPTIoP
D031.370.150.510.39
D041.390.250.520.38
D051.970.920.290.79
D061.410.230.590.48
D071.10.0140.111.02
D081.430.170.580.62
D111.820.381.010.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, Z.; Niu, Q.; You, I.; Pau, G. Acceleration Feature Extraction of Human Body Based on Wearable Devices. Energies 2021, 14, 924. https://doi.org/10.3390/en14040924

AMA Style

Huang Z, Niu Q, You I, Pau G. Acceleration Feature Extraction of Human Body Based on Wearable Devices. Energies. 2021; 14(4):924. https://doi.org/10.3390/en14040924

Chicago/Turabian Style

Huang, Zhenzhen, Qiang Niu, Ilsun You, and Giovanni Pau. 2021. "Acceleration Feature Extraction of Human Body Based on Wearable Devices" Energies 14, no. 4: 924. https://doi.org/10.3390/en14040924

APA Style

Huang, Z., Niu, Q., You, I., & Pau, G. (2021). Acceleration Feature Extraction of Human Body Based on Wearable Devices. Energies, 14(4), 924. https://doi.org/10.3390/en14040924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop