Next Article in Journal
Prediction of Motion Intentions as a Novel Method of Upper Limb Rehabilitation Support
Next Article in Special Issue
Prediction of Postoperative Complications for Patients of End Stage Renal Disease
Previous Article in Journal
Nanozymes for Environmental Pollutant Monitoring and Remediation
Previous Article in Special Issue
TumorGAN: A Multi-Modal Data Augmentation Framework for Brain Tumor Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application of Skeleton Data and Long Short-Term Memory in Action Recognition of Children with Autism Spectrum Disorder

1
School of Communication and Electronic Engineering, East China Normal University, Shanghai 200241, China
2
Experimental Teaching Center for Teacher Education, East China Normal University, Shanghai 200241, China
3
Department of Rehabilitation Sciences, East China Normal University, Shanghai 200062, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2021, 21(2), 411; https://doi.org/10.3390/s21020411
Submission received: 7 December 2020 / Revised: 25 December 2020 / Accepted: 6 January 2021 / Published: 8 January 2021
(This article belongs to the Special Issue Healthcare Monitoring and Management with Artificial Intelligence)

Abstract

:
The recognition of stereotyped action is one of the core diagnostic criteria of Autism Spectrum Disorder (ASD). However, it mainly relies on parent interviews and clinical observations, which lead to a long diagnosis cycle and prevents the ASD children from timely treatment. To speed up the recognition process of stereotyped actions, a method based on skeleton data and Long Short-Term Memory (LSTM) is proposed in this paper. In the first stage of our method, the OpenPose algorithm is used to obtain the initial skeleton data from the video of ASD children. Furthermore, four denoising methods are proposed to eliminate the noise of the initial skeleton data. In the second stage, we track multiple ASD children in the same scene by matching distance between current skeletons and previous skeletons. In the last stage, the neural network based on LSTM is proposed to classify the ASD children’s actions. The performed experiments show that our proposed method is effective for ASD children’s action recognition. Compared to the previous traditional schemes, our scheme has higher accuracy and is almost non-invasive for ASD children.

1. Introduction

Autism, or Autism Spectrum Disorder (ASD), is a neurodevelopmental disorder, characterized by persistent deficits in social communication and interaction as well as restricted and repetitive behaviors [1]. The cause of ASD is extremely complex, involving genetic and environmental factors [2,3]. In recent years, the number of children diagnosed with ASD has been increasing [4,5]. It is estimated that one in 160 children have autism [6]. This puts great pressure on society.
Studies have shown that early diagnosis and intervention are the most effective clinical treatment methods for ASD [7,8]. At present, the diagnosis of ASD is complicated and mainly relies on diagnostic tools, including parent interviews and clinical observations. These measures have an extend period, leading to delays in ASD intervention and treatment, causing ASD children to miss the window period. Moreover, parents of ASD children will be under tremendous pressure and lose confidence in the relevant professional medical staff after a long diagnosis delay [9]. Howlin et al. [10] surveyed over 1200 parents of ASD children and found that about half of the families were “not too” or “not at all” satisfied with the ASD diagnosis process. Studies have confirmed that dyskinesias are common in children with ASD [11,12,13]. They often have uncontrolled repetitive actions, such as clapping hands, shaking the body, and repeatedly fiddling with toys and objects. In the United States (DSM–V) [1] and Europe (ICD–10) [14], stereotyped action is the core diagnostic criteria in the professional clinical practice of ASD. Therefore, accurate and rapid recognition of ASD children’s action is an important idea to accelerate diagnosis.
For the action recognition of ASD children, direct observation is the most widely used method [15,16,17], i.e., rehabilitation experts directly observe the behavior of ASD children, then record and analyze their stereotyped actions [18]. However, ASD children move swiftly, rehabilitation experts cannot accurately observe and record all stereotyped actions. It is also difficult to determine the start and end time.
In addition to direct observation, wearable sensor-based methods provide a promising solution for action recognition in ASD children. Before the Deep Neural Network (DNN) is fully developed [19,20,21], manual extraction of features from sensor data for action classification is the most common. Gonçalves et al. [22] used the acceleration sensor worn on the right arm of ASD patients to collect action’s data and then analyzed the statistical features such as mean, variance, peak number, and root mean square to detect the stereotyped actions. Crippa et al. [23] designed a simple ball-grabbing task. They used an optoelectronic system to obtain the kinematics data of ASD children when they completed the actions such as reach, grasp, and drop, then extracted 17 kinematics indicators as features. Finally, the SVM classifier was used to classify different actions. The above works were designed to quickly and accurately obtain the action information of ASD children from wearable sensor data. The sensor’s advantages are simplicity, stability, and high sensitivity, but the disadvantages are also obvious. On the one hand, the sensor cannot analyze the complex actions of the human body. On the other hand, it is too invasive for ASD children. Wearing sensors will distract them, which will change their behavior and affect the accuracy of their actions [24,25]. Moreover, manual selection and extraction of features rely on researchers’ professional knowledge, and the omission of essential features related to the task will cause the system to fail. Compared with the wearable sensor-based methods, video-based automatic analysis methods are almost non-invasive. They have been widely used with the development of computer vision technology, such as football videos [26,27], basketball videos [28,29], tennis videos [30,31], and taekwondo videos [32,33]. However, due to the uncertainties and noisy backgrounds of ASD children’s action, there are few attempts to analyze ASD children’s videos.
In recent years, with the application of deep learning in the biomedical field [34,35,36,37], especially the development of Deep Neural Network and hardware computing capabilities, a suitable method is to use neural networks to extract action information from multi-dimensional data and integrate them into the action recognition of ASD children. Rad et al. [38] used a three-layer CNN network to automatically extract features from ASD patients’ data collected by acceleration sensors and ordinary inertial measurement units (IMU), then they used an SVM classifier for classification. The results showed that neural networks are superior to traditional manual feature extraction methods. Cook et al. [39] used the OpenPose algorithm to extract the skeleton data of ASD children’s upper limbs from the RGB image, then they calculated each key point’s speed to extract the features manually. Their method could recognize clapping, swinging back and forth, and repeatedly playing with the toy. They achieved 71% accuracy using the Decision Tree as a classifier. The results of various studies show that the action recognition of ASD children is still a challenging problem.
Considering this, to accelerate the diagnosis process, this paper proposes an effective ASD children action recognition method. Our method uses the OpenPose [40] algorithm to extract the skeleton data of ASD children from the video and then recognizes the action through the Long Short-Term Memory networks (LSTMs). The main contributions of this paper include:
  • We propose four measures to eliminate the initial skeleton data’s noise, which improve recognition accuracy and calculation efficiency.
  • The multi-person tracking method based on skeleton data is proposed to track multiple ASD children in the video. Unlike the latest multiple objects tracking technologies, our method does not need additional GPUs and is suitable for multi-person tracking in a fixed scene.
  • An action recognition model based on LSTMs is proposed in this paper, and the end-to-end deep learning-based framework eliminates the need to extract features manually. The experimental results show that the Precision, Recall, and F1-score are improved, and the proposed model outperforms other manual feature extraction-based methods on our ASD children dataset.
  • We evaluate the impact of the input data time steps and the number of hidden states on the LSTM network’s accuracy, which is crucial for the recognition of ASD children’s actions.
The rest of this paper is organized as follows: Section 2 introduces the proposed method, including three parts: the generation of de-noised skeleton data for ASD children, multi-person tracking based on skeleton data, and ASD children’s action recognition based on LSTM network. Section 3 describes the experiments and results, mainly including the introduction of the experimental dataset, evaluation of important parameters, system performance evaluation, and comparison with other methods. Finally, conclusions are drawn in Section 4.

2. Proposed Method

The flow chart of our ASD children action recognition method is shown in Figure 1. It consists of three stages. In stage 1, the OpenPose algorithm obtains the skeleton data of ASD children in the video sequence. However, the original skeleton data will be lost due to occlusion or overlap. Therefore, we denoise the skeleton data by discarding frames with missing important information and deleting the head data that has little effect on this task (detailed in Section 2.1). In stage 2, to track multi-person in real time, the skeleton data in different frames obtained by stage 1 are matched to obtain each person’s continuous skeleton data (detailed in Section 2.2). Finally, each ASD children’s skeleton data obtained by stage 2 as the input of the LSTM model to recognize actions. With all stages finish, we finally achieve three models: de-noised human skeleton data generation model, multi-person tracking model based on skeleton data, and multi-person action recognition model.
Based on these three models, we build a system for ASD children’s action recognition. To make our flow chart clearer, Figure 2 shows the visual flow of the system recognizing five different actions.

2.1. Generation of De-Noised Human Skeleton Data

In this paper, OpenPose [40] is used to obtain original human skeleton data. However, there are two problems with the original skeleton data: On the one hand, ASD children are sometimes occluded, which leads to a lack of skeleton data. On the other hand, ASD children’s actions are highly random, making the acquired skeleton data not obvious, and it is difficult to recognize the actions accurately. For this, we have taken four measures to eliminate noise: First, we scale the coordinates of the key points outputted by OpenPose to the same unit. Second, we remove the five joints on the head. Third, we discard frames without skeleton data or missing important joints. Finally, we use the relative joint positions in adjacent frames to fill in the unrecognized joint positions.
Figure 3 illustrates the overall pipeline of OpenPose. The system takes, as input, a color image of size w × h and produces, as output, the 2D location of anatomical key points for each person in the image. It uses the first 10 layers of VGG-19 [41] as a feature extractor, generating a set of feature maps F that is input to a feedforward network. First, the feedforward network simultaneously predicts a set of 2D confidence maps S of body part locations and a set of 2D vector fields L of part affinities, which encode the degree of association between parts. The set S = ( S 1 , S 2 , , S J ) has J confidence maps, one per part, where S j R w , j { 1 J } . The set L = ( L 1 , L 2 , , L C ) has C vector fields, one per limb, where L c R w × 2 , c { 1 C } , each image location in L c encodes a 2D vector. Finally, the confidence maps and the affinity fields are parsed by greedy inference to output the 2D key points for all people in the image. The output includes ears, eyes, nose, neck, shoulders, elbows, wrists, knees, hips, and ankles, as shown in Figure 4.
The original joint positions outputted by OpenPose has a different unit for the X co-ordinate and Y co-ordinate. As shown in Figure 5a, we scaled them to the same unit to deal with the images with different height/width ratios. Moreover, the head’s position helps little for the action classification. What matters is the configuration of the body and limbs. Thus, we manually removed the five joints (Nose, Right eye, Left eye, Right ear, Left ear) on the head, and the Neck becomes the 0th joint, as shown in Figure 5b. If in a frame there is no human skeleton detected by OpenPose or the detected skeleton has no Neck or Thigh (top of Figure 5c), then this frame is considered invalid and will be discarded. On the contrary, frames missing other joints will be retained (bottom of Figure 5c).
In some cases, OpenPose might fail to detect a complete human skeleton from the image, causing some blank in the joint positions. These joints must be filled with some values to maintain a fixed position for the following automatic feature extraction procedure. Two bad solutions are:
  • Discard this frame. However, in this way, the algorithm will barely recognize the action when the person is not facing the camera.
  • Fill in the joint positions with some value outside a reasonable range. In theory, when the algorithm is strong enough, this method could work. However, this requires a lot of experimentation and more calculations.
In this paper, we will set the missing joints as their relative positions to the Neck in the previous frame. Suppose a joint ( x i , y i ) is missing in the current frame, the position of it can be expressed as:
x i _ c u r r = x N e c k _ c u r r + ( x i _ p r e v x N e c k _ p r e v ) ,
y i _ c u r r = y N e c k _ c u r r + ( y i _ p r e v y N e c k _ p r e v ) ,
Figure 5d shows an example of using our method to fill in the missing joint.
The four measures to eliminate noise are described above. It is noteworthy that the skeleton data we currently obtain is for everyone in a frame. Therefore, a multi-person tracking technique should be used to obtain different persons’ id in a frame and match their skeleton data.

2.2. Tracking Multi-Person by Skeleton Data

In an ASD rehabilitation scenario, there are usually multiple children. So, tracking them is needed. It should be mentioned that for multiple object tracking (MOT), a usual consideration is that some commonly used algorithms such as SORT [42] and Deep SORT [43]. Through the calculation of the Convolutional Neural Network (CNN), they can achieve good accuracy. However, these methods have a high computational burden, which is hardly put into a real-time system. In this paper, we design an algorithm to track multiple people through human skeleton data in video sequences.
First, after the skeleton data is de-noised in Section 2.1, the multi-person skeleton data S n of each frame is obtained. Any two consecutive frames can be expressed as previous skeleton data S p r e v and current skeleton data S c u r r . The skeleton data of each person in each frame is expressed as:
S p r e v [ i ] = { ( x i _ 0 , y i _ 0 ) , ( x i _ 1 , y i _ 1 ) , . . . , ( x i _ 12 , y i _ 12 ) } ,
S c u r r [ j ] = { ( x j _ 0 , y j _ 0 ) , ( x j _ 1 , y j _ 1 ) , . . . , ( x j _ 12 , y j _ 12 ) } ,
where i , j represents different people in previous and current frame, respectively. Then, all skeletons are sorted based on the distance between the neck ( x 0 , y 0 ) and the image center ( x c e n t e r , y c e n t e r ) , from small to large. This step provides convenience for matching skeletons between current and previous. The calculation formula is as follows:
D = ( x 0 x c e n t e r ) 2 + ( y 0 y c e n t e r ) 2 .
It should be noted that if S p r e v is the first frame containing skeleton data, then the human id of everyone will be initialized according to the distance from the center of the frame.
Finally, matching the distance between current skeletons and previous skeletons. If S c u r r [ j ] and S p r e v [ i ] are matched, for S p r e v [ i ] , S c u r r [ j ] is the nearest skeleton in S c u r r and for S c u r r [ j ] , S p r e v [ i ] is the most nearest skeleton in S p r e v as well. Moreover, the distance between the two matched people’s joints should be smaller than our pre-set threshold. For unmatched skeletons in S c u r r , they are considered to be new people appeared in the video.
Through the above steps, we can track the input skeletons by matching them with previous skeletons and then obtain their corresponding human id.

2.3. Action Recognition Using LSTM Net

The skeleton data of different actions are time series. In this paper, the Long Short-Term Memory (LSTM) network automatically extracts features from skeleton data. The LSTM [44] is a special type of Recurrent Neural Network (RNN), which adds prior knowledge in the hidden layer: input gate, forget gate, and output gate. These gates process the inter-layer information at different moments and the input information at a certain moment more transparently, which can effectively pass the past information to the current calculation and can overcome the defect that the RNN structure cannot pass far apart information [45].
The structure of the LSTM unit is shown in Figure 6, the function expressions of input gate, forget gate, and output gate can be obtained as follows:
i ( t ) = s i g m o i d ( W x i x ( t ) + W h i h ( t 1 ) + b i ) ,
f ( t ) = s i g m o i d ( W x f x ( t ) + W h f h ( t 1 ) + b f ) ,
o ( t ) = s i g m o i d ( W x o x ( t ) + W h o h ( t 1 ) + b o ) ,
where W x is the input weight matrix, W h is the hidden layer state weight matrix at time t 1 , b is the bias term. The self-connected unit state c ( t ) and hidden layer state h ( t ) at time t are expressed as:
c ( t ) = f ( t ) · c ( t 1 ) + i ( t ) · tanh ( W x c x t + W h c h t 1 + b c ) ,
h ( t ) = o ( t ) · tanh ( c ( t ) ) .
Analyzing (6)–(10), it can be found that by adjusting the weight matrix W of each gate, the input gate i ( t ) can control the amount of information flowing into the self-connected unit state c ( t ) , and the forget gate f ( t ) can control the amount of information c ( t 1 ) contained in the self-connected unit state c ( t ) at the current moment, i.e., how much information is forgotten. The output gate o ( t ) controls the information of self-connected unit state c ( t ) that can flow into the current hidden layer state h ( t ) . Among them, the role of the self-connected unit state c ( t ) is to complete the accumulation of historical information, and its accumulation method is:
S e t : i n f o = tanh ( W x c x t + W h c h t 1 + b c ) ,
where i n f o is the source of information to be accumulated, substituting (11) into (9) to obtain:
c ( t ) = f ( t ) · c ( t 1 ) + i ( t ) · i n f o .
From (12), it can be known that when the self-connected unit state c ( t ) accumulates historical information, it relies on the forget gate f ( t ) to limit the information transmitted at the previous moment c ( t 1 ) , and at the same time, the input gate i ( t ) constrain the newly entered information. According to (10), the current hidden layer state h ( t ) is constrained by the output gate. Since it is updated linearly, the t a n h function with nonlinear is added.
The information source of the entire LSTM unit is the current input x ( t ) , the hidden layer state h ( t 1 ) at the previous moment, and the linear self-connected unit state c ( t 1 ) at the previous moment. Since c ( t 1 ) is calculated according to (9), the control basis of the three gate units actually comes from the current input x ( t ) and the hidden layer state h ( t 1 ) at the previous moment.
Compared with the traditional methods [22,23,46], using a Recurrent Neural Network (RNN) with Long short-term memory (LSTM) units requires almost no feature engineering, and the data can be directly fed into the neural network to model the problem correctly.
Figure 7 shows the action recognition network used in our method. The input skeleton data has a shape of n × 13 × 2, which denotes the n sequential frames with 13 key points having X and Y coordinates each. First is a fully connected layer activated by ReLU, and then two stacked LSTM layers are applied to the skeleton data of each frame. A many-to-one architecture of LSTM is used in this paper. LSTM leverages the sequential nature of the input sequences to identify temporal changes in skeleton data. The LSTM layers’ output is passed to a fully connected layer with SoftMax activation and five outputs. Each of these five outputs provides the probability of the corresponding action in terms of cross-entropy. Thresholding is applied to this output to detect when the ASD children are performing specific actions.

3. Experiments and Results

In this section, we first introduce the ASD children dataset used in this paper. Next, two important parameters are evaluated. Then, we verify the effectiveness of the proposed skeleton data de-noised methods. Finally, we compare the proposed approach with several methods of manually extracting features.

3.1. Experiment Dataset

There is currently no public dataset that can be used for the action recognition of ASD children. The dataset used in this paper was collected in our cooperative ASD rehabilitation institution. We arranged four 5 million pixels HIKVISION remote cameras in the classroom to collect videos of ASD children from four different angles, then transferred the images to the database through the POE recorder. We can view the real-time situation in the classroom through a smart phone, and adjust the shooting angle of the camera remotely to ensure that the image with the best angle can be collected without disturbing the ASD children. The data collection system is shown in Figure 8.
Our purpose was to collect ASD children’s videos in various real environments, to mark some stereotyped actions. The children with ASD selected by us are about 5 to 10 years old. It is worth noting that not all children are severely autistic, some children have milder symptoms, and their actions are similar to normal children. We mainly focused on their repetitive actions. These actions are usually very fast, such as shaking hands, shaking their bodies, and standing up suddenly, they are easily ignored by ASD children’s busy parents.
In this paper, the dataset we made contains video clips of 5 actions: sit, stand, squat, shake the body, and shake hands. Each clip is a sequence of images about 1–3 s (25 frames per second), a total of 1062 sequences. Each sequence contains the complete process of a single action. Table 1 below describes an overview of the dataset, including the name of each action, the number of frames, and the number of sequences.
After obtaining the dataset, we normalized the image size to 656 × 368, then sent them to the de-noised human skeleton data generation model proposed in Section 2.1. It should be noted that when extracting skeleton data for training the LSTM network, there may be multiple ASD children in some frames. For this, we only keep the skeleton data of the ASD child closest to the image center.

3.2. Evaluation of Parameters

Model network structure and the input data structure have a fundamental impact on the effect of action recognition. In this section, two important parameters that affect the accuracy of the LSTM network were evaluated. One is the time steps of the input sequence, which indicates how many frames constitute an action sequence. It is related to the structure of the input data. The other is the number of hidden states of the LSTM network, which is related to the network structure.
The time steps of the input sequence is one of the important parameters in making a training dataset. On the one hand, action sequences with small time steps cannot guarantee that it covers all the process of an action. On the other hand, action sequence with too large time steps will not only increase the pressure of network computing, but also introduce some interference features affecting action recognition of ASD children. Thus, to verify the effect of input data structure on network performance, datasets with time steps from 20–50 were trained under other conditions unchanged. As shown in Figure 9.
The number of hidden states, one of the basic parameters of the LSTM network structure, represents the number of nodes used to remember and store the past state, which determines how much information is remembered and how much is forgotten. From the construction principle of the neural network, the more hidden states, the more data features and the higher adaptability of the network. However, too many hidden states will overfit the neural network. Therefore, the extracted dataset was used to test the impact of the number of hidden states on recognition accuracy. As shown in Figure 9.
The extracted dataset from the ASD children dataset we made in Section 3.1, including 151 sequences of sit, 149 sequences of stand, 133 sequences of squat, 150 sequences of shake the body, and 141 sequences of shake hands. 70% of the complete action sequences were used to train the model, and the remaining 30% were used to test the recognition model. It is worth noting that in the original dataset proposed in Section 3.1, the length of a single action is 1–3 s (25–75 frames). Therefore, to evaluate the effect of the sequence’s time step, the original action sequence should be scaled to 20–50 frames. The methods we used as following:
  • For the sequences that are less than 20 frames, we fill them with all-zero skeleton data at the end of the sequence.
  • For the sequences that are more than 50 frames, we randomly select frames without repeating and make them form a shortened frame sequence in the original order.
According to Figure 9, we first analyze the impact of time steps on recognition accuracy. When the number of hidden states is 4 to 16, the network is too simple. At this time, the time steps have almost no effect on the accuracy of the model. When the number of hidden states between 32 and 512, increasing the time steps from 20 to 32 improves the model’s accuracy, and the accuracy almost unchanged after more than 32 frames. The decrease in accuracy at a time steps of 50 means that the time span is so long that the model learns the interference features. The above results indicate that for our dataset, the time steps of 32 frames (about 1.3 s) can summarize the action features of ASD children.
Then analyze the impact of hidden states, when the number of hidden states from 4 to 16, the network structure is so simple that the model cannot learn enough features from the data. At this time, the accuracy is low. When the number of hidden states from 32 to 512, the accuracy of the model improves significantly, which indicates that the network has learned more features to classify actions. When the number of hidden states is 128, the model reaches the best accuracy. It is worth noting that when the number of hidden states changes from 256 to 512, the accuracy decreases, mainly due to the over-fitting phenomenon caused by the complex network structure, which significantly reduces the generalization ability of the model.
Considering these factors comprehensively, we got the following best parameters: the time steps is 32, the number of hidden states is 128, and the learning rate is 0.001.

3.3. Performance Evaluation

To verify the impact of the skeleton data de-noised methods proposed in Section 2.1, we used the ASD children dataset to experiment with different time steps and the number of hidden states.
In the experiment, we selected 749 action sequences from the ASD children dataset made in Section 3.1 (144 for sit, 151 for stand, 129 for squat, 171 for shake the body, and 154 for shake hands). 70% of the complete sequences are used for training, and the remaining 30% are used for testing. After processing, we got two groups of skeleton data. One group was obtained after denoising using proposed methods, and the other group was the original skeleton data extracted by OpenPose. The results are shown in Figure 10. After denoising, the recognition accuracy under different time steps and hidden states is higher than that of unprocessed original data, about 10%. It shows that our skeleton data de-noised methods can make the model learn more useful features and reduce the influence of interference features. It is worth noting that same as the best parameters in Section 3.2, the model achieves a balance between recognition accuracy and calculation pressure under the time steps of 32 and the hidden state of 128, which shows that the best parameters obtained are indeed suitable for our experiments.
Two kinds of confusion matrices, obtained from original and de-noised skeleton data using the best parameters, are presented in Figure 11. The row of a confusion matrix represents true classes of samples while the column represents the predicted classes. The diagonal entries represent the true classification rate of each class, the shallower the color is, the higher the true classification rate is. Figure 11a is the confusion matrix obtained from the original skeleton data extracted by OpenPose, while Figure 11b uses the de-noised skeleton data. Obviously, skeleton data denoising processing improves action recognition accuracy, especially shake the body, squat, and sit. The features of these three actions are more complex. However, the denoising processing enables the network to learn more useful features and reduces invalid features’ interference. It is worth noting that for the two experiments, shake the body was confused with sit. A very important reason is that in our dataset, ASD children shaking the body are done in a sitting position, so the features of these two actions are similar. In this case, our denoising methods achieves a recognition rate of 90.29%. Similarly, squat and sit have been misrecognized due to decreased body height, but our denoising method finally gets a recognition rate of 90.39% and 93.55%, respectively.
Figure 12 shows some screenshots from real-time tests. Through the multi-person tracking method based on skeleton data proposed in Section 2.2, our system supports real-time action recognition of multiple ASD children. The threshold we set is 0.9. When the probability of action is greater than 0.9, the system will determine that the ASD children are performing this action and display the label of the action on the character selection box.

3.4. Comparison and Analysis

To further verify the accuracy of the proposed method, manual feature extraction-based methods were introduced to recognize ASD children’s actions. In the experiments, we selected 844 action sequences from the ASD children dataset made in Section 3.1 (160 for sit, 155 for stand, 151 for squat, 190 for shake the body, 188 for shake hands). 70% of the complete sequences were used for training and the remaining 30% were used for testing.
Manual feature extraction is very popular in motion analysis. In the field of ASD children’s action recognition, the most common is to extract features from sensor data. Gonçalves et al. [22] and Albinali et al. [47,48,49] both used acceleration sensors to collect action data, then captured time and frequency domain features from them, finally recognized the actions of ASD children. With the development of human pose estimation technology based on RGB images (such as OpenPose and AlphaPose [50]), extracting features from skeleton data can also achieve good performance. Like our work, [39] used OpenPose to extract the skeleton data of ASD children and then manually extract features from it for action recognition.
To compare the recognition effects of the proposed LSTM-based method and manual feature extraction methods more comprehensively, we re-set two feature extraction methods. Method 1 was based on [39], and the extracted features included the maximum magnitude, mean magnitude, and standard deviation of the autocorrelogram maxima. These features were calculated from each key point’s three velocity measures (velocity change in the X-axis, Y-axis, and overall position). It should be noted that [39] only used six key points of the upper body to extract features: Right shoulder, Right elbow, Right wrist, Left shoulder, Left elbow, and Left wrist. It means that their method was only sensitive to the upper body actions of ASD children. However, our dataset includes some lower body actions, so based on the original features, we added all normalized joints except the head positions as additional features to ensure the effectiveness of action recognition.
Method 2 included three features: normalized joint positions X, moving velocity of the body V b o d y , and joint velocities V j o i n t s . Compared with method 1, these features contained information on all key points and can better summarize the actions of ASD children. A summary of the computed features is shown in Table 2. In addition, method 1 used the Decision Tree mentioned in [39] as classifier, and method 2 used DNN (3 layers, 100 × 100 × 100), Linear SVM, and Random Forests (depth 40, trees 100) as classifiers. The implementation of these classifiers was from the Python library “sklearn”.
The experiments were carried out under the original and de-noised skeleton data, respectively. Figure 13 shows the recognition accuracy of different methods. Simultaneously, to further verify the performance of the proposed method, Table 3 reports the Precision, Recall, and F1-score of each action computed over all the methods, which are defined as follows:
P r e c i s i o n : P = T P T P + F P ,
R e c a l l : R = T P T P + F N ,
F 1 s c o r e = 2 × P × R P + R ,
where True Positive (TP) represents the number of actions correctly recognized, False Positives (FP) represents the number of actions falsely recognized, and False Negatives (FN) represents the number of missed actions.
By analyzing the performance indicators in Figure 13 and Table 3, we can find that using the de-noised skeleton data, the recognition accuracy of all actions has been significantly improved, and an overall improvement of more than 5% was gained in the F1-score. At the same time, we can find that the proposed method achieved the highest accuracy in all five classes of actions. Among the other four manual feature extraction-based methods, the DNN classifier got the highest accuracy, while the Decision Tree classifier got the lowest. It is worth noting that for shaking hands, [39] has a smaller gap with other methods. A crucial reason is that [39] mainly focuses on the upper body features, especially the hands, so the actions of hands are more sensitive. Therefore, we can conclude that the methods based on manual feature extraction rely more on the professionalism of feature selection, and the omission of essential features will make the model invalid. On the contrary, the proposed method based on the LSTM network can automatically learn features from the time series ASD children’s skeleton data, then recognize the actions with higher accuracy.

4. Conclusions

In this paper, an action recognition method based on skeleton data and LSTM network was proposed for ASD children. It can recognize single or multiple ASD children’s action from the real-time or pre-recorded video. First, the OpenPose algorithm is used to extract skeleton data from the continuous image sequences, then proposed denoising methods eliminate the noise of the skeleton data. Next, the multi-person tracking model based on the skeleton data tracks multiple ASD children in the previous and subsequent frames. Finally, the LSTM network analyzes each ASD children’s skeleton data to recognize different actions.
Experimental results show that the recognition performance of the proposed method is better than traditional manual feature extraction methods, and the denoising of skeleton data can also improve the action recognition accuracy. It must be noted that our approach does not require special hardware, and the image taken by the conventional RGB camera can be used as the input of the system, which is almost non-invasive for ASD children. In addition, the end-to-end deep learning-based framework eliminates the need to manually extract features, allowing for the addition of new actions by just retraining the model with new data. This method can assist rehabilitation experts in recognizing the actions of ASD children automatically, to shorten the diagnosis cycle. Moreover, the number of stereotyped actions recognized by the proposed method can be used as an indicator to evaluate the effect of ASD rehabilitation.
In future research, more actions of ASD children and a larger dataset can be included. Also, we plan to transplant the proposed method to the embedded platform and conduct some real-time tests to verify the reliability of the system.

Author Contributions

Conceptualization, Y.Z. and Y.T.; methodology, Y.Z. and Y.T.; software, Y.Z.; validation, Y.Z., Y.T. and D.C.; formal analysis, Y.Z. and Y.T.; investigation, Y.Z., Y.T. and P.W.; resources, Y.T. and D.C.; data curation, Y.Z.; writing—original draft preparation, Y.Z. and Y.T.; writing—review and editing, Y.Z., Y.T., P.W. and D.C.; visualization, Y.Z.; supervision, Y.T. and D.C.; project administration, Y.Z., Y.T. and D.C.; funding acquisition, D.C., All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the East China Normal University Project under Grant 2019ECNU-XFZH012 and in part by the Ministry of Education of China Project under Grant 17YJAZH008.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Ethics Committee of East China Normal University (HR 400-2020, approved on 30 September 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the privacy of our participants and the requirements of our UCHRP (HR 400-2020).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Arbanas, G. Diagnostic and Statistical Manual of Mental Disorders (DSM-5). Alcohol. Psychiatry Res. 2015, 51, 61–64. [Google Scholar]
  2. Hallmayer, J.; Cleveland, S.; Torres, A.; Phillips, J.; Cohen, B.; Torigoe, T.; Miller, J.; Fedele, A.; Collins, J.; Smith, K.; et al. Genetic Heritability and Shared Environmental Factors Among Twin Pairs With Autism. Arch. Gen. Psychiatry 2011, 68, 1095–1102. [Google Scholar] [CrossRef] [PubMed]
  3. Sandin, S.; Lichtenstein, P.; Kuja-Halkola, R.; Larsson, H.; Hultman, C.M.; Reichenberg, A. The familial risk of autism. JAMA 2014, 311, 1770–1777. [Google Scholar] [CrossRef] [PubMed]
  4. Jouaiti, M.; Henaff, P. Robot-Based Motor Rehabilitation in Autism: A Systematic Review. Int. J. Soc. Robot. 2019, 11, 753–764. [Google Scholar] [CrossRef]
  5. Dave, D.M.; Fernandez, J.M. Rising Autism Prevalence: Real or Displacing Other Mental Disorders? Evidence from Demand for Auxiliary Healthcare Workers in California. Econ. Inq. 2015, 53, 448–468. [Google Scholar] [CrossRef]
  6. Elsabbagh, M.; Divan, G.; Koh, Y.J.; Kim, Y.S.; Kauchali, S.; Marcín, C.; Montiel-Nava, C.; Patel, V.; Paula, C.S.; Wang, C.; et al. Global prevalence of autism and other pervasive developmental disorders. Autism Res. 2012, 5, 160–179. [Google Scholar] [CrossRef] [Green Version]
  7. Bradshaw, J.; Steiner, A.M.; Gengoux, G.; Koegel, L.K. Feasibility and Effectiveness of Very Early Intervention for Infants At-Risk for Autism Spectrum Disorder: A Systematic Review. J. Autism Dev. Disord. 2015, 45, 778–794. [Google Scholar] [CrossRef]
  8. Howlin, P.; Magiati, I.; Charman, T. Systematic review of early intensive behavioral interventions for children with autism. Ajidd-Am. J. Intellect. Dev. Disabil. 2009, 114, 23–41. [Google Scholar] [CrossRef]
  9. Crane, L.; Chester, J.W.; Goddard, L.; Henry, L.A.; Hill, E. Experiences of autism diagnosis: A survey of over 1000 parents in the United Kingdom. Autism 2016, 20, 153–162. [Google Scholar] [CrossRef]
  10. Howlin, P.; Moore, A. Diagnosis in Autism A Survey of Over 1200 Patients in the UK. Autism 1997, 1, 135–162. [Google Scholar] [CrossRef]
  11. Manjiviona, J.; Prior, M. Comparison of Asperger syndrome and high-functioning autistic children on a Test of Motor Impairment. J. Autism Dev. Disord. 1995, 25, 23–39. [Google Scholar] [CrossRef] [PubMed]
  12. Ghaziuddin, M.; Butler, E. Clumsiness in autism and Asperger syndrome: A further report. J. Intellect. Disabil. Res. 1998, 42, 43–48. [Google Scholar] [PubMed]
  13. Stott, D. A General Test of Motor Impairment for Children. Dev. Med. Child Neurol. 1966, 8, 523–531. [Google Scholar] [PubMed]
  14. World Health Organization. The ICD-10 classification of mental and behavioural disorders: Clinical descriptions and diagnostic guidelines. Wkly. Epidemiol. Rec. Relevé épidémiologique Hebdomadaire 1992, 67, 227. [Google Scholar]
  15. MacDonald, R.; Anderson, J.; Dube, W.V.; Geckeler, A.; Green, G.; Holcomb, W.; Mansfield, R.; Sanchez, J. Behavioral assessment of joint attention: A methodological report. Res. Dev. Disabil. 2006, 27, 138–150. [Google Scholar] [CrossRef]
  16. Freeman, B.; Ritvo, E.; Schroth, P. Behavior Assessment of the Syndrome of Autism: Behavior Observation System. J. Am. Acad. Child Psychiatry 1984, 23, 588–594. [Google Scholar] [CrossRef]
  17. Harrop, C.; McConachie, H.; Emsley, R.; Leadbitter, K.; Green, J. Restricted and Repetitive Behaviors in Autism Spectrum Disorders and Typical Development: Cross-Sectional and Longitudinal Comparisons. J. Autism Dev. Disord. 2014, 44, 1207–1219. [Google Scholar] [CrossRef]
  18. Gardenier, N.C.; MacDonald, R.; Green, G. Comparison of direct observational methods for measuring stereotypic behavior in children with autism spectrum disorders. Res. Dev. Disabil. 2004, 25, 99–118. [Google Scholar] [CrossRef]
  19. Zemouri, R.; Zerhouni, N.; Racoceanu, D. Deep Learning in the Biomedical Applications: Recent and Future Status. Appl. Sci. 2019, 9, 1526. [Google Scholar] [CrossRef] [Green Version]
  20. Cao, C.; Liu, F.; Tan, H.; Song, D.; Shu, W.; Li, W.; Zhou, Y.; Bo, X.; Xie, Z. Deep Learning and Its Applications in Biomedicine. Genom. Proteom. Bioinform. 2018, 16, 17–32. [Google Scholar] [CrossRef]
  21. Jones, W.; Alasoo, K.; Fishman, D.; Parts, L. Computational biology: Deep learning. Emerg. Top. Life Sci. 2017, 1, 257–274. [Google Scholar]
  22. Goncalves, N.; Rodrigues, J.L.; Costa, S.; Soares, F. Automatic detection of stereotyped hand flapping movements: Two different approaches. In Proceedings of the 2012 IEEE RO-MAN: The 21st IEEE International Symposium on Robot and Human Interactive Communication, Paris, France, 9–13 September 2012; pp. 392–397. [Google Scholar]
  23. Crippa, A.; Salvatore, C.; Perego, P.; Forti, S.; Nobile, M.; Molteni, M.; Castiglioni, I. Use of Machine Learning to Identify Children with Autism and Their Motor Abnormalities. J. Autism Dev. Disord. 2015, 45, 2146–2156. [Google Scholar] [CrossRef] [PubMed]
  24. Minnen, D.; Starner, T.; Ward, J.A.; Lukowicz, P.; Troster, G. Recognizing and discovering human actions from on-body sensor data. In Proceedings of the 2005 IEEE International Conference on Multimedia and Expo, Amsterdam, The Netherlands, 6 July 2005; pp. 1545–1548. [Google Scholar]
  25. Min, C.H.; Tewfik, A.H.; Kim, Y.; Menard, R. Optimal sensor location for body sensor network to detect self-stimulatory behaviors of children with autism spectrum disorder. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; Volume 2009, pp. 3489–3492. [Google Scholar]
  26. Duh, D.J.; Chang, S.Y.; Chen, S.Y.; Kan, C.C. Automatic Broadcast Soccer Video Analysis, Player Detection, and Tracking Based on Color Histogram. In Intelligent Technologies and Engineering Systems; Springer: New York, NY, USA, 2013; pp. 123–130. [Google Scholar]
  27. Wang, Z.; Yu, J.; He, Y.; Guan, T. Affection arousal based highlight extraction for soccer video. Multimed. Tools Appl. 2014, 73, 519–546. [Google Scholar] [CrossRef]
  28. Chen, C.M.; Chen, L.H. A novel method for slow motion replay detection in broadcast basketball video. Multimed. Tools Appl. 2015, 74, 9573–9593. [Google Scholar] [CrossRef]
  29. Liu, J.; Carr, P.; Collins, R.T.; Liu, Y. Tracking Sports Players with Context-Conditioned Motion Models. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1830–1837. [Google Scholar]
  30. Archana, M.; Geetha, M.K. An Efficient Ball and Player Detection in Broadcast Tennis Video. In Intelligent Systems Technologies and Applications; Springer: Cham, Switzerland, 2016; pp. 427–436. [Google Scholar]
  31. FarajiDavar, N.; de Campos, T.; Kittler, J.; Yan, F. Transductive transfer learning for action recognition in tennis games. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011; pp. 1548–1553. [Google Scholar]
  32. Kong, Y.; Wei, Z.; Huang, S. Automatic analysis of complex athlete techniques in broadcast taekwondo video. Multimed. Tools Appl. 2018, 77, 13643–13660. [Google Scholar] [CrossRef]
  33. Pinto, T.; Faria, E.; Cunha, P.; Soares, F.; Carvalho, V.; Carvalho, H. Recording of occurrences through image processing in Taekwondo training: First insights. In European Congress on Computational Methods in Applied Sciences and Engineering; Springer: Cham, Switzerland, 2017; Volume 27, pp. 427–436. [Google Scholar]
  34. Mahmud, M.; Kaiser, M.S.; Hussain, A.; Vassanelli, S. Applications of Deep Learning and Reinforcement Learning to Biological Data. IEEE Trans. Neural Netw. 2018, 29, 2063–2079. [Google Scholar] [CrossRef] [Green Version]
  35. Ravi, D.; Wong, C.; Deligianni, F.; Berthelot, M.; Andreu-Perez, J.; Lo, B.; Yang, G.Z. Deep Learning for Health Informatics. Biomed. Health Inform. 2017, 21, 4–21. [Google Scholar] [CrossRef] [Green Version]
  36. Min, S.; Lee, B.; Yoon, S. Deep learning in bioinformatics. Brief. Bioinform. 2016, 18, 851–869. [Google Scholar] [CrossRef] [Green Version]
  37. Angermueller, C.; Pärnamaa, T.; Parts, L.; Stegle, O. Deep learning for computational biology. Mol. Syst. Biol. 2016, 12, 878. [Google Scholar] [CrossRef]
  38. Rad, N.M.; Furlanello, C. Applying Deep Learning to Stereotypical Motor Movement Detection in Autism Spectrum Disorders. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining Workshops (ICDMW), Barcelona, Spain, 12–15 December 2016; pp. 1235–1242. [Google Scholar]
  39. Cook, A.; Mandal, B.; Berry, D.; Johnson, M. Towards Automatic Screening of Typical and Atypical Behaviors in Children with Autism. In Proceedings of the 2019 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Washington, DC, USA, 5–8 October 2019; pp. 504–510. [Google Scholar]
  40. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  41. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the ICLR 2015: International Conference on Learning Representations 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  42. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  43. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  44. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  45. Gers, F.A.; Schmidhuber, J.A.; Cummins, F.A. Learning to Forget: Continual Prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef] [PubMed]
  46. Nakai, M.; Tsunoda, Y.; Hayashi, H.; Murakoshi, H. Prediction of Basketball Free Throw Shooting by OpenPose. In JSAI International Symposium on Artificial Intelligence; Springer: Cham, Switzerland, 2018; pp. 435–446. [Google Scholar]
  47. Albinali, F.; Goodwin, M.S.; Intille, S. Fast track article: Detecting stereotypical motor movements in the classroom using accelerometry and pattern recognition algorithms. Pervasive Mob. Comput. 2012, 8, 103–114. [Google Scholar] [CrossRef]
  48. Albinali, F.; Goodwin, M.S.; Intille, S.S. Recognizing stereotypical motor movements in the laboratory and classroom: A case study with children on the autism spectrum. In Proceedings of the 11th International Conference on Ubiquitous Computing, Orlando, FL, USA, 30 September–3 October 2009; pp. 71–80. [Google Scholar]
  49. Goodwin, M.S.; Intille, S.S.; Albinali, F.; Velicer, W.F. Automated Detection of Stereotypical Motor Movements. J. Autism Dev. Disord. 2011, 41, 770–782. [Google Scholar] [CrossRef] [PubMed]
  50. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. RMPE: Regional Multi-person Pose Estimation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2353–2362. [Google Scholar]
Figure 1. Flow chart of proposed ASD children action recognition method.
Figure 1. Flow chart of proposed ASD children action recognition method.
Sensors 21 00411 g001
Figure 2. Our method first extracts the de-noised skeleton data (body key points coordinates) of ASD children in the video sequence and then analyzes the skeleton data through the LSTM network. Finally, the actions of ASD children are divided into five categories.
Figure 2. Our method first extracts the de-noised skeleton data (body key points coordinates) of ASD children in the video sequence and then analyzes the skeleton data through the LSTM network. Finally, the actions of ASD children are divided into five categories.
Sensors 21 00411 g002
Figure 3. The pipeline of OpenPose.
Figure 3. The pipeline of OpenPose.
Sensors 21 00411 g003
Figure 4. Key points detected using OpenPose.
Figure 4. Key points detected using OpenPose.
Sensors 21 00411 g004
Figure 5. Proposed four de-noised measures. (a) Scale the co-ordinate. (b) Remove all joints on the head. (c) Discard frames without Neck or Hip. (d) Fill in the missing joints.
Figure 5. Proposed four de-noised measures. (a) Scale the co-ordinate. (b) Remove all joints on the head. (c) Discard frames without Neck or Hip. (d) Fill in the missing joints.
Sensors 21 00411 g005
Figure 6. Structure diagram of the LSTM unit.
Figure 6. Structure diagram of the LSTM unit.
Sensors 21 00411 g006
Figure 7. Schematic diagram of the ASD children action recognition network.
Figure 7. Schematic diagram of the ASD children action recognition network.
Sensors 21 00411 g007
Figure 8. Data collection system.
Figure 8. Data collection system.
Sensors 21 00411 g008
Figure 9. The recognition accuracy of the LSTM model under different time steps and hidden states.
Figure 9. The recognition accuracy of the LSTM model under different time steps and hidden states.
Sensors 21 00411 g009
Figure 10. Action recognition accuracy of the model under original and de-noised skeleton data. (a) Varying the number of time steps. (b) Varying the number of hidden state.
Figure 10. Action recognition accuracy of the model under original and de-noised skeleton data. (a) Varying the number of time steps. (b) Varying the number of hidden state.
Sensors 21 00411 g010
Figure 11. Confusion matrices obtained from original and de-noised skeleton data using the best parameters. (a) The matrices with original skeleton data. (b) The matrices with de-noised skeleton data.
Figure 11. Confusion matrices obtained from original and de-noised skeleton data using the best parameters. (a) The matrices with original skeleton data. (b) The matrices with de-noised skeleton data.
Sensors 21 00411 g011
Figure 12. Sample images in real-time tests. (a) Sit. (b) Stand. (c) Squat. (d) Shake the body. (e) Shake hands. (f) Multi-person.
Figure 12. Sample images in real-time tests. (a) Sit. (b) Stand. (c) Squat. (d) Shake the body. (e) Shake hands. (f) Multi-person.
Sensors 21 00411 g012
Figure 13. Action recognition accuracy of five different methods.(a) Using original skeleton data. (b) Using de-noised skeleton data.
Figure 13. Action recognition accuracy of five different methods.(a) Using original skeleton data. (b) Using de-noised skeleton data.
Sensors 21 00411 g013
Table 1. Dataset description.
Table 1. Dataset description.
Action No.Action NameNo. of Frames No. of Sequences
1Sit7245 189
2Stand7903207
3Squat5997178
4Shake the body10,613251
5Shake hands9118237
Total40,8761062
Table 2. Summary of the computed features. Bold represents the features selected for experiments.
Table 2. Summary of the computed features. Bold represents the features selected for experiments.
FeatureMeaning
Method 1 [39] m e a n δ x , m e a n δ y , m e a n δ P Mean velocity change in X co-ordinate, Y co-ordinate, and overall position of each key point
σ δ x , σ δ y , σ δ P Standard deviation of velocity change in X co-ordinate, Y co-ordinate, and overall position of each key point
max ρ m Maximum magnitude of the autocorrelogram maxima computed for each of the three velocity measures tracked for each key point
mean ρ m Mean deviation of the autocorrelogram maxima computed for each of the three velocity measures tracked for each key point
σ ρ m Standard deviation of the autocorrelogram maxima computed for each of the three velocity measures tracked for each key point
X Normalized joint positions
Method 2 X s Concatenation of key points’ positions of N frames
HIn X s : Average height of the skeleton. It equals the length from Neck to Thigh.
V b o d y In X s : Velocity of the body
X Normalized joint positions
V j o i n s In X : Velocity of all joints
Table 3. P, R, and F1-score of five different methods under the original and de-noised skeleton data.
Table 3. P, R, and F1-score of five different methods under the original and de-noised skeleton data.
SitStandSquatShake the BodyShake Hands
OriginalDe-NoisedOriginalDe-NoisedOriginalDe-NoisedOriginalDe-NoisedOriginalDe-Noised
Method 1 + Decision Tree [39]P0.44980.5240.86340.88990.55820.61690.78960.84170.75840.8235
R0.6190.67620.73750.80830.63890.68890.64850.70910.72860.8
F1-score0.5210.59050.79550.84710.59580.65090.71210.76970.74320.8116
Method 2 + linear SVMP0.56920.61090.84230.90950.64620.66670.83450.89550.80630.8339
R0.68570.74760.77920.83750.70.75560.71820.77880.72860.8071
F1-score0.6220.67240.80950.8720.6720.70840.77190.83310.76550.8203
Method 2 + Random ForestsP0.57740.65740.8950.93340.63640.70050.88410.92630.80440.8856
R0.72860.78570.81670.87920.73890.80560.73940.80.77860.8517
F1-score0.64420.71580.85410.90550.68380.74940.80530.85850.79130.8711
Method 2 + DNNP0.58240.67980.91480.94370.66030.74380.92360.96130.86920.9145
R0.75710.8190.850.90830.76670.83890.76970.82730.80710.8786
F1-score0.65840.74290.88120.92570.70950.78850.83970.88930.8370.8962
Proposed methodP0.68160.84480.91840.98350.73870.88520.93910.98350.92270.9821
R0.79520.93330.93750.99580.81670.90.79390.9030.89640.9786
F1-score0.7340.88680.92790.98960.77570.89250.86040.94150.90940.9803
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Tian, Y.; Wu, P.; Chen, D. Application of Skeleton Data and Long Short-Term Memory in Action Recognition of Children with Autism Spectrum Disorder. Sensors 2021, 21, 411. https://doi.org/10.3390/s21020411

AMA Style

Zhang Y, Tian Y, Wu P, Chen D. Application of Skeleton Data and Long Short-Term Memory in Action Recognition of Children with Autism Spectrum Disorder. Sensors. 2021; 21(2):411. https://doi.org/10.3390/s21020411

Chicago/Turabian Style

Zhang, Yunkai, Yinghong Tian, Pingyi Wu, and Dongfan Chen. 2021. "Application of Skeleton Data and Long Short-Term Memory in Action Recognition of Children with Autism Spectrum Disorder" Sensors 21, no. 2: 411. https://doi.org/10.3390/s21020411

APA Style

Zhang, Y., Tian, Y., Wu, P., & Chen, D. (2021). Application of Skeleton Data and Long Short-Term Memory in Action Recognition of Children with Autism Spectrum Disorder. Sensors, 21(2), 411. https://doi.org/10.3390/s21020411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop