Next Article in Journal
Water–Air Interface to Mimic In Vitro Tumoral Cell Migration in Complex Micro-Environments
Next Article in Special Issue
A Pervasive Pulmonary Function Estimation System with Six-Minute Walking Test
Previous Article in Journal
New Horizons for MXenes in Biosensing Applications
Previous Article in Special Issue
Recent Advances in Stretchable and Wearable Capacitive Electrophysiological Sensors for Long-Term Health Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Applications of Metaheuristics for Human Activity Recognition and Fall Detection Using Wearable Sensors: A Comprehensive Analysis

by
Mohammed A. A. Al-qaness
1,*,
Ahmed M. Helmi
2,3,
Abdelghani Dahou
4,5 and
Mohamed Abd Elaziz
6,7,8,9
1
College of Physics and Electronic Information Engineering, Zhejiang Normal University, Jinhua 321004, China
2
College of Engineering and Information Technology, Buraydah Private Colleges, Buraydah 51418, Saudi Arabia
3
Computer and Systems Engineering Department, Faculty of Engineering, Zagazig University, Zagazig 44519, Egypt
4
Mathematics and Computer Science Department, University of Ahmed DRAIA, Adrar 01000, Algeria
5
LDDI Laboratory, Faculty of Science and Technology, University of Ahmed DRAIA, Adrar 01000, Algeria
6
Department of Mathematics, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
7
Artificial Intelligence Research Center (AIRC), College of Engineering and Information Technology, Ajman University, Ajman 346, United Arab Emirates
8
Faculty of Computer Science and Engineering, Galala University, Suez 435611, Egypt
9
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 13-5053, Lebanon
*
Author to whom correspondence should be addressed.
Biosensors 2022, 12(10), 821; https://doi.org/10.3390/bios12100821
Submission received: 31 July 2022 / Revised: 25 September 2022 / Accepted: 28 September 2022 / Published: 3 October 2022
(This article belongs to the Special Issue Wearable Sensing for Health Monitoring)

Abstract

:
In this paper, we study the applications of metaheuristics (MH) optimization algorithms in human activity recognition (HAR) and fall detection based on sensor data. It is known that MH algorithms have been utilized in complex engineering and optimization problems, including feature selection (FS). Thus, in this regard, this paper used nine MH algorithms as FS methods to boost the classification accuracy of the HAR and fall detection applications. The applied MH were the Aquila optimizer (AO), arithmetic optimization algorithm (AOA), marine predators algorithm (MPA), artificial bee colony (ABC) algorithm, genetic algorithm (GA), slime mold algorithm (SMA), grey wolf optimizer (GWO), whale optimization algorithm (WOA), and particle swarm optimization algorithm (PSO). First, we applied efficient prepossessing and segmentation methods to reveal the motion patterns and reduce the time complexities. Second, we developed a light feature extraction technique using advanced deep learning approaches. The developed model was ResRNN and was composed of several building blocks from deep learning networks including convolution neural networks (CNN), residual networks, and bidirectional recurrent neural networks (BiRNN). Third, we applied the mentioned MH algorithms to select the optimal features and boost classification accuracy. Finally, the support vector machine and random forest classifiers were employed to classify each activity in the case of multi-classification and to detect fall and non-fall actions in the case of binary classification. We used seven different and complex datasets for the multi-classification case: the PAMMP2, Sis-Fall, UniMiB SHAR, OPPORTUNITY, WISDM, UCI-HAR, and KU-HAR datasets. In addition, we used the Sis-Fall dataset for the binary classification (fall detection). We compared the results of the nine MH optimization methods using different performance indicators. We concluded that MH optimization algorithms had promising performance in HAR and fall detection applications.

1. Introduction

1.1. Motivation

Many context-awareness services allow computers to track and recognize human motion and activities. It is clear that smart environments can be investigated due to the increased usage of smart devices in homes as well as Internet of things (IoT)-supported devices [1]. Human activity recognition (HAR) can be defined as the process of identifying a person’s actions from a series of measurements captured by different mechanisms such as cameras (computer vision mechanism) [2], interior sensors [3,4], radars [5], wireless signals [6], and others.
This study focuses only on the HAR applications that use sensor data. Sensor-based HAR mechanisms have the advantage of being able to collect sensing data at any time and in any location as well as obtain information that is unique to each user. The disadvantage is that each user must own the sensor equipment; however, the widespread popularity of smartphones and smartwatches has solved this problem. In addition, there are still some issues, such as different measurement conditions, e.g., device kind; the installed position of the sensors; the wearing technique; and the measuring applications, which differ from one user to the next and from one measurement date to the next. To solve these challenges, in this study, we use different and complex datasets that cover all types of sensor data proposed for HAR applications.
The classification process in HAR applications is a challenge due to the complexity of the sensors’ datasets. Thus, the feature selection process plays a significant role in HAR applications to reduce computation time and select only the optimal feature set. The selected MH optimization algorithms have been widely employed for feature selection in different domains. For example, the AO was utilized as an FS technique to enhance the intrusion detection system in the IoT and cloud environments [7,8]. In [9], the AO was employed as an FS for COVID-19 image classification. The AOA was utilized in various FS tasks such as osteosarcoma detection [10] and our previous study of HAR [11]. The MPA was also utilized in different applications, such as COVID-19 CT image classification [12], breast cancer classification [13], and wind power forecasting [14]. The SMA was adopted in different FS applications such as medical data classification [15], parameter identification of photovoltaic systems [16], and crude oil forecasting. The GA was also utilized for different applications such as lung cancer classification [17], data mining [18], and credit risk assessment [19]. The GWO was adopted in advanced applications such as mammogram image analysis [20], Parkinson’s disease diagnosis [21], and chronic disease prediction [22]. The WOA was applied in several FA applications such as email spam detection [23], software fault prediction [24], and medical data classification [25]. The ABC was adopted in different FS tasks such as colon cancer detection [26], medical image classification [27], and IDS [28]. The PSO is one of earlier methods that was employed for different FS such as data mining in the oil industry [29], breast cancer recurrence prediction [30], and intrusion detection system (IDS) [31].

1.2. Paper—Main Contributions

In this paper, we propose an efficient HAR approach that can be utilized for multi-classification (for different daily activities) and binary classification (for fall detection, including fall or non-fall action). The developed approach depends on two main processes. The first is feature extraction in which a light deep learning (DL) approach called ResRNN is developed to extract a subset of features that represent human motion from the sensor data. The ResRNN is built based on several building blocks from DL networks including convolution neural networks (CNN), residual networks, and bidirectional recurrent neural network (BiRNN). The second approach is to leverage recent advances in MH optimization algorithms in feature selection. We tested nine MH algorithms in feature selection to build an efficient HAR system, namely the Aquila optimizer (AO), arithmetic optimization algorithm (AOA), marine predators algorithm (MPA), artificial bee colony (ABC) algorithm, genetic algorithm (GA), slime mold algorithm (SMA), grey wolf optimizer (GWO), whale optimization algorithm (WOA), and particle swarm optimization algorithm (PSO). The main problems of sensor-based HAR approaches are the different positions of sensors in the human body, the types of motions, the complexities of human activities, the number of activities, and the number of tested users (the people who implemented the tested activities). To solve those challenges and comprehensively analyze the MH algorithms’ applications in HAR applications, we used seven datasets that covered all the mentioned challenges, namely, UCI-HAR, WISDM, UniMiB SHAR, OPPORTUNITY, KU-HAR, Sis-Fall, and PAMMP2. To sum up, we present the following contributions to the field of HAR:
  • We studied the impacts of metaheuristic (MH) optimization algorithms on human activity recognition (HAR) and fall detection using body-attached sensor data. We tested nine MH algorithms and compared their performances.
  • We developed a light feature extraction approach called ResRNN using several deep learning models, such as convolution neural networks (CNN), residual networks, and bidirectional recurrent neural network (BiRNN), to expose the related features from the collected signal data.
  • We examined the suggested feature selection methods based on MH algorithms using different and complex datasets that covered all the aspects of sensor data for HAR and fall-detection applications.

1.3. Paper—Organization

In Section 2, we describe several recently published studies for sensor-based human activity recognition using different datasets. In Section 3, we describe the preliminaries of the applied methods including the nine MH optimization algorithms, the basic definitions of the seven datasets, the developed feature extraction method, and the feature optimization process. Moreover, the evaluation experiments’ settings, evaluation, comparison, and results are presented in Section 4. We conclude the paper in Section 6.

2. Related Work

In this section, we present a quick review of the previous studies proposed for senor-based HAR applications using different public datasets. Deep learning models have been widely utilized in different fields [32] including HAR and fall detection.
In [33], the authors proposed a new deep learning method using a hybrid gated recurrent unit (GRU) and LSTM -based RNN model for HAR. They used the TRECVID dataset to test the performance of the proposed deep learning model, which showed significant performance. Wang et al. [34] applied the traditional CNN model with an attention mechanism to extract sensor data features. They used the well-known UCI-HAR dataset to assess the performance of the CNN-based HAR model. Xia et al. [35] applied a hybrid LSTM and CNN to classify different human activities using three public datasets: OPPORTUNITY, UCI-HAR, and WISDM. The main goal of this combined model was to automatically extract features from the sensor data with fewer parameters to reduce the computation complexity. Sikder et al. [36] employed a sequential deep learning method to classify human activities using three public datasets: UCI-HAR, KU-HAR, and WISDM. They applied a specific preprocessing method using a specific matrix formulated from sensor data. The classification outcomes achieved high accuracy rates for the three datasets. Kumar and Suresh [37] proposed a new HAR model called DeepTransHHAR using deep learning with heterogeneous deep transfer learning. They used two well-known datasets, KU-HAR and HHAR, to evaluate the developed DeepTransHHAR, achieving acceptable classification results. Dua et al. [38] presented a combined deep learning model using the CNN and GUR to extract features automatically and recognize implemented activities. They evaluated the combined CNN-GRU with three public datasets, PAMAP2, WISDM, and UCI-HAR. Khatun et al. [39] proposed a combined deep learning model (CNN-LSTM) with a self-attention mechanism for HAR smartphone applications. They assessed the quality of the combined model using two benchmark datasets, UCI-HAR and MHEALTH. Ghate et al. [40] applied a hybrid of deep CNN with the random forest classifier to classify human activities using two datasets, WISDM and UCI-HAR. It was compared to other models, such as LSTM, GRU, and CNN, and they found that CNN with RF obtained the best classification accuracy. Ronald et al. [41] developed a deep leering model based on Inception-ResNet architecture called iSPLInception to classify different human activities. They assessed the quality of the proposed model with four well-known HAR datasets, UCI HAR, OPPORTUNITY, PAMAP2, and Daphnet freezing of the gait dataset. Tufek et al. [42] studied the capability of the LSTM and CNN models to classify human activities based on sensor data. They used the UCI-HAR dataset to evaluate both the LSTM and CNN models. They found that the LSTM model had better classification accuracy than the CNN model. Gao et al. [43] developed a new HAR model based on a multi-branch CNN architecture using a selective-kernel mechanism. They employed four datasets to test the performance of the developed CNN model, namely UCI-HAR, WISDM, UniMiB SHAR, OPPORTUNITY, and PAMAP2. Huang et al. [44] employed a shallow CNN model with channel selectivity for HAR applications. They used five benchmark datasets to verify the performance of the CNN-based HAR model called OPPORTUNITY, namely the UCI-HAR, PAMAP2, WISDM, and UniMiB SHAR datasets. Gao et al. [45] suggested a dual attention deep learning approach, namely DanHAR, for HAR applications using sensor data. This model can blend channel and temporal attention on the residual network to enhance feature representation capability. The WISDM, UniMiB SHAR, PAMAP2, and OPPORTUNITY datasets were utilized to assess the classification capability of the developed DanHAR model, which obtained higher accuracy rates than several other models. Tang et al. [46] presented a CNN-based HAR method. The main idea was to boost the multi-scale feature representation capability using one feature layer. The UniMiB SHAR, PAMAP2, UCI-HAR, and WISDM datasets were utilized to assess the method performance with comparisons to several CNN models. The outcomes showed that the applied CNN-based feature extraction approach boosted the classification accuracy of all datasets.
It is worth mentioning that MH optimization algorithms including SI methods have limited applications in HAR systems. Almost all of the MH and SI method applications were in computer vision-based HAR applications such as the PSO [47,48] and genetic algorithm (GA) [49]. For sensor data, we carried out a simple implementation for the arithmetic optimization algorithm (AOA) with the KU-HAR, UCI-HAR, and WISDM datasets from [11] and the grey wolf optimizer using WISDM and UCI-HAR [50].
Unlike previous approaches, this paper presents a comprehensive analysis of MH optimization algorithms in HAR applications using sensor data. This is the first comprehensive study that investigates nine MH algorithms in HAR applications with seven public datasets, including comprehensive as well as normal and complex daily activities. The paper opens a new direction for further investigating MH optimization algorithms to boost HAR applications based on smart devices installed with embedded sensors.

3. Materials and Methods

In this section, we describe the preliminaries and the backgrounds of the used HAR datasets (KU-HAR, UCI-HAR, WISDM, PAMAP2, OPPORTUNITY, SiS-Fall, and UniMiB SHAR) and the applied feature selection methods based on different MH optimization algorithms, namely the Aquila optimizer (AO), arithmetic optimization algorithm (AOA), marine predators algorithm (MPA), artificial bee colony (ABC) algorithm, genetic algorithm (GA), slime mold algorithm (SMA), grey wolf optimizer (GWO), whale optimization algorithm (WOA), and particle swarm optimization algorithm (PSO). The main workflow of the HAR approach described in this paper can be seen in Figure 1. It consists of four main stages. The first is the data collection from the body-attached sensors. In this study, we used seven public datasets containing comprehensive and complex activities in different environments and were collected by different people in different countries. The second stage is data prepossessing. We used different cleaning and filtering methods to obtain clean data that can expose the impact of human activities on the collected signals. The third stage is the feature extraction stage. In this paper, we developed a new deep learning model to extract relevant features from the prepossessed datasets. The model is called ResRNN. It consists of several building blocks using different deep learning structures. (BiRNN). The fourth stage is the classification stage. In this stage, MH optimization methods are employed to select optimal features, reduce computation complexity, and optimize the classification process. Then, a classifier can be employed to recognize the implemented activities. This study tested and compared two well-known classifiers, the SVM and RF classification models.

3.1. Experimental HAR Datasets

3.1.1. KU-HAR

Skider and Nahid [51] collected and built a new dataset for human daily activities and actions, called the KU-HAR dataset. They used the gyroscopes and accelerometers of smartphones (See Figure 2a). They collected 1945 raw samples for 18 different classes (human activities) that were implemented by 90 users aged between 18 and 34. The collected samples contained 20,750 subsamples, each with a period of 3 s.

3.1.2. OPPORTUNITY (Oppo)

The OPPORTUNITY dataset was collected by [52] using wearable inertia measurement units (IMUs) that were placed on human users’ bodies at seven different positions (see Figure 2b). Four human users implemented 18 daily motions and activities in a simulation room, such as “close/open doors/fridges/drawers, clear tables, toggle switches, and sip from cups,” and others. The collected samples had 77-dimensional attribute columns that characterized the activity signals and have been sampled at a frequency of 30 Hz. It is worth mentioning that the “emphNULL” class represents 72% of all collected samples, which represents a non-relevant activity. More details can be found in [52].

3.1.3. PAMAP2

The PAMAP2 dataset was collected by [53] with nine human users who implemented different daily activities. The users were asked to wear three IMUs at the ankle, chest, and wrist positions, as clarified in Figure 2c. During data collection, for each activity (one motion), the magnetometer, gyroscope, and accelerometer reported 27-dimensional signal spaces at a frequency rate of 100 Hz. A total of 2,844,868 raw samples for 12 different human activities and motions were collected during the data collection period.

3.1.4. Sis-Fall

The SiS-Fall dataset was collected by [54] using an embedded unit with a Kinets MKL25Z128VLK4 microcontroller, which had two accelerometers (MMA8451Q and ADXL345) and one ITG3200 gyroscope (see Figure 2d). It contained two types of human motions; the first was called activities of daily living (ADLs) and the second was called fall actions. The ADL type contained 19 activities: jogging, walking, walking downstairs and upstairs, standing, jumping, sitting, stumbling while walking, and different sitting motions. The fall type contained 15 fall actions such as fall forward, fall backward, lateral fall, fall while jogging, fall while getting up, fall while sitting down, and others. A total of 38 human users implemented the activities, with 15 elderly users (7 female and 8 male) and 23 young people (12 female and 11 male).

3.1.5. UCI-HAR

In the collection stage of the UCI-HAR dataset, a Samsung Galaxy SII smartphone was used [55] (see Figure 2e). A total of 30 users implemented six daily activities: walking, walking upstairs, walking downstairs, sitting, standing, and lying. The gyroscope and accelerometer records were collected separately. Thus, a total of 10,299 samples were collected at a rate of 50 Hz, with a sliding window of 2.56 s.

3.1.6. UniMiB SHAR

In [56], a new HAR dataset was collected based on the accelerometers of smartphones. During the data collection stage, the users were asked to put the smartphone in their pockets (right and left), as shown in Figure 2f. The UniMiB SHAR had two types of human motions: daily activities and fall actions. The first contained nine activities, (“Running, Walking, Standing UPFS (Standing up from sitting), Jumping, Standing UPFL (Standing up from lying), Going Ups (Going upstairs), Going Downs (Going downstairs), sitting down, and Lying DownFS (Lying down from standing)”). The second type had eight fall actions (“Hitting Obstacle, Falling right, Falling Back, Falling Left, Falling BackSC (Falling backward sitting chair), Falling with protection strategy (Falling withPS), Falling Forward, and Syncope”). A total of 30 human users were asked to implement the mentioned motions, and 11,771 raw samples were collected during the experiments.

3.1.7. WISDM

The wireless sensor data mining (WISDM) dataset was collected by [57] using Android smart devices (i.e., smartwatches and smartphones; see Figure 2g) ). A total of 36 users implemented six physical activities, such as jogging, walking, sitting, upstairs, downstairs, and standing. A total of 1,098,207 raw samples were collected during the experiments, where each sample indicated an accelerometer data measurement (at 20 Hz).

3.2. Applied Metaheuristic Optimization Algorithms

3.2.1. Aquila Optimizer (AO)

The AO algorithm was developed by [58]. It is a population-based MH optimization method. The main idea of the AO is to mimic aquilas’ natural behavior in catching prey in the wild. The workflow and detailed description of the AO can be found in [58].

3.2.2. Arithmetic Optimization Algorithm (AOA)

The AOA optimization method was developed by [59] and was inspired by basic mathematical operations (i.e., , + , * , and /). The basic steps and detailed description of the workflow of the AOA can be found in [59].

3.2.3. Marine Predators Algorithm (MPA)

The MPA was developed by [60] and was inspired by the natural behavior of predators and prey. The prey and predators can be considered search agents, whereas a predator searches for prey and a prey searches for food. Similar to other MH techniques, it starts by generating a set of solutions (agents) as an initialization. After that, the agents can be modified depending on the main workflow of the algorithm. More details and mathematical descriptions can be found in [60].

3.2.4. Slime Mold Algorithm

The SMA was developed by [61] as a natural-inspired MH technique and belongs to swarm intelligence-inspired algorithms. The main idea of the SMA is to mimic the natural behaviors of slime mold oscillations and their propagation wave feedback depending on the bio-oscillator. It generates the optimum routes to connect food. More details can be found in [61].

3.2.5. Whale Optimization Algorithm (WOA)

The WOA was proposed by [62] as a natural swarm intelligence method. It was inspired by the behavior of humpback whales in nature. In contrast, the whale’s position indicates the agent’s solution to a problem and it can be updated depending on the whale’s behavior when attacking prey. Two attacking techniques are used in the mathematical definition of the WOA. More details can be found in [62].

3.2.6. Artificial Bee Colony (ABC) Algorithm

The ABC algorithm is a swarm intelligence-based MH method developed by Karaboga in 2005 [63] to solve complex optimization, numerical, and engineering problems. The ABC was inspired by the intelligent foraging behavior of honey bees. The ABC generally depends on the model developed by [64] based on the foraging behavior of honey bee colonies. This model has three phases, employed and unemployed foraging bees and food sources. The employed and unemployed bees are searching for food sources. This behavior can be represented in the ABC algorithm by generating a population that has agents. Thus, an agent (a colony of artificial forager bees) searches for a good solution (a good food source) (good solutions for a given problem). More details about the ABC algorithm can be found in [65]

3.2.7. Grey Wolf Optimizer (GWO)

The GWO is a swarm intelligence-based method inspired by the natural behavior of Canis lupus (grey wolves). It was developed by [66] and received wide attention for solving different optimization problems. It simulates the leadership hierarchy’s natural behavior and grey wolves’ hunting processes. Four kinds of grey wolves (including alpha, beta, omega, and delta) are applied to simulate the leadership hierarchy. Additionally, there are three main steps in the hunting process: searching for prey, encircling prey, and attacking prey. The mathematical definition and a more detailed description can be found in [66].

3.2.8. Genetic Algorithm

The GA was presented by [67] as a population-based MH algorithm. It can be considered an evolution-based algorithm. In the definition of the GA, each individual (agent) in the population is referred to as the solution. It has three phases that can be utilized for updating the solutions (agents) called the selection, crossover, and mutation processes. In the first phase, two agents can be randomly selected to boost the diversity of the population. The second phase, the crossover mechanism, can generate new agents (individuals) from the selected agents (parents). After that, the mutation process can be utilized to replace randomly selected agents with random values belonging to the search domain. Lastly, the current population can be updated based on the fitness values of the newly initialized agents and their parents. Furthermore, the population can be updated using the three processes mentioned (selection, crossover, and mutation) until the stop criteria are met. The mathematical definition and a detailed description of the GA can be found in [68].

3.2.9. Particle Swarm Optimization (PSO)

The PSO is a population-based intelligence method proposed by [69]. It is one of the earliest swarm intelligence methods and has received much attention in previous decades. The PSO’s main workflow and mathematical details can be found in [69].

3.3. Data Cleaning, Filtration, and Segmentation

In general, HAR applications are considered real-time applications in which the recognition process must be performed immediately. Thus, the time window is generally set between 2 and 10 s [70,71,72]. In this section, we briefly describe the data preparation for all seven employed datasets.
KU-HAR. The KU-HAR dataset samples contain 300 data points; each segment is 3 s with non-overlapping, as described in the original study of the KU-HAR dataset [51]. At the same time, only one accelerometer was used to collect the tracing human activity samples. The Butterworth lowpass filter was applied to remove noise and filter the collected signals.
OPPORTUNITY (Oppo). In the Oppo dataset, there are many NAN values. The collected samples were gathered using wearable IMUs at a low-frequency of 30 Hz. Thus, a light linear interpolation was performed to prepare the collected samples. Following [45,73], the time window was set to 2.133 s, which had 64 data points. Furthermore, overlapping was employed for the segmentation process at a rate of 50%.
PAMAP2. To process the PAMAP2 collected samples, a lowpass filter called the Butterworth filter was used with a cut-off frequency of 20 Hz. The time window for activity segmentation was set to 5 s [45,74]. Overlapping also was used at a rate of 50% for data segmentation.
SisFall. Following previous studies such as [75], for this dataset, the time window was set to 3 s, in which 600 points were collected for each activity sample. A lowpass filter was also utilized to remove noise, as suggested by the original study of the Sis-Fall dataset [54].
UCI-HAR. The samples of the UCI-HAR dataset were segmented with a time window of 2.56 s, with a frequency of 50 Hz. Also, overlapping with 50% was performed. More details can be found in the original study [55].
UniMiB SHAR. The segmentation of the UniMiB SHAR dataset was set by the original study [56]. A time window of 3 s was used, with 150 data points for each activity sample.
WISDM. Similar to the other datasets, the Butterworth lowpass filter was employed to remove the noise from the collected samples, with a cut–off frequency of 10 Hz. For the collected samples, a time window with a length of 128 s was applied. Also, overlapping with 50% was implemented to train the model, as carried out by [71,76].
It is worth mentioning that we performed standardization on the raw data before feeding it to the DL model, where the input signals were scaled to unit variance along with removing the mean as in Equation (1).
z = ( x u ) s
Where the mean and standard deviation of sample data, x, are represented by u and s, respectively. And thus z becomes the corresponding standard score.

3.4. Feature Extraction

In this section, we describe the DL model used to extract features from preprocessed input data to learn better representations of the raw input data. The proposed DL model named ResRNN was composed of several building blocks from various DL networks including convolution neural networks (CNN), residual networks, and bidirectional recurrent neural networks (BiRNN). The ResRNN used parallel and sequential alignment of the components based on their structures. For instance, the CNN layers aligned parallel to extract different feature maps from the input data and learn activity nature-related features. The BiRNN layers were aligned in sequence to learn more complex and temporal-related features based on the previous layer’s output (CNN output). The input data were composed of triaxial signals (X, Y, Z), which represented each data sample based on the collected IMU signals from sensors such as the accelerometer, gyroscope, and magnetometer. At the beginning of the ResRNN, a set of convolution blocks aligned in parallel received the input data and extracted the feature maps. Based on the IMU signals collected from each sensor, the signals were distributed over the parallel convolution block with the following structure: ( C o n v B N R e L U R C M a x p o o l i n g ) . The Conv represented a convolution layer with a kernel size of ( 1 × 3 ), a stride of 1, padding of 1, and several filters of 64, 128, and 256 for the triaxial signals. The BN layer was a batch normalization layer. The ReLU layer was a Rectified Linear Unit activation layer. The RC layer was a residual connection similar to the skip connection mechanism in residual networks. The max-pooling layer was a max-pooling layer with a kernel size of 2 and a stride of 2. The learned feature maps from all parallel CNN blocks with different output channels were concatenated and fed to a BiRNN block with gated recurrent units (GRU) as the RNN structure and attention mechanism. The ResRNN benefited from the DL techniques as it could learn more complex representations and lower the computation complexity, thus generating a small model size, overcoming over-fitting and reducing the training time. The learned features from the BiGRU with the attention block were concatenated and fed to a fully connected layer (FC) with 128 neurons, which served as our feature extraction layer. Thus, the fine-tuned features in the FC layer were extracted and inputted into the feature selection algorithm. Each input sample was represented by a feature vector of size 128 extracted from the FC layer. The ResRNN’s weight was optimized for the activity classification task using the Adam optimizer with a 1 × 10 4 learning rate. A softmax layer was placed at the top of the model to perform the classification. In addition, the ResRNN was trained for 350 epochs with dropout (0.3) and early stopping set to monitor the validation loss. Figure 3 shows the architecture of the ResRNN.

3.5. Feature Optimization

Let F represent the feature vector extracted by the proposed DL model, then Equation (2) can be applied to characterize the model performance using the optimized feature vector S according to both the classification rate improvement and feature set reduction.
f ( S ) = α C E S + ( 1 α ) | S | | F | ,
where C E S represents model classification error for S, α [ 0 , 1 ] is a commonly applied balance factor to control the effect of C E S and the feature set optimization | S | | F | , where | F | denotes the feature set cardinality. Thus, a lower value, i.e.,  f 0 , in Equation (2) refers to the higher performance of the working model and vise-versa. In this work, C E S is set as the model classification accuracy defined by Equation (3).
A c c u r a c y ( A c c ) = T P T P + T N + F P + F N
where TP, TN, FP, and FN are the model rates of true positive, true negative, false positive, and false negative, respectively. Algorithm 1 illustrates the designed cost function to work under the applied MH algorithms here, where α is set to 0.99 and t h r = 0 .
Algorithm 1Cost Function ( S , T t r , T t s t , α , t h r )
1:
Input: Solution S , Train T t r and test T t s t sets, control parameter α and threshold t h r .
2:
Output: Solution cost (f)
3:
f
4:
Obtain the equivalent binary solution S B S > t h r
5:
Select the features indicated by ones in S B in each of T t r and T t s t
6:
Train the classification layer in the model using the selected features in T t r
7:
Calculate the classification predictions of the trained model for T t s t
8:
Calculate f using Equation (2)

4. Experiments

4.1. Experiments Setup

Table 1 presents the parameter settings of each MH algorithm applied in this work. For efficient computational overheads, the search settings were set to a population size of 20 and a total number of iterations of 20. To characterize the average performance of each applied MH algorithm, each experiment was repeated for 20 independent runs. Both binary and multi-class classification tasks were carried out using the SVM algorithm [77] and RF algorithm [78] under the ’Bag’ learning method, where the size of decision trees was set to 50.
In addition to Acc, precision, recall, and the F1 measure defined in Equations (3)–(6) were used here to characterize the performance of the proposed model.
P r e c i s i o n ( P r e c ) = T P ( T P + F P )
R e c a l l ( R e c ) = T P ( T P + F N )
F 1 m e a s u r e ( F 1 m ) = 2 × P r e c × R e c ( P r e c + R e c )
The classification results were validated based on previous studies where the KU-HAR, UCI-HAR, UniMiB SHAR, and WISDM datasets were randomly split into 70% training and 30% testing subsets. For the OPPORTUNITY dataset, the data records of the second, third, and fourth trials of participants 1, 2, and 3 were used only for testing. For PAMAP2, the testing set was also selected for the data records of subjects 5 and 6, which were never introduced to the model during training. For Sis-Fall, 10-fold cross-validation was applied.

4.2. Results

In this section, the performances of all the applied MH methods are discussed and analyzed. The comparison results are given in Table 2, Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12. In addition, Table 2 illustrates the average of the classification accuracy obtained using RF and SVM as classifiers to assess the different MH techniques. Additionally, the convergence curve for compared MH optimization methods using all datasets are displayed in Figure 4.
It can be seen in Table 2 that the WOA had the highest accuracy for PAMAP2, SisFallB (binary classification for fall detection), UCI-HAR, and WISDM. This was followed by the GWO, which had the best accuracy using three datasets Oppo, UniMiB SHAR, and WISDM. Meanwhile, the AO and PSO algorithms achieved the highest accuracy on two datasets, and the MPA and SMA provided better accuracy in one dataset only. In addition, we observed that RF provided better accuracy than SVM among the tested datasets and competitive algorithms.
Furthermore, Table 3 depicts the accuracy of the classification obtained using the best model (i.e., RF or SVM). Highest accuracy for KUHAR, OPPO, PAMAP2, SisFallB (binary classification), SisFallM (multi-classification), UCI-HAR, UniMiB, and WISDM was achieved with MPA, GWO, WOA, AO/WOA, AO, WOA, AO, and AO/WOA, respectively. Moreover, from these results, we can conclude that RF still performs better than SVM.
Table 4 illustrates the average reduction rate of the features obtained by each MH technique and both classifiers (i.e., RF and SVM). From these values, it can be seen that using the AO with RF and the AOA with SVM was better than other techniques on the KU-HAR dataset and the PSO with RF and the MOA with SVM were the better MH techniques on the Oppo dataset. In addition, the GWO and PSO provided better results than the other models on the PAMAP2 dataset using RF; however, by using SVM, the WOA and MPA were the best techniques. The SMA reduced the number of selected features better than the others on the SisFallB dataset using either RF or SVM. From the results of the algorithms on the SisFallM dataset, one can see that the PSO was better than the compared MH algorithms when the RF classifier was used, but by using SVM, we can see that the SMA was the best MH algorithm. For the last three datasets, we can see that the SMA had the smallest number of features on the UCI-HAR dataset with RF/SVM, UniMiB with SVM, and WISDM using SVM. However, the PSO with RF and the WOA with RF were the better algorithms according to their feature reduction rates on the UniMiB and WISDM datasets, respectively.
Table 5 lists the different classification metrics used to assess the best performing MH and classification models using the KU-HAR dataset. The AO and RF techniques showed the best performance among the other techniques on the KU-HAR dataset. As shown in Table 5, most of the activities scored more than a 90% classification score in terms of the F1 score. Activities such as Standing, Jumping, Pushing-up, and Walking-backward scored less than 86% in terms of the F1 score due to the data imbalance. Table 6 lists the results of classifying the Oppo dataset activities using the GWO and RF (SVM), where the NULL class had the majority of samples dominate the results with the highest F1 score of 96.9%. The F1 scores for the Close Drawer 1 and Close Drawer 2 activities were the lowest among all the activities and were 33.33% and 33.96%, respectively. Table 7 lists the results using the WOA and SVM on the PAMAP2 dataset. The Vacuuming activity scored lower than the other activities in F1 score and Recall. Other activity scores were close to each other due to the balanced nature of the dataset. Table 8 lists the AO and RF results on the SisFallB dataset, which is a binary classification task. The RF model showed excellent results, reaching an F1 score of 100% on both classes, Fall and No Fall. In terms of multiclass classifications of the SisFall dataset, Table 9 shows the results of SisFallM using the AO and RF on all regular activities and the Fall activity. The RF classifiers with fewer selected features using the AO compared to the original amount of features performed well on the classification of regular activities and scored 99.44% in terms of the F1 score on Fall activity. Table 10 lists the results of the WAO and RF on the UCI-HAR dataset, where the Sitting and Standing activities were less accurately classified compared to the other activities due to the similarity between the two activity signals. Table 11 lists the AO and RF results on the UniMiB SHAR dataset, where most falling activities were classified with high F1 scores. It can be seen that the Falling Rightward activity, Falling with protection, Falling backward-sitting chair, and Syncope fall had lower F1 values due to the difficult nature of the activities. Table 12 lists the WISDM dataset results using the AO and RF, which shows almost perfect scores on all classification metrics, where the RF classifier accurately differentiated between activities.
Additionally, to test the validity of the combination of MH optimization algorithms with the feature selection based on deep learning, Figure 5 presents a comparison of each applied technique (i.e., deep learning (DL) model, DL model using RF and SVM (best model (initialize results) and the optimized DL model (best model (Opt.)), which is a combination of MH and DL). The figure shows that the combination of MH and DL achieved the best results. In summary, we noticed that the applications of MH optimization algorithms showed superior performance and can be further investigated for human activity recognition and fall detection applications using body-attached sensors.

5. Comparison with Previous Related Studies

The accuracy reported in this simulation for the different considered datasets are compared in Table 13 to some relevant results obtained in previous studies.
From the table, we can see that the proposed approach recorded the best results in all datasets, except for PAMAP2 in which Gao et al. [45] achieved the best results, as reported in their study. The KUHAR dataset is so recent and there is not as much rich literature as the above-mentioned datasets. Binary analysis for the SisFall dataset (accuracy of SisFallB reaches 99.98%, see Table 2) revealed the excellent performance of the proposed model regarding the fall detection task, similar to the results reported by Mrozek et al. [79]. To the best of our knowledge, our results are the first multiclass analysis of the SisFall dataset.
Table 13. Results of related models’ accuracies ( % ) for common datasets
Table 13. Results of related models’ accuracies ( % ) for common datasets
OPPOPAMAP2UCI-HARUniMiBWISDM
Multi-ResAtt [80]86.85Multi-ResAtt [80]90.08Daho et al. [11]95.23Multi-ResAtt [80]74.94LSTM-CNN [35]95.01
Gao et al. [45]82.75Gao et al. [45]93.16LSTM-CNN [35]95.31DanHAR [45]79.03U-Net [81]96.40
Teng et al. [73]81Teng et al. [73]92.97DSmT [82]95.31Teng et al. [73]78.07MHCA [83]96.40
iSPLInception [41]88.14DanHAR [45]93.16Net-att3-pc-tanh [34]93.83Predsim ResNet [84]80.33DanHAR [45]98.85
Proposed93.90Proposed92.63Proposed95.51Proposed86.25Proposed98.95

6. Conclusions

This paper focused on applying metaheuristic (MH) optimization algorithms for daily human activity recognition and fall detection using wearable sensors such as smartphones and body-attached sensors. In general, we developed an efficient HAR and fall detection model using an integration of deep learning and MH algorithms. A new feature extraction technique based on an efficient deep learning model called ResRNN was developed to expose the related feature from the collected sensor data. Afterward, the feature selection process was implemented using nine MH algorithms, namely, the Aquila optimizer (AO), arithmetic optimization algorithm (AOA), slime mold algorithm (SMA), marine predators algorithm (MPA), genetic algorithm (GA), grey wolf optimizer (GWO), whale optimization algorithm (WOA), artificial bee colony (ABC) algorithm, and particle swarm optimization (PSO) algorithm. These MH methods were implemented to obtain optimal features, reduce computation time, and boost classification accuracy. The proposed approach was implemented in two main areas. The first was daily activity recognition (multi-classification) and the second was fall detection (binary classification). For comprehensive analysis and study, we used seven datasets that contained different and complex activities. The used datasets were UCI-HAR, Sis-Fall, WISDOM, UNIMIB-SHAR, OPPO, KU-HAR, and PAMMP2. Extensive experiments were conducted using two classifiers, support vector machine (SVM) and random forest (RF). The findings of this study showed that the MH optimization algorithms showed significant performance in HAR and fall detection applications. This study opens the possibilities of the MH to be applied for sensing applications. We suggest using more advanced methods, such as modified MH optimization methods integrated with intelligent search mechanisms, that may provide more robust results, such as levy flight, opposition-based learning, and a hybridization of two MH algorithms.

Author Contributions

Conceptualization, M.A.E.; Data curation, A.M.H.; Formal analysis, M.A.A.A.-q.; Funding acquisition, M.A.A.A.-q.; Investigation, M.A.A.A.-q.; Methodology, M.A.A.A.-q., A.M.H. and A.D.; Software, M.A.A.A.-q., A.M.H. and A.D.; Validation, M.A.E.; Visualization, A.M.H. and A.D.; Writing—original draft, M.A.A.A.-q., A.M.H., A.D. and M.A.E.; Writing—review and editing, M.A.E. All authors contributed equally to this paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 62150410434), by LIESMARS Special Research Funding, and in part by the Scientific Research Center at Buraydah Private Colleges under the research project # BPC-SRC/2022-010.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are publicly available as described in the main text.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hasegawa, T. Smartphone sensor-based human activity recognition robust to different sampling rates. IEEE Sens. J. 2020, 21, 6930–6941. [Google Scholar] [CrossRef]
  2. Beddiar, D.R.; Nini, B.; Sabokrou, M.; Hadid, A. Vision-based human activity recognition: A survey. Multimed. Tools Appl. 2020, 79, 30509–30555. [Google Scholar] [CrossRef]
  3. Gu, F.; Chung, M.H.; Chignell, M.; Valaee, S.; Zhou, B.; Liu, X. A survey on deep learning for human activity recognition. ACM Comput. Surv. (CSUR) 2021, 54, 1–34. [Google Scholar] [CrossRef]
  4. Bouchabou, D.; Nguyen, S.M.; Lohr, C.; LeDuc, B.; Kanellos, I. A survey of human activity recognition in smart homes based on IoT sensors algorithms: Taxonomies, challenges, and opportunities with deep learning. Sensors 2021, 21, 6037. [Google Scholar] [CrossRef]
  5. Li, X.; He, Y.; Jing, X. A survey of deep learning-based human activity recognition in radar. Remote Sens. 2019, 11, 1068. [Google Scholar] [CrossRef] [Green Version]
  6. Al-Qaness, M.A.; Abd Elaziz, M.; Kim, S.; Ewees, A.A.; Abbasi, A.A.; Alhaj, Y.A.; Hawbani, A. Channel state information from pure communication to sense and track human motion: A survey. Sensors 2019, 19, 3329. [Google Scholar] [CrossRef] [Green Version]
  7. Fatani, A.; Dahou, A.; Al-Qaness, M.A.; Lu, S.; Elaziz, M.A. Advanced feature extraction and selection approach using deep learning and Aquila optimizer for IoT intrusion detection system. Sensors 2021, 22, 140. [Google Scholar] [CrossRef]
  8. Dahou, A.; Abd Elaziz, M.; Chelloug, S.A.; Awadallah, M.A.; Al-Betar, M.A.; Al-qaness, M.A.; Forestiero, A. Intrusion Detection System for IoT Based on Deep Learning and Modified Reptile Search Algorithm. Comput. Intell. Neurosci. 2022, 2022, 1–15. [Google Scholar] [CrossRef]
  9. Li, L.; Pan, J.S.; Zhuang, Z.; Chu, S.C. A Novel Feature Selection Algorithm Based on Aquila Optimizer for COVID-19 Classification. In Proceedings of the International Conference on Intelligent Information Processing, Bucharest, Romania, 29–30 September 2022; pp. 30–41. [Google Scholar]
  10. Bansal, P.; Gehlot, K.; Singhal, A.; Gupta, A. Automatic detection of osteosarcoma based on integrated features and feature selection using binary arithmetic optimization algorithm. Multimed. Tools Appl. 2022, 81, 8807–8834. [Google Scholar] [CrossRef]
  11. Dahou, A.; Al-qaness, M.A.; Abd Elaziz, M.; Helmi, A. Human activity recognition in IoHT applications using arithmetic optimization algorithm and deep learning. Measurement 2022, 199, 111445. [Google Scholar] [CrossRef]
  12. Sahlol, A.T.; Yousri, D.; Ewees, A.A.; Al-Qaness, M.A.; Damasevicius, R.; Elaziz, M.A. COVID-19 image classification using deep features and fractional-order marine predators algorithm. Sci. Rep. 2020, 10, 15364. [Google Scholar] [CrossRef] [PubMed]
  13. Rajinikanth, V.; Kadry, S.; Taniar, D.; Damaševičius, R.; Rauf, H.T. Breast-cancer detection using thermal images with marine-predators-algorithm selected features. In Proceedings of the 2021 Seventh International Conference on Bio Signals, Images, and Instrumentation (ICBSII), Chennai, India, 25–27 March 2021; pp. 1–6. [Google Scholar]
  14. Al-qaness, M.A.; Ewees, A.A.; Fan, H.; Abualigah, L.; Abd Elaziz, M. Boosted ANFIS model using augmented marine predator algorithm with mutation operators for wind power forecasting. Appl. Energy 2022, 314, 118851. [Google Scholar] [CrossRef]
  15. Wazery, Y.M.; Saber, E.; Houssein, E.H.; Ali, A.A.; Amer, E. An efficient slime mould algorithm combined with k-nearest neighbor for medical classification tasks. IEEE Access 2021, 9, 113666–113682. [Google Scholar] [CrossRef]
  16. Liu, Y.; Heidari, A.A.; Ye, X.; Liang, G.; Chen, H.; He, C. Boosting slime mould algorithm for parameter identification of photovoltaic models. Energy 2021, 234, 121164. [Google Scholar] [CrossRef]
  17. Maleki, N.; Zeinali, Y.; Niaki, S.T.A. A k-NN method for lung cancer prognosis with the use of a genetic algorithm for feature selection. Expert Syst. Appl. 2021, 164, 113981. [Google Scholar] [CrossRef]
  18. Abualigah, L.; Dulaimi, A.J. A novel feature selection method for data mining tasks using hybrid sine cosine algorithm and genetic algorithm. Clust. Comput. 2021, 24, 2161–2176. [Google Scholar] [CrossRef]
  19. Lappas, P.Z.; Yannacopoulos, A.N. A machine learning approach combining expert knowledge with genetic algorithms in feature selection for credit risk assessment. Appl. Soft Comput. 2021, 107, 107391. [Google Scholar] [CrossRef]
  20. Sathiyabhama, B.; Kumar, S.U.; Jayanthi, J.; Sathiya, T.; Ilavarasi, A.; Yuvarajan, V.; Gopikrishna, K. A novel feature selection framework based on grey wolf optimizer for mammogram image analysis. Neural Comput. Appl. 2021, 33, 14583–14602. [Google Scholar] [CrossRef]
  21. Rajammal, R.R.; Mirjalili, S.; Ekambaram, G.; Palanisamy, N. Binary Grey Wolf Optimizer with Mutation and Adaptive K-nearest Neighbour for Feature Selection in Parkinson’s Disease Diagnosis. Knowl. Based Syst. 2022, 246, 108701. [Google Scholar] [CrossRef]
  22. Preeti, P.; Deep, K. A Random Walk Grey Wolf Optimizer based on dispersion factor for feature selection on Chronic Disease Prediction. Expert Syst. Appl. 2022, 206, 117864. [Google Scholar] [CrossRef]
  23. Mohammadzadeh, H.; Gharehchopogh, F.S. A novel hybrid whale optimization algorithm with flower pollination algorithm for feature selection: Case study Email spam detection. Comput. Intell. 2021, 37, 176–209. [Google Scholar] [CrossRef]
  24. Hassouneh, Y.; Turabieh, H.; Thaher, T.; Tumar, I.; Chantar, H.; Too, J. Boosted whale optimization algorithm with natural selection operators for software fault prediction. IEEE Access 2021, 9, 14239–14258. [Google Scholar] [CrossRef]
  25. Moorthy, U.; Gandhi, U.D. A novel optimal feature selection technique for medical data classification using ANOVA based whale optimization. J. Ambient Intell. Humaniz. Comput. 2021, 12, 3527–3538. [Google Scholar] [CrossRef]
  26. Shafi, A.; Molla, M.; Jui, J.J.; Rahman, M.M. Detection of colon cancer based on microarray dataset using machine learning as a feature selection and classification techniques. SN Appl. Sci. 2020, 2, 1243. [Google Scholar] [CrossRef]
  27. Agrawal, V.; Chandra, S. Feature selection using Artificial Bee Colony algorithm for medical image classification. In Proceedings of the 2015 Eighth International Conference on Contemporary Computing (IC3), Noida, India, 20–22 August 2015; pp. 171–176. [Google Scholar]
  28. Rani, M.; Gagandeep. Employing Artificial Bee Colony Algorithm for Feature Selection in Intrusion Detection System. In Proceedings of the 2021 8th International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 17–19 March 2021; pp. 496–500.
  29. Sharkawy, R.; Ibrahim, K.; Salama, M.; Bartnikas, R. Particle swarm optimization feature selection for the classification of conducting particles in transformer oil. IEEE Trans. Dielectr. Electr. Insul. 2011, 18, 1897–1907. [Google Scholar] [CrossRef]
  30. Sakri, S.B.; Rashid, N.B.A.; Zain, Z.M. Particle swarm optimization feature selection for breast cancer recurrence prediction. IEEE Access 2018, 6, 29637–29647. [Google Scholar] [CrossRef]
  31. Kunhare, N.; Tiwari, R.; Dhar, J. Particle swarm optimization and feature selection for intrusion detection system. Sādhanā 2020, 45, 109. [Google Scholar] [CrossRef]
  32. Joshi, G.; Walambe, R.; Kotecha, K. A review on explainability in multimodal deep neural nets. IEEE Access 2021, 9, 59800–59821. [Google Scholar] [CrossRef]
  33. Deotale, D.; Verma, M.; Suresh, P.; Kotecha, K. Optimized hybrid RNN model for human activity recognition in untrimmed video. J. Electron. Imaging 2022, 31, 051409. [Google Scholar] [CrossRef]
  34. Wang, K.; He, J.; Zhang, L. Attention-based convolutional neural network for weakly labeled human activities’ recognition with wearable sensors. IEEE Sens. J. 2019, 19, 7598–7604. [Google Scholar] [CrossRef]
  35. Xia, K.; Huang, J.; Wang, H. LSTM-CNN architecture for human activity recognition. IEEE Access 2020, 8, 56855–56866. [Google Scholar] [CrossRef]
  36. Sikder, N.; Ahad, M.A.R.; Nahid, A.A. Human Action Recognition Based on a Sequential Deep Learning Model. In Proceedings of the 2021 Joint 10th International Conference on Informatics, Electronics & Vision (ICIEV) and 2021 5th International Conference on Imaging, Vision & Pattern Recognition (icIVPR), Krakow, Poland, 16–18 June 2021; pp. 1–7. [Google Scholar]
  37. Kumar, P.; Suresh, S. DeepTransHHAR: Inter-subjects Heterogeneous Activity Recognition Approach in the Non-identical Environment Using Wearable Sensors. Natl. Acad. Sci. Lett. 2022, 45, 317–323. [Google Scholar] [CrossRef]
  38. Dua, N.; Singh, S.N.; Semwal, V.B. Multi-input CNN-GRU based human activity recognition using wearable sensors. Computing 2021, 103, 1461–1478. [Google Scholar] [CrossRef]
  39. Khatun, M.A.; Yousuf, M.A.; Ahmed, S.; Uddin, M.Z.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Khan, A.; Azad, A.; Moni, M.A. Deep CNN-LSTM with Self-Attention Model for Human Activity Recognition using Wearable Sensor. IEEE J. Transl. Eng. Health Med. 2022, 10, 2700316. [Google Scholar] [CrossRef] [PubMed]
  40. Ghate, V.; Hemalatha, S. Hybrid deep learning approaches for smartphone sensor-based human activity recognition. Multimed. Tools Appl. 2021, 80, 35585–35604. [Google Scholar] [CrossRef]
  41. Ronald, M.; Poulose, A.; Han, D.S. iSPLInception: An inception-ResNet deep learning architecture for human activity recognition. IEEE Access 2021, 9, 68985–69001. [Google Scholar] [CrossRef]
  42. Tufek, N.; Yalcin, M.; Altintas, M.; Kalaoglu, F.; Li, Y.; Bahadir, S.K. Human action recognition using deep learning methods on limited sensory data. IEEE Sens. J. 2019, 20, 3101–3112. [Google Scholar] [CrossRef] [Green Version]
  43. Gao, W.; Zhang, L.; Huang, W.; Min, F.; He, J.; Song, A. Deep neural networks for sensor-based human activity recognition using selective kernel convolution. IEEE Trans. Instrum. Meas. 2021, 70, 2512313. [Google Scholar] [CrossRef]
  44. Huang, W.; Zhang, L.; Teng, Q.; Song, C.; He, J. The convolutional neural networks training with channel-selectivity for human activity recognition based on sensors. IEEE J. Biomed. Health Inform. 2021, 25, 3834–3843. [Google Scholar] [CrossRef] [PubMed]
  45. Gao, W.; Zhang, L.; Teng, Q.; He, J.; Wu, H. DanHAR: Dual attention network for multimodal human activity recognition using wearable sensors. Appl. Soft Comput. 2021, 111, 107728. [Google Scholar] [CrossRef]
  46. Tang, Y.; Zhang, L.; Min, F.; He, J. Multi-scale deep feature learning for human activity recognition using wearable sensors. IEEE Trans. Ind. Electron. 2022. [Google Scholar] [CrossRef]
  47. Berlin, S.J.; John, M. Particle swarm optimization with deep learning for human action recognition. Multimed. Tools Appl. 2020, 79, 17349–17371. [Google Scholar] [CrossRef]
  48. Zhang, R. Sports action recognition based on particle swarm optimization neural networks. Wirel. Commun. Mob. Comput. 2022, 2022, 6912315. [Google Scholar] [CrossRef]
  49. Guha, R.; Khan, A.H.; Singh, P.K.; Sarkar, R.; Bhattacharjee, D. CGA: A new feature selection model for visual human action recognition. Neural Comput. Appl. 2021, 33, 5267–5286. [Google Scholar] [CrossRef]
  50. Helmi, A.M.; Al-Qaness, M.A.; Dahou, A.; Damaševičius, R.; Krilavičius, T.; Elaziz, M.A. A novel hybrid gradient-based optimizer and grey wolf optimizer feature selection method for human activity recognition using smartphone sensors. Entropy 2021, 23, 1065. [Google Scholar] [CrossRef]
  51. Sikder, N.; Nahid, A.A. KU-HAR: An open dataset for heterogeneous human activity recognition. Pattern Recognit. Lett. 2021, 146, 46–54. [Google Scholar] [CrossRef]
  52. Chavarriaga, R.; Sagha, H.; Calatroni, A.; Digumarti, S.T.; Tröster, G.; Millán, J.D.R.; Roggen, D. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition. Pattern Recognit. Lett. 2013, 34, 2033–2042. [Google Scholar] [CrossRef] [Green Version]
  53. Reiss, A.; Stricker, D. Introducing a new benchmarked dataset for activity monitoring. In Proceedings of the 2012 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012; pp. 108–109. [Google Scholar]
  54. Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall: A fall and movement dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef] [Green Version]
  55. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A public domain dataset for human activity recognition using smartphones. In Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
  56. Micucci, D.; Mobilio, M.; Napoletano, P. Unimib shar: A dataset for human activity recognition using acceleration data from smartphones. Appl. Sci. 2017, 7, 1101. [Google Scholar] [CrossRef] [Green Version]
  57. Weiss, G.M.; Lockhart, J. The impact of personalization on smartphone-based activity recognition. In Proceedings of the Workshops at the Twenty-Sixth AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012. [Google Scholar]
  58. Abualigah, L.; Yousri, D.; Abd Elaziz, M.; Ewees, A.A.; Al-Qaness, M.A.; Gandomi, A.H. Aquila optimizer: A novel meta-heuristic optimization algorithm. Comput. Ind. Eng. 2021, 157, 107250. [Google Scholar] [CrossRef]
  59. Abualigah, L.; Diabat, A.; Mirjalili, S.; Abd Elaziz, M.; Gandomi, A.H. The arithmetic optimization algorithm. Comput. Methods Appl. Mech. Eng. 2021, 376, 113609. [Google Scholar] [CrossRef]
  60. Faramarzi, A.; Heidarinejad, M.; Mirjalili, S.; Gandomi, A.H. Marine Predators Algorithm: A nature-inspired metaheuristic. Expert Syst. Appl. 2020, 152, 113377. [Google Scholar] [CrossRef]
  61. Li, S.; Chen, H.; Wang, M.; Heidari, A.A.; Mirjalili, S. Slime mould algorithm: A new method for stochastic optimization. Future Gener. Comput. Syst. 2020, 111, 300–323. [Google Scholar] [CrossRef]
  62. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  63. Karaboga, D. An Idea Based on Honey Bee Swarm for Numerical Optimization; Technical Report-tr06; Erciyes University: Kayseri, Turkey, 2005. [Google Scholar]
  64. Tereshko, V.; Loengarov, A. Collective decision making in honey-bee foraging dynamics. Comput. Inf. Syst. 2005, 9, 1. [Google Scholar]
  65. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  66. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef] [Green Version]
  67. Goldberg, D.E.; Richardson, J. Genetic algorithms with sharing for multimodal function optimization. In Proceedings of the Second International Conference on Genetic Algorithms, Cambridge, MA, USA, 28–31 July 1987; Volume 4149. [Google Scholar]
  68. Goldberg, D.E. Genetic Algorithms; Pearson Education India: Bengaluru, India, 2013. [Google Scholar]
  69. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  70. Preece, S.J.; Goulermas, J.Y.; Kenney, L.P.; Howard, D. A comparison of feature extraction methods for the classification of dynamic activities from accelerometer data. IEEE Trans. Biomed. Eng. 2008, 56, 871–879. [Google Scholar] [CrossRef] [PubMed]
  71. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J. Complex human activity recognition using smartphone and wrist-worn motion sensors. Sensors 2016, 16, 426. [Google Scholar] [CrossRef]
  72. Voicu, R.A.; Dobre, C.; Bajenaru, L.; Ciobanu, R.I. Human physical activity recognition using smartphone sensors. Sensors 2019, 19, 458. [Google Scholar] [CrossRef] [Green Version]
  73. Teng, Q.; Wang, K.; Zhang, L.; He, J. The layer-wise training convolutional neural networks using local loss for sensor-based human activity recognition. IEEE Sens. J. 2020, 20, 7265–7274. [Google Scholar] [CrossRef]
  74. Ma, H.; Li, W.; Zhang, X.; Gao, S.; Lu, S. AttnSense: Multi-level Attention Mechanism For Multimodal Human Activity Recognition. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3109–3115. [Google Scholar]
  75. Sousa Lima, W.; Souto, E.; El-Khatib, K.; Jalali, R.; Gama, J. Human activity recognition using inertial sensors in a smartphone: An overview. Sensors 2019, 19, 3213. [Google Scholar] [CrossRef] [Green Version]
  76. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. Human activity recognition on smartphones using a multiclass hardware-friendly support vector machine. In Proceedings of the International Workshop on Ambient Assisted Living, Vitoria-Gasteiz, Spain, 3–5 December 2012; pp. 216–223. [Google Scholar]
  77. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  78. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  79. Mrozek, D.; Koczur, A.; Małysiak-Mrozek, B. Fall detection in older adults with mobile IoT devices and machine learning in the cloud and on the edge. Inf. Sci. 2020, 537, 132–147. [Google Scholar] [CrossRef]
  80. Al-qaness, M.A.; Dahou, A.; Abd Elaziz, M.; Helmi, A. Multi-ResAtt: Multilevel Residual Network with Attention for Human Activity Recognition Using Wearable Sensors. IEEE Trans. Ind. Inform. 2022. [Google Scholar] [CrossRef]
  81. Zhang, Y.; Zhang, Z.; Zhang, Y.; Bao, J.; Zhang, Y.; Deng, H. Human activity recognition based on motion sensor using u-net. IEEE Access 2019, 7, 75213–75226. [Google Scholar] [CrossRef]
  82. Li, X.; Dezert, J.; Khyam, M.O.; Noor-A-Rahim, M.; Ge, S.S.; Dezert, J.; Dong, Y. DSmT-Based Fusion Strategy for Human Activity Recognition in Body Sensor Networks. IEEE Trans. Ind. Inform. 2020, 16, 7138–7149. [Google Scholar]
  83. Zhang, H.; Xiao, Z.; Wang, J.; Li, F.; Szczerbicki, E. A novel IoT-perceptive human activity recognition (HAR) approach using multihead convolutional attention. IEEE Internet Things J. 2019, 7, 1072–1080. [Google Scholar] [CrossRef]
  84. Teng, Q.; Zhang, L.; Tang, Y.; Song, S.; Wang, X.; He, J. Block-wise training residual networks on multi-channel time series for human activity recognition. IEEE Sens. J. 2021, 21, 18063–18074. [Google Scholar] [CrossRef]
Figure 1. The main workflow of HAR application using the integration of deep learning and MH optimization algorithms.
Figure 1. The main workflow of HAR application using the integration of deep learning and MH optimization algorithms.
Biosensors 12 00821 g001
Figure 2. Sensor placement on the subject’s body. A waist-mounted smartphone was used for KUHAR, UCI-HAR, and WISDOM. WISDOM was collected using additional smartwatches. IMU units were used to collect the signals for OPPO (positions: back, right/left upper/lateral parts of the arm and right/left shoe ), PAMAP2 (positions: chest, wrist, and ankle), and Sis-Fall (position: waist).
Figure 2. Sensor placement on the subject’s body. A waist-mounted smartphone was used for KUHAR, UCI-HAR, and WISDOM. WISDOM was collected using additional smartwatches. IMU units were used to collect the signals for OPPO (positions: back, right/left upper/lateral parts of the arm and right/left shoe ), PAMAP2 (positions: chest, wrist, and ankle), and Sis-Fall (position: waist).
Biosensors 12 00821 g002
Figure 3. The proposed ResRNN model for feature extraction.
Figure 3. The proposed ResRNN model for feature extraction.
Biosensors 12 00821 g003
Figure 4. Convergence curve for applied MH optimizers for each studied dataset using RF and SVM classifiers.
Figure 4. Convergence curve for applied MH optimizers for each studied dataset using RF and SVM classifiers.
Biosensors 12 00821 g004aBiosensors 12 00821 g004b
Figure 5. Comparison of the average accuracy of the applied DL model, DL model using RF and SVM (best model (Init.)), and the optimized DL model (best model (Opt.).
Figure 5. Comparison of the average accuracy of the applied DL model, DL model using RF and SVM (best model (Init.)), and the optimized DL model (best model (Opt.).
Biosensors 12 00821 g005
Table 1. Parameter settings of used MH algorithms.
Table 1. Parameter settings of used MH algorithms.
AlgorithmParameters
ABC a = 1
AO α = 0.1 , δ = 0.1 , u = 0.0265 , r 0 = 10 , ω = 0.005 , ϕ 0 = 3 π / 2 ,
AOA M O P M a x = 1 , M O P M i n = 0.2 , α = 2.5 , μ = 0.1
GACrossover = 40 % , Mutation = 15 %
GWO a : 2 0
MPA F A D s = 0.2 , P = 0.5
PSOInertia weight (w) = 1, w d a m p = 0.99 , c 1 = 2 , c 2 = 2
SMA z = 0.03
WOA a : 2 0 , a2: 1 2 , b = 1
Table 2. Average classification accuracy (%) of applied MH optimizers
Table 2. Average classification accuracy (%) of applied MH optimizers
KUHAROPPOPAMAP2SisFallBSisFallMUCI-HARUniMiBWISDM
RFSVMRFSVMRFSVMRFSVMRFSVMRFSVMRFSVMRFSVM
ABC88.0386.6193.7393.8492.0692.1399.9799.9789.5989.5094.8594.7685.7486.0298.8398.83
AO88.386.8393.8193.7692.4492.299.9899.9789.9289.3395.3795.0786.1785.7598.9598.87
AOA87.8986.6693.6493.6192.2391.4499.9799.9789.8289.2695.1294.98685.6998.7998.85
GA87.9586.793.7493.8991.6792.2599.9799.9789.4789.6295.1594.7585.8786.198.8598.82
GWO88.486.6993.9093.992.592.3199.9799.9789.9389.6595.1394.7786.2586.198.9598.83
MPA88.4186.693.8893.8692.5892.0299.9799.9789.9689.5095.2794.7586.2286.0298.9498.82
PSO88.3286.7193.8793.9192.5892.199.9799.9790.0089.5495.2194.7886.2186.0598.9598.82
SMA87.7786.3893.7293.7691.6391.3499.9799.9789.5289.3794.6294.8285.7185.8298.8598.8
WOA88.3186.793.8193.7592.5792.6399.9899.9789.9189.3395.5195.2886.1485.8598.9598.87
Table 3. Classification accuracy (%) of the best model for each dataset
Table 3. Classification accuracy (%) of the best model for each dataset
KUHAROPPOPAMAP2SisFallBSisFallMUCI-HARUniMiBWISDM
RFRF/SVMSVMRFRFRFRFRF
ABC88.3793.7992.3699.9789.9295.1185.9698.85
AO88.5393.9292.4810090.1395.4986.4498.99
AOA88.0993.7391.5799.9789.8695.2286.1398.81
GA88.0793.7892.3899.9789.5995.3986.398.91
GWO88.4993.9592.4299.9789.9895.4286.3398.97
MPA88.5293.9192.399.9790.0495.3986.3598.97
PSO88.4493.9192.3299.9790.195.2586.3398.97
SMA88.0693.8691.7199.9789.6894.7486.0498.87
WOA88.4293.8392.9510089.9895.5986.1898.99
Table 4. Average percentage of feature reduction (%) of applied MH optimizers
Table 4. Average percentage of feature reduction (%) of applied MH optimizers
KUHAROPPOPAMAP2SisFallBSisFallMUCI-HARUniMiBWISDM
RFSVMRFSVMRFSVMRFSVMRFSVMRFSVMRFSVMRFSVM
ABC51.41 ± 4.650 ± 6.0460.31 ± 5.0760.94 ± 3.8757.66 ± 4.4461.72 ± 3.3973.13 ± 2.5173.75 ± 2.9751.56 ± 3.6753.13 ± 8.1556.72 ± 3.5866.09 ± 4.0454.22 ± 7.555.78 ± 7.8363.91 ± 5.6369.22 ± 2.07
AO39.69 ± 14.717.97 ± 14.5385.47 ± 8.9687.19 ± 4.9363.13 ± 13.3382.97 ± 8.4498.44 ± 1.2298.75 ± 0.5551.09 ± 3.9153.91 ± 12.5990.78 ± 5.0293.28 ± 2.0766.41 ± 16.469.84 ± 13.8592.81 ± 2.9592.5 ± 3.44
AOA51.88 ± 8.448.59 ± 24.685.31 ± 10.7190.47 ± 2.9555.31 ± 2.5962.34 ± 24.1698.91 ± 0.5598.44 ± 1.4152.66 ± 3.2151.88 ± 5.4176.25 ± 27.5278.13 ± 28.5264.69 ± 13.3363.44 ± 12.9593.44 ± 4.195.31 ± 1.87
GA48.75 ± 5.1346.56 ± 4.4559.53 ± 2.3962.66 ± 3.8354.06 ± 4.9758.59 ± 4.4774.06 ± 2.0573.75 ± 4.2856.25 ± 3.6151.41 ± 4.3255.63 ± 8.762.66 ± 5.5453.44 ± 4.7256.72 ± 4.5165.63 ± 3.9467.34 ± 2.39
GWO50.47 ± 3.7846.25 ± 2.5964.53 ± 4.7762.34 ± 3.4251.88 ± 6.9159.69 ± 2.371.88 ± 3.4673.59 ± 4.2755.47 ± 4.951.72 ± 4.7659.84 ± 7.1361.72 ± 4.0652.5 ± 2.1753.13 ± 4.4767.5 ± 2.1969.06 ± 6.35
MPA47.81 ± 1.345.31 ± 4.359.38 ± 4.6457.5 ± 5.6453.28 ± 11.2356.56 ± 2.4170 ± 5.3769.38 ± 2.1755.16 ± 2.7951.72 ± 5.7657.81 ± 3.2462.97 ± 3.2153.13 ± 1.8754.69 ± 4.5360.31 ± 3.765.47 ± 1.64
PSO46.88 ± 3.6744.69 ± 6.6157.5 ± 8.4461.09 ± 5.6351.88 ± 2.1959.84 ± 4.8373.13 ± 4.2872.81 ± 1.6450.78 ± 11.1151.72 ± 5.9358.28 ± 6.7364.22 ± 4.1550 ± 5.9656.88 ± 3.1161.72 ± 2.7470.16 ± 5.59
SMA46.88 ± 2.7446.09 ± 5.5258.59 ± 4.359.69 ± 3.4455.31 ± 5.557.97 ± 6.363.75 ± 2.5162.19 ± 1.5253.75 ± 3.2750.47 ± 5.5953.59 ± 4.3456.09 ± 10.4755.78 ± 5.3253.13 ± 5.7960.62 ± 2.762.5 ± 3
WOA47.81 ± 1.345.31 ± 4.359.38 ± 4.6457.5 ± 5.6453.28 ± 11.2356.56 ± 2.4170 ± 5.3769.38 ± 2.1755.16 ± 2.7961.09 ± 11.4357.81 ± 3.2462.97 ± 3.2153.13 ± 1.8754.69 ± 4.5360.31 ± 3.765.47 ± 1.64
Table 5. Classification rates (%) of the best model of the KUHAR dataset using RF.
Table 5. Classification rates (%) of the best model of the KUHAR dataset using RF.
ActivityPreRecF1
Standing76.2973.3274.77
Sitting98.395.796.98
Talking-Sit94.896.2395.51
Talking-Stand92.8695.7994.3
Standing-Sit9692.3194.12
Laying93.8594.3894.12
Laying-Stand95.291.2193.16
Picking95.7195.395.5
Jumping53.3691.8167.5
Pushing-up89.8281.8285.63
Sitting-up97.1195.8996.5
Walking98.4899.2398.86
Walking-backward80.6233.6447.47
Walking-circle97.1998.1197.65
Running96.529796.76
Stairs up96.949595.96
Stairs down97.8494.4496.11
Table 6. Classification rates (%) of the best model of the OPPO dataset using either RF or SVM.
Table 6. Classification rates (%) of the best model of the OPPO dataset using either RF or SVM.
ActivityPreRecF1
NULL95.3798.4996.9
Open Door 194.0380.7786.9
Open Door 297.590.793.98
Close Door 191.2381.2585.95
Close Door 296.6793.5595.08
Open Fridge89.0483.8786.38
Close Fridge88.9971.8579.51
Open Dishwasher79.6369.3574.14
Close Dishwasher68.5266.0767.27
Open Drawer 146.8862.553.57
Close Drawer 146.6725.9333.33
Open Drawer 257.146460.38
Close Drawer 227.274533.96
Open Drawer 368.8579.2573.68
Close Drawer 397.1470.8381.93
Clean Table81.457.3867.31
Drink from Cup91.9458.7671.7
Toggle Switch87.560.8771.79
Table 7. Classification rates (%) of the best model of the PAMAP2 dataset using SVM.
Table 7. Classification rates (%) of the best model of the PAMAP2 dataset using SVM.
ActivityPreRecF1
Lying95.7793.6194.68
Sitting9795.6796.33
Standing84.0691.3487.55
Normal Walking86.1996.9891.27
Running99.7898.799.24
Cycling96.4399.6197.99
Nordic Walking95.7195.7195.71
Ascending Stairs91.2195.8293.46
Descending Stairs95.293.8394.51
Vacuuming98.5861.575.75
Ironing83.9990.7987.26
Rope Jumping97.1599.0998.11
Table 8. Classification rates (%) of the best model of the SisFallB dataset using RF.
Table 8. Classification rates (%) of the best model of the SisFallB dataset using RF.
ActivityPreRecF1
Fall100100100
No Fall100100100
Table 9. Classification rates (%) of the best model of the SisFallM dataset using RF.
Table 9. Classification rates (%) of the best model of the SisFallM dataset using RF.
ActivityPreRecF1
Walking Slowly98.7198.7198.71
Walking Quickly100100100
Jogging Slowly9898.3998.2
Jogging Quickly98.6497.7598.19
Walking upstairs and downstairs slowly92.2592.5892.42
Walking upstairs and downstairs quickly77.7881.2979.5
Sitting in a half-height chair, waiting a moment, and getting up slowly79.8677.0878.45
Sitting in a half-height chair, waiting a moment, and getting up quickly76.2881.5178.81
Sitting in a low-height chair, waiting a moment, and getting up slowly7074.3472.1
Sitting in a low-height chair, waiting a moment, and getting up quickly8067.873.39
Sitting a moment, trying to get up, and collapsing into a chair86.5580.4783.4
Sitting a moment, lying slowly, waiting a moment, and sitting again93.9193.9193.91
Sitting a moment, lying quickly, waiting a moment, and sitting again88.4684.1586.25
Being on one’s back, changing to lateral position, waiting a moment, and changing to one’s back95.9795.9795.97
Standing, slowly bending at knees, and getting up86.1578.8782.35
Standing, slowly bending without bending knees, and getting up87.0788.2887.67
Standing, getting into a car, remaining seated, and getting out of the car82.4991.7686.88
Stumbling while walking97.6295.3596.47
Gently jumping without falling (trying to reach a high object)88.5182.885.56
Falling99.4499.4499.44
Table 10. Classification rates (%) of the best model of the UCI-HAR dataset using RF.
Table 10. Classification rates (%) of the best model of the UCI-HAR dataset using RF.
ActivityPreRecF1
Walking99.5997.9898.78
Walking Upstairs98.3198.5198.41
Walking Downstairs96.7499.0597.88
Sitting92.0584.9388.35
Standing88.4393.4290.86
Laying Down99.0810099.54
Table 11. Classification rates (%) of the best model of the UniMiB-SHAR dataset using RF.
Table 11. Classification rates (%) of the best model of the UniMiB-SHAR dataset using RF.
ActivityPreRecF1
Standing Up From Sitting64.3368.2466.23
Standing Up From Lying87.577.2182.03
Walking76.0775.6175.84
Running64.3365.5864.95
Going UpS77.770.5973.97
Jumping70.0672.3771.2
Going DownS94.4196.4395.41
Lying DownFS92.0688.5590.27
Sitting Down92.7588.6190.63
Falling Forward98.1498.698.37
Falling Rightward69.7973.6371.66
Falling Backward99.1499.3199.22
Hitting Obstacle73.6873.6873.68
Falling With Protection72.3777.4674.83
Falling Backward-Sitting Chair81.826069.23
Syncope Fall52.1761.1556.3
Falling Leftward98.2897.7298
Table 12. Classification rates (%) of the best model of the WISDM dataset using RF.
Table 12. Classification rates (%) of the best model of the WISDM dataset using RF.
ActivityPreRecF1
Walking97.0996.0196.54
Walking Upstairs10098.8199.4
Walking Downstairs98.4610099.22
Sitting95.8395.6795.75
Standing10099.9599.97
Jogging99.3899.7599.57
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Al-qaness, M.A.A.; Helmi, A.M.; Dahou, A.; Elaziz, M.A. The Applications of Metaheuristics for Human Activity Recognition and Fall Detection Using Wearable Sensors: A Comprehensive Analysis. Biosensors 2022, 12, 821. https://doi.org/10.3390/bios12100821

AMA Style

Al-qaness MAA, Helmi AM, Dahou A, Elaziz MA. The Applications of Metaheuristics for Human Activity Recognition and Fall Detection Using Wearable Sensors: A Comprehensive Analysis. Biosensors. 2022; 12(10):821. https://doi.org/10.3390/bios12100821

Chicago/Turabian Style

Al-qaness, Mohammed A. A., Ahmed M. Helmi, Abdelghani Dahou, and Mohamed Abd Elaziz. 2022. "The Applications of Metaheuristics for Human Activity Recognition and Fall Detection Using Wearable Sensors: A Comprehensive Analysis" Biosensors 12, no. 10: 821. https://doi.org/10.3390/bios12100821

APA Style

Al-qaness, M. A. A., Helmi, A. M., Dahou, A., & Elaziz, M. A. (2022). The Applications of Metaheuristics for Human Activity Recognition and Fall Detection Using Wearable Sensors: A Comprehensive Analysis. Biosensors, 12(10), 821. https://doi.org/10.3390/bios12100821

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop