Next Article in Journal
Design of an Edge-Detection CMOS Image Sensor with Built-in Mask Circuits
Next Article in Special Issue
LARa: Creating a Dataset for Human Activity Recognition in Logistics Using Semantic Attributes
Previous Article in Journal
UAV Positioning Mechanisms in Landing Stations: Classification and Engineering Design Review
Previous Article in Special Issue
Sensory Interactive Table (SIT)—Development of a Measurement Instrument to Support Healthy Eating in a Social Dining Setting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance †

by
Sebastian Scheurer
1,*,
Salvatore Tedesco
2,
Brendan O’Flynn
1,2,3 and
Kenneth N. Brown
1,*
1
Insight Centre for Data Analytics, School of Computer Science and Information Technology, University College Cork, T12 XF62 Cork, Ireland
2
Tyndall National Institute, University College Cork, T12 R5CP Cork, Ireland
3
CONNECT Centre for Future Networks and Communications, Tyndall National Institute, University College Cork, T12 R5CP Cork, Ireland
*
Authors to whom correspondence should be addressed.
This paper is an extended version of our paper published in Scheurer, S.; Tedesco, S.; Brown, K.N.; O’Flynn, B. Subject-dependent and-independent human activity recognition with person-specific and-independent models. In Proceedings of the 6th international Workshop on Sensor-based Activity Recognition and Interaction, Rostock, Germany, 16–17 September 2019; pp. 1–7
Sensors 2020, 20(13), 3647; https://doi.org/10.3390/s20133647
Submission received: 31 May 2020 / Revised: 23 June 2020 / Accepted: 24 June 2020 / Published: 29 June 2020
(This article belongs to the Special Issue Sensor-Based Activity Recognition and Interaction)

Abstract

:
The distinction between subject-dependent and subject-independent performance is ubiquitous in the human activity recognition (HAR) literature. We assess whether HAR models really do achieve better subject-dependent performance than subject-independent performance, whether a model trained with data from many users achieves better subject-independent performance than one trained with data from a single person, and whether one trained with data from a single specific target user performs better for that user than one trained with data from many. To those ends, we compare four popular machine learning algorithms’ subject-dependent and subject-independent performances across eight datasets using three different personalisation–generalisation approaches, which we term person-independent models (PIMs), person-specific models (PSMs), and ensembles of PSMs (EPSMs). We further consider three different ways to construct such an ensemble: unweighted, κ -weighted, and baseline-feature-weighted. Our analysis shows that PSMs outperform PIMs by 43.5% in terms of their subject-dependent performances, whereas PIMs outperform PSMs by 55.9% and κ -weighted EPSMs—the best-performing EPSM type—by 16.4% in terms of the subject-independent performance.

1. Introduction

This is an extended version of a paper we presented at the 6th international Workshop on Sensor-based Activity Recognition and Interaction (iWOAR’19) [1]. In addition to the experiments, results, and analyses presented in that paper, this paper covers more of the literature in more depth, considers additional ways to combine person-specific models into ensembles, sharpens and deepens the statistical analysis, and expands the discussion by relating our findings to the pertinent literature.
Human activity recognition (HAR) systems are typically evaluated for their ability to generalise to either unknown users (people not represented in the HAR algorithm’s training data) or to known users (people represented in the training data); the former is known as subject-independent and the latter as subject-dependent performance. The subject-independent performance can be estimated by performing a leave-one-subject-out cross-validation across all users in the dataset, and the subject-dependent performance by performing a separate k-fold cross-validation for each user. Which performance should be optimised when developing a HAR system depends on how it is going to be commissioned and deployed. If commissioning a HAR system entails obtaining examples of the activities of interest from its end users—the people whose activities are to be recognised by the deployed system—then we should optimise the subject-dependent performance. Intuitively, this suggests that we train a personalised HAR inference model for each user. We refer to models obtained in this manner as person-specific models (PSMs), because they are tuned for a specific person. If, on the other hand, the system is to be deployed without prior commissioning (i.e., without being trained on data from its end users), then it must ship with a HAR model that has been pre-trained on data from a (presumably representative) sample of users. We refer to a model obtained in this manner as a person-independent model (PIM), because its performance is assumed to be independent of the person using it. PIMs are usually evaluated on subject-independent performance (i.e., unknown users), but it is not uncommon to see them evaluated on subject-dependent performance (known users), an approach that corresponds to a scenario wherein it is possible to obtain sample data from (some of) the system’s end users during commissioning, but not possible to identify users (and hence the appropriate PSM) once the system has been deployed. In addition to PIMs and PSMs, we also consider three different types of ensembles of PSMs to assess whether such ensembles can compete with PIMs in terms of subject-independent performance. The advantage of an ensemble of PSMs (EPSM) is that an EPSM can be improved by exploiting data from new users without accessing the dataset that was used to develop the initial model, simply by fitting a PSM for each new user and adding them to the ensemble. This is not the case for a PIM, which requires the dataset used to develop the initial model to exploit data from new users. The distinction between subject-dependent and subject-independent performance is ubiquitous in the HAR literature, and most empirical evaluations of HAR algorithms make it clear which performance measure is being assessed. Intuitively, we hypothesise that methods ought to do better in terms of subject-dependent performance than in terms of subject-independent performance; that PIMs outperform PSMs in subject-independent performance; and that PSMs outperform PIMs in subject-dependent performance. Unfortunately, not many HAR papers report results for more than one combination of the personalisation–generalisation approach (e.g., PIM or PSM), and subject-dependent and subject-independent performance, and none of them report results for all four combinations, making it impossible to verify whether these hypotheses are correct. This paper presents the first empirical comparison of the subject-dependent and subject-independent performances achieved with PIMs and PSMs on multiple (eight) HAR datasets, using four popular machine learning algorithms that have been used extensively and successfully in the HAR literature. The remainder of this paper proceeds as follows. Section 2 summarises the literature pertinent to the topic of subject-dependent and subject-independent performance in HAR, and the state of the art for the benchmark datasets used in this paper. Section 3 describes our methods, including the benchmark datasets, pre-processing, segmentation, feature extraction, and activity inference and evaluation. Section 4 presents the results of our experiments and their analysis. Section 5 discusses our findings in the context of the related works from Section 2, and Section 6 concludes the paper.

2. Related Work

In a 2019 survey of 56 papers on deep learning—deep neural, convolutional, and recurrent neural networks, auto-encoders, and restricted Boltzmann machines—for sensor-based human activity recognition, Wang et al. [2] concluded that there is no single “model that outperforms all others in all situations.” Comparing the results from the original studies for three HAR datasets, among them the Opportunity [3] dataset also employed in this paper, their survey identifies four papers [4,5,6,7] as the state of the art. The following paragraph summarises, in chronological order, the results from these four papers with respect to the predictive performance on the HAR datasets also used in this paper. The first of the four papers is the work by Jiang and Yin [4]. The deep convolutional neural network (DCNN) they proposed in 2015, termed DCNN+, recognises human activities from signal and activity images which are obtained by transforming the signals from a single inertial measurement unit (IMU) via the discrete Fourier transform or 2D wavelets. To disambiguate between pairs of classes with confused predictions (i.e., classes with similar large predicted probabilities) they employ binary SVM classifiers. The three considered models (DCNN, DCNN+, and SVM) all achieved a subject-dependent accuracy of >99% on the FUSION dataset [8]. The second paper, published in the same year by Zhang et al. [5], proposes a different DNN for human activity recognition. This DNN recognises human activities from the raw signals acquired from a wearable IMU and the signal magnitude of the accelerometer’s combined three axes. In an empirical comparison, the authors pit their DNN against traditional (i.e., not deep learning) machine learning algorithms with default, untuned hyper-parameters that operate on five statistical features (mean, standard deviation, energy, spectral entropy, and pairwise correlations between the accelerometer axes) extracted from the raw IMU signals. Their results show their DNN achieving a subject-dependent error rate of 18% on the Opportunity dataset, SVM being a close runner-up with an error rate of 19%. The third paper is the work by Ordóñez and Roggen [6], in which they propose a deep convolutional long short-term memory cell (LSTM) model for human activity recognition. Their LSTM outperformed the baseline convolutional neural network (CNN) in terms of subject-dependent performance (F-score: 93% vs. 91%) on the Opportunity dataset. The fourth paper was published in 2016 by Hammerla et al. [7]. In it, the authors compare CNNs, DNNs, and three different types of LSTMs across three benchmark HAR datasets, all of them consisting of data from multiple IMUs per user. Among them are two datasets, the Opportunity and PAMAP2 [9] datasets, which we also use in this paper. This paper is particularly elucidating because of its exploration of how sensitive models are to the various hyper-parameters that determine a deep learning model’s architecture, learning, and regularisation of parameters. The authors explore the hyper-parameter space by randomly sampling hyper-parameter configurations in hundreds or thousands of experiments. The results from these experiments clearly show that deep neural networks are extremely sensitive to hyper-parameter settings, which is illustrated by the differences between each model’s median and best performance. On the Opportunity dataset, the best model’s (LSTM) median score is 17 percentage points lower than its best score, and on the PAMAP2 dataset the best model’s (CNN) median score is seven percentage points lower than its best score. This latter number is the smallest discrepancy between best and median scores across all models and datasets. In their conclusions, the authors concur with Wang et al. [2] in that no single model dominates across all datasets. The best subject-dependent performance on the Opportunity dataset, an F-score of 93%, was achieved with a bi-directional LSTM, whereas the best subject-dependent performance on the PAMAP2 dataset, an F-score of 94%, was achieved with a CNN. In 2019, Jordao et al. [10] evaluated seven state-of-the-art methods published between 2010 and 2016, including the work by Jiang and Yin [4] discussed above, on six publicly available HAR benchmark datasets, including the PAMAP2 and MHEALTH [11] datasets, which are used in this paper. Three of the seven methods entirely rely on deep neural networks, while the remaining four use handpicked features as inputs to classification algorithms. Each method was evaluated with a different data segmentation strategy (overlapping and non-overlapping sliding windows), and its predictive performance estimated via stratified k-fold, leave-one-subject-out, and leave-trials-out cross-validation. Leave-trials-out cross-validation, which is discussed in more detail in Section 3, segments data into overlapping sliding windows one trial at a time. A trial corresponds to one individual’s performance of one activity or sequence of activities during the data acquisition. The results show that naïve resampling methods, such as stratified k-fold cross-validation, tend to inflate the predictive performance when used with overlapping sliding windows, because the overlap between subsequent windows can appear in both the training and test data. Based on the (unbiased) subject-dependent performance, estimated via leave-trials-out cross-validation, and the subject-independent performance, estimated via leave-one-person-out cross-validation, the authors identify two state-of-the-art methods for HAR with wearable sensor data. The first is an ensemble classifier, consisting of a decision tree, logistic regression, and a shallow multi-layer perceptron, proposed in 2015 by Catal et al. [12]. This ensemble achieved subject-dependent accuracy of 92% and 81% on MHEALTH and PAMAP2, respectively, with an average of 76% across all six datasets, and subject-independent accuracy of 95% and 85% on MHEALTH and PAMAP2, respectively, with an average of 69%. The second method that emerged as the state-of-the-art is the convolutional neural network (CNN) proposed in 2015 by Chen and Xue [13]. This CNN achieved subject-dependent accuracy of 90% and 82% on MHEALTH and PAMAP2, respectively, with an average of 83%, and subject-independent accuracy of 89% and 83% on MHEALTH and PAMAP2, respectively, with an average of 78%. Note that the 7–10 percentage points difference between the averages is largely due to the ensemble’s poor performance on the dataset to which the study’s authors could not apply the CNN. The authors’ analysis concludes that, although deep neural networks such as CNNs have achieved remarkable results in HAR with wearable sensor data, machine learning algorithms operating on handpicked features can in many cases achieve comparable results. In 2019, Abdu-Aguye and Gomaa [14] proposed an approach to feature extraction for sensor-based HAR. Their approach applies a wavelet transform and subjects the resulting decompositions to spatial pyramid pooling [15] to obtain fixed-length features which preserve both local and global patterns of the input signals. They used these features as inputs to a random forest, and compared the performance against a CNN. They estimated the subject-dependent performance by averaging over 15 repeated 75%/25% train/test splits. Their method achieved a subject-dependent accuracy of 89% and an F-score of 90%, while the CNN achieved an accuracy of 87% and an F-score of 87% on the REALWORLD dataset [16]. In 2020, Vakili et al. [17] evaluated seven machine learning algorithms, presumably operating on a set of extracted features, an artificial and a convolutional neural network, and a long short-term memory (LSTM) model on a range of internet-of-things datasets. The datasets’ application domains range from occupancy detection from environmental sensors, to rain prediction from weather station data, to HAR from wearable IMUs. They found that random forests outperformed the other methods on the SIMFALL [18] dataset with an average subject-dependent accuracy of 76%, estimated via 10-fold cross-validation. In 2018, Alharbi and Farrahi [19] proposed a CNN to recognise smoking activities from a smartphone and smartwatch’s inertial measurement units. They evaluated their approach on the UTSMOKE [20] dataset, but excluded the walking, sitting, and standing activities “because they were very simple to classify using the CNN.” They report an average subject-independent F1 score of 90%, estimated via a 70%/15%/15% train/validation/test split, but do not clarify how these splits can be construed in a subject-independent manner. This concludes our review of the state of the art in human activity recognition with respect to the datasets also used in this paper. We conclude our discussion of related work by reviewing the few papers that directly investigate the relationship between subject-dependent and subject-independent HAR performance.
Bao and Intille [21] assessed the subject-dependent performance of PSMs for recognising 20 activities of daily living (ADLs) across 20 users by training four learning algorithms on a set of semi-controlled laboratory data and evaluating them on a set of semi-naturalistic data, and the subject-independent performance of a PIM by performing a leave-one-subject-out cross-validation on the combined data from both sets. In a second experiment, they assess the subject-dependent performance of a PSM trained on three additional users’ laboratory data, and the subject-independent performance of a PIM trained on five different users’ laboratory data, using the three new users’ semi-naturalistic data for evaluation. Unfortunately, the differences in the protocols for estimating subject-dependent and subject-independent performance in the first experiment means that we cannot compare them directly (the latter accuracies are 18% to 50% higher than the former). Their second experiment, which affords a fairer comparison, directly contradicts these findings: the subject-dependent PSM accuracy (77%) exceeds the subject-independent PIM accuracy (73%) by 6%. Weiss and Lockhart [22] assessed the subject-independent and subject-dependent performances of PIMs, and the subject-dependent performances of PSMs for recognising six ADLs across 59 users and eight learning algorithms. They report that PSMs outperform a PIM by 2% to 30% on subject-dependent accuracy, and that a PIM achieves an 11% to 41% higher subject-dependent than subject-independent accuracy. These results suggest that, all else being equal, HAR methods will indeed perform better on data from known users than on data from unknown users. However, they tell us little about the size of the difference for a given personalisation–generalisation approach or about how the trade-off between subject-dependent and independent performance relates to the personalisation–generalisation approach.

3. Methods

We follow the standard approach to human activity recognition comprised of data pre-processing, segmentation into windows, feature extraction from those windows, and activity inference on them based on their features [23]—where the inference step is implemented with machine learning algorithms. We estimate and compare the performance of four popular machine learning algorithms—L2 (Ridge) regularised logistic regression, k-nearest neighbours (kNN), support vector machines (SVM), and a gradient boosted ensemble of decision trees (GBT)—using a set of features extracted from eight HAR datasets, which are summarised in Table 1.
For each dataset, Table 1 cites the relevant publication, and lists the number of activities (act) and people (ind), and the average number of trials per activity (±standard error) and sampling frequency (Hz). We chose datasets that were acquired via wearable inertial measurement units (IMU) comprised of an acceleration and angular velocity sensor, and worn either on the chest or the wrist. When sensors were to be worn on both wrists we chose the one associated with the right wrist. Unfortunately, the information about whether a user is right or left handed is unavailable for most datasets, making it impossible to choose the dominant wrist consistently. All datasets, except REALWORLD and SAFESENS which used a chest-worn sensor only, used a wrist-worn sensor, and only two datasets—PAMAP2 and SIMFALL—employed both a wrist and a chest-worn sensor. Figure 1 illustrates how the instances—each of which corresponds to the features extracted from one window—are distributed among the activities. Note that instead of distinguishing falls from ADLs in the SIMFALL dataset, which [18] were able to do with sensitivity, specificity, and accuracy all >99%, we focus on the 16 ADLs shown in the figure. Most of the activity labels are self-explanatory, but some of the activities in the UTSMOKE dataset merit further explanation. “SmokeST” denotes “Smoke Sitting”—smoking (presumably a cigarette) while sitting down—while “SmokeSD” denotes “Smoke Standing”—smoking while standing up. Similarly, “DrinkST” and “DrinkSD” denote “Drink Sitting” (drinking while sitting down) and “Drink Standing” (drinking while standing up), respectively.
The various cross-validation strategies, machine learning algorithms, and calculations of performance measures were implemented in Python (version 3.7.3), using the scipy ([25] version 1.1.0), numpy ([26] version 1.16.2), pandas ([27] version 0.23.3), and sklearn ([28] version 0.20.2) libraries, and parallelised via GNU parallel ([29] version 20161222). Analysis—t-tests, mixed effects models, and estimated marginal means—and all visualisations, were implemented in R ([30] version 3.6.1), wherein we used the mixed effects model implementation from the lme4 library by ([31] version 1.1), and the estimated marginal means implementation from the emmeans library by ([32] version 1.3.2).
We propose and consider another personalisation–generalisation approach in addition to the person-independent model (PIM) and person-specific model (PSM), which we term an ensemble of PSMs (EPSM). An EPSM maintains a PSM for each known user. When an instance for a known user needs to be classified, an EPSM simply applies that user’s PSM, but when an instance originates with an unknown user, it applies each user’s PSM to obtain confidence scores (e.g., the estimated probability) for each activity of interest. Then the EPSM calculates each activity’s mean score, and classifies the instance to the activity with the maximum mean score. To deal with the (very few) users for whom the data do not cover all the activities of interest, and whose PSMs are therefore unaware of some activities and hence unable to generate a confidence score for those activities, we assume that those activities have a probability of zero. This is not unreasonable if we accept that some people will never perform certain activities (e.g., smoking, military crawling).
In addition to the basic, unweighted EPSM described above, this paper proposes two flavours of weighted EPSMs. A weighted EPSM makes predictions for known users in the same manner as the basic EPSM described above, but when making predictions for unknown users, a WEPSM combines its constituent models’ predictions via a weighted average. The two types of weighted EPSMs proposed in this paper differ in how the weights are determined. The first type is the κ -weighted EPSM (WEPSM). A WEPSM weights its constituents’ predictions according to each PSM’s average κ across all the other training users—i.e., all users except the one whose data are held out for testing and the one whose data (were) used for fitting the PSM. The second type is the baseline-feature-weighted EPSM (WEPSMbf). A baseline-feature-weighted EPSM weights its constituents’ predictions according to the mean euclidean distance between the PSM (training) user’s baseline features and the test user’s baseline features. A user’s baseline features are the features extracted from a random instance (window) of standing or sitting. To obtain the weight for a given training-user and test-user, we calculate the pairwise distances between the two users’ baseline features, and take the mean of these distances to weight the training user’s PSM predictions for the test user when aggregating them.
Figure 2 provides a graphical overview of the experiment we conducted, whose details are explained in the following subsections. The raw sensor signals from each dataset and sensor are first subjected to pre-processing. The pre-processed data—extracted features and labels, along with metadata identifying the originating dataset, sensor, and user—are then used to evaluate each learning algorithm’s ability to infer human activities in terms of subject-dependent (the figure’s left-hand branch) and subject-independent (the right-hand branch) performance. The figure also illustrates the difference between PSMs, unweighted EPSMs, κ -weighted EPSMs (WEPSM), and baseline-feature-weighted EPSMs (labelled "WEPSM_BF" in the figure).

3.1. Pre-Processing, Segmentation, and Feature Extraction

Some datasets come with a constant timestamp for each trial—presumably introduced when attempting to store POSIX® epoch timestamps in (sub)millisecond resolution in Microsoft® Excel® spreadsheets. For these datasets we generate timestamps with a fixed inter-arrival time equal to the dataset’s nominal sampling frequency. Then, we (automatically) separate the raw data into non-overlapping natural trials by splitting the signal whenever the activity (label) changes or the inter-arrival time (i.e., the time between two subsequent samples) exceeds 1.5 s. To ensure that we have at least two trials per user and activity, each of the natural trials is then split into non-overlapping batches of 15 s. Next, the body and gravity components of each trial’s accelerometer signal are separated by the elliptical infinite-impulse response (IIR) low pass filter separates described by [33]. After discarding the original accelerometer data—which do not contain any information beyond that in the gravity and body components—we are left with three tri-axial signals: the gyroscope signal, the body acceleration signal, and the gravity acceleration signal. Finally, a set of time and frequency-domain features is extracted along a sliding 3 s window with 50% (1.5 s) overlap from each trial. From the angular velocity signal and both acceleration components we extract the mean, standard deviation, skew, and kurtosis, and from the angular velocity and body acceleration signal, the spectral power entropy, peak-power frequency, signal magnitude area, and the pairwise correlations between each signal’s axes. This amounts to a total of 84 features that are extracted from each window.

3.2. Activity Inference and Evaluation

We use logistic ridge regression with C = 0.98, a kNN classifier with k = 2 and weighted voting, a SVM classifier with a radial basis function with kernel coefficient γ = 0.001 and cost penalty C = 316, and a GBT with a learning rate α = 0.02 and comprised of 750 trees. The parameters for kNN, SVM, and GBT are taken from [24], who tuned them for subject-independent performance on the 17 activities in the SAFESENS dataset. The ridge parameter of C = 0.98 corresponds to weak regularisation, and was chosen to counteract the impact of correlated features. All features are standardised ( [ x x ¯ ] / s ) according to each feature’s mean ( x ¯ ) and standard deviation (s) in the training data. We use Cohen’s kappa ( κ ) to quantify the predictive performance because—unlike other performance metrics such as sensitivity, specificity, and accuracy—it corrects for the probability of obtaining the observed level of agreement between the ground truth and predicted labels by chance, and because it is designed to measure predictive performance for multi-class classification.
To estimate an algorithm’s subject-dependent performance, the trials are used to generate the folds in a k-fold cross validation, a method we call leave-trials-out cross-validation [10]. Leave-trials-out cross-validation ensures that the raw data used to derive an instance in a training split are never used to derive the instances that constitute the corresponding test split, an issue that is bound to occur when working with instances derived from partially overlapping sliding windows [10], as we do here. PIM performance for known users is estimated by carrying out a k-fold leave-trials-out cross-validation across all the users in each dataset, and PSM performance by carrying out a separate k-fold leave-trials-out cross-validation for each user. In both cases k = n , where n denotes the number of people in the dataset. To estimate the subject-independent performance, we carry out a leave-m-users-out cross-validation with m = 1 for EPSM and PIM, and m = n 1 for PSM.

4. Results and Analysis

Figure 3 illustrates the trade-off between the subject-dependent performance—i.e., the performance for users who were represented in the data used for training the model—on the horizontal axis, and the subject-independent performance—the performance for users who were not represented in the training data—on the vertical axis. In this figure, each datum corresponds to a single person (user), except in the case of person-specific models, where it corresponds to the median performance a model trained on data from the known user achieved on the other users in the dataset. The symbol and colour indicate which personalisation–generalisation approach (PIM, PSM, EPSM, WEPSM, or WEPSM bf ) was used. Table 2 summarises the results depicted in Figure 3, but using the PSM performance for all rather than, as shown in the figure, only that for the average unknown user. The table lists the mean κ (in %) ± standard error for each personalisation–generalisation approach, machine learning algorithm, dataset, and sensor location. To make the results more comparable with other results for these same datasets, the appendix presents the same tables with the accuracy (Table A1), error rate (Table A2), and weighted F-score (Table A3).
Inspecting these results, it is clear that the subject-independent performance is systematically and substantially worse than the subject-dependent performance. It is also clear that PSMs perform worse than PIMs in terms of their subject-independent performance. Furthermore, it is clear that a gradient boosted ensemble of decision trees (GBT) generally outperforms logistic regression (logreg) and k-nearest neighbours (kNN), with few exceptions. The most notable of these is the subject-independent performance with PSM, EPSM, WEPSM, and WEPSM bf on the SAFESENS dataset, where both kNN and logistic regression outperform gradient boosted trees, in some cases by more than one standard error. However, things are less clear when it comes to comparing PIMs and PSMs in terms of their subject-dependent performance, comparing PIMs and EPSMs in terms of their subject-independent performance, or comparing the different types of EPSMs against each other. To elucidate these matters, and to quantify the obvious differences mentioned above, we turn to statistical analyses which are discussed in the remainder of this section.

4.1. Analysis of the Subject-Dependent Performance

We can pair the performance when a person-specific model (PSM) is combined with a machine learning algorithm and applied to the data from a known person for a given dataset and sensor, to the performance when the same algorithm is combined with a person-independent model (PIM) and applied to the same dataset, sensor, and person. A paired t-test of these data yields a 95% confidence interval (C.I.) of 4.1 to 5.2 percentage points (hereafter, points) for the difference between the κ achieved with PSMs and that achieved with PIM, with a mean difference of 4.6 points ( t 442 = 16.2 , p < 2.2 × 10 16 ), suggesting that we can be 95% confident that a PSM outperforms a PIM on data from known users by 4.1 to 5.2 points. However, it is unlikely that the t-test’s underlying assumption of identically and independently distributed (IID) data is met, because the difference in the subject-dependent performance between PIM and PSM might depend not only on the dataset—which is expected due to the different activities of interest, and evident in Figure 3—but also on the learning algorithm.
A more appropriate tool for analysing data, such as these, which are not IID, is the linear mixed effects model (LMM). A LMM extends linear regression with so-called random effects which allow us to impose structure on the residuals. We can, for example, specify that the performances within datasets are correlated, or even that the difference in performance between personalisation–generalisation approaches varies depending on the dataset. The random effects are assumed to add up to zero, and hence the fixed effects (which are analogous to linear regression coefficients) can be estimated via (restricted) maximum likelihood. A LMM, like the linear regression model which it is based on, is built on the assumption of normally distributed residuals. The generalised linear mixed effects model (GLMM) extends the LMM to non-normal data, analogously to the generalised linear model (GLM). Like the GLM, it employs a link function, such as the logit, and error distribution (from the exponential family) to model non-normal data. We use logistic GLMMs—a GLMM with a logistic link function, and a binomial error distribution—to analyse the subject-dependent performance and subject-independent performance, and the difference between them. The explanatory variables considered as fixed effects were those of the personalisation–generalisation approach (PGA)—e.g., PIM, PSM, EPSM, and WEPSM; those of the machine learning algorithm (MLA); and the random effects of the dataset and sensor. For a detailed treatment of LMMs and GLMMs we refer interested readers to [34].
We use a GLMM to model the subject-dependent performance as a combination of (fixed) effects for the machine learning algorithm and personalisation–generalisation approach—either PIM or PSM, since subject-dependent EPSM performance is identical to that of its constituent PSMs—and a random effect to control for the variation of the personalisation–generalisation approach effect between datasets. This model explains the observed variation in the response with a residual standard deviation of 1.0157 between datasets, and with a standard deviation of 0.0255 between datasets’ sensors. This model reveals that the (random) effect of applying PSM varies with a standard deviation of 0.2929 between datasets, where it is very weakly negatively correlated (−0.11) with PIM performance, and with a standard deviation of 0.0407 between datasets’ sensors, where it is strongly positively correlated (0.75) with PIM performance. This confirms the intuition that a PSM likely confers less advantage when applied to a dataset on which a PIM performs well. The maximum likelihood estimates of the fixed effects, which are shown in Table 3, indicate that GBT—with an estimated κ of 89.4% and a 95% C.I. of 80.6% to 94.4% when used as a PIM—outperforms SVM by 19.5% (18.5% to 20.5%), logistic regression by 45.5% (44.9% to 46.2%), and kNN by 46% (45.3% to 46.6%), regardless of whether they are combined with a PIM or a PSM. They further show that PSMs outperform the corresponding PIM by 43.5% (17% to 76%, p = 0.00058) on subject-dependent performance.
Both the paired t-test and the GLMM agree that PSMs outperform PIMs in terms of subject-dependent performance. The GLMM estimates that PSMs outperform PIMs by 17% to 76% (with a mean of 43.5%) on subject-dependent performance in terms of the odds-ratio, the t-test estimates that difference as 4.1 to 5.2 points with a mean of 4.6 points. These estimates are consistent with each other. The GLMM estimate for GBT with PSM is κ = 92.3% with a C.I. of 85.5% to 96.1%. With PIM it is κ = 89.4% with a C.I. of 80.6% to 94.4%; 89.4% + 4.6 points equates to 94%, which is only 1.7 points above the GLMM estimate and well within the C.I. of 85.5% to 96.1% postulated by the GLMM. For kNN, the GLMM estimates the κ achieved with PSM at 86.7% with a C.I. of 76.2% to 93%, while the κ achieved with PIM is estimated at 82% with a C.I. of 69.2% to 90.2%; 82% + 4.6 equates to 86.6%, which not only lies well within the C.I. postulated by the GLMM, but is exceedingly close to the point estimate of 86.7%. It is similar for logistic regression: the GLMM estimates the κ achieved with PSM at 86.8% with a C.I. of 76.3% to 93.1%, and with PIM at 82.1% with a C.I. of 69.4% to 90.3%; 82.1% + 4.6 equates to 86.7%, which is only 0.1 point below the GLMM point estimate (and well withing the C.I. postulated by the GLMM). For SVM, the GLMM estimates the κ achieved with PSM at 90.7% with a C.I. of 82.6% to 95.2%, and the κ achieved with PIM at 87.1% with a C.I. of 77% to 93.2%; 87.1% + 4.6 equates to 91.7%, which is only one point above the GLMM point estimate, and well within the 95% C.I. postulated by the GLMM.
Both the t-test and GLMM rely on statistical assumptions. To compare the evaluated methods without relying on statistical assumptions, we rank the methods by their κ within each user, dataset, and sensor. Figure 4 illustrates the distribution of each method’s (learning algorithm + PSM or PIM) ranks across all datasets, sensors, and users. While no single method dominates across all users, the figure shows that person-specific models with gradient boosted ensembles of decision trees (GBT) perform better than all other methods for nearly 80% of users, and better than all but one method for over 10% of users, which confirms our statistical analysis. It is also clear that PSMs more often than not outperform a PIM for any given algorithm.

4.2. Analysis of the Subject-Independent Performance

A binomial logistic GLMM with fixed effects for the learning algorithm and personalisation–generalisation approach, and a random effect for sensors nested within datasets explains the variation in subject-independent performance with a residual error that varies with a standard deviation of 0.781 between datasets and with a standard deviation of 0.0218 between datasets’ sensors. The maximum likelihood estimates of the fixed effects, which are shown in Table 4, indicate that GBT—with an estimated κ of 71.3% (59.2% to 81.0%, p = 0.001) when used as a PIM—outperforms kNN by 20.2% (19.7% to 20.6%), logistic regression by 21.7% (21.3% to 22.2%), and SVM by 15.8% (15.3% to 16.3%). PIM outperforms PSM by 55.9% (55.3% to 56.2%), EPSM by 17.5% (17% to 18.1%), WEPSM by 16.4% (15.8% to 16.9%), and WEPSM bf by 18.4% (17.8% to 18.9%). All p-values < 2 × 10 16 . This analysis clearly shows that PIM performs better for unknown users (i.e., on subject-independent performance) than the other personalisation–generalisation approaches, and that ensembles of PSMs perform better than a PSM. However, because the fixed effects for the different EPSMs—unweighted (EPSM), κ -weighted (WEPSM), or baseline-feature-weighted (WEPSM bf )—estimate the difference between the particular EPSM and PIMs, and because their estimates are quite similar, this analysis on its own cannot compare the different types of EPSMs. To compare the different types of EPSMs we employ paired t-tests and estimated marginal means (also known as least-squares means) analysis.
According to the estimated marginal means, which are shown in Figure 5, the odds achieved by κ -weighted EPSMs are 1.4% higher than those achieved by unweighted EPSMs (p = 0.0009), which in turn are 1.1% higher than those achieved by EPSMs weighted by the inverse distance between the train and test user’s baseline features (p = 0.0159). A paired t-test of the difference in the subject-independent performance between WEPSM and EPSM yields a mean difference of 0.32 points with a 95% C.I. of 0.20 to 0.44 points, and a t-value of 17 on 443 degrees of freedom, which corresponds to a p-value of 3.48 × 10 7 . This shows that a weighted EPSM significantly (albeit by less than one third of a point) outperforms an unweighted EPSM. A paired t-test of the difference in the subject-independent performance between EPSM and WEPSM bf yields a mean difference of 0.38 points, a 95% C.I. of 0.21 to 0.55 points, and a t-value of 4.45 on 435 degrees of freedom, corresponding to a p-value of 1.1 × 10 5 . This shows that using baseline features for weighting the PSM predictions performs significantly worse (albeit by little more than one third of a point) than an unweighted EPSM. Both the estimated marginal means and paired t-tests lead to the conclusion that κ -weighted EPSMs significantly outperform unweighted EPSMs, which in turn significantly outperform EPSMs that are weighted by the inverse mean distance between the train and test user’s baseline features.
Analogously to the subject-dependent ranks shown in Figure 4, Figure 6 illustrates the distribution of each method’s (personalisation–generalisation approach + learning algorithm) subject-independent ranks across all datasets, sensors, and users. The figure clearly shows that PIM + GBT ranks first or second for over 60% of users—more often than any other method—and third or fourth for over 20% of users, which confirms the statistical analysis’s finding that PIM + GBT outperforms other methods on subject-independent performance. They also show that PIMs tend to perform better than other personalisation–generalisation approaches, in particular PSM, whose ranks are mostly in the bottom third. We can also see how a κ -weighted EPSM tends to shift the rather flat distribution of unweighted EPSMs slightly towards the left, towards the better (lower) ranks, particularly when combined with kNN or GBT.

4.3. Comparison of Subject-Dependent and Independent Performance

We use a binomial logistic GLMM with a fixed effect for the performance type (subject-dependent or subject-independent), one for the personalisation–generalisation approach (PIM, PSM, EPSM, etc.), and one for the interaction between them. There are two random effects, one for sensors nested within datasets, and one for the four learning algorithms. The reference level (intercept) corresponds to the subject-dependent performance achieved by PIMs. This model shows that the subject-dependent performance varies with a standard deviation of 0.1770 between learning algorithms, 0.825 between datasets, and 0.0382 between datasets’ sensors. The model’s estimates for the fixed effects are shown in Table 5, along with their 95% C.I.s and p-values. According to these estimates, PIMs achieve a subject-dependent κ , averaged over learning algorithms, of 82.2% (71.8% to 89.3%, p = 4.8 × 10 7 ) and PSMs outperform PIMs by 46% (44.9% to 47.2%, p < 2 × 10 16 ) in terms of the subject-dependent odds, with an estimated mean κ of 87.1% and a C.I. of 78.9% to 89.3% (according to the estimated marginal means shown in Figure 7). The subject-independent odds of PIMs are estimated at 48.1% (47.7% to 48.4%) of their subject-dependent odds, with a κ of 69% and an (estimated marginal means) C.I. of 55.1% to 80.1%. The subject-independent odds of PSMs are 13.6% of their subject-dependent odds, with a κ of 47.9% and an estimated marginal means C.I. of 33.7% to 62.4%. The subject-independent odds of EPSMs are 27% of their subject-dependent odds, with a κ of 64.6% and estimated marginal means C.I. of 50.2% to 76.8%. The subject-independent odds of WEPSMs are 27.4% of their subject-dependent odds, with a κ of 64.9% and an estimated marginal means C.I. of 50.6% to 77%. The subject-independent odds of WEPSM bf are 26.8% of their subject-dependent odds, with a κ of 64.4% and an estimated marginal means C.I. of 50% to 76.6%.

5. Discussion

Our analysis of the results shows that, on average, the best subject-dependent performance is achieved with PSMs and the best subject-independent performance with PIMs. Hence, in order to simultaneously optimise subject-dependent and subject-independent performance, we should use a PIM for unknown users and PSMs for known users wherever possible. If we use a PIM, rather than a PSM, to make predictions for known users we forego an expected improvement of over 43% in terms of the odds of a correct classification. For the datasets and models investigated in this paper, this corresponds to 4.1 to 5.6 percentage points difference in Cohen’s κ . If, on the other hand, we use a PSM rather than a PIM for unknown users we forego an expected improvement of nearly 56% in terms of the odds of a correct classification. If a PIM is not practicable—e.g., because we do not have access to the original training data when the time comes to integrate new users’ data into the HAR model—then an ensemble of PSMs can be employed. Among the three approaches for forming an EPSM which we considered in this paper, the κ -weighted EPSM emerged as a slightly, but significantly, better method for forming an EPSM than an unweighted EPSM, or a baseline-feature-weighted EPSM. Although a PIM performs significantly better than a κ -weighted EPSM (WEPSM), the difference in odds is estimated at a mere 16.4%, which is only 0.6 percentage points bigger than the difference in subject-independent κ between GBT and SVM, the two best-performing learning algorithms. Our analysis further shows that GBT significantly outperforms kNN, L 2 -regularised logistic regression, and SVM, but the latter by only 15.2%.
The state of the art for the FUSION dataset achieves a subject-dependent accuracy of > 99% [4]. Our GBT PIM achieves an accuracy of 98.3% ± 0.3 and our GBT PSM an accuracy of 98% ± 0.3 on this dataset. On the OPPORT dataset, the state of the art achieves a subject-dependent error rate of 18% [5] and F-score of 93% [6,7]. Our GBT PSM achieves a subject-dependent error rate of 10.9% ± 1.5 (GBT PIM: 12.2% ± 1.9) and F-score of 89.1% ± 1.5 (GBT PIM: 87.8% ± 1.9) on this dataset. According to Jordao et al. [10], the state of the art for the MHEALTH dataset achieves a subject-dependent accuracy of 92% (with a C.I. of 88% to 96%), and a subject-independent accuracy of 95% (C.I.: 91% to 98%). Our GBT PIM achieves a subject-dependent accuracy of 97.8% ± 0.7 (GBT PSM: 97.5% ± 1.1), and a subject-independent accuracy of 84% ± 3.3 on this dataset. Following the same paper [10], the state of the art for the PAMAP2 dataset achieves a subject-dependent accuracy of 82% (C.I.: 77% to 88%) and a subject-independent accuracy of 85% (C.I.: 76% to 94%). Our GBT PSM achieves a subject-dependent accuracy of 88.9% ± 0.6 (GBT PIM: 88.8% ± 0.4) when using the chest-mounted sensor, and our GBT PIM a subject-independent accuracy of 80.6% ± 2.6 when using the wrist-mounted sensor from this dataset. On the REALWORLD dataset, the state of the art achieves a subject-dependent accuracy of 89% and F-score of 90% [16]. Our GBT PSM achieves a subject-dependent accuracy of 96.9% ± 0.3 (GBT PIM: 94.5% ± 0.5) and F-score of 96.8% ± 0.5 (GBT PIM: 94.7% ± 0.5) on this dataset. The state of the art for the SIMFALL dataset achieves a subject-dependent accuracy of 76% [17]. Our GBT PSM achieves a subject-dependent accuracy of 68.1% ± 1.2 (GBT PIM: 60% ± 1.1) when using the chest-mounted sensor and 65.1% ± 1.4 (GBT PIM: 58.3% ± 1.4) when using the wrist-mounted sensor from this dataset. On the UTSMOKE dataset, the state of the art achieves a subject-independent accuracy of 90% [19]. Our GBT PIM achieves a subject-independent accuracy of 73.2% ± 2.5 (SVM PIM: 73.6% ± 2.3) on this dataset. To summarise, our approach performs comparable (i.e., within the margin of error) to the state of the art for all but one (UTSMOKE) dataset in terms of the subject-independent performance. In terms of the subject-dependent performance, our approach performs clearly worse than the state of the art (by about 8 points) for one dataset (SIMFALL), within one percentage point for another (FUSION), and better than the state of the art for four datasets—namely the REALWORLD, UTSMOKE, MHEALTH, and PAMAP2 datasets. This shows that our gradient boosted ensemble of decision trees, combined with PSMs for subject-dependent or a PIM for subject-independent performance, performs comparably to the state-of-the-art for a wide range of HAR problems, without problem-specific feature engineering or tuning of model hyper-parameters.

6. Conclusions

This paper compared the subject-dependent and subject-independent performances of person-independent models (PIMs), person-specific models (PSMs), and three types of ensembles of PSMs (EPSMs)—unweighted, κ -weighted, and baseline-feature-weighted—when combined with four popular HAR algorithms across eight publicly available HAR datasets. An analysis with generalised linear mixed effects models (GLMM) showed that GBT significantly outperforms the other algorithms on both subject-dependent and subject-independent performance; that PSMs outperform a PIM by 43.5% (in terms of the odds of correct versus incorrect classification) in subject-dependent performance; and that a PIM outperforms PSMs by 55.9% and κ -weighted EPSM by 16.4% on subject-independent performance. Furthermore, our analysis of the subject-independent performance shows that κ -weighted EPSMs significantly outperform unweighted EPSMs, albeit by as little as 0.32 percentage points and 1.4% in terms of the odds, and that unweighted EPSMs in turn significantly outperform baseline-feature-weighted EPSMs by about the same amount.
Our approach—a gradient boosted ensemble of decision trees combined with person-specific models for known users and with a PIM for unknown users—performs comparably to the state of the art on one dataset and outperforms the state of the art on four datasets in terms of subject-dependent performance. In terms of the subject-independent performance, our approach performs comparably to the state of the art on all but one of the datasets, for which published results that include the subject-independent performance exist.

Author Contributions

Conceptualisation, S.S.; methodology, S.S. and S.T.; software, S.S.; validation, S.S. and S.T.; formal analysis, S.S.; investigation, S.S.; resources, K.N.B. and B.O.; data curation, S.S.; writing—original draft preparation, S.S.; writing—review and editing, S.T., K.N.B., and B.O.; visualisation, S.S.; supervision, K.N.B. and B.O.; project administration, S.S. and S.T.; funding acquisition, K.N.B. and B.O. All authors have read and agreed to the published version of the manuscript.

Funding

This publication has emanated from research conducted with the financial support of Science Foundation Ireland (SFI) under grant number 12/RC/2289-P2, the European Regional Development Fund under grant number 13/RC/2077-CONNECT, and the European-funded project SAFESENS under the ENIAC program in association with Enterprise Ireland under grant number IR20140024.

Conflicts of Interest

The authors declare no conflict of interest. The funding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Appendix A. Alternative Performance Metrics

Table A1 lists the mean subject-dependent and -independent accuracy score (in percent, ± standard error) achieved with PIM, PSM, unweighted EPSM, and EPSMs weighted by their constituents’ estimated subject-independent κ (WEPSM), and by the mean distance between constituent’s baseline features and the test user’s baseline features (WEPSM bf ). Table A2 lists the mean error rate (zero-one loss) and Table A3 the mean weighted F-score for same.
Table A1. Subject-dependent and subject-independent accuracy (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM) a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weigthed EPSM (WEPSM bf ).
Table A1. Subject-dependent and subject-independent accuracy (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM) a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weigthed EPSM (WEPSM bf ).
Subject-DependentSubject-Independent
DatasetSensormlaPIM(E)PSMPIMPSMEPSMWEPSMWEPSM bf
FUSIONwristgbt98.3 ± 0.398.0 ± 0.393.5 ± 1.884.1 ± 1.891.9 ± 2.292.1 ± 2.291.7 ± 2.2
knn94.9 ± 0.895.0 ± 0.987.9 ± 1.778.5 ± 2.289.3 ± 2.489.3 ± 2.489.1 ± 2.4
logreg97.2 ± 0.597.8 ± 0.393.0 ± 1.882.3 ± 2.391.0 ± 2.491.0 ± 2.490.6 ± 2.3
svm98.2 ± 0.398.1 ± 0.392.2 ± 1.882.9 ± 2.091.7 ± 2.391.8 ± 2.391.4 ± 2.3
MHEALTHwristgbt97.8 ± 0.797.5 ± 1.184.0 ± 3.363.2 ± 2.174.7 ± 3.274.9 ± 3.174.2 ± 3.2
knn93.6 ± 1.294.3 ± 1.378.3 ± 2.860.0 ± 1.974.0 ± 2.374.2 ± 2.475.3 ± 2.3
logreg93.8 ± 1.296.1 ± 1.380.9 ± 3.058.3 ± 2.272.8 ± 2.672.9 ± 2.673.4 ± 2.6
svm95.9 ± 0.997.0 ± 0.983.6 ± 2.461.8 ± 1.974.5 ± 3.674.7 ± 3.674.1 ± 3.6
OPPORTwristgbt87.8 ± 1.989.1 ± 1.579.8 ± 4.672.4 ± 3.478.7 ± 4.978.5 ± 5.078.7 ± 4.8
knn80.7 ± 1.883.2 ± 1.867.7 ± 2.761.6 ± 2.270.0 ± 3.469.5 ± 3.170.1 ± 2.8
logreg81.5 ± 2.584.6 ± 1.973.8 ± 4.363.9 ± 2.172.0 ± 4.071.8 ± 4.272.2 ± 3.8
svm87.0 ± 1.787.4 ± 1.677.4 ± 4.365.6 ± 1.875.3 ± 4.075.1 ± 4.175.6 ± 3.7
PAMAP2chestgbt88.8 ± 0.488.9 ± 0.679.6 ± 3.959.2 ± 2.775.2 ± 3.875.6 ± 3.974.8 ± 3.5
knn78.0 ± 0.980.6 ± 1.167.3 ± 2.354.1 ± 1.771.0 ± 3.071.4 ± 3.169.8 ± 2.8
logreg84.3 ± 1.086.9 ± 0.875.0 ± 3.453.9 ± 2.272.4 ± 4.472.4 ± 4.571.7 ± 4.4
svm87.4 ± 0.786.6 ± 0.776.3 ± 4.154.6 ± 2.572.4 ± 4.772.9 ± 4.771.6 ± 4.7
PAMAP2wristgbt88.1 ± 1.087.4 ± 0.880.6 ± 2.661.1 ± 2.274.6 ± 2.475.0 ± 2.474.8 ± 2.6
knn79.7 ± 1.381.1 ± 1.468.7 ± 3.752.6 ± 2.571.4 ± 3.571.7 ± 3.670.9 ± 3.4
logreg84.2 ± 1.685.2 ± 1.277.2 ± 3.754.8 ± 3.372.0 ± 4.572.4 ± 4.572.1 ± 4.4
svm86.5 ± 1.185.0 ± 1.175.6 ± 4.851.6 ± 2.871.5 ± 4.172.0 ± 4.171.3 ± 3.9
REALWORLDchestgbt94.5 ± 0.596.9 ± 0.376.4 ± 3.747.1 ± 1.768.7 ± 3.469.9 ± 3.369.2 ± 3.4
knn88.1 ± 1.192.9 ± 0.866.1 ± 2.947.2 ± 2.168.3 ± 3.269.2 ± 3.168.5 ± 3.3
logreg86.8 ± 1.496.3 ± 0.466.4 ± 5.540.7 ± 2.064.0 ± 3.865.3 ± 3.864.0 ± 4.0
svm93.6 ± 0.696.4 ± 0.468.1 ± 4.640.6 ± 2.062.0 ± 4.263.3 ± 4.062.2 ± 4.3
SAFESENSchestgbt94.7 ± 0.797.3 ± 0.871.2 ± 2.633.3 ± 1.953.1 ± 4.552.7 ± 5.357.9 ± 3.1
knn83.0 ± 1.889.0 ± 1.459.9 ± 3.635.8 ± 1.858.9 ± 3.358.9 ± 3.458.0 ± 2.5
logreg80.9 ± 1.693.7 ± 1.067.7 ± 2.833.1 ± 1.758.6 ± 2.558.0 ± 2.458.0 ± 2.6
svm89.2 ± 1.295.7 ± 0.870.2 ± 2.635.8 ± 1.856.7 ± 2.457.8 ± 2.356.1 ± 2.8
SIMFALLchestgbt60.0 ± 1.168.1 ± 1.247.5 ± 1.524.6 ± 0.637.8 ± 1.437.8 ± 1.437.0 ± 1.4
knn48.6 ± 0.952.8 ± 1.234.8 ± 0.725.0 ± 0.637.7 ± 1.037.5 ± 1.036.5 ± 0.9
logreg42.3 ± 0.855.4 ± 1.138.8 ± 1.120.0 ± 0.433.6 ± 0.933.6 ± 0.932.7 ± 0.9
svm53.4 ± 0.752.8 ± 1.542.2 ± 1.319.7 ± 0.430.7 ± 0.931.0 ± 0.929.6 ± 0.9
SIMFALLwristgbt58.3 ± 1.465.1 ± 1.444.6 ± 2.224.4 ± 0.937.3 ± 2.037.4 ± 2.036.4 ± 2.1
knn48.2 ± 1.152.1 ± 1.133.8 ± 1.424.4 ± 0.936.2 ± 1.536.3 ± 1.535.4 ± 1.6
logreg41.3 ± 1.352.9 ± 1.237.0 ± 1.921.5 ± 0.832.3 ± 1.632.8 ± 1.632.1 ± 1.7
svm51.5 ± 1.249.4 ± 1.440.1 ± 2.219.2 ± 0.731.6 ± 1.431.9 ± 1.531.1 ± 1.6
UTSMOKEwristgbt83.6 ± 1.392.1 ± 0.873.2 ± 2.561.3 ± 1.670.3 ± 2.870.4 ± 2.870.2 ± 2.7
knn79.7 ± 1.183.9 ± 1.067.1 ± 2.157.8 ± 1.566.3 ± 2.466.4 ± 2.566.1 ± 2.3
logreg73.3 ± 1.886.4 ± 1.068.5 ± 2.257.6 ± 1.465.1 ± 2.165.2 ± 2.265.3 ± 2.0
svm86.0 ± 1.190.7 ± 0.873.6 ± 2.359.6 ± 1.668.8 ± 2.568.9 ± 2.569.0 ± 2.3
Table A2. Subject-dependent and subject-independent error rate (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Table A2. Subject-dependent and subject-independent error rate (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Subject-DependentSubject-Independent
DatasetSensormlaPIM(E)PSMPIMPSMEPSMWEPSMWEPSM bf
FUSIONwristgbt1.8 ± 0.32.0 ± 0.392.7 ± 2.382.3 ± 2.190.6 ± 2.890.8 ± 2.890.4 ± 2.8
knn5.1 ± 0.85.0 ± 0.987.7 ± 1.877.4 ± 2.588.2 ± 2.988.4 ± 2.988.2 ± 2.9
logreg2.8 ± 0.52.2 ± 0.392.5 ± 2.280.7 ± 2.589.8 ± 2.989.9 ± 2.989.4 ± 2.9
svm1.8 ± 0.31.9 ± 0.391.6 ± 2.181.2 ± 2.390.4 ± 2.990.5 ± 3.090.1 ± 2.9
MHEALTHwristgbt2.2 ± 0.72.5 ± 1.181.9 ± 4.058.3 ± 2.269.8 ± 3.670.1 ± 3.569.2 ± 3.6
knn6.4 ± 1.25.7 ± 1.376.6 ± 2.856.3 ± 1.970.9 ± 2.771.2 ± 2.772.6 ± 2.6
logreg6.2 ± 1.23.9 ± 1.378.9 ± 3.353.0 ± 2.167.6 ± 3.267.8 ± 3.268.2 ± 3.2
svm4.1 ± 0.93.0 ± 0.981.9 ± 2.756.6 ± 2.070.0 ± 4.170.2 ± 4.169.3 ± 4.1
OPPORTwristgbt12.2 ± 1.910.9 ± 1.579.3 ± 5.071.4 ± 4.277.3 ± 5.777.0 ± 5.877.4 ± 5.5
knn19.3 ± 1.816.9 ± 1.867.2 ± 2.961.1 ± 2.369.2 ± 3.668.7 ± 3.469.5 ± 3.0
logreg18.5 ± 2.515.5 ± 1.972.7 ± 4.963.0 ± 2.670.3 ± 4.770.2 ± 4.970.8 ± 4.5
svm13.0 ± 1.712.6 ± 1.676.6 ± 4.864.5 ± 2.273.9 ± 4.773.6 ± 5.074.4 ± 4.4
PAMAP2chestgbt11.2 ± 0.411.1 ± 0.679.2 ± 4.656.2 ± 3.374.3 ± 4.574.8 ± 4.673.9 ± 4.2
knn21.9 ± 0.919.4 ± 1.167.7 ± 2.552.4 ± 2.170.9 ± 3.471.3 ± 3.569.5 ± 3.2
logreg15.7 ± 1.013.1 ± 0.874.1 ± 4.150.4 ± 2.771.4 ± 5.171.4 ± 5.270.4 ± 5.1
svm12.6 ± 0.713.4 ± 0.775.9 ± 4.651.5 ± 3.071.2 ± 5.671.7 ± 5.570.3 ± 5.5
PAMAP2wristgbt11.9 ± 1.012.6 ± 0.879.9 ± 3.058.2 ± 2.473.1 ± 2.673.5 ± 2.673.3 ± 2.8
knn20.3 ± 1.318.9 ± 1.468.2 ± 4.150.5 ± 2.870.3 ± 3.970.7 ± 3.969.5 ± 3.8
logreg15.8 ± 1.614.8 ± 1.276.5 ± 4.251.5 ± 3.770.0 ± 5.370.6 ± 5.270.1 ± 5.2
svm13.6 ± 1.114.9 ± 1.175.0 ± 5.348.3 ± 3.369.7 ± 4.870.3 ± 4.869.4 ± 4.7
REALWORLDchestgbt5.5 ± 0.53.1 ± 0.376.5 ± 3.643.4 ± 1.668.4 ± 3.470.0 ± 3.268.7 ± 3.2
knn11.9 ± 1.17.1 ± 0.867.5 ± 2.945.3 ± 2.167.9 ± 3.569.0 ± 3.368.4 ± 3.4
logreg13.2 ± 1.43.7 ± 0.466.3 ± 5.637.2 ± 2.262.0 ± 3.963.8 ± 3.962.1 ± 3.9
svm6.4 ± 0.63.6 ± 0.467.8 ± 4.936.9 ± 2.059.4 ± 4.461.5 ± 4.259.8 ± 4.4
SAFESENSchestgbt5.3 ± 0.72.7 ± 0.870.8 ± 2.629.0 ± 1.752.5 ± 3.951.3 ± 4.855.1 ± 3.6
knn17.0 ± 1.811.0 ± 1.461.0 ± 4.033.3 ± 2.259.7 ± 3.959.6 ± 4.057.1 ± 3.1
logreg19.1 ± 1.66.3 ± 1.067.7 ± 3.029.9 ± 2.158.1 ± 2.857.6 ± 2.956.3 ± 3.1
svm10.8 ± 1.24.3 ± 0.870.5 ± 3.031.8 ± 2.256.8 ± 2.957.7 ± 3.054.0 ± 3.3
SIMFALLchestgbt40.0 ± 1.131.9 ± 1.246.9 ± 1.522.4 ± 0.634.7 ± 1.534.9 ± 1.533.9 ± 1.5
knn51.4 ± 0.947.2 ± 1.234.6 ± 0.723.7 ± 0.636.0 ± 1.036.1 ± 1.134.9 ± 0.9
logreg57.7 ± 0.844.6 ± 1.138.2 ± 1.117.7 ± 0.531.0 ± 1.031.1 ± 1.030.0 ± 1.0
svm46.6 ± 0.747.2 ± 1.541.8 ± 1.317.4 ± 0.427.8 ± 1.128.3 ± 1.026.4 ± 1.1
SIMFALLwristgbt41.7 ± 1.434.9 ± 1.444.2 ± 2.122.1 ± 0.935.7 ± 1.935.9 ± 1.934.7 ± 1.9
knn51.8 ± 1.147.9 ± 1.133.5 ± 1.423.5 ± 0.935.5 ± 1.535.7 ± 1.534.5 ± 1.6
logreg58.7 ± 1.347.1 ± 1.236.5 ± 1.919.6 ± 0.830.8 ± 1.531.3 ± 1.629.9 ± 1.6
svm48.5 ± 1.250.6 ± 1.439.8 ± 2.217.3 ± 0.629.1 ± 1.229.5 ± 1.328.2 ± 1.4
UTSMOKEwristgbt16.4 ± 1.37.9 ± 0.872.3 ± 2.659.2 ± 1.669.1 ± 2.869.1 ± 2.869.0 ± 2.7
knn20.3 ± 1.116.1 ± 1.066.6 ± 2.155.6 ± 1.564.0 ± 2.564.1 ± 2.563.7 ± 2.4
logreg26.7 ± 1.813.6 ± 1.067.2 ± 2.254.7 ± 1.462.7 ± 2.262.8 ± 2.263.0 ± 2.2
svm14.0 ± 1.19.3 ± 0.873.0 ± 2.357.1 ± 1.667.0 ± 2.567.1 ± 2.667.2 ± 2.4
Table A3. Weighted subject-dependent and subject-independent F-score (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Table A3. Weighted subject-dependent and subject-independent F-score (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Subject-DependentSubject-Independent
DatasetSensormlaPIM(E)PSMPIMPSMEPSMWEPSMWEPSM bf
FUSIONwristgbt98.2 ± 0.398.0 ± 0.392.7 ± 2.382.3 ± 2.190.6 ± 2.890.8 ± 2.890.4 ± 2.8
knn94.9 ± 0.895.0 ± 0.987.7 ± 1.877.4 ± 2.588.2 ± 2.988.4 ± 2.988.2 ± 2.9
logreg97.2 ± 0.597.8 ± 0.392.5 ± 2.280.7 ± 2.589.8 ± 2.989.9 ± 2.989.4 ± 2.9
svm98.2 ± 0.398.1 ± 0.391.6 ± 2.181.2 ± 2.390.4 ± 2.990.5 ± 3.090.1 ± 2.9
MHEALTHwristgbt97.7 ± 0.797.5 ± 1.181.9 ± 4.058.3 ± 2.269.8 ± 3.670.1 ± 3.569.2 ± 3.6
knn93.5 ± 1.294.1 ± 1.376.6 ± 2.856.3 ± 1.970.9 ± 2.771.2 ± 2.772.6 ± 2.6
logreg93.7 ± 1.396.1 ± 1.378.9 ± 3.353.0 ± 2.167.6 ± 3.267.8 ± 3.268.2 ± 3.2
svm95.9 ± 0.997.0 ± 0.981.9 ± 2.756.6 ± 2.070.0 ± 4.170.2 ± 4.169.3 ± 4.1
OPPORTwristgbt87.8 ± 1.989.1 ± 1.579.3 ± 5.071.4 ± 4.277.3 ± 5.777.0 ± 5.877.4 ± 5.5
knn80.7 ± 1.883.2 ± 1.867.2 ± 2.961.1 ± 2.369.2 ± 3.668.7 ± 3.469.5 ± 3.0
logreg81.3 ± 2.484.5 ± 1.972.7 ± 4.963.0 ± 2.670.3 ± 4.770.2 ± 4.970.8 ± 4.5
svm87.0 ± 1.887.4 ± 1.676.6 ± 4.864.5 ± 2.273.9 ± 4.773.6 ± 5.074.4 ± 4.4
PAMAP2chestgbt89.0 ± 0.489.1 ± 0.679.2 ± 4.656.2 ± 3.374.3 ± 4.574.8 ± 4.673.9 ± 4.2
knn78.2 ± 1.080.6 ± 1.167.7 ± 2.552.4 ± 2.170.9 ± 3.471.3 ± 3.569.5 ± 3.2
logreg84.3 ± 1.187.0 ± 0.874.1 ± 4.150.4 ± 2.771.4 ± 5.171.4 ± 5.270.4 ± 5.1
svm87.6 ± 0.786.6 ± 0.775.9 ± 4.651.5 ± 3.071.2 ± 5.671.7 ± 5.570.3 ± 5.5
PAMAP2wristgbt88.3 ± 1.087.6 ± 0.879.9 ± 3.058.2 ± 2.473.1 ± 2.673.5 ± 2.673.3 ± 2.8
knn79.7 ± 1.381.0 ± 1.468.2 ± 4.150.5 ± 2.870.3 ± 3.970.7 ± 3.969.5 ± 3.8
logreg84.2 ± 1.685.3 ± 1.276.5 ± 4.251.5 ± 3.770.0 ± 5.370.6 ± 5.270.1 ± 5.2
svm86.6 ± 1.185.1 ± 1.175.0 ± 5.348.3 ± 3.369.7 ± 4.870.3 ± 4.869.4 ± 4.7
REALWORLDchestgbt94.7 ± 0.596.8 ± 0.376.5 ± 3.643.4 ± 1.668.4 ± 3.470.0 ± 3.268.7 ± 3.2
knn88.8 ± 1.092.9 ± 0.867.5 ± 2.945.3 ± 2.167.9 ± 3.569.0 ± 3.368.4 ± 3.4
logreg87.5 ± 1.496.2 ± 0.466.3 ± 5.637.2 ± 2.262.0 ± 3.963.8 ± 3.962.1 ± 3.9
svm93.9 ± 0.596.3 ± 0.467.8 ± 4.936.9 ± 2.059.4 ± 4.461.5 ± 4.259.8 ± 4.4
SAFESENSchestgbt95.1 ± 0.797.3 ± 0.870.8 ± 2.629.0 ± 1.752.5 ± 3.951.3 ± 4.855.1 ± 3.6
knn83.5 ± 1.988.9 ± 1.461.0 ± 4.033.3 ± 2.259.7 ± 3.959.6 ± 4.057.1 ± 3.1
logreg81.4 ± 1.893.6 ± 1.067.7 ± 3.029.9 ± 2.158.1 ± 2.857.6 ± 2.956.3 ± 3.1
svm89.6 ± 1.295.6 ± 0.870.5 ± 3.031.8 ± 2.256.8 ± 2.957.7 ± 3.054.0 ± 3.3
SIMFALLchestgbt59.8 ± 1.168.3 ± 1.246.9 ± 1.522.4 ± 0.634.7 ± 1.534.9 ± 1.533.9 ± 1.5
knn48.7 ± 0.952.9 ± 1.234.6 ± 0.723.7 ± 0.636.0 ± 1.036.1 ± 1.134.9 ± 0.9
logreg41.9 ± 0.955.3 ± 1.238.2 ± 1.117.7 ± 0.531.0 ± 1.031.1 ± 1.030.0 ± 1.0
svm53.6 ± 0.752.7 ± 1.541.8 ± 1.317.4 ± 0.427.8 ± 1.128.3 ± 1.026.4 ± 1.1
SIMFALLwristgbt58.2 ± 1.465.2 ± 1.444.2 ± 2.122.1 ± 0.935.7 ± 1.935.9 ± 1.934.7 ± 1.9
knn48.2 ± 1.152.2 ± 1.133.5 ± 1.423.5 ± 0.935.5 ± 1.535.7 ± 1.534.5 ± 1.6
logreg41.0 ± 1.352.8 ± 1.236.5 ± 1.919.6 ± 0.830.8 ± 1.531.3 ± 1.629.9 ± 1.6
svm51.8 ± 1.249.4 ± 1.439.8 ± 2.217.3 ± 0.629.1 ± 1.229.5 ± 1.328.2 ± 1.4
UTSMOKEwristgbt83.3 ± 1.392.2 ± 0.872.3 ± 2.659.2 ± 1.669.1 ± 2.869.1 ± 2.869.0 ± 2.7
knn79.5 ± 1.283.8 ± 1.066.6 ± 2.155.6 ± 1.564.0 ± 2.564.1 ± 2.563.7 ± 2.4
logreg72.4 ± 1.986.2 ± 1.167.2 ± 2.254.7 ± 1.462.7 ± 2.262.8 ± 2.263.0 ± 2.2
svm85.9 ± 1.190.7 ± 0.873.0 ± 2.357.1 ± 1.667.0 ± 2.567.1 ± 2.667.2 ± 2.4

References

  1. Scheurer, S.; Tedesco, S.; Brown, K.N.; O’Flynn, B. Subject-dependent and-independent human activity recognition with person-specific and-independent models. In Proceedings of the 6th international Workshop on Sensor-based Activity Recognition and Interaction, Rostock, Germany, 16–17 September 2019; pp. 1–7. [Google Scholar]
  2. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  3. Chavarriaga, R.; Sagha, H.; Calatroni, A.; Digumarti, S.T.; Tröster, G.; Millán, J.d.R.; Roggen, D. The Opportunity challenge: A benchmark database for on-body sensor-based activity recognition. Pattern Recognit. Lett. 2013, 34. [Google Scholar] [CrossRef] [Green Version]
  4. Jiang, W.; Yin, Z. Human Activity Recognition Using Wearable Sensors by Deep Convolutional Neural Networks. In Proceedings of the 23rd ACM International Conference on Multimedia, New York, NY, USA, October 2015; pp. 1307–1310. [Google Scholar] [CrossRef]
  5. Zhang, L.; Wu, X.; Luo, D. Recognizing Human Activities from Raw Accelerometer Data Using Deep Neural Networks. In Proceedings of the 2015 IEEE 14th International Conference on Machine Learning and Applications (ICMLA), Miami, FL, USA, 9–11 December 2015; pp. 865–870. [Google Scholar] [CrossRef]
  6. Ordóñez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Hammerla, N.Y.; Halloran, S.; Plötz, T. Deep, Convolutional, and Recurrent Models for Human Activity Recognition Using Wearables. arXiv 2016, arXiv:1604.08880. [Google Scholar]
  8. Shoaib, M.; Bosch, S.; Incel, O.D.; Scholten, H.; Havinga, P.J.M. Fusion of Smartphone Motion Sensors for Physical Activity Recognition. Sensors 2014, 14. [Google Scholar] [CrossRef] [PubMed]
  9. Reiss, A.; Stricker, D. Introducing a New Benchmarked Dataset for Activity Monitoring. In Proceedings of the 2012 16th International Symposium on Wearable Computers, Newcastle, UK, 18–22 June 2012. [Google Scholar] [CrossRef]
  10. Jordao, A.; Nazare, A.C., Jr.; Sena, J.; Schwartz, W.R. Human Activity Recognition Based on Wearable Sensor Data: A Standardization of the State-of-the-Art. arXiv 2019, arXiv:1806.05226. [Google Scholar]
  11. Banos, O.; Garcia, R.; Holgado-Terriza, J.A.; Damas, M.; Pomares, H.; Rojas, I.; Saez, A.; Villalonga, C. mHealthDroid: A Novel Framework for Agile Development of Mobile Health Applications. In International Workshop on Ambient Assisted Living; Pecchia, L., Chen, L.L., Nugent, C., Bravo, J., Eds.; Springer: Cham, Switzerland, 2014. [Google Scholar] [CrossRef]
  12. Catal, C.; Tufekci, S.; Pirmit, E.; Kocabag, G. On the Use of Ensemble of Classifiers for Accelerometer-Based Activity Recognition. Appl. Soft Comput. 2015, 37, 1018–1022. [Google Scholar] [CrossRef]
  13. Chen, Y.; Xue, Y. A Deep Learning Approach to Human Activity Recognition Based on Single Accelerometer. In Proceedings of the 2015 IEEE International Conference on Systems, Kowloon, China, 9–12 October 2015; pp. 1488–1492. [Google Scholar] [CrossRef]
  14. Abdu-Aguye, M.G.; Gomaa, W. Competitive Feature Extraction for Activity Recognition based on Wavelet Transforms and Adaptive Pooling. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar] [CrossRef]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Sztyler, T.; Stuckenschmidt, H. On-body localization of wearable devices: An investigation of position-aware activity recognition. In Proceedings of the 2016 IEEE International Conference on Pervasive Computing and Communications (PerCom), Sydney, Australia, 14–19 March 2016. [Google Scholar] [CrossRef]
  17. Vakili, M.; Ghamsari, M.; Rezaei, M. Performance Analysis and Comparison of Machine and Deep Learning Algorithms for IoT Data Classification. arXiv 2020, arXiv:2001.09636. [Google Scholar]
  18. Özdemir, A.T.; Barshan, B. Detecting Falls with Wearable Sensors Using Machine Learning Techniques. Sensors 2014, 14. [Google Scholar] [CrossRef] [PubMed]
  19. Alharbi, F.; Farrahi, K. A Convolutional Neural Network for Smoking Activity Recognition. In Proceedings of the 2018 IEEE 20th International Conference on e-Health Networking, Applications and Services (Healthcom), Ostrava, Czech Republic, 17–20 September 2018; pp. 1–6. [Google Scholar] [CrossRef]
  20. Shoaib, M.; Scholten, H.; Havinga, P.J.M.; Incel, O.D. A hierarchical lazy smoking detection algorithm using smartwatch sensors. In Proceedings of the 2016 IEEE 18th International Conference on e-Health Networking, Applications and Services (Healthcom), Munich, Germany, 14–16 September 2016. [Google Scholar] [CrossRef]
  21. Bao, L.; Intille, S.S. Activity Recognition from User-Annotated Acceleration Data. In International Conference on Pervasive Computing; Springer: Berlin/Heidelberg, Germany, 2004; pp. 1–17. [Google Scholar] [CrossRef]
  22. Weiss, G.M.; Lockhart, J.W. The Impact of Personalization on Smartphone-Based Activity Recognition. In Workshops at the Twenty-Sixth AAAI Conference on Artificial Intelligence; AAAI: Palo Alto, CA, USA, 2012. [Google Scholar]
  23. Bulling, A.; Blanke, U.; Schiele, B. A Tutorial on Human Activity Recognition Using Body-worn Inertial. Sensors 2014, 46. [Google Scholar] [CrossRef]
  24. Scheurer, S.; Tedesco, S.; Brown, K.N.; O’Flynn, B. Human Activity Recognition for Emergency First Responders via Body-Worn Inertial Sensors. In Proceedings of the 2017 IEEE 14th International Conference on Wearable and Implantable Body Sensor Networks (BSN), Eindhoven, The Netherlands, 9–12 May 2017; pp. 5–8. [Google Scholar]
  25. Jones, E.; Oliphant, T.; Peterson, P. SciPy: Open Source Scientific Tools for Python. Available online: https://www.scipy.org (accessed on 27 May 2020).
  26. Van der Walt, S.; Colbert, S.C.; Varoquaux, G. The NumPy Array: A Structure for Efficient Numerical Computation. Comput. Sci. Eng. 2011, 13, 22–30. [Google Scholar] [CrossRef] [Green Version]
  27. McKinney, W. Data Structures for Statistical Computing in Python. In Proceedings of the 9th Python in Science Conference, Austin, TX, USA, June 28–July 2010; pp. 51–56. [Google Scholar]
  28. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  29. Tange, O. GNU Parallel—The Command-Line Power Tool. Login Usenix Mag. 2011, 36, 42–47. [Google Scholar]
  30. R Core Team. R: A Language and Environment for Statistical Computing. Available online: https://www.r-project.org (accessed on 20 May 2020).
  31. Bates, D.; Mächler, M.; Bolker, B.; Walker, S. Fitting Linear Mixed-Effects Models Using lme4. J. Stat. Softw. 2015, 67. [Google Scholar] [CrossRef]
  32. Lenth, R. emmeans: Estimated Marginal Means, aka Least-Squares Means. 2019. Available online: https://cran.r-project.org/package=emmeans (accessed on 10 May 2020).
  33. Karantonis, D.M.; Narayanan, M.R.; Mathie, M.; Lovell, N.H.; Celler, B.G. Implementation of a real-time human movement classifier using a triaxial accelerometer for ambulatory monitoring. IEEE Trans. Inf. Technol. Biomed. 2006, 10, 156–167. [Google Scholar] [CrossRef] [PubMed]
  34. Gelman, A.; Hill, J. Data Analysis Using Regression and Multilevel/Hierarchical Models; Cambridge University Press: Cambridge, MA, USA, 2006. [Google Scholar]
Figure 1. Number of instances per activity for each dataset.
Figure 1. Number of instances per activity for each dataset.
Sensors 20 03647 g001
Figure 2. Graphical summary of the experiment. Each dataset is pre-processed once, and then used to obtain the subject-dependent and subject-independent predictions with each learning algorithm and personalisation–generalisation approach (person-independent models (PIMs), person-specific models (PSMs), etc.). Note that an ensemble of PSMs is identical to a PSM in the case of subject-dependent predictions (i.e., for known users).
Figure 2. Graphical summary of the experiment. Each dataset is pre-processed once, and then used to obtain the subject-dependent and subject-independent predictions with each learning algorithm and personalisation–generalisation approach (person-independent models (PIMs), person-specific models (PSMs), etc.). Note that an ensemble of PSMs is identical to a PSM in the case of subject-dependent predictions (i.e., for known users).
Sensors 20 03647 g002
Figure 3. Subject-independent (vertical axis) versus subject-dependent (horizontal axis) κ (%) across all learning algorithms. Axes have been scaled to encompass the data for improved visibility. Note how the subject-dependent performance of (E)PSMs tends to be better (further to the right) than that of PIMs, and there is a clear difference in subject-independent performance between PIMs and PSMs.
Figure 3. Subject-independent (vertical axis) versus subject-dependent (horizontal axis) κ (%) across all learning algorithms. Axes have been scaled to encompass the data for improved visibility. Note how the subject-dependent performance of (E)PSMs tends to be better (further to the right) than that of PIMs, and there is a clear difference in subject-independent performance between PIMs and PSMs.
Sensors 20 03647 g003
Figure 4. Distribution of the user-wise subject-dependent ranks of each learning algorithm + personalisation–generalisation approach. A person-specific gradient boosted ensemble of decision trees (GBT + PSM) outperforms the other methods for over 70% of users.
Figure 4. Distribution of the user-wise subject-dependent ranks of each learning algorithm + personalisation–generalisation approach. A person-specific gradient boosted ensemble of decision trees (GBT + PSM) outperforms the other methods for over 70% of users.
Sensors 20 03647 g004
Figure 5. Estimated marginal means for comparing the subject-independent performance of personalisation–generalisation approaches across learning algorithms. Person-independent models (PIM) clearly outperform the other personalisation–generalisation approaches for any learning algorithm, and ensembles of person-specific models (particularly WEPSMs) clearly outperform its constituent person-specific models (PSM).
Figure 5. Estimated marginal means for comparing the subject-independent performance of personalisation–generalisation approaches across learning algorithms. Person-independent models (PIM) clearly outperform the other personalisation–generalisation approaches for any learning algorithm, and ensembles of person-specific models (particularly WEPSMs) clearly outperform its constituent person-specific models (PSM).
Sensors 20 03647 g005
Figure 6. Distribution of the user-wise subject-independent ranks of each personalisation–generalisation approach + learning algorithm. PIM + GBT outperforms the other methods for over 60% of users, and performs second-best for over 20% of users.
Figure 6. Distribution of the user-wise subject-independent ranks of each personalisation–generalisation approach + learning algorithm. PIM + GBT outperforms the other methods for over 60% of users, and performs second-best for over 20% of users.
Sensors 20 03647 g006
Figure 7. Estimated marginal means for the (average) difference between subject-dependent (SD) and subject-independent (SI) performance across personalisation–generalisation approaches. Subject-dependent performance is clearly better than subject-independent performance, a discrepancy that is minimised with person-independent models (PIM). The confidence intervals for the subject-independent performances of PIMs and ensembles of PSMs, particularly WEPSMs, overlap.
Figure 7. Estimated marginal means for the (average) difference between subject-dependent (SD) and subject-independent (SI) performance across personalisation–generalisation approaches. Subject-dependent performance is clearly better than subject-independent performance, a discrepancy that is minimised with person-independent models (PIM). The confidence intervals for the subject-independent performances of PIMs and ensembles of PSMs, particularly WEPSMs, overlap.
Sensors 20 03647 g007
Table 1. Number of (act)ivities and (ind)ividuals, trials/activity (±standard error), and sampling frequency (Hz) for each of the datasets.
Table 1. Number of (act)ivities and (ind)ividuals, trials/activity (±standard error), and sampling frequency (Hz) for each of the datasets.
DatasetActIndTrials/ActHz
[8]FUSION71090 ± 050
[11]MHEALTH111038 ± 050
[3]OPPORT44590 ± 25830
[9]PAMAP212981 ± 8100
[16]REALWORLD815318 ± 4250
[24]SAFESENS171191 ± 1333
[18]SIMFALL1617128 ± 825
[20]UTSMOKE711859 ± 750
Table 2. Subject-dependent and subject-independent κ (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Table 2. Subject-dependent and subject-independent κ (%) ± standard error when learning algorithms (mla) are combined with a person-independent model (PIM), a person-specific model (PSM), an unweighted ensemble of PSMs (EPSM), a κ -weighted EPSM (WEPSM), or a baseline-feature-weighted EPSM (WEPSM bf ).
Subject-DependentSubject-Independent
DatasetSensormlaPIM(E)PSMPIMPSMEPSMWEPSMWEPSM bf
FUSIONwristgbt97.9 ± 0.397.6 ± 0.492.4 ± 2.281.4 ± 2.190.6 ± 2.690.7 ± 2.590.2 ± 2.6
knn94.0 ± 0.994.2 ± 1.085.9 ± 2.074.9 ± 2.687.5 ± 2.887.5 ± 2.887.3 ± 2.8
logreg96.7 ± 0.697.4 ± 0.491.9 ± 2.179.4 ± 2.789.5 ± 2.889.5 ± 2.889.0 ± 2.7
svm98.0 ± 0.397.8 ± 0.490.9 ± 2.180.0 ± 2.390.3 ± 2.790.4 ± 2.790.0 ± 2.6
MHEALTHwristgbt97.5 ± 0.897.2 ± 1.282.4 ± 3.659.5 ± 2.372.2 ± 3.572.4 ± 3.471.5 ± 3.5
knn92.9 ± 1.393.7 ± 1.476.1 ± 3.156.0 ± 2.171.4 ± 2.671.6 ± 2.672.8 ± 2.5
logreg93.2 ± 1.495.8 ± 1.478.9 ± 3.354.1 ± 2.470.0 ± 2.970.2 ± 2.970.8 ± 2.9
svm95.5 ± 1.096.8 ± 1.082.0 ± 2.658.0 ± 2.172.0 ± 3.972.1 ± 3.971.4 ± 4.0
OPPORTwristgbt81.5 ± 2.883.5 ± 2.469.0 ± 7.257.9 ± 5.866.5 ± 8.166.2 ± 8.466.7 ± 7.9
knn71.1 ± 2.774.8 ± 2.751.4 ± 4.242.9 ± 3.354.5 ± 5.453.6 ± 5.154.8 ± 4.6
logreg71.9 ± 3.776.7 ± 3.059.9 ± 7.046.2 ± 3.456.7 ± 6.756.4 ± 7.057.2 ± 6.4
svm80.3 ± 2.681.0 ± 2.465.4 ± 6.748.4 ± 2.861.7 ± 6.661.3 ± 7.062.3 ± 6.2
PAMAP2chestgbt87.5 ± 0.587.7 ± 0.677.4 ± 4.354.7 ± 2.972.4 ± 4.172.9 ± 4.272.0 ± 3.8
knn75.5 ± 1.078.4 ± 1.263.7 ± 2.549.0 ± 1.867.7 ± 3.268.2 ± 3.466.3 ± 3.1
logreg82.5 ± 1.085.4 ± 0.972.2 ± 3.848.8 ± 2.469.4 ± 4.969.4 ± 4.968.6 ± 4.8
svm86.0 ± 0.785.1 ± 0.873.7 ± 4.549.7 ± 2.669.4 ± 5.170.0 ± 5.168.5 ± 5.1
PAMAP2wristgbt86.8 ± 1.186.0 ± 0.978.5 ± 2.856.8 ± 2.371.7 ± 2.772.2 ± 2.672.0 ± 2.8
knn77.4 ± 1.578.9 ± 1.665.2 ± 4.147.5 ± 2.768.1 ± 3.968.5 ± 4.067.6 ± 3.8
logreg82.4 ± 1.783.5 ± 1.374.7 ± 4.149.9 ± 3.568.8 ± 4.969.3 ± 4.969.0 ± 4.8
svm84.9 ± 1.383.3 ± 1.373.1 ± 5.146.6 ± 3.068.3 ± 4.568.8 ± 4.568.2 ± 4.3
REALWORLDchestgbt93.3 ± 0.696.1 ± 0.471.7 ± 4.437.5 ± 1.962.7 ± 4.064.1 ± 3.863.2 ± 3.9
knn85.3 ± 1.591.3 ± 1.059.3 ± 3.437.9 ± 2.461.8 ± 3.762.8 ± 3.662.0 ± 3.8
logreg83.8 ± 1.895.4 ± 0.560.6 ± 5.830.3 ± 2.157.1 ± 4.258.7 ± 4.257.1 ± 4.3
svm92.0 ± 0.795.5 ± 0.462.2 ± 5.130.4 ± 2.154.8 ± 4.556.5 ± 4.355.0 ± 4.7
SAFESENSchestgbt93.9 ± 0.997.0 ± 0.867.6 ± 3.327.9 ± 2.048.9 ± 4.848.7 ± 5.453.9 ± 3.3
knn81.3 ± 1.987.8 ± 1.555.7 ± 3.530.2 ± 1.754.7 ± 3.254.6 ± 3.354.2 ± 2.7
logreg78.7 ± 1.693.1 ± 1.064.1 ± 3.027.4 ± 1.854.0 ± 2.753.3 ± 2.754.1 ± 2.8
svm88.1 ± 1.295.2 ± 0.866.9 ± 2.729.9 ± 1.851.9 ± 2.653.0 ± 2.451.9 ± 2.9
SIMFALLchestgbt57.2 ± 1.265.9 ± 1.343.9 ± 1.619.3 ± 0.733.5 ± 1.633.5 ± 1.632.6 ± 1.5
knn45.0 ± 0.949.5 ± 1.230.3 ± 0.719.8 ± 0.633.3 ± 1.033.2 ± 1.132.2 ± 0.9
logreg38.3 ± 0.952.3 ± 1.234.5 ± 1.114.5 ± 0.529.1 ± 1.029.1 ± 1.028.2 ± 1.0
svm50.1 ± 0.749.5 ± 1.638.2 ± 1.414.1 ± 0.425.9 ± 1.026.2 ± 0.924.8 ± 1.0
SIMFALLwristgbt55.4 ± 1.562.7 ± 1.540.8 ± 2.319.1 ± 1.032.9 ± 2.233.0 ± 2.232.0 ± 2.2
knn44.6 ± 1.248.8 ± 1.229.2 ± 1.519.2 ± 1.031.8 ± 1.631.9 ± 1.631.0 ± 1.6
logreg37.3 ± 1.449.6 ± 1.332.7 ± 2.116.2 ± 0.927.9 ± 1.728.4 ± 1.727.6 ± 1.8
svm48.2 ± 1.345.9 ± 1.536.0 ± 2.313.7 ± 0.727.1 ± 1.527.3 ± 1.626.5 ± 1.7
UTSMOKEwristgbt80.9 ± 1.590.8 ± 0.968.7 ± 2.954.8 ± 1.865.4 ± 3.265.4 ± 3.365.3 ± 3.1
knn76.3 ± 1.381.2 ± 1.261.6 ± 2.450.8 ± 1.760.7 ± 2.860.8 ± 2.960.5 ± 2.7
logreg68.9 ± 2.184.1 ± 1.263.2 ± 2.550.5 ± 1.659.4 ± 2.559.4 ± 2.559.6 ± 2.4
svm83.6 ± 1.389.1 ± 0.969.2 ± 2.752.9 ± 1.863.6 ± 2.963.8 ± 3.063.8 ± 2.7
Table 3. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects on subject-dependent performance associated with learning algorithms and personalisation–generalisation approaches.
Table 3. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects on subject-dependent performance associated with learning algorithms and personalisation–generalisation approaches.
Coefficient2.5%β97.5%p
(Intercept)1.4252.1292.8333.1 × 10−9
kNN−0.628−0.616−0.604<2.0 × 10−16
logreg−0.620−0.608−0.596<2.0 × 10−16
SVM−0.229−0.217−0.204<2.0 × 10−16
PSM0.1550.3610.5665.78 × 10−4
Table 4. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects on subject-independent performance associated with learning algorithms and personalisation–generalisation approaches.
Table 4. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects on subject-independent performance associated with learning algorithms and personalisation–generalisation approaches.
Coefficient2.5%β97.5%p
(Intercept)0.3670.9071.4481.00 × 10−3
kNN−0.231−0.225−0.219<2.0 × 10−16
logreg−0.251−0.245−0.239<2.0 × 10−16
SVM−0.178−0.172−0.166<2.0 × 10−16
PSM−0.825−0.818−0.812<2.0 × 10−16
EPSM−0.199−0.193−0.186<2.0 × 10−16
WEPSM−0.186−0.179−0.172<2.0 × 10−16
WEPSM bf −0.210−0.203−0.196<2.0 × 10−16
Table 5. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects associated with subject-independent (SI) performance and personalisation–generalisation approaches.
Table 5. GLMM (Generalised Linear Mixed Effects Model) estimates ( β ), 95% confidence intervals, and p-values of the fixed effects associated with subject-independent (SI) performance and personalisation–generalisation approaches.
Coefficient2.5%β97.5%p
(Intercept)0.9301.5302.1305.83 × 10−7
PSM/(E)PSM ( bf ) 0.3710.3790.387<2.0 × 10−16
SI−0.739−0.732−0.725<2.0 × 10−16
SI + EPSM−0.586−0.575−0.564<2.0 × 10−16
SI + PSM−1.272−1.261−1.250<2.0 × 10−16
SI + WEPSM−0.572−0.561−0.551<2.0 × 10−16
SI + WEPSM bf −0.595−0.584−0.574<2.0 × 10−16

Share and Cite

MDPI and ACS Style

Scheurer, S.; Tedesco, S.; O’Flynn, B.; Brown, K.N. Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance. Sensors 2020, 20, 3647. https://doi.org/10.3390/s20133647

AMA Style

Scheurer S, Tedesco S, O’Flynn B, Brown KN. Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance. Sensors. 2020; 20(13):3647. https://doi.org/10.3390/s20133647

Chicago/Turabian Style

Scheurer, Sebastian, Salvatore Tedesco, Brendan O’Flynn, and Kenneth N. Brown. 2020. "Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance" Sensors 20, no. 13: 3647. https://doi.org/10.3390/s20133647

APA Style

Scheurer, S., Tedesco, S., O’Flynn, B., & Brown, K. N. (2020). Comparing Person-Specific and Independent Models on Subject-Dependent and Independent Human Activity Recognition Performance. Sensors, 20(13), 3647. https://doi.org/10.3390/s20133647

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop