Next Article in Journal
Development of Measurement Method for Temperature and Velocity Field with Optical Fiber Sensor
Next Article in Special Issue
Visualizing the Landscape of Home IoT Research: A Bibliometric Analysis Using VOSviewer
Previous Article in Journal
Algorithm for Mobile Platform-Based Real-Time QRS Detection
Previous Article in Special Issue
Decentralized Policy Coordination in Mobile Sensing with Consensual Communication
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Graph-Attention-Based Method for Single-Resident Daily Activity Recognition in Smart Homes

1
Shien-Ming Wu School of Intelligent Engineering, South China University of Technology, Guangzhou 511442, China
2
Department of Rehabilitation Sciences, The Hong Kong Polytechnic University, Hong Kong
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(3), 1626; https://doi.org/10.3390/s23031626
Submission received: 24 December 2022 / Revised: 22 January 2023 / Accepted: 25 January 2023 / Published: 2 February 2023
(This article belongs to the Special Issue Internet of Things for Smart Homes Ⅲ)

Abstract

:
In ambient-assisted living facilitated by smart home systems, the recognition of daily human activities is of great importance. It aims to infer the household’s daily activities from the triggered sensor observation sequences with varying time intervals among successive readouts. This paper introduces a novel deep learning framework based on embedding technology and graph attention networks, namely the time-oriented and location-oriented graph attention (TLGAT) networks. The embedding technology converts sensor observations into corresponding feature vectors. Afterward, TLGAT provides a sensor observation sequence as a fully connected graph to the model’s temporal correlation as well as the sensor’s location correlation among sensor observations and facilitates the feature representation of each sensor observation through receiving other sensor observations and weighting operations. The experiments were conducted on two public datasets, based on the diverse setups of sensor event sequence length. The experimental results revealed that the proposed method achieved favorable performance under diverse setups.

1. Introduction

With the development of Internet of Things (IoT) technology and the progress of deep learning algorithms, there is currently widespread interest in smart home systems using sensors for improving the daily lives of humans. For example, detecting dementia-related abnormal behaviours [1,2], analysing context [3], controlling indoor equipment automatically [4], and helping the disabled or patients to live independently better [5,6]. Human activity recognition (HAR) has been one of the research hotspots since it plays a significant role in ambient-assisted living (AAL). HAR is the process of utilizing sensors to understand human motion processes or behavioural logic, so as to infer human daily activities. If the occupant’s daily activities can be identified with smart home systems, it can satisfy the demands for smart home systems to provide rational and automated services to residents.
In recent years, several studies have focused on deep neural networks (DNNs), convolutional neural networks (CNNs), and recurrent neural networks (RNNs) to investigate the solutions for HAR [7,8,9]. For the purpose of performing the primary task of HAR, deep learning algorithms can be employed to build a model, either in a supervised or unsupervised learning way. They are among data-driven methods, taking advantage of the massive volume of data collected using the sensors available in smart home systems. With the development of IoT-based smart home platforms, this study focuses more on HAR based on environmental sensor data instead of video or wearable sensor data. The research description of IoT-based HAR is presented in Section 2.
In general, the existing deep learning methods pay more attention to inferring the relationship between daily activities and sensor observations. Indeed, the type and location of sensors, as well as user habits, will affect the performance of HAR tasks. Several practical user cases collect data using multi-type sensors such as the CASAS dataset [10]. For smart home systems with multi-type sensors that can be activated by residents or at a specific frequency, ambient sensors are irregularly activated, and the time interval between each two successive activated sensors is varied. Irregularity means that the correlation between the observed values of every two continuously activated sensors may be inconsistent. Thus, HAR challenges still remain because of the sparsity and irregularity of the data [11]. Additionally, in smart home systems, HAR is expected to assist with other tasks; therefore, real-time performance is important. However, in the state-of-the-art options available, it is difficult to maintain high accuracy when identifying activities with fewer sensor events, which will cause delays in the smart home system.
In the proposed method, we sought to model the time interval between two sensor activation events and the relationship between their location information, thus automatically extracting features to better understand the context of sensor signals, so as to deal with sparsity and irregularity. Furthermore, we aim to introduce more relevant information to represent sensor events and further improve the feature representation of sensor event sequences using the proposed algorithm. The contributions of this paper are summarised as follows:
  • A novel representation method incorporates sensor location information and time-slice information in smart homes and generates features for each sensor event through an embedding process;
  • We first apply the time-encoding method and vector operation to automatically capture the correlations between the sensor events relying on their timestamps;
  • Two novel parallel modules based on graph attention (GAT), namely a location-oriented GAT module and a time-oriented GAT module, are proposed to capture the correlations between different sensor events based on location information and the activated time information. These modules automatically improve the feature representation of the sensor event sequences;
  • A new end-to-end novel framework, namely the time-oriented and location-oriented graph attention (TLGAT) network, is established to address HAR issues. The experimental results reveal that such a model achieves superior performance on public datasets compared with other state-of-the-art methods.
In the rest of the paper, Section 2 gives a brief overview of the state-of-the-art research based on deep learning, Section 3 explains the preparation process of data from two public datasets and describes the basic principles of several modules that make up our proposed framework. Section 4 presents the experiment setup and shows the evaluation results. Lastly, we summarise the results, highlighting the superiority of the method, and draw brief conclusions in Section 5.

2. Related Work

In this section, we will summarise the related research of deep learning methods for HAR in smart homes and describe graph neural networks in more detail.

2.1. Human Activity Recognition

The task of human activity recognition has been studied for decades. In smart home systems, there are typically two HAR models: wearable-sensor-based HAR and ambient-sensor-based HAR [12]. Wearable-sensor-based HAR models use integrated sensors on mobile devices, clothes, or other portable devices, such as accelerometers, gyroscopes, and magnetometers. Ambient sensors are commonly installed at fixed locations to recognise activities, mainly including infrared sensors (e.g., RFID and PIR) and temperature sensors. In contrast, ambient sensors can detect multi-occupant activities in smart homes in a non-invasive manner for users, but it is harder to achieve accuracy using these sensors due to many influencing factors (e.g., people’s movement, people’s identity, etc.).
Currently, there are several kinds of methods of ambient-sensor-based HAR [13,14,15,16,17], including naive Bayes, hidden Markov models, support vector machines, conditional random fields, etc. Nevertheless, these traditional methods highly rely on handcrafted feature extraction techniques. With the inevitable development of deep learning in many fields, deep-learning-based studies on HAR are conducted to address the problems of feature extraction.

2.2. Deep Learning Algorithms

Deep learning is a promising field of artificial intelligence that is able to learn high-level features from a great volume of raw data to achieve satisfactory results in HAR. In smart home systems, the existing HAR methods focus on the preprocessing strategy of transforming a series of sensor events into specific feature representations, which further model the spatial and temporal dependencies to predict the daily activities of occupants.
Deep-learning-based methods comprise CNN-based and RNN-based methods; the former is useful to capture local dependency, i.e., the importance of nearby observations correlated with sensor events. In addition, its variants can capture either one-dimensional (1D) or two-dimensional (2D) features of event sequences with different receptive field sizes and steps. Singh et al. [18] exploited a 1D-CNN structure of raw data sequences for activity recognition and extracted the local dependencies of activity signals in an invariant feature scale. Bouchabou et al. introduced natural language-processing technology to automatically generate feature representations of raw sensor observations and applied fully 1D-CNN to improve performance [19]. In these works, the way of reducing the receptive field of convolution kernel layers by layer can achieve better outcomes. Although these works focused more on the initial representations of raw sequences, they ignored the spatial and temporal characteristics of sensor event sequences. In addition, some researchers proposed a strategy of converting sensor event sequences into 2D images for corresponding investigations. In a smart home with only binary sensors, Gochoo et al. transformed activity sequences into binary images or RGB images, to apply a 2D-CNN model [7,20]. Later, they used coloured images to represent sensor event sequences containing non-binary sensors, such as temperature sensors [20]. In comparison, Mohmed et al. utilised grayscale images instead of coloured images and selected the AlexNet structure in the feature extraction phase [21,22]. Their experiments showed that the image construction method and 2D-CNN structures cannot achieve better performance. Such constructed images have sparse pixel information, which makes it difficult for 2D-CNNs to extract local features.
RNN is another advantageous deep-learning approach to deal with sequential dependencies. Previously, Arifoglu et al. explored and compared the performance of many types of RNN models in activity recognition for elderly people with dementia [23]. Liciotti et al. summarised several types of long-short term memory variants and evaluated these models in the activity recognition task of a smart home scenario [24], proving that RNN-based approaches have excellent effects in HAR. Furthermore, Bouchabou used the NLP word-encoding method and LSTM to encode sensor observation [25]. They could achieve slight improvement with the help of word-encoding technology, and the dependency of RNNs on capturing dynamic time-varying sequences is still limited.
As introduced above, deep-learning-based methods have shown superiority in HAR. Generally speaking, the above-mentioned methods statically represent the characteristics observed with each sensor, but none of them take into account time-specific and location-specific characteristics of sensor events. In contrast, this study takes the activation moment and spatial location of the sensors into consideration, which automatically refreshes feature representations of sensor observations based on their data.

2.3. Graph Attention Network

The latest development, known as the graph attention network, is increasingly prevalent and has extensive applications in various fields [26], owing to its effectiveness and simplicity in transmitting and aggregating messages from neighbouring nodes in the graph structure. For instance, the data in the recommendation system can be represented as a graph structure to construct the user–item interaction [27]. Social networks can accordingly be denoted as topological graphs, to describe the relationships between people based on different factors (e.g., emails, transactions, etc.) [28,29]. Moreover, a sequence can be treated as a sequence graph, where each graph node can be connected with one or more nodes [30]. Subsequently, on the basis of the prior progress of graph convolutional networks (GCN) [31], GAT is able to produce node representations with an attention mechanism applied in each pair of nodes to distinguish the contribution of each node on the target node. A variant of the algorithm based on the key-value attention mechanism can be viewed as an extension and application of the GAT framework in activity recognition in a smart home setting. In our case, we formulate the links between sensor events as graphs, where nodes refer to sensor events and edges are the intensity of these connections. The details are described in Section 3.3.2. The variant takes advantage of GAT to determine the specific importance of neighbouring nodes for the central node, to improve the representation of each node.

3. Methodology

This section describes the proposed method in detail. The overall framework involves four steps, namely (i) the preprocessing of several open datasets in the CASAS, (ii) data representation, (iii) model architecture, and (iv) training and evaluation.
The flowchart of the proposed method is shown in Figure 1. The implementation process is described in detail in this section.

3.1. Dataset Description

In this study, two publicly available datasets in the CASAS project [10], namely ARUBA and MILAN, were used to evaluate the proposed approach. Each dataset is a file containing more than one million raw sensor events, which have been continuously recorded in the respective smart home testbed for several months. Table 1 summarises the types of sensors, the number of residents, the number of categories involved in daily activities, and the total raw sensor events in these two datasets [32,33].
Figure 2 illustrates samples of the ARUBA’s raw dataset, including annotations. One sensor event comprises a date, time, sensor ID, and sensor status or indication, which implies that a sensor is activated and changes in indication or status. In particular, the activity annotation will indicate the corresponding sensor events at the beginning or end of the activity.
Moreover, the opening datasets in the CASAS project provide the sensor layout in the smart home with image files.

3.2. Data Preprocessing

After acquiring the publicly available datasets, we preprocessed the data for validation experiments, and the preparation process can be divided into several phases.

3.2.1. Activity Segmentation

Based on the labels provided by the dataset, a series of activity samples were extracted, each of which represented a sensor event sequence recorded from the beginning to the end of an activity. For example, a total of 6413 annotated activity segments can be segmented from the ARUBU dataset. However, the activity segments are composed of different numbers of sensor events. Some activity segments last longer than others, so the maximum and minimum number of events are 2341 and 20, respectively.

3.2.2. Data Representation

Index Representation. The sensor ID, sensor status, or indication of a sensor event in the raw dataset are represented as one word. For example, in the ARUBA dataset, the motion sensor ID is represented as a word with four characters from ’M001’ to ’M031’. The complete word list forms a limited-size word list. To make such discrete data usable using a neural network, we transformed each word into an index. Specifically, an N s -length word list, for recording the unique sensor ID, and an N v -length word list for recording the unique sensor status or indication were generated, respectively. The index tokens would then be passed through an embedding layer for the continuous feature vectors learned in an unsupervised manner.
Moreover, the layouts of sensors vary in different smart home settings. As shown in Figure 3, for both datasets, the index representations based on the layout of sensors in different rooms were manually defined. For example, the layout indexes of the sensors located in the same room should be the same. The key idea is to introduce sensor layout information based on the minimal semantic information location information of smart homes ( e.g., kitchen or bathroom), as there is a strong correlation between daily indoor activities and the continuous activation of multiple sensors in a specific space. First, a smart home is divided into several subspaces according to the room layout image file provided by the dataset, and then an index is defined for each subspace. Finally, layout index information is added to each sensor event. This idea is also consistent with the human understanding of a housing layout.
Furthermore, in this study, we considered the fact that temporal features play a certain role in activity recognition because of the regularity of household daily activities. However, it is more accurate to state that there is a relationship between the activities within a certain time slice. In this method, the timeframe of 24 h is divided into 1440 time windows, and an index is assigned based on the time sequence. If the time recorded in the sensor event falls within a time slice, the assigned index is the position order of the time slice, which is defined as the time slice representation, avoiding the isolation of the temporal information contained within the slices. In general, daily household activities start at a certain time of the day and are continuously performed for a period of time, meaning that the characteristics of the activities in two different time windows may be related. Therefore, it is reasonable to use a time-slice representation to segment and analyse the temporal information.
Time Encoding. Herein, a high-dimension feature is generated to represent the time point but not the time slice. For each given fine timestamp t, it can be transformed into a multi-dimensional feature vector p t as follows [34]:
p 2 k t = s i n ( t T 2 k / ϵ )
p 2 k + 1 t = c o s ( t T 2 k / ϵ )
where ϵ N + refers to the dimensionality of the generated time representation, and T is the time scale of the time-feature transformation, that is, the specific frequency. We set ϵ = 32 and T = 1440 (unit: minute) in all the experimental settings.

3.2.3. Sliding Windows

It is known that a smart home system will generate an uncertain length of sensor event sequences whenever a user performs any activity. Therefore, a series of event streams should be split through a fixed-size, event-count-based window. The window size is denoted as N, which represents the event count in a window. It is assumed that the residents perform daily activities to generate a sensor event sequence a i and denote an event as e j . As shown in Figure 4, the sliding-window method can transform an event sequence a i = { e 1 , e 2 , , e l } into a i = { S 1 , S 2 , , S l N + 1 } , where S k = { e k , e k + 1 , , e k + N 1 } .
Thus, using the window per sensor event generates a large number of window samples, which is helpful for obtaining more data. Obviously, setting a different window size N, the complexity and accuracy of the recognition model also have significant differences. Commonly, the length of the window refers to the time latency of the prediction using past sensor events. The smaller the window length is, the lower the model precision is, but the better the real-time performance of the model.

3.3. Model Architecture

This section introduces the proposed model architecture in detail with the purpose of learning an appropriate sensor observation embedding for a given sample S k to predict an activity label.

3.3.1. Embedding Layer

In the first step, the prepared index representations of a sensor event e i i n d = { s i i i n d , o i i n d , l i i n d , t s i i n d } are passed through four embedding layers to a fixed-dimensional feature vector:
s i e m b = E D s i ( s i i n d )
o e m b = E D o ( o i n d )
l e m b = E D l ( l i n d )
t s e m b = E D t s ( t s i n d )
where s i , o , l , t s represent sensor ID, sensor observation, sensor location, and time slice, respectively. The superscript i n d and e m b are applied to distinguish between the indexed representation and the embedded representation.

3.3.2. Graph Attention Layer

This section introduces the core graph attention in detail, which was used in the proposed model. Overall, the embedding process can be learned using a hierarchical architecture composed of four levels to model the irregular intervals and latent spaces of the sensor events in the samples.
Generally, a graph attention layer is adopted to model the relationships between nodes in an arbitrary graph [35]. Therefore, for a given graph and one of its nodes u, it is proposed that a time-oriented graph attention layer and a location-oriented graph attention layer can borrow information from node u’s neighbour nodes.
Firstly, a directed weighted graph G i = { V , E i } is built. For every sample, S k , node V represents a sensor event, and an edge E i describes the dependencies between the sensor events in sample S k . It can be determined that there is a bidirectional asymmetric information exchange between all sensor event nodes, so the graph is a kind of complete graph. The aforementioned two proposed graph attention layers use the same way of building the weighted graph.
Secondly, let u indicate a sensor event, whose original observation is X = { s i e m b , o e m b } and location information l u e m b at time t with time-slide representation t s u e m b and time-encoding representation p u . The time-oriented and location-oriented production of observation embeddings T u and L u for sensor event u, as described in Figure 5, are determined using the following equations:
O u = G A T ( J u ) , [ J u , O u ] { [ p u , T u ] , [ l u e m b , L u ] }
Specifically, first, the time-oriented graph attention layer is used to capture the relationship between u and its adjacent nodes to estimate the enhanced observation embedding T u . There are N nodes in each sample S k , which indicates that an N-th attention weight α u = { α u , v α u , v [ 0 , 1 ] } will be generated. The intermediate LeakyReLU nonlinearity operations are omitted (the negative input slope is 0.2), and dropout operations occur (the dropout rate is 0.2), followed by a normalisation operation; thus, the attention weight is calculated using the following equation:
α u , v = σ ( A [ p u p v ] T )
where A is a trainable weight matrix shared to every two nodes u and v. p u and p v are multi-dimensional vectors using the time-encoding method. σ is a SoftMax function for the normalisation of α u , v . Later, the time-oriented GAT layer computes the enhanced representation for each node as follows:
X u p = σ ( v = 1 N α u , v X v )
The process of the location-oriented graph attention layer is similar to that of the time-oriented graph attention layer, which can be described with the following equations:
β u , v = σ ( B [ l u e m b l v e m b ] T )
X u l = σ ( v = 1 N β u , v X v )
where B is the trainable weight matrix in the location-oriented graph attention layer. It applies the location information to estimate the appropriate observation embeddings L u .
Finally, high-level sensor event representations are obtained by concatenating X u p and X u l .

3.4. Overall Network Architecture

The overall network architecture based on GATs is presented in Figure 6, which includes several modules that are arranged in sequence. More details are provided in this section.
  • The proposed method applies a total of four embedding layers [ E D s i ,   E D o , E D t s ,   E D l ] in the first layer, each of which converts each index representation into a 32-dimensional vector;
  • The original embedding representation of sensor ID and sensor observation is processed with two parallel multiple GAT layers, which automatically optimise the characteristic of sensor events considering time-wise and location-wise specificity;
  • The outputs of two parallel multiple GAT layers are connected with the location representation and time-slice representation. Then, they are passed through a four-layer 1-D convolution layer with a kernel size of 5, followed by an average pooling layer to capture the high-level features of the sensor event sequence for the inference of daily activities;
  • The outputs of the average pooling layer are fed to the fully connected layer to infer activity categories.

3.5. Model Complexity Analysis

This section introduces the computational complexity of the proposed model in detail. As mentioned in Section 3.2, the sensor event count in a window is denoted to N, which refers to the length of the sensor event sequence. In addition, O is commonly used to describe the asymptotic behaviour of the functions that define the time cost of the algorithms in the computer science field. F i and F i are used to represent the dimension of input and output vectors of the i t h layer, where i = 1 , 2 , 3 refers to the embedding layer, the GAT layer, and the convolutional layer, respectively. C refers to a known constant.
Firstly, the embedding layer converts indexes into vectors. Since the number of indexes is constant, the computational complexity of an embedding layer is O ( N F 1 ) . Afterwards, the GAT changes the feature dimension of input vectors through matrix operation and computes the coefficient between every two sensor events, so its computational complexity is O ( N F 2 F 2 + N 2 F 2 ) . In addition, the computational complexity of a 1D convolutional layer is quadratic with kernel size (K), the numbers of channels ( F 3 and F 3 ) as well as the length of input vector (N), and its computational complexity is actually O ( N F 3 F 3 K 2 ) . Therefore, the total computational complexity of the proposed model is described as follows:
α 1 × ( N F 1 ) + α 2 × O ( N F 2 F 2 + N 2 F 2 ) + α 3 × O ( N F 3 F 3 K 2 )
where α refers to the coefficient of each formula.

4. Experiments

4.1. Experimental Setups

First, as previously described, the datasets were segmented using the sliding-window method with different window sizes. In general, the cost of the proposed methods and other existing methods are mostly related to the sequence length. Therefore, the smaller the sequence length, the faster an activity can be recognised. To verify the performance of such models under small sequences, the window sizes were set to 20, 25, 30, and 35 to verify the model performance in various latencies.
Next, to evaluate the proposed method, the data samples of each category were split into two parts, i.e., 70% for training and 30 % for testing. Random-shuffle processing ensured that each model could be trained to achieve better generalisation performance.
Then, the proposed TLGAT method was compared with the state-of-the-art models for human daily activity recognition in publicly available smart home systems, including embedding-based FCNs (E-FCNs) [19], ImageDCNN [7,20], and ELMoBiLSTM [25]. As far as our method is concerned, the hidden dimension sizes of the embedding layer and the convolutional layer were empirically set to 32 and 128. Other hyperparameter settings are presented in Table 2. In particular, the Adam optimiser was used to train all the models in a specific number of iterations with an initial learning rate (LR) of 0.01 , which was delayed at a specific rate during training. Each experiment was performed 30 times, with a trust gap of 99% using the statistical calculation based on three-sigma ( 3 σ ) limits.
Eventually, all the experiments were carried out on a server, with an 11th Gen Inter(R) Core(TM) i7-11700K @3.60GHz, 32GB RAM, and an NVIDIA GeForce RTX 3070(8GB). Torch frameworks were used for the algorithm’s implementation.

4.2. Training and Evaluation

The overall loss for training was computed using the cross-entropy function, which is a suitable measurement option for the difference between the predicted and true probability distributions.
In this task, the sample share of each category was imbalanced. The performance of the models was measured using the F 1 - s c o r e (mean).
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
where T P , T N , and F N refer to true positives, true negatives, and false positives, respectively.

4.3. Comparison with State-of-the-Art Models

To show the satisfying performance of the proposed model, we compared the F 1 - s c o r e of this model and other state-of-the-art models with the Aruba dataset. Table 3 and Table 4 show the overall results of the overall performance under different sensor event window sizes. In total, the proposed model outperformed the traditional state-of-the-art models in HAR under the delay of a different number of sensor events. The F 1 - s c o r e increased as the sensor event window size increased. When the length of the sensor event sequences was no more than 35, longer sequences could provide more information for these models to infer the activities. Furthermore, it seems that TLGAT can automatically generate more relevant features than other state-of-the-art methods on small sequences and can further provide a better recognition performance.
Furthermore, the F 1 - s c o r e values of all activities were determined except for “Enter Home” and “Leave Home” of the ARUBA dataset as well as “Bed to Toilet” and “Eve Meds” of the MILAN dataset (for which there were a low number of samples for testing); the results are shown in Table 5 and Table 6. It can be observed that this model performed well in the classification of each activity. In particular, it seems that it was more difficult to infer some activities using the small sensor event sequences with lengths of 20 and 25 because these activities were more complex than others. If more than 30 sensor events were used, the TLGAT could better represent the sensor event sequence and improve the performance of inferring complex activities.

4.4. Ablation Experiments

In this work, the proposed model consists of several modules based on GAT and CNN. To validate the effect and importance of the proposed GAT-based modules, i.e., the time-oriented and location-oriented GAT modules (measured with the F 1 -score on the test set), we separately trained four kinds of models, namely TLGAT, LGAT, TGAT, and baseline, in the ARUBA and MILAN datasets. Among them, the first model is our final model architecture, the second and third models are the versions of the first one after removing the time-oriented GAT or the location-oriented GAT module, respectively, and the last model removes both of these two modules. Once the architectures of the above models were identified, ablation studies were performed on the datasets.
Each trial was performed using the parameters defined in Section 4.1, so as to obtain the model with optimal performance through training. Table 7 and Table 8 demonstrate the experimental results of the four models under different sensor event window sizes in the ARUBA and MILAN datasets separately. Obviously, compared with the baseline CNN model, the improvement in the F 1 -score of the models using the time-oriented or location-oriented GAT module could slightly provide positive improvement in recognizing activities to some extent. These two modules can learn the temporal and spatial correlations between different sensor events from the sensor sequence, so as to improve the expression of the sensor sequence. Moreover, merging these two modules with a feature concatenation strategy can effectively leverage the superiority of both modules. Therefore, TLGAT was found to have favourable performance in HAR.

5. Conclusions and Discussion

This paper introduces a novel GAT-based model, namely TLGAT, for HAR in IoT-based smart home systems. TLGAT is able to learn the time-oriented and location-oriented dependencies between sensor events and leverages a feature concatenation strategy, thus significantly improving the recognition precision of HAR tasks. The ability of the graph attention structure enables TLGAT to naturally capture the correlations between different sensor events about irregularly activated time intervals and the layout of the sensors. We validated the effectiveness of the proposed architecture on two public datasets, and extensive experiments showed that it consistently outperformed three state-of-the-art models. In addition, through ablation studies, it was found that this framework applied rational strategies to integrate the proposed GAT modules as well as achieved remarkable improvement. Therefore, GAT can provide a better understanding of HAR tasks.
In addition, there are still several challenges in multi-resident HAR tasks because of data complexity, redundancy, and heterogeneity. We plan to explore the GAT-based methods of understanding data and address these challenges. Moreover, we aim to evaluate other methods such as fuzzy Windows methods [36], to explore a novel framework to improve HAR tasks in smart homes.

Author Contributions

Conceptualisation, J.Y. and J.Z.; methodology, J.Y.; validation, J.Y., H.J. and J.Z.; formal analysis, H.J. and J.Z.; investigation, J.Y.; resources, H.J.; writing—original draft preparation, J.Y.; writing—review and editing, J.Y., H.J. and J.Z.; supervision, J.Z. and H.J.; project administration, J.Z. and H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Foundation of China (No. 42177440); the Germany/Hong Kong Joint Research Scheme sponsored by the Research Grants Council of Hong Kong; the German Academic Exchange Service of Germany (Ref. No. G-PolyU505/22), PolyU Start-up Grant: ZVUY-P0035417, CD5E-P0043422 and WZ09-P0043123.

Institutional Review Board Statement

Ethical review and approval were waived for this study due to experiments conducted on already published public datasets, without the immediate participation of humans.

Informed Consent Statement

Patient consent was waived due to experiments conducted on already published public datasets, without the immediate participation of humans.

Data Availability Statement

CASAS datasets are publicly available. CASAS dataset can be found at casas.wsu.edu/datasets/ (accessed on 23 December 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lotfi, A.; Langensiepen, C.; Mahmoud, S.M.; Akhlaghinia, M.J. Smart homes for the elderly dementia sufferers: Identification and prediction of abnormal behaviour. J. Ambient Intell. Humaniz. Comput. 2012, 3, 205–218. [Google Scholar] [CrossRef]
  2. Ye, J.; Stevenson, G.; Dobson, S. Detecting abnormal events on binary sensors in smart home environments. Pervasive Mob. Comput. 2016, 33, 32–49. [Google Scholar] [CrossRef]
  3. Rocha Filho, G.P.; Brandão, A.H.; Nobre, R.A.; Meneguette, R.I.; Freitas, H.; Gonçalves, V.P. HOsT: Towards a Low-Cost Fog Solution via Smart Objects to Deal with the Heterogeneity of Data in a Residential Environment. Sensors 2022, 22, 6257. [Google Scholar] [CrossRef] [PubMed]
  4. Ziadeh, A.; Abualigah, L.; Elaziz, M.A.; Şahin, C.B.; Almazroi, A.A.; Omari, M. Augmented grasshopper optimization algorithm by differential evolution: A power scheduling application in smart homes. Multimed. Tools Appl. 2021, 80, 31569–31597. [Google Scholar] [CrossRef]
  5. Ramlee, R.; Tang, D.; Ismail, M. Smart home system for disabled people via wireless bluetooth. In Proceedings of the 2012 International Conference on System Engineering and Technology (ICSET), Bandung, Indonesia, 11–12 September 2012; pp. 1–4. [Google Scholar]
  6. Torres Neto, J.R.; Rocha Filho, G.P.; Mano, L.Y.; Villas, L.A.; Ueyama, J. Exploiting offloading in IoT-based microfog: Experiments with face recognition and fall detection. Wirel. Commun. Mob. Comput. 2019, 2019, 2786837. [Google Scholar] [CrossRef]
  7. Gochoo, M.; Tan, T.H.; Liu, S.H.; Jean, F.R.; Alnajjar, F.S.; Huang, S.C. Unobtrusive activity recognition of elderly people living alone using anonymous binary sensors and DCNN. IEEE J. Biomed. Health Inform. 2018, 23, 693–702. [Google Scholar] [CrossRef]
  8. Lee, S.M.; Yoon, S.M.; Cho, H. Human activity recognition from accelerometer data using Convolutional Neural Network. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju Island, Republic of Korea, 13–16 February 2017; pp. 131–134. [Google Scholar]
  9. Murad, A.; Pyun, J.Y. Deep recurrent neural networks for human activity recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef]
  10. Cook, D.J.; Crandall, A.S.; Thomas, B.L.; Krishnan, N.C. CASAS: A smart home in a box. Computer 2012, 46, 62–69. [Google Scholar] [CrossRef]
  11. Bouchabou, D.; Nguyen, S.M.; Lohr, C.; LeDuc, B.; Kanellos, I. A survey of human activity recognition in smart homes based on IoT sensors algorithms: Taxonomies, challenges, and opportunities with deep learning. Sensors 2021, 21, 6037. [Google Scholar] [CrossRef]
  12. Gu, F.; Chung, M.H.; Chignell, M.; Valaee, S.; Zhou, B.; Liu, X. A survey on deep learning for human activity recognition. ACM Comput. Surv. (CSUR) 2021, 54, 1–34. [Google Scholar] [CrossRef]
  13. Tapia, E.M.; Intille, S.S.; Larson, K. Activity recognition in the home using simple and ubiquitous sensors. In Proceedings of the International Conference on Pervasive Computing, Linz/Vienna, Austria, 21–23 April 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 158–175. [Google Scholar]
  14. Van Kasteren, T.; Englebienne, G.; Kröse, B.J. Activity recognition using semi-Markov models on real world smart home datasets. J. Ambient Intell. Smart Environ. 2010, 2, 311–325. [Google Scholar] [CrossRef]
  15. Wu, H.; Pan, W.; Xiong, X.; Xu, S. Human activity recognition based on the combined SVM&HMM. In Proceedings of the 2014 IEEE International Conference on Information and Automation (ICIA), Hailar, China, 28–30 July 2014; pp. 219–224. [Google Scholar]
  16. Nazerfard, E.; Das, B.; Holder, L.B.; Cook, D.J. Conditional random fields for activity recognition in smart environments. In Proceedings of the 1st ACM International Health Informatics Symposium, Arlington, VA, USA, 11–12 November 2010; pp. 282–286. [Google Scholar]
  17. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human activity recognition using inertial, physiological and environmental sensors: A comprehensive survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef]
  18. Singh, D.; Merdivan, E.; Hanke, S.; Kropf, J.; Geist, M.; Holzinger, A. Convolutional and recurrent neural networks for activity recognition in smart environment. In Towards Integrative Machine Learning and Knowledge Extraction; Springer: Berlin/Heidelberg, Germany, 2017; pp. 194–205. [Google Scholar]
  19. Bouchabou, D.; Nguyen, S.M.; Lohr, C.; Leduc, B.; Kanellos, I. Fully convolutional network bootstrapped by word encoding and embedding for activity recognition in smart homes. In Proceedings of the International Workshop on Deep Learning for Human Activity Recognition, Kyoto, Japan, 8 January 2020; Springer: Singapore, 2021; pp. 111–125. [Google Scholar]
  20. Tan, T.H.; Gochoo, M.; Huang, S.C.; Liu, Y.H.; Liu, S.H.; Huang, Y.F. Multi-resident activity recognition in a smart home using RGB activity image and DCNN. IEEE Sens. J. 2018, 18, 9718–9727. [Google Scholar] [CrossRef]
  21. Mohmed, G.; Lotfi, A.; Pourabdollah, A. Employing a deep convolutional neural network for human activity recognition based on binary ambient sensor data. In Proceedings of the 13th ACM International Conference on PErvasive Technologies Related to Assistive Environments, Virtual, 30 June–3 July 2020; pp. 1–7. [Google Scholar]
  22. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  23. Arifoglu, D.; Bouchachia, A. Activity recognition and abnormal behaviour detection with recurrent neural networks. Procedia Comput. Sci. 2017, 110, 86–93. [Google Scholar] [CrossRef]
  24. Liciotti, D.; Bernardini, M.; Romeo, L.; Frontoni, E. A sequential deep learning application for recognising human activities in smart homes. Neurocomputing 2020, 396, 501–513. [Google Scholar] [CrossRef]
  25. Bouchabou, D.; Nguyen, S.M.; Lohr, C.; LeDuc, B.; Kanellos, I. Using language model to bootstrap human activity recognition ambient sensors based in smart homes. Electronics 2021, 10, 2498. [Google Scholar] [CrossRef]
  26. Zhou, J.; Cui, G.; Hu, S.; Zhang, Z.; Yang, C.; Liu, Z.; Wang, L.; Li, C.; Sun, M. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar] [CrossRef]
  27. Wang, X.; He, X.; Cao, Y.; Liu, M.; Chua, T.S. Kgat: Knowledge graph attention network for recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 950–958. [Google Scholar]
  28. Kosaraju, V.; Sadeghian, A.; Martín-Martín, R.; Reid, I.; Rezatofighi, H.; Savarese, S. Social-bigat: Multimodal trajectory forecasting using bicycle-gan and graph attention networks. arXiv 2019, arXiv:1907.03395. [Google Scholar]
  29. Wu, Y.; Lian, D.; Xu, Y.; Wu, L.; Chen, E. Graph convolutional networks with markov random field reasoning for social spammer detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 1054–1061. [Google Scholar]
  30. Li, L.; Gan, Z.; Cheng, Y.; Liu, J. Relation-aware graph attention network for visual question answering. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10313–10322. [Google Scholar]
  31. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  32. Cook, D.J. Learning setting-generalized activity models for smart spaces. IEEE Intell. Syst. 2010, 2010, 1. [Google Scholar] [CrossRef] [PubMed]
  33. Cook, D.J.; Schmitter-Edgecombe, M. Assessing the quality of activities in a smart environment. Methods Inf. Med. 2009, 48, 480–485. [Google Scholar] [PubMed] [Green Version]
  34. Horn, M.; Moor, M.; Bock, C.; Rieck, B.; Borgwardt, K. Set functions for time series. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 4353–4363. [Google Scholar]
  35. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  36. Hamad, R.A.; Hidalgo, A.S.; Bouguelia, M.R.; Estevez, M.E.; Quero, J.M. Efficient activity recognition in smart homes using delayed fuzzy temporal windows on binary sensors. IEEE J. Biomed. Health Inform. 2019, 24, 387–395. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Sensors 23 01626 g001
Figure 2. A sample of ARUBA’s raw dataset.
Figure 2. A sample of ARUBA’s raw dataset.
Sensors 23 01626 g002
Figure 3. The sensor layout and location definition; blue circles are the motion sensors; yellow circles are the temperature sensors; and orange circles are the door sensors. Arrows indicate that the location indexes of the sensor in the specified area are set to the corresponding numbers.
Figure 3. The sensor layout and location definition; blue circles are the motion sensors; yellow circles are the temperature sensors; and orange circles are the door sensors. Arrows indicate that the location indexes of the sensor in the specified area are set to the corresponding numbers.
Sensors 23 01626 g003
Figure 4. Sliding-window method.
Figure 4. Sliding-window method.
Sensors 23 01626 g004
Figure 5. The architecture of J-oriented graph attention layer. In the time-oriented graph attention layer, it is assumed that [ J , J u , J v , O ] is the time-encoding feature vector [ p , p u , p v , T ] , while in the location-oriented graph attention layer, we assume [ J , J u , J v , O ] to be the location-embedding representation [ l e m b , l u e m b , l v e m b , L ] .
Figure 5. The architecture of J-oriented graph attention layer. In the time-oriented graph attention layer, it is assumed that [ J , J u , J v , O ] is the time-encoding feature vector [ p , p u , p v , T ] , while in the location-oriented graph attention layer, we assume [ J , J u , J v , O ] to be the location-embedding representation [ l e m b , l u e m b , l v e m b , L ] .
Sensors 23 01626 g005
Figure 6. The overall network architecture of the proposed TLGAT.
Figure 6. The overall network architecture of the proposed TLGAT.
Sensors 23 01626 g006
Table 1. Dataset description (D = door switch sensor, M = motion sensor, T = temperature sensor).
Table 1. Dataset description (D = door switch sensor, M = motion sensor, T = temperature sensor).
DatasetThe Kind of SensorsThe Number of ResidentsThe Number of Raw Sensor EventsThe Number of Activity Categories
ARUBAD/M/T11,719,55811
MILAND/M/T1+pet433,66515
Table 2. Training hyperparameter settings.
Table 2. Training hyperparameter settings.
Training HyperparametersValues
Batch size256
Dropout rate0.2
Channels of embedding layers32
Channels of convolutional layers128
Learning rate 1 × 10 2
IterationsARUBA: 90; MILAN: 40
Decayed epochs of LRARUBA: 24/48/72; MILAN: 20
Decayed rate of LR5
Number of replications30
Table 3. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of ELMoBiLSTM, ImageDCNN, E-FCNs, and TLGAT in the ARUBA dataset.
Table 3. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of ELMoBiLSTM, ImageDCNN, E-FCNs, and TLGAT in the ARUBA dataset.
ModelN = 20N = 25N = 30N = 35
ELMoBiLSTM [25]85.307 ± 0.80586.867 ± 0.80589.564 ± 2.21790.515 ± 2.855
ImageDCNN [7,20]84.668 ± 1.86487.618 ± 2.60592.453 ± 2.98495.711 ± 1.654
E-FCNs [19]96.987 ± 0.43698.463 ± 0.25099.108 ± 0.09899.442 ± 0.128
TLGAT97.651 ± 0.25498.942 ± 0.10499.514 ± 0.06299.760 ± 0.199
Table 4. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of ELMoBiLSTM, ImageDCNN, E-FCNs, and TLGAT in the MILAN dataset.
Table 4. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of ELMoBiLSTM, ImageDCNN, E-FCNs, and TLGAT in the MILAN dataset.
ModelN = 20N = 25N = 30N = 35
ELMoBiLSTM [25]76.847 ± 1.68585.357 ± 2.22889.975 ± 2.15692.476 ± 1.706
ImageDCNN [7,20]67.382 ± 1.69676.408 ± 2.78084.932 ± 2.58990.857 ± 1.726
E-FCNs [19]92.901 ± 0.25294.656 ± 0.21295.621 ± 0.25796.445 ± 0.154
TLGAT97.569 ± 0.13498.081 ± 0.18198.465 ± 0.17498.736 ± 0.165
Table 5. The F 1 - s c o r e ( % ) (mean ± 3 σ ) of different activities of TLGAT in the ARUBA dataset.
Table 5. The F 1 - s c o r e ( % ) (mean ± 3 σ ) of different activities of TLGAT in the ARUBA dataset.
IDActivityN = 20N = 25N = 30N = 35
1Eating95.167 ± 2.26698.138 ± 0.65398.984 ± 0.36799.328 ± 0.211
2Housekeeping94.397 ± 1.26598.088 ± 0.85598.613 ± 0.74199.593 ± 0.422
3Meal Preparation95.384 ± 0.89097.813 ± 1.14198.940 ± 0.21699.524 ± 0.320
4Other98.021 ± 1.24799.143 ± 0.23799.591 ± 0.10399.835 ± 0.037
5Relax98.877 ± 0.46599.564 ± 0.13699.811 ± 0.07499.925 ± 0.011
6Respirate100.00099.978 ± 0.01899.578 ± 0.034100.000
7Sleeping99.611 ± 0.05699.852 ± 0.04399.914 ± 0.01299.980 ± 0.021
8Wash Dishes84.513 ± 3.26492.104 ± 2.15796.745 ± 1.49098.266 ± 1.873
9Work96.786 ± 0.34498.901 ± 0.02199.443 ± 0.18799.780 ± 0.054
Table 6. The F 1 - s c o r e ( % ) (mean ± 3 σ ) of different activities of TLGAT in the MILAN dataset.
Table 6. The F 1 - s c o r e ( % ) (mean ± 3 σ ) of different activities of TLGAT in the MILAN dataset.
IDActivityN = 20N = 25N = 30N = 35
1Chores99.653 ± 0.21299.751 ± 0.10299.671 ± 0.04899.888 ± 0.011
2Desk Activity99.768 ± 0.20399.817 ± 0.01299.864 ± 0.10199.944 ± 0.024
3Dining Rm Activity99.687 ± 0.16499.342 ± 0.18799.795 ± 0.105100.000
4Guest Bathroom99.909 ± 0.15899.594 ± 0.24899.805 ± 0.12499.827 ± 0.233
5Kitchen Activity97.999 ± 0.12997.994 ± 0.15398.403 ± 0.62098.535 ± 0.311
6Leave Home99.900 ± 0.08999.730 ± 0.180100.00099.893 ± 0.071
7Master Bathroom98.098 ± 0.37597.681 ± 0.94797.708 ± 1.18597.794 ± 0.674
8Master Bedroom96.945 ± 0.76396.432 ± 0.12596.822 ± 0.41797.113 ± 0.312
9Meditate100.000100.000100.000100.000
10Morning Meds94.480 ± 4.16294.881 ± 2.964.97.992 ± 1.76595.928 ± 1.178
11Other98.277 ± 0.11198.140 ± 0.22398.493 ± 0.18998.641 ± 0.126
12Read98.702 ± 0.21998.576 ± 0.36799.018 ± 0.16999.162 ± 0.155
13Sleep99.514 ± 0.11099.286 ± 0.08799.573 ± 0.06199.630 ± 0.127
14Watch TV98.806 ± 0.18898.431 ± 0.21798.590 ± 0.26698.785 ± 0.235
Table 7. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of TLGAT, LGAT, TGAT, and baseline in the ARUBA dataset.
Table 7. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of TLGAT, LGAT, TGAT, and baseline in the ARUBA dataset.
ModelN = 20N = 25N = 30N = 35
baseline96.999 ± 0.15498.694 ± 0.22999.262 ± 0.14599.524 ± 0.076
LGAT97.519 ± 0.24998.834 ± 0.19999.469 ± 0.13699.710 ± 0.246
TGAT97.496 ± 0.47998.979 ± 0.26699.491 ± 0.14099.737 ± 0.242
TLGAT97.651 ± 0.25498.942 ± 0.10499.514 ± 0.06299.760 ± 0.199
Table 8. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of TLGAT, LGAT, TGAT, and baseline in the MILAN dataset.
Table 8. The overall F 1 - s c o r e ( % ) (mean ± 3 σ ) of TLGAT, LGAT, TGAT, and baseline in the MILAN dataset.
ModelN = 20N = 25N = 30N = 35
baseline96.381 ± 0.20297.211 ± 0.21197.758 ± 0.20798.161 ± 0.184
LGAT97.233 ± 0.12597.826 ± 0.11598.239 ± 0.17698.439 ± 0.151
TGAT97.149 ± 0.28797.843 ± 0.15298.404 ± 0.10798.620 ± 0.161
TLGAT97.569 ± 0.13498.081 ± 0.18198.465 ± 0.17498.736 ± 0.065
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, J.; Jiang, H.; Zhong, J. A Graph-Attention-Based Method for Single-Resident Daily Activity Recognition in Smart Homes. Sensors 2023, 23, 1626. https://doi.org/10.3390/s23031626

AMA Style

Ye J, Jiang H, Zhong J. A Graph-Attention-Based Method for Single-Resident Daily Activity Recognition in Smart Homes. Sensors. 2023; 23(3):1626. https://doi.org/10.3390/s23031626

Chicago/Turabian Style

Ye, Jiancong, Hongjie Jiang, and Junpei Zhong. 2023. "A Graph-Attention-Based Method for Single-Resident Daily Activity Recognition in Smart Homes" Sensors 23, no. 3: 1626. https://doi.org/10.3390/s23031626

APA Style

Ye, J., Jiang, H., & Zhong, J. (2023). A Graph-Attention-Based Method for Single-Resident Daily Activity Recognition in Smart Homes. Sensors, 23(3), 1626. https://doi.org/10.3390/s23031626

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop