Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework
Abstract
:1. Introduction
- A systematic evaluation of different augmentation techniques is presented to solve the inadequacy problem of labeled time series data.
- An automatic feature-learning technique is proposed to recognise multimodal data of different wearable smart devices.
- A detailed overview of existing techniques and their categorisation is presented.
- An extensive experimental evaluation is proposed on two benchmark datasets.
2. Literature Review
2.1. Human Activity Recognition
2.1.1. Handcrafted Feature-Based Techniques
2.1.2. Codebook-Based Feature-Encoding Techniques
2.1.3. Automatic Feature-Based Techniques
2.2. Data Augmentation
2.2.1. Magnitude Domain Transformations (MDTs)
2.2.2. Time Domain Transformations (TDTs)
2.2.3. Mixing Patterns (MPs)
2.2.4. Deep Learning-Based Generative Models
3. Proposed Method
3.1. Augmentation Techniques
3.1.1. Augmentation Based on MDTs
- Jittering: This is a process of introducing noise to time series. It is an example of a transformation-based data-augmentation method that is both straightforward and efficient. Equation (1) represents mathematical notation of jittering.
- Scaling: A random scalar number is used in scaling to alter a time series’ global magnitude or intensity. Scaling involves multiplying the scaling parameter by the total time series. Equation (2) represents mathematical notation of scaling.The scaling parameter is selected via a Gaussian distribution (N) with 1.5. Figure 6 demonstrates the actual and transformed data after applying scaling on CogAge atomic (bending) activity.
3.1.2. Augmentation Based on TDTs
- Time warping: This refers to the process of altering a pattern in the temporal dimension. This task accomplished by employing a seamless distortion trajectory [62,85]. Equation (3) represents the mathematical notation of time warping.Here, denotes a warping function that adjusts the time steps based on a smooth curve. The curve’s smoothness is dictated by a cubic spline with knots . The knot heights are determined from . This transformation manipulates the time axis by compressing or expanding it at various points in the time series, which introduces diversity and enhances the dataset. Figure 7 demonstrates the actual and transformed data after applying time warping on CogAge atomic (bending) activity.
- Linear Interpolation: This calculates new values by fitting a straight line between neighboring data points. The interpolated value between two existing data points and at any time t can be calculated by the Equation (4).
- Exponential Moving Median Smoothing: This provides smooth data by exponentially diminishing weights. To be resistant against outliers, it computes a weighted median rather than a weighted average. Mathematically, it can be expressed by Equation (5).
- Channel Permutation: This shuffles the channels (or features) of the complete sequence without changing the values inside each channel. Mathematically, it is represented by Equation (6).X is the rearranged version of the original data, where is a function that changes the order of the channels. This transformation preserves the chronological order of the data but adds diversity by reorganising the channels. Figure 10 demonstrates the actual and transformed data after applying channel permutation on CogAge atomic (bending) activity.
- Rolling Window Averaging: This helps to smooth or denoise the data while maintaining its underlying structure. This method entails utilising a moving average process on the time series data by employing a sliding window. Equation (7) represents the mathematical notation of it.
3.1.3. Mixing Pattern
- Sub Averaging: This process involves the averaging of two patterns to produce a unique pattern. This method includes averaging the temporal values of two sequences belonging to the same class to generate a novel time series pattern, similar to the mixup approach [33]. We integrate different subjects within the same class. Mathematically, it is formulated as in Equation (8)The new time series pattern is calculated by taking the average of the corresponding values in and . This method increases the diversity of the dataset and can help the generalisation of the model by generating a wide range of training samples. Figure 12 demonstrates the actual and transformed data after applying sub averaging on CogAge atomic (bending) activity.
- Sub Cutmix: This technique involves the random replacement of segments from two different sequences belonging to different subjects. It improves dataset variety by combining information from various time series, which can strengthen the resilience and generalisation capabilities of time series models. Mathematically, it can be expressed by Equation (9)
- AugMix: This uses three simultaneous augmentation chains with randomly selected augmentation operations. Three transformed sequences (smoothing, scaling, time warping) are created by consecutively applying these operations to the input sequence. The altered data are mixed with original data to create a new sequence. Incorporation of different transformation techniques directly into data increases the variability and robustness of models. Figure 14 demonstrates the actual and transformed data after applying AugMix on CogAge atomic (bending) activity.
- Mixup: This is used to combine two randomly selected time series to create new sequences. During data blending, the mixing factor sets the fraction of values from each sequence [32]. It is mathematically represented by Equation (10):Here, represents the augmented sequences generated by blending sequences and based on the mixing factor . The values for are typically drawn from a beta distribution. Figure 15 demonstrates the actual and transformed data after applying mixup on CogAge atomic (bending) activity.
- Cutmix: This technique substitutes arbitrarily shaped segments from one image for the other [75,76]. This can be stated mathematically by Equation (11).and are two initial data patterns. The new data pattern, , is produced by combining and with a mixing coefficient of . The amount of original patterns that are kept in the blended pattern is controlled by which is empirically set to 0.2. By adding variability to the data, this method increases the diversity of datasets and may also strengthen the generalisation and robustness of the models. Figure 16 demonstrates the actual and transformed data after applying cutmix on CogAge atomic (bending) activity.
- Hide and Seek: This splits up sequences into a predetermined number of segments or intervals using the hide and seek strategy. Next, a random selection process is used to mask each segment with a specific probability, thus concealing its information. Random parts of the time series are eliminated by substituting the average of all the data points in the dataset for the masked segments. By replicating missing or noisy data, this approach increases variability and can improve the robustness of time series models. Figure 17 demonstrates the actual and transformed data after applying hide and seek on CogAge atomic (bending) activity.
3.1.4. Other Techniques
3.2. Model Architecture Design
3.2.1. Convolutional Neural Network
3.2.2. Long Short-Term Memory
3.2.3. Multi-Branch Hybrid Conv-LSTM (MHyCoL)
4. Experiments and Results
4.1. Dataset Description
4.1.1. CogAge
- Smartphone Accelerometer (sp-acc);
- Smartphone Gyroscope (sp-gyro);
- Smartphone Gravity (sp-grav);
- Smartphone Linear Accelerometer (sp-linAcc);
- Smartphone Magnetometer (sp-magn);
- Smartwatch Accelerometer (sw-acc);
- Smartwatch Gyroscope (sw-gyro);
- Smartglasses Accelerometer (sg-acc);
- Smartglasses Gyroscope (sg-gyro).
4.1.2. UniMiBSHAR
4.2. Pre-Processing
4.3. Datasetting
4.4. Experimental Setup
4.5. Results and Discussion
4.6. Performance Comparison with Existing State-of-the-Art Techniques
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Bisio, I.; Lavagetto, F.; Marchese, M.; Sciarrone, A. Smartphone-centric ambient assisted living platform for patients suffering from co-morbidities monitoring. IEEE Commun. Mag. 2015, 53, 34–41. [Google Scholar] [CrossRef]
- Batool, S.; Khan, M.H.; Farid, M.S. An ensemble deep learning model for human activity analysis using wearable sensory data. Appl. Soft Comput. 2024, 159, 111599. [Google Scholar] [CrossRef]
- Li, F.; Shirahama, K.; Nisar, M.A.; Köping, L.; Grzegorzek, M. Comparison of feature learning methods for human activity recognition using wearable sensors. Sensors 2018, 18, 679. [Google Scholar] [CrossRef] [PubMed]
- Logan, B.; Healey, J.; Philipose, M.; Tapia, E.M.; Intille, S. A long-term evaluation of sensing modalities for activity recognition. In Proceedings of the International Conference on Ubiquitous Computing; Springer: Berlin/Heidelberg, Germany, 2007; pp. 483–500. [Google Scholar]
- Yang, R.; Wang, B. PACP: A position-independent activity recognition method using smartphone sensors. Information 2016, 7, 72. [Google Scholar] [CrossRef]
- Li, F.; Shirahama, K.; Nisar, M.A.; Huang, X.; Grzegorzek, M. Deep transfer learning for time series data based on sensor modality classification. Sensors 2020, 20, 4271. [Google Scholar] [CrossRef]
- Miotto, R.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J.T. Deep learning for healthcare: Review, opportunities and challenges. Brief. Bioinform. 2018, 19, 1236–1246. [Google Scholar] [CrossRef]
- Avati, A.; Jung, K.; Harman, S.; Downing, L.; Ng, A.; Shah, N.H. Improving palliative care with deep learning. BMC Med. Inform. Decis. Mak. 2018, 18, 55–64. [Google Scholar] [CrossRef]
- Ismail Fawaz, H.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019, 33, 917–963. [Google Scholar] [CrossRef]
- Lim, B.; Zohren, S. Time-series forecasting with deep learning: A survey. Philos. Trans. R. Soc. A 2021, 379, 20200209. [Google Scholar] [CrossRef]
- Choi, K.; Yi, J.; Park, C.; Yoon, S. Deep learning for anomaly detection in time series data: Review, analysis, and guidelines. IEEE Access 2021, 9, 120043–120065. [Google Scholar] [CrossRef]
- Khan, M.H.; Farid, M.S.; Grzegorzek, M. A nonlinear view transformations model for cross-view gait recognition. Neurocomputing 2020, 402, 100–111. [Google Scholar] [CrossRef]
- Naaz, F.; Herle, A.; Channegowda, J.; Raj, A.; Lakshminarayanan, M. A generative adversarial network-based synthetic data augmentation technique for battery condition evaluation. Int. J. Energy Res. 2021, 45, 19120–19135. [Google Scholar] [CrossRef]
- Khan, M.H.; Azam, H.; Farid, M.S. Automatic multi-gait recognition using pedestrian’s spatiotemporal features. J. Supercomput. 2023, 79, 19254–19276. [Google Scholar] [CrossRef]
- Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
- Olson, M.; Wyner, A.; Berk, R. Modern neural networks generalise on small datasets. Adv. Neural Inf. Process. Syst. 2018, 31, 3623–3632. [Google Scholar]
- Blagus, R.; Lusa, L. SMOTE for high-dimensional class-imbalanced data. BMC Bioinform. 2013, 14, 106. [Google Scholar] [CrossRef] [PubMed]
- Hasibi, R.; Shokri, M.; Dehghan, M. Augmentation scheme for dealing with imbalanced network traffic classification using deep learning. arXiv 2019, arXiv:1901.00204. [Google Scholar]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; University of Toronto: Toronto, ON, Canada, 2009. [Google Scholar]
- Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
- Wen, Q.; Sun, L.; Yang, F.; Song, X.; Gao, J.; Wang, X.; Xu, H. Time series data augmentation for deep learning: A survey. arXiv 2020, arXiv:2002.12478. [Google Scholar]
- Fields, T.; Hsieh, G.; Chenou, J. Mitigating drift in time series data with noise augmentation. In Proceedings of the 2019 International Conference on Computational Science and Computational Intelligence (CSCI), Las Vegas, NV, USA, 5–7 December 2019; pp. 227–230. [Google Scholar]
- Um, T.T.; Pfister, F.M.; Pichler, D.; Endo, S.; Lang, M.; Hirche, S.; Fietzek, U.; Kulić, D. Data augmentation of wearable sensor data for parkinson’s disease monitoring using convolutional neural networks. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, Glasgow, UK, 13–17 November 2017; pp. 216–220. [Google Scholar]
- Nweke, H.F.; Teh, Y.W.; Al-Garadi, M.A.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
- Min, J.; Tu, J.; Xu, C.; Lukas, H.; Shin, S.; Yang, Y.; Solomon, S.A.; Mukasa, D.; Gao, W. Skin-interfaced wearable sweat sensors for precision medicine. Chem. Rev. 2023, 123, 5049–5138. [Google Scholar] [CrossRef] [PubMed]
- Jaitly, N.; Hinton, G.E. Vocal tract length perturbation (VTLP) improves speech recognition. Proc. ICML Workshop Deep. Learn. Audio Speech Lang. 2013, 117, 21. [Google Scholar]
- Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
- Inoue, H. Data augmentation by pairing samples for images classification. arXiv 2018, arXiv:1801.02929. [Google Scholar]
- Nisar, M.A.; Shirahama, K.; Li, F.; Huang, X.; Grzegorzek, M. Rank pooling approach for wearable sensor-based ADLs recognition. Sensors 2020, 20, 3463. [Google Scholar] [CrossRef] [PubMed]
- Micucci, D.; Mobilio, M.; Napoletano, P. Unimib shar: A dataset for human activity recognition using acceleration data from smartphones. Appl. Sci. 2017, 7, 1101. [Google Scholar] [CrossRef]
- Urwyler, P.; Rampa, L.; Stucki, R.; Büchler, M.; Müri, R.; Mosimann, U.P.; Nef, T. Recognition of activities of daily living in healthy subjects using two ad-hoc classifiers. Biomed. Eng. Online 2015, 14, 54. [Google Scholar] [CrossRef]
- Khan, M.H. Human Activity Analysis in Visual Surveillance and Healthcare; Logos Verlag Berlin GmbH: Berlin, Germany, 2018; Volume 45. [Google Scholar]
- Fan, S.; Jia, Y.; Jia, C. A feature selection and classification method for activity recognition based on an inertial sensing unit. Information 2019, 10, 290. [Google Scholar] [CrossRef]
- Amjad, F.; Khan, M.H.; Nisar, M.A.; Farid, M.S.; Grzegorzek, M. A comparative study of feature selection approaches for human activity recognition using multimodal sensory data. Sensors 2021, 21, 2368. [Google Scholar] [CrossRef] [PubMed]
- Sargano, A.B.; Angelov, P.; Habib, Z. A comprehensive review on handcrafted and learning-based action representation approaches for human activity recognition. Appl. Sci. 2017, 7, 110. [Google Scholar] [CrossRef]
- Hsu, W.C.; Sugiarto, T.; Liao, Y.Y.; Lin, Y.J.; Yang, F.C.; Hueng, D.Y.; Sun, C.T.; Chou, K.N. Can trunk acceleration differentiate stroke patient gait patterns using time-and frequency-domain features? Appl. Sci. 2021, 11, 1541. [Google Scholar] [CrossRef]
- Liang, H.; Sun, X.; Sun, Y.; Gao, Y. Text feature extraction based on deep learning: A review. EURASIP J. Wirel. Commun. Netw. 2017, 2017, 211. [Google Scholar] [CrossRef] [PubMed]
- Fatima, R.; Khan, M.H.; Nisar, M.A.; Doniec, R.; Farid, M.S.; Grzegorzek, M. A Systematic Evaluation of Feature Encoding Techniques for Gait Analysis Using Multimodal Sensory Data. Sensors 2023, 24, 75. [Google Scholar] [CrossRef]
- Khan, M.H.; Farid, M.S.; Grzegorzek, M. A comprehensive study on codebook-based feature fusion for gait recognition. Inf. Fusion 2023, 92, 216–230. [Google Scholar] [CrossRef]
- Lagodzinski, P.; Shirahama, K.; Grzegorzek, M. Codebook-based electrooculography data analysis towards cognitive activity recognition. Comput. Biol. Med. 2018, 95, 277–287. [Google Scholar] [CrossRef]
- Shirahama, K.; Grzegorzek, M. On the generality of codebook approach for sensor-based human activity recognition. Electronics 2017, 6, 44. [Google Scholar] [CrossRef]
- Köping, L.; Shirahama, K.; Grzegorzek, M. A general framework for sensor-based human activity recognition. Comput. Biol. Med. 2018, 95, 248–260. [Google Scholar] [CrossRef]
- Khan, M.H.; Farid, M.S.; Grzegorzek, M. A generic codebook based approach for gait recognition. Multimed. Tools Appl. 2019, 78, 35689–35712. [Google Scholar] [CrossRef]
- Khan, M.H.; Li, F.; Farid, M.S.; Grzegorzek, M. Gait recognition using motion trajectory analysis. In Proceedings of the 10th International Conference on Computer Recognition Systems CORES 2017 10, Polanica Zdroj, Poland, 22–24 May 2017; Springer: Berlin/Heidelberg, Germany, 2018; pp. 73–82. [Google Scholar]
- Zhang, J.; Wu, F.; Wei, B.; Zhang, Q.; Huang, H.; Shah, S.W.; Cheng, J. Data augmentation and dense-LSTM for human activity recognition using WiFi signal. IEEE Internet Things J. 2020, 8, 4628–4641. [Google Scholar] [CrossRef]
- Bianchi, V.; Bassoli, M.; Lombardo, G.; Fornacciari, P.; Mordonini, M.; De Munari, I. IoT wearable sensor and deep learning: An integrated approach for personalized human activity recognition in a smart home environment. IEEE Internet Things J. 2019, 6, 8553–8562. [Google Scholar] [CrossRef]
- Anagnostis, A.; Benos, L.; Tsaopoulos, D.; Tagarakis, A.; Tsolakis, N.; Bochtis, D. Human activity recognition through recurrent neural networks for human–robot interaction in agriculture. Appl. Sci. 2021, 11, 2188. [Google Scholar] [CrossRef]
- Bu, C.; Zhang, L.; Cui, H.; Yang, G.; Wu, H. Dynamic inference via localizing semantic intervals in sensor data for budget-tunable activity recognition. IEEE Trans. Ind. Inform. 2023, 20, 3801–3813. [Google Scholar] [CrossRef]
- Cheng, D.; Zhang, L.; Bu, C.; Wang, X.; Wu, H.; Song, A. ProtoHAR: Prototype guided personalized federated learning for human activity recognition. IEEE J. Biomed. Health Inform. 2023, 27, 3900–3911. [Google Scholar] [CrossRef]
- Huang, W.; Zhang, L.; Wu, H.; Min, F.; Song, A. Channel-Equalization-HAR: A light-weight convolutional neural network for wearable sensor based human activity recognition. IEEE Trans. Mob. Comput. 2022, 22, 5064–5077. [Google Scholar] [CrossRef]
- Saha, J.; Chowdhury, C.; Roy Chowdhury, I.; Biswas, S.; Aslam, N. An ensemble of condition based classifiers for device independent detailed human activity recognition using smartphones. Information 2018, 9, 94. [Google Scholar] [CrossRef]
- Kolkar, R.; Geetha, V. Human activity recognition in smart home using deep learning techniques. In Proceedings of the 2021 13th International Conference on Information & Communication Technology and System (ICTS), Surabaya, Indonesia, 20–21 October 2021; pp. 230–234. [Google Scholar]
- Dua, N.; Singh, S.N.; Semwal, V.B. Multi-input CNN-GRU based human activity recognition using wearable sensors. Computing 2021, 103, 1461–1478. [Google Scholar] [CrossRef]
- Khatun, M.A.; Yousuf, M.A.; Ahmed, S.; Uddin, M.Z.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Khan, A.; Azad, A.; Moni, M.A. Deep CNN-LSTM with self-attention model for human activity recognition using wearable sensor. IEEE J. Transl. Eng. Health Med. 2022, 10, 2700316. [Google Scholar] [CrossRef]
- Perez-Gamboa, S.; Sun, Q.; Zhang, Y. Improved sensor based human activity recognition via hybrid convolutional and recurrent neural networks. In Proceedings of the 2021 IEEE International Symposium on Inertial Sensors and Systems (INERTIAL), Kailua-Kona, HI, USA, 22–25 March 2021; pp. 1–4. [Google Scholar]
- Nisar, M.A.; Shirahama, K.; Irshad, M.T.; Huang, X.; Grzegorzek, M. A Hierarchical Multitask Learning Approach for the Recognition of Activities of Daily Living Using Data from Wearable Sensors. Sensors 2023, 23, 8234. [Google Scholar] [CrossRef]
- Pfister, F.M.; Um, T.T.; Pichler, D.C.; Goschenhofer, J.; Abedinpour, K.; Lang, M.; Endo, S.; Ceballos-Baumann, A.O.; Hirche, S.; Bischl, B.; et al. High-resolution motor state detection in Parkinson’s disease using convolutional neural networks. Sci. Rep. 2020, 10, 5860. [Google Scholar] [CrossRef]
- Iwana, B.K.; Uchida, S. An empirical survey of data augmentation for time series classification with neural networks. PLoS ONE 2021, 16, e0254841. [Google Scholar] [CrossRef] [PubMed]
- Rashid, K.M.; Louis, J. Activity identification in modular construction using audio signals and machine learning. Autom. Constr. 2020, 119, 103361. [Google Scholar] [CrossRef]
- Steven Eyobu, O.; Han, D.S. Feature representation and data augmentation for human activity classification based on wearable IMU sensor data using a deep LSTM neural network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef]
- Fawaz, H.I.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Data augmentation using synthetic data for time series classification with deep residual networks. arXiv 2018, arXiv:1808.02455. [Google Scholar]
- Kalouris, G.; Zacharaki, E.I.; Megalooikonomou, V. Improving CNN-based activity recognition by data augmentation and transfer learning. In Proceedings of the 2019 IEEE 17th International Conference on Industrial Informatics (INDIN), Helsinki, Finland, 22–25 July 2019; Volume 1, pp. 1387–1394. [Google Scholar]
- Wang, J.; Zhu, T.; Gan, J.; Chen, L.L.; Ning, H.; Wan, Y. Sensor data augmentation by resampling in contrastive learning for human activity recognition. IEEE Sens. J. 2022, 22, 22994–23008. [Google Scholar] [CrossRef]
- Tran, L.; Choi, D. Data augmentation for inertial sensor-based gait deep neural network. IEEE Access 2020, 8, 12364–12378. [Google Scholar] [CrossRef]
- Tsinganos, P.; Cornelis, B.; Cornelis, J.; Jansen, B.; Skodras, A. Data augmentation of surface electromyography for hand gesture recognition. Sensors 2020, 20, 4892. [Google Scholar] [CrossRef] [PubMed]
- Jeong, C.Y.; Shin, H.C.; Kim, M. Sensor-data augmentation for human activity recognition with time-warping and data masking. Multimed. Tools Appl. 2021, 80, 20991–21009. [Google Scholar] [CrossRef]
- Cheng, D.; Zhang, L.; Bu, C.; Wu, H.; Song, A. Learning hierarchical time series data augmentation invariances via contrastive supervision for human activity recognition. Knowl.-Based Syst. 2023, 276, 110789. [Google Scholar] [CrossRef]
- Uchitomi, H.; Ming, X.; Zhao, C.; Ogata, T.; Miyake, Y. Classification of mild Parkinson’s disease: Data augmentation of time series gait data obtained via inertial measurement units. Sci. Rep. 2023, 13, 12638. [Google Scholar] [CrossRef]
- Guo, P.; Yang, H.; Sano, A. Empirical Study of Mix-based Data Augmentation Methods in Physiological Time Series Data. In Proceedings of the 2023 IEEE 11th International Conference on Healthcare Informatics (ICHI), Houston, TX, USA, 26–29 June 2023; pp. 206–213. [Google Scholar]
- Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–November 2019; pp. 6023–6032. [Google Scholar]
- Cauli, N.; Reforgiato Recupero, D. Survey on videos data augmentation for deep learning models. Future Internet 2022, 14, 93. [Google Scholar] [CrossRef]
- Zhang, B.; Gu, S.; Zhang, B.; Bao, J.; Chen, D.; Wen, F.; Wang, Y.; Guo, B. Styleswin: Transformer-based gan for high-resolution image generation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11304–11314. [Google Scholar]
- Zhao, Z.; Kunar, A.; Birke, R.; Chen, L.Y. Ctab-gan: Effective table data synthesizing. In Proceedings of the Asian Conference on Machine Learning, PMLR 2021, Virtual, 17–19 November 2021; pp. 97–112. [Google Scholar]
- Lou, H.; Qi, Z.; Li, J. One-dimensional data augmentation using a Wasserstein generative adversarial network with supervised signal. In Proceedings of the 2018 Chinese Control And Decision Conference (CCDC), Shenyang, China, 9–11 June 2018; pp. 1896–1901. [Google Scholar]
- Chen, G.; Zhu, Y.; Hong, Z.; Yang, Z. EmotionalGAN: Generating ECG to enhance emotion state classification. In Proceedings of the 2019 International Conference on Artificial Intelligence and Computer Science, Wuhan, China, 12–13 July 2019; pp. 309–313. [Google Scholar]
- Fons, E.; Dawson, P.; Zeng, X.j.; Keane, J.; Iosifidis, A. Adaptive weighting scheme for automatic time series data augmentation. arXiv 2021, arXiv:2102.08310. [Google Scholar]
- Huang, J.; Lin, S.; Wang, N.; Dai, G.; Xie, Y.; Zhou, J. TSE-CNN: A two-stage end-to-end CNN for human activity recognition. IEEE J. Biomed. Health Inform. 2019, 24, 292–299. [Google Scholar] [CrossRef] [PubMed]
- Oh, C.; Han, S.; Jeong, J. Time-series data augmentation based on interpolation. Procedia Comput. Sci. 2020, 175, 64–71. [Google Scholar] [CrossRef]
- Shi, W.; Fang, X.; Yang, G.; Huang, J. Human activity recognition based on multichannel convolutional neural network with data augmentation. IEEE Access 2022, 10, 76596–76606. [Google Scholar] [CrossRef]
- Le Guennec, A.; Malinowski, S.; Tavenard, R. Data augmentation for time series classification using convolutional neural networks. In Proceedings of the ECML/PKDD Workshop on Advanced Analytics and Learning on Temporal Data, Grenoble, France, 19–23 September 2022. [Google Scholar]
- Wen, T. tsaug. 2019. Available online: https://tsaug.readthedocs.io/en/stable/ (accessed on 20 March 2024).
- Olah, C. Understanding LSTM Networks. 2015. Available online: http://colah.github.io/posts/2015-08-Understanding-LSTMs/ (accessed on 16 April 2024).
- Augustinov, G.; Nisar, M.A.; Li, F.; Tabatabaei, A.; Grzegorzek, M.; Sohrabi, K.; Fudickar, S. Transformer-based recognition of activities of daily living from wearable sensor data. In Proceedings of the 7th International Workshop on Sensor-Based Activity Recognition and Artificial Intelligence, Rostock, Germany, 19–20 September 2022; pp. 1–8. [Google Scholar]
- Qian, H.; Pan, S.J.; Miao, C. Latent independent excitation for generalisable sensor-based cross-person activity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; Volume 35, pp. 11921–11929. [Google Scholar]
- Al-Qaness, M.A.; Helmi, A.M.; Dahou, A.; Elaziz, M.A. The applications of metaheuristics for human activity recognition and fall detection using wearable sensors: A comprehensive analysis. Biosensors 2022, 12, 821. [Google Scholar] [CrossRef]
Ref. | Activity Type | Modality Type | Features |
---|---|---|---|
Amjad et al. [39] | atomic | sp, sw, sg | Handcrafted |
Sargano et al. [40] | atomic | videos | Handcrafted |
Hsu et al. [41] | atomic | IMU sensor | Handcrafted |
Lagodzinski et al. [45] | composite | sg | Codebook |
Nisar et al. [34] | atomic, composite | sp, sw, sg | Codebook |
Koping et al. [47] | composite | sp, sw, sg | Codebook |
Zhang et al. [50] | atomic, composite | wifi devices | LSTM |
Bianchi et al. [51] | composite | IMUs | CNN |
Anagnostis et al. [52] | atomic | IMUs | LSTM |
Bu et al. [53] | atomic | IMUs | CNN |
Huang et al. [55] | atomic | IMUs | CNN |
Kolkar et al. [57] | composite | IMUs | CNN + GRU |
Dua et al. [58] | atomic | IMUs | CNN + GRU |
Khatun et al. [59] | atomic | IMUs | CNN + LSTM |
Nisar et al. [61] | atomic, composite | sp, sw, sg | CNN + LSTM |
Reference | Augmentation Technique | Category | Dataset | Activity Type | Accuracy Increase |
---|---|---|---|---|---|
zhang et al. [50] | Time stratching, Spectrum shifting, Scaling, Frequency filtering, Jittring | T, M | CSI data | atomic | 11% |
Rashid et al. [64] | Jittering, Scaling, Rotation and Time warping | T, M | Equipment Activity Recognition | atomic | w 97.9% |
Steven et al. [65] | Jittering, Permuting, Scaling, Rotating, Time-warping, Magnitude warping | T, M | UCI HAR | atomic | 2% |
Uchitomi et al. [73] | Rotation, Jittering, Scaling, Magnitude warping, Permutation, Time warping and Cropping | T, M | Parkinson’s disease data | atomic | w 86.4% |
huang et al. [82] | Linear interpolation. | T | WISDM | atomic | w 95.7% |
oh et al. [83] | Linear interpolation | T | 85 UCR Archive datasets | 1% | |
cheng et al. [72] | Rotation, Jittering, Scaling, Permutation, Flipping and Resampling | T, M | UCR Archive datasets | atomic | 5–10% |
shi et al. [84] | Feature window | M | WISDM and MHEALTH | atomic | 5% |
guo et al. [74] | Manifold Mixup, Cutmix, Mixup, Cutout, Rotation, Jittering, Scaling, Permutation | T, M and P | 5 UCR Archive datasets | atomic | 2% |
CogAge Dataset | Activity Type | Classes |
---|---|---|
Atomic | Postural | Standing, Sitting, Lying, Squatting, Walking, Bending |
Behavioural | Sit down, Stand up, Lie down, Get up, Squat down, Stand up from squatting, Open door, Close door, Open drawer, Close drawer, Open small box, Close small box, Open big box, Close big box, Open lid by rotation, Close lid by rotation, Open other lid, Close other lid, Open bag, Take from floor, Put on floor, Bring, Put on high position, Take from high position, Take out, Eat small thing, Drink, Scoop and put, Plug in, Unplug, Rotate, Throw out, Hang, Unhang, Wear jacket, Take off jacket, Read, Write, Type, Talk using telephone, Touch smartphone screen, Open tap water, Close tap water, Put from tap water, Put from bottle, Throw out water, Gargle, Rub hands, Dry off hands by shake, Dry off hands, Press from top, Press by grasp, Press switch/button, Clean surface, Clean floor. | |
Composite | Brushing Teeth, Cleaning Room, Handling Medication, Preparing Food, Styling Hair, Using Phone, Washing Hands |
Activity Type | Classes |
---|---|
Fall | Falling forward, Falling right, Falling backward, Falling left, Hitting obstacle, Falling with protection strategies, Falling backward sitting on chair, Syncope |
ADLs | Standing up from sitting, Standing up from lying, Walking, Running, Going upstairs, Jumping, Going downstairs, Lying down from standing, Sitting down |
Experiment | Description |
---|---|
AF-17-1 | This experiment involves classifying 17 different activities, including falls and activities of daily living (ADLs), with a data distribution of 20–10. |
AF-17-2 | This experiment falls under the classification of 17 activities, encompassing both fall and ADL classes, with a data distribution of 60–40. |
ADL-9 | This task evaluates the performance of recognising activity sequences from 9 ADL classes, with a data distribution of 60–40. |
F-8 | This task evaluates the performance of recognising activity sequences from 8 different Fall classes, using a data distribution of 60–40. |
A-17 | This experiment was conducted on a total of 17 activities. The testing was conducted by dividing into two sets: 8 Fall classes (AF8) and 9 ADL classes (AF9), with a 60–40 distribution. |
Experiment | Description |
---|---|
CC-7 | This experiment falls under the classification of 7 composite activities with an equal distribution of subject data. |
CA-61 | This experiment falls under the classification of 61 atomic activities. |
CAB-55 | This task evaluates the performance of recognising atomic activity sequences across 55 different behaviour classes. |
CAP-6 | This task evaluates the performance of recognising atomic activity sequences across 6 different posture classes. |
CA-2 | This experiment was conducted on 61 atomic activities and subsequently tested on 6 posture classes (CA-6) and 55 behaviour classes (CA-55) individually. |
Augmentation Techniques | CAP-6 | CAB-55 | CA-61 | CC-7 |
---|---|---|---|---|
Actual | 92.71 | 62.45 | 55.04 | 79.29 |
Interpolation | 99.75 | 71.53 | 76.09 | 82.43 |
Jittering | 96.84 | 74.66 | 73.34 | 80.65 |
Scaling | 99.51 | 86.58 | 80.68 | 84.12 |
Median Smoothing | 99.75 | 70.78 | 76.32 | 81.28 |
Rowling Mean Smoothing | 93.8 | 62.51 | 60.58 | 76.19 |
Time Warping | 98.54 | 92.5 | 84.73 | 82.16 |
AugMix | 91.05 | 64.04 | 60.4 | 73.2 |
Cutmix | 91.29 | 57.53 | 56.97 | 79.52 |
Hide and Seek | 89.83 | 57.01 | 51.43 | 75.24 |
Mixup | 87.89 | 59.43 | 63.18 | 78.87 |
Sequential Transformation | 90.56 | 62.11 | 63.42 | 78.23 |
TSaug | 91.29 | 50.87 | 51.26 | 80.38 |
Subject Mix | 85.67 | 63.85 | 62.74 | 78.34 |
Subject Cutmix | 23.72 | 38.23 | 40.88 | 63.53 |
Augmentation Techniques | AF-17-1 | AF-17-2 | ADL-9 | F-8 |
---|---|---|---|---|
Actual | 68.64 | 84.47 | 67.33 | 96.71 |
Interpolation | 69.27 | 80.79 | 63.83 | 93.17 |
Jittering | 70.71 | 92 | 89.44 | 98.89 |
Scaling | 71.31 | 91.63 | 88.61 | 98.81 |
Median Smoothing | 72.41 | 88.1 | 81.53 | 97.61 |
Rowling Mean Smoothing | 70.69 | 86.6 | 84.23 | 96.05 |
Time Warping | 71.3 | 88.01 | 75.32 | 97.81 |
AugMix | 71.56 | 85.97 | 86.02 | 93.14 |
Cutmix | 70.02 | 89.17 | 78.78 | 97.64 |
Hide and Seek | 70.59 | 74.84 | 71.82 | 86.11 |
Mixup | 69.76 | 79.57 | 80.03 | 65.89 |
Sequential Transformation | 68.71 | 79.86 | 64.49 | 87.95 |
TSaug | 69.79 | 85.04 | 70 | 97.15 |
Subject Mix | 68.36 | 89.22 | 80.28 | 96.69 |
Subject Cutmix | 62.33 | 78.82 | 50.32 | 85.23 |
Augmentation Techniques | AF9 | AF8 | ADL-9 | F-8 |
---|---|---|---|---|
Actual | 67.16 | 94.61 | 67.33 | 96.71 |
Interpolation | 61.29 | 91.89 | 63.83 | 93.17 |
Jittering | 83.58 | 96.79 | 89.44 | 98.89 |
Scaling | 80.86 | 97.36 | 88.61 | 98.81 |
Median Smoothing | 73.96 | 96.16 | 81.53 | 97.61 |
Rowling Mean Smoothing | 69.51 | 95.39 | 84.23 | 96.05 |
Time Warping | 60.65 | 91.75 | 75.32 | 97.81 |
AugMix | 75.18 | 92.11 | 86.02 | 93.14 |
Cutmix | 77.1 | 96.34 | 78.78 | 97.64 |
Hide and Seek | 61.84 | 81.44 | 71.82 | 86.11 |
Mixup | 68.76 | 64.67 | 80.03 | 65.89 |
Sequential Transformation | 60.67 | 91.82 | 64.49 | 87.95 |
TSaug | 68.08 | 94.98 | 70 | 97.15 |
Subject Mix | 77.74 | 96.05 | 80.28 | 96.69 |
Subject Cutmix | 56.75 | 88.32 | 50.32 | 85.23 |
Augmentation Techniques | CAP-6 | CAB-55 | CA-6 | CA-55 |
---|---|---|---|---|
Actual | 92.71 | 62.45 | 62.14 | 54.25 |
Interpolation | 99.75 | 71.53 | 79.36 | 75.53 |
Jittering | 96.84 | 74.66 | 75.24 | 73.13 |
Scaling | 99.51 | 86.58 | 71.35 | 81.7 |
Median Smoothing | 99.75 | 70.78 | 77.71 | 76.19 |
Rowling Mean Smoothing | 93.8 | 62.51 | 63.62 | 59.14 |
Time Warping | 98.54 | 92.5 | 77.42 | 85.53 |
AugMix | 91.05 | 64.04 | 59.82 | 57.41 |
Cutmix | 91.29 | 57.53 | 61.92 | 56.4 |
Hide and Seek | 89.83 | 57.01 | 52.94 | 51.26 |
Mixup | 87.89 | 59.43 | 65.32 | 62.94 |
Sequential Transformation | 90.56 | 62.11 | 66.53 | 63.07 |
TSaug | 91.29 | 50.87 | 50.32 | 55.48 |
Subject Mix | 85.67 | 63.85 | 67.99 | 62.16 |
Subject Cutmix | 23.72 | 38.23 | 39.32 | 37.88 |
Method | Year | CogAge Atomic | CogAge Composite | UniMiB-SHAR |
---|---|---|---|---|
Transformer [88] | 2022 | – | 73.36% | – |
Random Forest [39] | 2021 | – | 79% | – |
Rank pooling + SVM [34] | 2020 | – | 68.65% | – |
CNN-transfer [6] | 2020 | state—95.94%, behaviour—71.8% | – | – |
GILE [89] | 2021 | – | – | 70.31% |
Fusion [3] | 2018 | – | – | 74.66% |
CNN [72] | 2023 | – | – | 78.83% |
HM + RF [90] | 2022 | – | – | 80.27% |
Proposed Model | 2024 | state—98.54%, behaviour—92.5% | 82.16% | 88.01% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ashfaq, N.; Khan, M.H.; Nisar, M.A. Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework. Information 2024, 15, 343. https://doi.org/10.3390/info15060343
Ashfaq N, Khan MH, Nisar MA. Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework. Information. 2024; 15(6):343. https://doi.org/10.3390/info15060343
Chicago/Turabian StyleAshfaq, Nazish, Muhammad Hassan Khan, and Muhammad Adeel Nisar. 2024. "Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework" Information 15, no. 6: 343. https://doi.org/10.3390/info15060343
APA StyleAshfaq, N., Khan, M. H., & Nisar, M. A. (2024). Identification of Optimal Data Augmentation Techniques for Multimodal Time-Series Sensory Data: A Framework. Information, 15(6), 343. https://doi.org/10.3390/info15060343