Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud
Abstract
:Featured Application
Abstract
1. Introduction
- We propose an integrated framework for multi-person action recognition based on millimeter-wave radar point clouds.
- Based on the characteristics of the millimeter-wave radar point clouds, we introduce a filtering method based on the inter-frame differences of the millimeter-wave point clouds.
- Using the collected multi-millimeter-wave radar human point cloud data, we achieved an action classification accuracy of 92.2% in multi-person scenarios.
2. System Framework Design
2.1. Overview of the System Framework
2.2. Data Preprocessing
- Calculate the statistical properties of the data set, including the mean and standard deviation. The mean is the average value of the data set, and the standard deviation is the dispersion of the data set values.
- Based on the mean and standard deviation, a threshold range is defined within which the data points are considered normal, while the data points outside this range are considered outliers.
- By comparing each data point’s mean and standard deviation with the corresponding variable, the data points that fall within the threshold range are retained, while those that fall outside the range are filtered out.
- Calculate the range difference in the spatial coordinates of the point cloud between the neighboring frames: For each point’s spatial coordinates (x, y, z), calculate the difference between the neighboring frames. This results in a vector representing the variation between the frames.
- Select the median as a benchmark: Since the human body is usually at rest in the beginning and end frames, the effect of these frames on overall variation may be significant. The median of the range differences of all the neighboring frames is chosen as the benchmark to eliminate this effect. The median can reduce the effect of outliers to a certain extent and, more accurately, reflect the general trend of change between the frames.
- Flag and delete the point clouds whose range of variation exceeds the benchmark: Iterate through all the neighboring frames, flag them as noisy, and delete the points whose range of point cloud spatial coordinates exceeds the benchmark. Doing so eliminates the points whose range of variation between frames is too extensive, as they may be caused by noise or other disturbances.
Algorithm 1: Filtering Method Based on Frame Differences |
Input: frames—a list containing multiple frames Output: filtered_frames—a list of frames after filtering Calculate the range difference of x, y, z coordinates between adjacent frames FUNCTION calculate_range_difference (frame1, frame2): range1 = calculate_range (frame1) range2 = calculate_range (frame2) difference = absolute_value (range2–range1) RETURN difference Calculate the range of a frame FUNCTION calculate_range (frame): min_coords = min (frame, axis=0) max_coords = max (frame, axis=0) range = max_coords − min_coords RETURN range Calculate the differences between all adjacent frames FUNCTION frame_differences (frames): differences = [] for i from 1 to length (frames) − 1: difference = calculate_range_difference (frames[i − 1], frames[i]) differences.append (difference) RETURN differences Get the median of all adjacent frame range differences as the threshold FUNCTION get_median_threshold (differences): sorted_differences = sort (differences) median_diff = median (sorted_differences) RETURN median_diff Filter noise in point cloud data FUNCTION filter_noise (frames, threshold): filtered_frames = [] for i from 1 to length (frames) − 1: difference = calculate_range_difference (frames[i − 1], frames[i]) IF all_elements_less_or_equal (difference, threshold): filtered_frames.append (frames[i]) RETURN filtered_frames |
2.3. Target Segmentation
2.4. Data Optimization and Dimensionality Reduction
2.5. CNN + LSTM Classification Network
3. Experimental Setup
3.1. Hardware Setup and Experimental Scenario
3.2. Model Parameter Settings
4. Experimental Results Analysis
4.1. Robustness Experiment
4.2. Ablation Study
4.3. Time Cost
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Jalal, A.; Kim, Y.H.; Kim, Y.J.; Kamal, S.; Kim, D. Robust human activity recognition from depth video using spatiotemporal multi-fused features. Pattern Recognit. 2017, 61, 295–308. [Google Scholar] [CrossRef]
- Attal, F.; Mohammed, S.; Dedabrishvili, M.; Chamroukhi, F.; Oukhellou, L.; Amirat, Y. Physical human activity recognition using wearable sensors. Sensors 2015, 15, 31314–31338. [Google Scholar] [CrossRef] [PubMed]
- Chen, C.; Jafari, R.; Kehtarnavaz, N. UTD-MHAD: A multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015. [Google Scholar]
- Hao, Q.; Hu, F.; Xiao, Y. Multiple human tracking and identification with wireless distributed pyroelectric sensor systems. IEEE Syst. J. 2009, 3, 428–439. [Google Scholar] [CrossRef]
- Han, J.; Bhanu, B. Human activity recognition in thermal infrared imagery. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05)-Workshops, San Diego, CA, USA, 21–23 September 2005. [Google Scholar]
- Ma, Y.; Zhou, G.; Wang, S. WiFi sensing with channel state information: A survey. ACM Comput. Surv. (CSUR) 2019, 52, 1–36. [Google Scholar] [CrossRef]
- Li, C.; Cao, Z.; Liu, Y. Deep AI enabled ubiquitous wireless sensing: A survey. ACM Comput. Surv. (CSUR) 2021, 54, 1–35. [Google Scholar] [CrossRef]
- Wang, Y.; Wu, K.; Ni, L.M. Wifall: Device-free fall detection by wireless networks. IEEE Trans. Mob. Comput. 2016, 16, 581–594. [Google Scholar] [CrossRef]
- Wang, W.; Liu, A.X.; Shahzad, M.; Ling, K.; Lu, S. Understanding and modeling of wifi signal based human activity recognition. In Proceedings of the 21st Annual International Conference on Mobile Computing and Networking, Paris, France, 7–11 September 2015; pp. 65–76. [Google Scholar]
- Iovescu, C.; Rao, S. The Fundamentals of Millimeter Wave Sensors; Texas Instruments: Dallas, TX, USA, 2017; pp. 1–8. [Google Scholar]
- Fairchild, D.P.; Narayanan, R.M. Multistatic micro-Doppler radar for determining target orientation and activity classification. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 512–521. [Google Scholar] [CrossRef]
- Singh, A.D.; Sandha, S.S.; Garcia, L.; Srivastava, M. Radhar: Human activity recognition from point clouds generated through a millimeter-wave radar. In Proceedings of the 3rd ACM Workshop on Millimeter-Wave Networks and Sensing Systems, Los Cabos, Mexico, 25 October 2019; pp. 51–56. [Google Scholar]
- Gong, P.; Wang, C.; Zhang, L. Mmpoint-gnn: Graph neural network with dynamic edges for human activity recognition through a millimeter-wave radar. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–7. [Google Scholar]
- Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [PubMed]
- Kong, L.; Khan, M.K.; Wu, F.; Chen, G.; Zeng, P. Millimeter-wave wireless communications for IoT-cloud supported autonomous vehicles: Overview, design, and challenges. IEEE Commun. Mag. 2017, 55, 62–68. [Google Scholar] [CrossRef]
- Heath, R.W.; Gonzalez-Prelcic, N.; Rangan, S.; Roh, W.; Sayeed, A.M. An overview of signal processing techniques for millimeter wave MIMO systems. IEEE J. Sel. Top. Signal Process. 2016, 10, 436–453. [Google Scholar] [CrossRef]
- Schubert, E.; Sander, J.; Ester, M.; Kriegel, H.P.; Xu, X. DBSCAN revisited, revisited: Why and how you should (still) use DBSCAN. ACM Trans. Database Syst. (TODS) 2017, 42, 1–21. [Google Scholar] [CrossRef]
- Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
- Khan, M.; Ahmad, J.; El Saddik, A.; Gueaieb, W.; De Masi, G.; Karray, F. Drone-HAT: Hybrid Attention Transformer for Complex Action Recognition in Drone Surveillance Videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 4713–4722. [Google Scholar]
- Mutegeki, R.; Han, D.S. A CNN-LSTM approach to human activity recognition. In Proceedings of the 2020 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Fukuoka, Japan, 19–21 February 2020; pp. 362–366. [Google Scholar]
- Han, X.-F.; Jin, J.S.; Wang, M.-J.; Jiang, W.; Gao, L.; Xiao, L. A review of algorithms for filtering the 3D point cloud. Signal Process. Image Commun. 2017, 57, 103–112. [Google Scholar] [CrossRef]
- Palipana, S.; Salami, D.; Leiva, L.A.; Sigg, S. Pantomime: Mid-air gesture recognition with sparse millimeter-wave radar point clouds. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2021, 5, 1–27. [Google Scholar] [CrossRef]
- Bevilacqua, A.; MacDonald, K.; Rangarej, A.; Widjaya, V.; Caulfield, B.; Kechadi, T. Human activity recognition with convolutional neural networks. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2018, Dublin, Ireland, September 10–14, 2018, Proceedings, Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; pp. 541–552. [Google Scholar]
- Pienaar, S.W.; Malekian, R. Human activity recognition using LSTM-RNN deep neural network architecture. In Proceedings of the 2019 IEEE 2nd Wireless Africa Conference (WAC), Pretoria, South Africa, 18–20 August 2019; pp. 1–5. [Google Scholar]
- Duran, B.S.; Odell, P.L. Cluster Analysis: A Survey; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013; Volume 100. [Google Scholar]
- Alex, S. Fundamentals of recurrent neural network (RNN) and long short-term memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar]
- Wang, F.; Tax, D.M. Survey on the attention based RNN model and its applications in computer vision. arXiv 2016, arXiv:1601.06823. [Google Scholar]
- Abdu, F.J.; Zhang, Y.; Fu, M.; Li, Y.; Deng, Z. Application of deep learning on millimeter-wave radar signals: A review. Sensors 2021, 21, 1951. [Google Scholar] [CrossRef] [PubMed]
- Texas Instruments. AWR1843BOOST and IWR1843BOOST Single-Chip mmWave Sensing Solution User’s Guide (Rev. B). 2020. Available online: https://www.ti.com/tool/IWR1843BOOST (accessed on 23 October 2023).
- Mishra, K.V.; Shankar, M.B.; Koivunen, V.; Ottersten, B.; Vorobyov, S.A. Toward millimeter-wave joint radar communications: A signal processing perspective. IEEE Signal Process. Mag. 2019, 36, 100–114. [Google Scholar] [CrossRef]
- Yujiri, L.; Shoucri, M.; Moffa, P. Passive millimeter wave imaging. IEEE Microw. Mag. 2003, 4, 39–50. [Google Scholar] [CrossRef]
- Zhang, Q.; Wang, C.; Wu, H.; Xin, C.; Phuong, T.V. GELU-Net: A Globally Encrypted, Locally Unencrypted Deep Neural Network for Privacy-Preserved Learning. In Proceedings of the International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 3933–3939. [Google Scholar]
- Xia, K.; Huang, J.; Wang, H. LSTM-CNN architecture for human activity recognition. IEEE Access 2020, 8, 56855–56866. [Google Scholar] [CrossRef]
Equipment Type | Resolution | Multiplayer | Privacy Protection | Affected by Weather |
---|---|---|---|---|
camera equipment | highest | yes | no | yes |
wearable equipment | high | no | yes | no |
infrared equipment | low | yes | yes | no |
millimeter-wave radar | high | yes | yes | no |
Activity | Time/ms |
---|---|
Boxing | 62 |
Leg lift | 75 |
Bowing | 67 |
Squat down | 69 |
Stand up | 59 |
Activity | Time/ms |
---|---|
Boxing | 756 |
Leg lift | 893 |
Bowing | 844 |
Squat down | 747 |
Stand up | 689 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Dang, X.; Fan, K.; Li, F.; Tang, Y.; Gao, Y.; Wang, Y. Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud. Appl. Sci. 2024, 14, 7253. https://doi.org/10.3390/app14167253
Dang X, Fan K, Li F, Tang Y, Gao Y, Wang Y. Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud. Applied Sciences. 2024; 14(16):7253. https://doi.org/10.3390/app14167253
Chicago/Turabian StyleDang, Xiaochao, Kai Fan, Fenfang Li, Yangyang Tang, Yifei Gao, and Yue Wang. 2024. "Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud" Applied Sciences 14, no. 16: 7253. https://doi.org/10.3390/app14167253
APA StyleDang, X., Fan, K., Li, F., Tang, Y., Gao, Y., & Wang, Y. (2024). Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud. Applied Sciences, 14(16), 7253. https://doi.org/10.3390/app14167253