Exploring Trends and Clusters in Human Posture Recognition Research: An Analysis Using CiteSpace
Abstract
:1. Introduction
2. Methods
2.1. CiteSpace as the Main Extraction and Screening Tool
Validation Through Multiple Metrics in CiteSpace
2.2. PRISMA as a Major Direction of Study Checklist and Guidance
3. Data Collection and Results
3.1. Screening Criteria
- Relevance to Human Posture Recognition: The study must focus on human posture recognition, motion capture, or activity recognition using either sensor-based or vision-based methods.
- Technological Focus: Articles focusing on hardware (sensors, IMUs, etc.) or software (deep learning, machine learning) innovations in posture detection were prioritized.
- Exclusion of Survey or Review Papers: Papers that were purely literature reviews or meta-analyses without original research data were excluded.
- Exclusion of Articles Lacking Experimental Data: Studies that did not present specific experimental setups or lacked detailed data collection methods were removed from the dataset.
- Research Novelty: Priority was given to papers that presented innovative methodologies or introduced new technologies in human posture recognition.
3.2. Statistics on the Total Number of Articles Collected
3.2.1. Logarithmic Analysis
3.2.2. Logistic Model Analysis
- is the cumulative number of publications at time;
- represents the upper limit or the maximum number of publications that the field can support (as growth slows down);
- and are constants that define the curve’s shape and the rate of growth; and
- is time (in years).
- is the actual number of publications in year ; and
- is the predicted number of publications for the same year from the fitted logistic model.
3.2.3. The Distribution of Disciplines in the Research Literature on Human Pose Recognition
3.3. Overlay Maps Analysis
3.4. Analysis of Co-Authorship Between Institutions
3.5. Analysis of Co-Authorship Between Countries and Regions
3.6. Co-Cited Literature Clustering Screening
- Opening Pruning, Pruning sliced networks and Pruning the Merged Network. Pathfinder is to simplify the network and highlight its important structural features. The advantage of Pathfinder is completeness (unique solution), but MST (Minimum Spanning Tree) does not [22].
- Title words are used as the cluster representation, and the figure shows the labels under the LSI algorithm.
- The screening period is 2014–2024 (through all data deletion, it is found that the research is mainly concentrated after 2015, and the early research has little influence on the clustering formation, so the selection here is just divided into 10 slices to specifically study the results of a decade).
Major Clusters
3.7. Supplementary Articles After Cluster Screening
4. Content Analysis of the Filtered Articles
4.1. Trend of Selected Articles
4.2. Installation Position and Use Environment of Sensor
4.2.1. Installation Position
4.2.2. Applicable Environment
4.3. Algorithms and Identification Types
4.4. Computational Complexity and Accuracy of the Algorithm
4.4.1. Deep Learning Algorithm
4.4.2. Statistical Models
4.4.3. Traditional Machine Learning Algorithm
4.4.4. Geometric Transformation and Coordinate System Transformation
4.4.5. Meta-Heuristic Search Algorithm
4.4.6. The Relationship Between Hardware and Deep Learning
4.5. Selection of Some Articles
5. Discussion
5.1. Limitations of CiteSpace and Its Implications
5.2. Recent Advancements in Human Posture Recognition
5.2.1. Sensor-Based Methods
5.2.2. Vision-Based Methods
5.3. Status of Interdisciplinary Cooperation in the Field
5.4. Research Trends and Relevance of Industry Practices and Commercial Products
6. Results
6.1. Summary of Characteristics of Each Sensor
6.2. Summary of Algorithm Characteristics
7. Conclusions
Funding
Conflicts of Interest
References
- Liu, H.; Li, D.; Wang, X.; Liu, L.; Zhang, Z.; Subramanian, S. Precise Head Pose Estimation on HPD5A Database for Attention Recognition Based on Convolutional Neural Network in Human-Computer Interaction. Infrared Phys. Technol. 2021, 116, 103740. [Google Scholar] [CrossRef]
- Brulin, D.; Benezeth, Y.; Courtial, E. Posture Recognition Based on Fuzzy Logic for Home Monitoring of the Elderly. IEEE Trans. Inform. Technol. Biomed. 2012, 16, 974–982. [Google Scholar] [CrossRef] [PubMed]
- Liu, Y.; Nie, L.; Liu, L.; Rosenblum, D.S. From Action to Activity: Sensor-Based Activity Recognition. Neurocomputing 2016, 181, 108–115. [Google Scholar] [CrossRef]
- Hassan, M.M.; Huda, S.; Uddin, M.Z.; Almogren, A.; Alrubaian, M. Human Activity Recognition from Body Sensor Data Using Deep Learning. J. Med. Syst. 2018, 42, 99. [Google Scholar] [CrossRef]
- Qiu, S.; Zhao, H.; Jiang, N.; Wang, Z.; Liu, L.; An, Y.; Zhao, H.; Miao, X.; Liu, R.; Fortino, G. Multi-Sensor Information Fusion Based on Machine Learning for Real Applications in Human Activity Recognition: State-of-the-Art and Research Challenges. Inf. Fusion 2022, 80, 241–265. [Google Scholar] [CrossRef]
- Clever, H.M.; Erickson, Z.; Kapusta, A.; Turk, G.; Liu, C.K.; Kemp, C.C. Bodies at Rest: 3D Human Pose and Shape Estimation from a Pressure Image Using Synthetic Data. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; IEEE Computer Society: Los Alamitos, CA, USA, 2020; pp. 6214–6223. [Google Scholar]
- Wan, Q.; Zhao, H.; Li, J.; Xu, P. Hip Positioning and Sitting Posture Recognition Based on Human Sitting Pressure Image. Sensors 2021, 21, 426. [Google Scholar] [CrossRef]
- Hu, F.; Wang, L.; Wang, S.; Liu, X.; He, G. A Human Body Posture Recognition Algorithm Based on BP Neural Network for Wireless Body Area Networks. China Commun. 2016, 13, 198–208. [Google Scholar] [CrossRef]
- Gjoreski, H.; Lustrek, M.; Gams, M. Accelerometer Placement for Posture Recognition and Fall Detection. In Proceedings of the 2011 Seventh International Conference on Intelligent Environments, Nottingham, UK, 25–28 July 2011; IEEE: Nottingham, UK, 2011; pp. 47–54. [Google Scholar]
- Zmitri, M.; Fourati, H.; Vuillerme, N. Vuillerme Human Activities and Postures Recognition: From Inertial Measurements to Quaternion-Based Approaches. Sensors 2019, 19, 4058. [Google Scholar] [CrossRef]
- Zhang, D.; Wu, Y.; Guo, M.; Chen, Y. Deep Learning Methods for 3D Human Pose Estimation under Different Supervision Paradigms: A Survey. Electronics 2021, 10, 2267. [Google Scholar] [CrossRef]
- Gong, W.; Zhang, X.; Gonzàlez, J.; Sobral, A.; Bouwmans, T.; Tu, C.; Zahzah, E. Human Pose Estimation from Monocular Images: A Comprehensive Survey. Sensors 2016, 16, 1966. [Google Scholar] [CrossRef]
- Chen, C. CiteSpace II: Detecting and Visualizing Emerging Trends and Transient Patterns in Scientific Literature. J. Am. Soc. Inf. Sci. 2006, 57, 359–377. [Google Scholar] [CrossRef]
- Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. BMJ-Br. Med. J. 2021, 372, n71. [Google Scholar] [CrossRef] [PubMed]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
- Bornmann, L.; Mutz, R. Growth Rates of Modern Science: A Bibliometric Analysis Based on the Number of Publications and Cited References. J. Assoc. Inf. Sci. Technol. 2015, 66, 2215–2222. [Google Scholar] [CrossRef]
- Nelis, J.L.D.; da Silva, G.R.; Ortuño, J.; Tsagkaris, A.S.; Borremans, B.; Haslova, J.; Colgrave, M.L.; Elliott, C.T. The General Growth Tendency: A Tool to Improve Publication Trend Reporting by Removing Record Inflation Bias and Enabling Quantitative Trend Analysis. PLoS ONE 2022, 17, e0268433. [Google Scholar] [CrossRef]
- Thelwall, M.; Sud, P. Scopus 1900–2020: Growth in Articles, Abstracts, Countries, Fields, and Journals. Quant. Sci. Stud. 2022, 3, 37–50. [Google Scholar] [CrossRef]
- Gu, X.; Blackmore, K.L. Recent Trends in Academic Journal Growth. Scientometrics 2016, 108, 693–716. [Google Scholar] [CrossRef]
- Chen, C.; Leydesdorff, L. Patterns of Connections and Movements in Dual-map Overlays: A New Method of Publication Portfolio Analysis. J. Assoc. Inf. Sci. Technol. 2014, 65, 334–351. [Google Scholar] [CrossRef]
- Larsen, P.O.; von Ins, M. The Rate of Growth in Scientific Publication and the Decline in Coverage Provided by Science Citation Index. Scientometrics 2010, 84, 575–603. [Google Scholar] [CrossRef]
- Chen, C.; Morris, S. Visualizing Evolving Networks: Minimum Spanning Trees versus Pathfinder Networks. In Proceedings of the IEEE Symposium on Information Visualization 2003 (IEEE Cat. No.03TH8714), Seattle, WA, USA, 19–21 October 2003; pp. 67–74. [Google Scholar]
- Chen, C.; Ibekwe-SanJuan, F.; Hou, J. The Structure and Dynamics of Cocitation Clusters: A Multiple-perspective Cocitation Analysis. J. Am. Soc. Inf. Sci. 2010, 61, 1386–1409. [Google Scholar] [CrossRef]
- Madokoro, H.; Nix, S.; Woo, H.; Sato, K. A Mini-Survey and Feasibility Study of Deep-Learning-Based Human Activity Recognition from Slight Feature Signals Obtained Using Privacy-Aware Environmental Sensors. Appl. Sci. 2021, 11, 11807. [Google Scholar] [CrossRef]
- Ezquerra, A.; Agen, F.; Toma, R.B.; Ezquerra-Romano, I. Using Facial Emotion Recognition to Research Emotional Phases in an Inquiry-Based Science Activity. Res. Sci. Technol. Educ. 2023, 1–24. [Google Scholar] [CrossRef]
- Ryu, J.; McFarland, T.; Banting, B.; Haas, C.T.; Abdel-Rahman, E. Health and Productivity Impact of Semi-Automated Work Systems in Construction. Autom. Constr. 2020, 120, 103396. [Google Scholar] [CrossRef]
- Pervaiz, M.; Ghadi, Y.Y.; Gochoo, M.; Jalal, A.; Kamal, S.; Kim, D.-S. A Smart Surveillance System for People Counting and Tracking Using Particle Flow and Modified SOM. Sustainability 2021, 13, 5367. [Google Scholar] [CrossRef]
- Fridriksdottir, E.; Bonomi, A.G. Accelerometer-Based Human Activity Recognition for Patient Monitoring Using a Deep Neural Network. Sensors 2020, 20, 6424. [Google Scholar] [CrossRef]
- Vinciarelli, A.; Pantic, M.; Heylen, D.; Pelachaud, C.; Poggi, I.; D’Errico, F.; Schroeder, M. Bridging the Gap between Social Animal and Unsocial Machine: A Survey of Social Signal Processing. IEEE Trans. Affective Comput. 2012, 3, 69–87. [Google Scholar] [CrossRef]
- R Core Team. R: A Language and Environment for Statistical Computing; Foundation for Statistical Computing: Vienna, Austria, 2013. [Google Scholar]
- Clark, R.A.; Bower, K.J.; Mentiplay, B.F.; Paterson, K.; Pua, Y.-H. Concurrent Validity of the Microsoft Kinect for Assessment of Spatiotemporal Gait Variables. J. Biomech. 2013, 46, 2722–2725. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Cao, Z.; Simon, T.; Wei, S.-E.; Sheikh, Y. Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Honolulu, HI, USA, 2017; pp. 1302–1310. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA, 2016; pp. 770–778. [Google Scholar]
- Robert-Lachaine, X.; Mecheri, H.; Larue, C.; Plamondon, A. Validation of Inertial Measurement Units with an Optoelectronic System for Whole-Body Motion Analysis. Med. Biol. Eng. Comput. 2017, 55, 609–619. [Google Scholar] [CrossRef]
- Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.-E.; Sheikh, Y. OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 172–186. [Google Scholar] [CrossRef]
- Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep Learning for Sensor-Based Activity Recognition: A Survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef]
- Wang, P.; Li, W.; Ogunbona, P.; Wan, J.; Escalera, S. RGB-D-Based Human Motion Recognition with Deep Learning: A Survey. Comput. Vis. Image Underst. 2018, 171, 118–139. [Google Scholar] [CrossRef]
- Kim, W.; Sung, J.; Saakes, D.; Huang, C.; Xiong, S. Ergonomic Postural Assessment Using a New Open-Source Human Pose Estimation Technology (OpenPose). Int. J. Ind. Ergon. 2021, 84, 103164. [Google Scholar] [CrossRef]
- Hu, Q.; Bai, Y.; He, L.; Huang, J.; Wang, H.; Cheng, G. Workers’ Unsafe Actions When Working at Heights: Detecting from Images. Sustainability 2022, 14, 6126. [Google Scholar] [CrossRef]
- Olivas-Padilla, B.E.; Manitsaris, S.; Menychtas, D.; Glushkova, A. Stochastic-Biomechanic Modeling and Recognition of Human Movement Primitives, in Industry, Using Wearables. Sensors 2021, 21, 2497. [Google Scholar] [CrossRef]
- Antwi-Afari, M.F.; Qarout, Y.; Herzallah, R.; Anwer, S.; Umer, W.; Zhang, Y.; Manu, P. Deep Learning-Based Networks for Automated Recognition and Classification of Awkward Working Postures in Construction Using Wearable Insole Sensor Data. Autom. Constr. 2022, 136, 104181. [Google Scholar] [CrossRef]
- Lee, W.; Seto, E.; Lin, K.-Y.; Migliaccio, G.C. An Evaluation of Wearable Sensors and Their Placements for Analyzing Construction Worker’s Trunk Posture in Laboratory Conditions. Appl. Ergon. 2017, 65, 424–436. [Google Scholar] [CrossRef]
- Carreira, J.; Zisserman, A. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Honolulu, HI, USA, 2017; pp. 4724–4733. [Google Scholar]
- Minh Dang, L.; Min, K.; Wang, H.; Jalil Piran, M.; Hee Lee, C.; Moon, H. Sensor-Based and Vision-Based Human Activity Recognition: A Comprehensive Survey. Pattern Recognit. 2020, 108, 107561. [Google Scholar] [CrossRef]
- Muangprathub, J.; Sriwichian, A.; Wanichsombat, A.; Kajornkasirat, S.; Nillaor, P.; Boonjing, V. A Novel Elderly Tracking System Using Machine Learning to Classify Signals from Mobile and Wearable Sensors. Int. J. Environ. Res. Public Health 2021, 18, 12652. [Google Scholar] [CrossRef]
- Joo, H.; Kim, H.; Ryu, J.-K.; Ryu, S.; Lee, K.-M.; Kim, S.-C. Estimation of Fine-Grained Foot Strike Patterns with Wearable Smartwatch Devices. Int. J. Environ. Res. Public Health 2022, 19, 1279. [Google Scholar] [CrossRef]
- Russell, B.; McDaid, A.; Toscano, W.; Hume, P. Moving the Lab into the Mountains: A Pilot Study of Human Activity Recognition in Unstructured Environments. Sensors 2021, 21, 654. [Google Scholar] [CrossRef] [PubMed]
- Dua, N.; Singh, S.N.; Semwal, V.B. Multi-Input CNN-GRU Based Human Activity Recognition Using Wearable Sensors. Computing 2021, 103, 1461–1478. [Google Scholar] [CrossRef]
- Milosevic, B.; Leardini, A.; Farella, E. Kinect and Wearable Inertial Sensors for Motor Rehabilitation Programs at Home: State of the Art and an Experimental Comparison. Biomed. Eng. Online 2020, 19, 25. [Google Scholar] [CrossRef] [PubMed]
- Beshara, P.; Chen, J.F.; Read, A.C.; Lagadec, P.; Wang, T.; Walsh, W.R. The Reliability and Validity of Wearable Inertial Sensors Coupled with the Microsoft Kinect to Measure Shoulder Range-of-Motion. Sensors 2020, 20, 7238. [Google Scholar] [CrossRef]
- Liu, C.-H.; Lee, P.; Chen, Y.-L.; Yen, C.-W.; Yu, C.-W. Study of Postural Stability Features by Using Kinect Depth Sensors to Assess Body Joint Coordination Patterns. Sensors 2020, 20, 1291. [Google Scholar] [CrossRef]
- Ma, Y.; Liu, D.; Cai, L. Deep Learning-Based Upper Limb Functional Assessment Using a Single Kinect v2 Sensor. Sensors 2020, 20, 1903. [Google Scholar] [CrossRef]
- Zhao, J.; Obonyo, E. Convolutional Long Short-Term Memory Model for Recognizing Construction Workers’ Postures from Wearable Inertial Measurement Units. Adv. Eng. Inform. 2020, 46, 101177. [Google Scholar] [CrossRef]
- Clark, R.A.; Pua, Y.-H.; Oliveira, C.C.; Bower, K.J.; Thilarajah, S.; McGaw, R.; Hasanki, K.; Mentiplay, B.F. Reliability and Concurrent Validity of the Microsoft Xbox One Kinect for Assessment of Standing Balance and Postural Control. Gait Posture 2015, 42, 210–213. [Google Scholar] [CrossRef]
- Akhavian, R.; Behzadan, A.H. Smartphone-Based Construction Workers’ Activity Recognition and Classification. Autom. Constr. 2016, 71, 198–209. [Google Scholar] [CrossRef]
- Xu, X.; Robertson, M.; Chen, K.B.; Lin, J.; McGorry, R.W. Using the Microsoft KinectTM to Assess 3-D Shoulder Kinematics during Computer Use. Appl. Ergon. 2017, 65, 418–423. [Google Scholar] [CrossRef]
- Xu, X.; McGorry, R.W. The Validity of the First and Second Generation Microsoft KinectTM for Identifying Joint Center Locations during Static Postures. Appl. Ergon. 2015, 49, 47–54. [Google Scholar] [CrossRef] [PubMed]
- Schall, M.C.; Fethke, N.B.; Chen, H.; Oyama, S.; Douphrate, D.I. Accuracy and Repeatability of an Inertial Measurement Unit System for Field-Based Occupational Studies. Ergonomics 2016, 59, 591–602. [Google Scholar] [CrossRef]
- Alwasel, A.; Sabet, A.; Nahangi, M.; Haas, C.T.; Abdel-Rahman, E. Identifying Poses of Safe and Productive Masons Using Machine Learning. Autom. Constr. 2017, 84, 345–355. [Google Scholar] [CrossRef]
- Mehta, D.; Sridhar, S.; Sotnychenko, O.; Rhodin, H.; Shafiei, M.; Seidel, H.-P.; Xu, W.; Casas, D.; Theobalt, C. VNect: Real-Time 3D Human Pose Estimation with a Single RGB Camera. ACM Trans. Graph. 2017, 36, 1–14. [Google Scholar] [CrossRef]
- Morrow, M.M.B.; Lowndes, B.; Fortune, E.; Kaufman, K.R.; Hallbeck, M.S. Validation of Inertial Measurement Units for Upper Body Kinematics. J. Appl. Biomech. 2017, 33, 227–232. [Google Scholar] [CrossRef] [PubMed]
- Bolink, S.A.A.N.; Naisas, H.; Senden, R.; Essers, H.; Heyligers, I.C.; Meijer, K.; Grimm, B. Validity of an Inertial Measurement Unit to Assess Pelvic Orientation Angles during Gait, Sit–Stand Transfers and Step-up Transfers: Comparison with an Optoelectronic Motion Capture System*. Med. Eng. Phys. 2016, 38, 225–231. [Google Scholar] [CrossRef]
- Jalal, A.; Kim, Y.-H.; Kim, Y.-J.; Kamal, S.; Kim, D. Robust Human Activity Recognition from Depth Video Using Spatiotemporal Multi-Fused Features. Pattern Recognit. 2017, 61, 295–308. [Google Scholar] [CrossRef]
- Jalal, A.; Khalid, N.; Kim, K. Automatic Recognition of Human Interaction via Hybrid Descriptors and Maximum Entropy Markov Model Using Depth Sensors. Entropy 2020, 22, 817. [Google Scholar] [CrossRef]
- Jalal, A.; Akhtar, I.; Kim, K. Human Posture Estimation and Sustainable Events Classification via Pseudo-2D Stick Model and K-Ary Tree Hashing. Sustainability 2020, 12, 9814. [Google Scholar] [CrossRef]
- Jalal, A.; Batool, M.; Kim, K. Sustainable Wearable System: Human Behavior Modeling for Life-Logging Activities Using K-Ary Tree Hashing Classifier. Sustainability 2020, 12, 10324. [Google Scholar] [CrossRef]
- Zhu, H.; Samtani, S.; Chen, H.; Nunamaker, J.F. Human Identification for Activities of Daily Living: A Deep Transfer Learning Approach. J. Manag. Inform. Syst. 2020, 37, 457–483. [Google Scholar] [CrossRef]
- Ordóñez, F.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed]
- Ignatov, A. Real-Time Human Activity Recognition from Accelerometer Data Using Convolutional Neural Networks. Appl. Soft Comput. 2018, 62, 915–922. [Google Scholar] [CrossRef]
- Arif, M.; Kattan, A. Physical Activities Monitoring Using Wearable Acceleration Sensors Attached to the Body. PLoS ONE 2015, 10, e0130851. [Google Scholar] [CrossRef]
- Tahir, S.B.U.D.; Jalal, A.; Kim, K. Wearable Inertial Sensors for Daily Activity Analysis Based on Adam Optimization and the Maximum Entropy Markov Model. Entropy 2020, 22, 579. [Google Scholar] [CrossRef]
- Jalal, A.; Batool, M.; Kim, K. Stochastic Recognition of Physical Activity and Healthcare Using Tri-Axial Inertial Wearable Sensors. Appl. Sci. 2020, 10, 7122. [Google Scholar] [CrossRef]
- Jalal, A.; Quaid, M.A.K.; Tahir, S.B.U.D.; Kim, K. A Study of Accelerometer and Gyroscope Measurements in Physical Life-Log Activities Detection Systems. Sensors 2020, 20, 6670. [Google Scholar] [CrossRef]
- Sun, P.; Wen, Y.; Han, R.; Feng, W.; Yan, S. GradientFlow: Optimizing Network Performance for Large-Scale Distributed DNN Training. IEEE Trans. Big Data 2022, 8, 495–507. [Google Scholar] [CrossRef]
- Yi, X.; Zhou, Y.; Xu, F. TransPose: Real-Time 3D Human Translation and Pose Estimation with Six Inertial Sensors. ACM Trans. Graph. 2021, 40, 86:1–86:13. [Google Scholar] [CrossRef]
- Li, J.; Bian, S.; Xu, C.; Chen, Z.; Yang, L.; Lu, C. HybrIK-X: Hybrid Analytical-Neural Inverse Kinematics for Whole-Body Mesh Recovery. arXiv 2023, arXiv:2304.05690. [Google Scholar] [CrossRef]
- Toshev, A.; Szegedy, C. DeepPose: Human Pose Estimation via Deep Neural Networks. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; IEEE: Columbus, OH, USA, 2014; pp. 1653–1660. [Google Scholar]
- Tompson, J.; Jain, A.; LeCun, Y.; Bregler, C. Joint Training of a Convolutional Network and a Graphical Model for Human Pose Estimation. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; Neural Information Processing Systems (nips): La Jolla, CA, USA, 2014; Volume 27. [Google Scholar]
Count | Centrality | Institutions |
---|---|---|
41 | 0.27 | University of California System |
39 | 0.08 | University of London |
25 | 0.36 | Centre National de la Recherche Scientifique (CNRS) |
22 | 0.02 | Max Planck Society |
20 | 0.32 | University College London |
12 | 0.1 | Harvard University |
12 | 0.14 | State University System of Florida |
11 | 0.02 | University of Toronto |
10 | 0.05 | University of Texas System |
9 | 0.04 | Istituto Italiano di Tecnologia—IIT |
Count | Centrality | Countries and Regions |
---|---|---|
345 | 0 | USA |
153 | 0.09 | ENGLAND |
130 | 0 | PEOPLES R CHINA |
114 | 0.05 | GERMANY |
92 | 0 | ITALY |
88 | 0.05 | CANADA |
70 | 0.05 | AUSTRALIA |
63 | 0.11 | FRANCE |
49 | 0.09 | SOUTH KOREA |
48 | 0.22 | SPAIN |
Cluster ID | Size (Cited) | Size (Citing) | Silhouette | Label (LSI) | Label (LLR) | Label (MI) | Average Year |
---|---|---|---|---|---|---|---|
1 | 15 | 15 | 0.973 | inquiry-based science activity | using applying facial emotion recognition (24.83, 1 × 10−4) | deep locality-preserving learning (0.79) | 2017 |
2 | 15 | 8 | 1 | new open-source | new open-source (26.48, 1 × 10−4) | new open-source (26.48, 1 × 10−4) | 2018 |
3 | 15 | 9 | 0.844 | human activity recognition | human activity recognition (21.13, 1 × 10−4) | using convolutional neural network (1.27) | 2017 |
4 | 14 | 10 | 0.923 | semi-automated work system | semi-automated work system (27.27, 1 × 10−4) | Microsoft Kinect (0.32) | 2016 |
5 | 10 | 2 | 1 | sustainable video surveillance system | locomotion (10.51, 0.005) | som (0.06) | 2019 |
12 | 7 | 6 | 0.95 | learning approach | deep transfer (22.2, 1 × 10−4) | gyroscope measurement (0.54) | 2015 |
Metrics | Remarks | Added Articles |
---|---|---|
Citation counts | High citation count signifies wide influence and recognition in the field | The top 10 articles in this category have been included in all five clusters. |
Burst | Identifies key articles experiencing sudden increases in citation, indicating emerging trends or pivotal research breakthroughs | id4b |
Degree | Articles with high degree values, showcasing strong connectivity and influence in the citation network | id7d, id8d, id9d |
Centrality | Reflects articles acting as critical bridges in the network, connecting various research topics | The top 10 articles in this category have been included in all five clusters. |
Sigma | Combines centrality and novelty, highlighting innovative and impactful studies | The top 10 articles in this category have been included in all five clusters. |
Category | Sensor or Input | Sensor Location | Sampling Rate |
---|---|---|---|
id2(1) [24] | fisheye lens camera (RealSenseT265) | In a meeting corner of the room | 120 Hz (120 fps) |
id2(2) [39] | ① 3 FLIR Blackfly S cameras ② Microsoft Kinect v2 ③ Xsens MVN Link inertial MoCap system | ① Located directly in front of the participant and 40 degrees in front of the left and right (camera calibration is required) ② 2.5 m directly in front of the participant ③ Located in 17 joints of the human body | ① 10 Hz ② 30 Hz ③ 240 Hz (Xsens system) |
id2(3) [40] | RGB image | N/A | N/A |
id2(4) [41] | BioMed bundle motion capture system (52 IMUs) | Spine, arms, legs | 90 Hz |
id2(7) [42] | OpenGo system (16 capacitive pressure sensors, a 3-axis gyroscope, and a 3-axis accelerometer) | Foot | 50 Hz |
id2(1)e [33] | RGB image | N/A | N/A |
id2(2)e [32] | RGB image | N/A | N/A |
id2(3)e [43] | ① ActiGraph GT9X Link ② Zephyr BioHarness™3 | ① Head, shoulder, center-waist, Non-dominant-side waist ② Chest, under armpit | 100 Hz |
id2(5)e [44] | Video | N/A | N/A |
id2(12)e [45] | RGB-D image | N/A | N/A |
id3(3) [46] | GPS module and GY-521 module (including angular acceleration measurement, angular velocity measurement, angle measurement, and temperature) | Right hip | Wifi reporting frequency: 1 Hz |
id3(5) [47] | Smartwatch FTW6024 by Fossil (Acceleration and angular velocity sensors) | Ankle | 50 Hz |
id3(10) [48] | BioHarness (3-axis acceleration) Garmin Forerunner GPS watch | Around the chest | BioHarness (100 Hz) Watch (1 Hz) |
id3(11)e [49] | Gyroscopes and accelerometers | N/A | N/A |
id4(2) [50] | ① Kinect v2 ② Wearable IMU (EXLs3 by Exel srl, Italy) | ① Was placed 3.5 m in front of the subject ② Thorax and of left thigh and shank | ① Kinect v2 30 Hz ② 100 Hz |
id4(3) [51] | Kinect v2 Three-axis gyroscope and three-axis accelerometer | Wrists and ankles | N/A |
id4(4) [52] | Kinect V2 | 2 m away in front of the subjects | 30 Hz |
id4(5) [53] | Kinect V2 | N/A | 30 Hz |
id4(6) [54] | Accelerometer and gyroscope | Multiple joint nodes | 20–90 Hz (40 Hz was used as a cut-off to remove lower-frequency window) |
id4(1)e [35] | ① Xsens ② Eight-camera Optotrak system | Located in 17 joints of the human body | ① 240 Hz ② 30 Hz |
id4(2)e [55] | ① Kinect V2 ② 3DMA system | ① 2.5 m directly in front of the participant | ① 30 Hz ② 100 Hz |
id4(3)e [56] | Accelerometer and gyroscope sensors built in smartphones | Upper arm | 100 Hz |
id4(4)e [57] | Kinect V2 | Optical axis parallel to participants’ sagittal plane (front), 15° to the left side of the participants, or 30° to the left of the participants | 30 Hz (Spline interpolation upsamples data to 60 Hz) |
id4(6)e [58] | ① Kinect V1 ② Kinect v2 | Directly in front of the subject | 60 Hz |
id4(7)e [59] | ① Accelerometer and gyroscope ② OMC system | N/A | 20 Hz ② 80 Hz (Linear interpolation down sampling to 20 Hz) |
id4(8)e [60] | Xsens MVN | The position of the center of 28 body joints | 100 Hz |
id4(9)e [61] | RGB image | N/A | N/A |
id4(11)e [62] | Gyroscopes, magnetometers, and accelerometers | Head, chest, bilateral upper arms and forearms | 80 Hz |
id4(12)e [63] | ① Accelerometers, gyroscopes, and magnetometers ② MOCAP (6 VICONMX-3+ and 2 VIcON MX-T20 cameras and one Kistler 9281A pressure plate) | Pelvis | ① 100 Hz ② 200 Hz |
id5(1)e [64] | Depth camera | N/A | N/A |
id5(3)e [65] | Depth Sensors | N/A | N/A |
id5(5)e [66] | RGB image | N/A | N/A |
id5(10)e [67] | Three IMU sensors (3-axis accelerometer, 3-axis gyroscope and 3-axis magnetometer) and an NRFL0 | chest, elbow and ankle | N/A |
id12(2) [68] | ① OPPO: accelerometer ② HANDY: accelerometer, gyroscope, magnetometer | N/A | ① 30 Hz ② 52 Hz |
id12(2)e [69] | 5 RS485-networked XSense(IMU) 2 InertiaCube3 inertial sensors 12 Bluetooth acceleration sensors | XSense (IMU) included in a custom-made motion jacket, 2 InertiaCube3 inertial sensors located on each foot 12 Bluetooth acceleration sensors on the limbs | 30 Hz |
id12(5)e [70] | Accelerometer and gyroscope | N/A | 50 Hz |
id12(7)e [71] | Accelerometer | Wrists, chest and ankles | 100 Hz |
id4b [31] | ① Kinect ② 3DMA (Vicon Nexus V1.5.2 with 12 Vicon MX cam) | ① Within about 3.5 m from Kinect | ① Fluctuation around 30 Hz ② 120 Hz |
id7d [72] | Gyroscope and accelerometer | N/A | N/A |
id8d [73] | Three GY-521 accelerometer sensors | Arms, legs and neck | N/A |
id9d [74] | Three-axis accelerometer, gyroscope and magnetometer | Chest, thighs and wrists | 100 Hz |
Category | Recognition Range | Recognition Type | Algorithm Category | Arithmetic |
---|---|---|---|---|
id2(1) [24] | Eating, reading, operating a smartphone, operating a laptop computer, and sitting | Classify limited content | DL | DL (VGG-16 + OpenPose) Optimization algorithms: Adam |
id2(2) [39] | Global 3D human skeleton posture | Accurate relative position of components to space | DL | ① open pose |
id2(3) [40] | Throwing, Relying, Lying, Jumping, Helmet | Classify limited content | DL | DL (Faste R-CNN) |
id2(4) [41] | ① TV assembly ② Airplane assembly ③ Glassblowing ④ Motions based on EAWS | Classify limited content | Statistical model | GOM model HMM |
id2(7) [42] | 5 different types of awkward working posture | Classify limited content | DL | ① LSTM ② Bi-LSTM ③ GRU |
id2(1)e [33] | Accurate to every movement of the human body | Accurate relative position of components to space | DL | DL: Part Affinity Fields (PAFs) |
id2(2)e [32] | Accurate to every movement of the human body | Accurate relative position of components to space | DL | DL: OpenPose |
id2(3)e [43] | Trunk posture | Classify limited content | N/A | N/A |
id2(5)e [44] | N/A | Classify limited content | CML | Two-Stream Inflated 3D ConvNets (I3D) |
id2(12)e [45] | Crossing, talking, walking, queueing, waiting | Classify limited content | CML | Preprocessing: (geodesic distance, 3D Cartesian plane, way-point trajectory and joints MOCAP) feature optimization: Particle Swarm Optimization (PSO) classifier: Neuro-Fuzzy Classifier (NFC) |
id3(3) [46] | Sit, stand, walk, run, fall, lay on one’s back, lay face down, lay on | Classify limited content | CML | CML (k-NN) |
id3(5) [47] | Standby Forefoot strike Midfoot strike Rearfoot strike | Accurate relative position of components to space | DL | Feature-Based Learning: NB RF SVM dl: LSTM GRU Conv1D Optimization algorithms: Adam |
id3(10) [48] | In the mountain environment Lay, sit, climb gate, walk, run | Classify limited content | DL | DL (CNN) |
id3(11)e [49] | Various daily activities, including walking, running, going up and down stairs, etc. | Classify limited content | DL | DL (CNN-GRU) |
id4(2) [50] | Accurate to every movement of the human body | Accurate relative position of components to space | ① CML ② Traditional signal processing and data processing algorithms | ① Kinect SDK ② Kalman filter (KF) Madgwick filter(MAD) Complementary filter (CF) |
id4(3) [51] | Shoulder range-of-motion | Accurate relative position of components to space | N/A | N/A |
id4(4) [52] | Body joint coordination patterns | Accurate relative position of components to space | N/A | N/A |
id4(5) [53] | Evaluate the joint angle of the upper limb during daily functional tasks | Accurate relative position of components to space | DL | dl (LSTM) |
id4(6) [54] | A variety of worker postures | Classify limited content | DL | dl (Convolutional LSTM) |
id4(1)e [35] | Accurate to every movement of the human body | Accurate relative position of components to space | Traditional signal processing and data processing algorithms | ① MVN model (Xsens built-in) |
id4(2)e [55] | Accurate to every movement of the human body | Accurate relative position of components to space | N/A | N/A |
id4(3)e [56] | Construction posture classification | Classify limited content | CML | ① Neural network ② Decision tree ③ K-nearest neighbor (KNN) ④ Logical regression ⑤ Support vector machine (SVM) |
id4(4)e [57] | Shoulder movement | Accurate relative position of components to space | CML | Kinect for Windows SDK 2.0 |
id4(6)e [58] | Joint center locations during static postures | Accurate relative position of components to space | CML | ① Kinect for Windows SDK 1.5 ② Kinect for Windows SDK 2.0 |
id4(7)e [59] | Accuracy and repeatability of measuring torso angular displacement and upper arm elevation | Accurate relative position of components to space | Traditional data processing | Complementary weighting algorithm |
id4(8)e [60] | Proficiency in brick-moving posture | Classify limited content | CML | SVM |
id4(9)e [61] | Global 3D human skeleton posture | Accurate relative position of components to space | DL | Vnect (CNN) |
id4(11)e [62] | 3D joint kinematics | Accurate relative position of components to space | Geometric transformation and coordinate system transformation | Quaternion rotation matrix and joint-specific Euler angles |
id4(12)e [63] | Measure dynamic pelvic positioning angle | Accurate relative position of components to space | N/A | N/A |
id5(1)e [64] | Accurate to every movement of the human body | Accurate relative position of components to space | Statistical model | Preprocessing: floor removal object segment 3-D CCL human detection human identification HMM |
id5(3)e [65] | Interactions between people, such as handshakes and hugs | Classify limited content | Statistical model | Feature extraction and optimization: codebook, GMM, fisher encoding, cross-entropy optimization function classification; MEMM |
id5(5)e [66] | Accurate to every movement of the human body | Accurate relative position of components to space | Traditional data processing | Pseudo-2D stick model Ray Optimization K-ary Tree Hashing |
id5(10)e [67] | Daily movements, such as lying or standing | Classify limited content | CML | Filter: Chebyshev, Elliptic and Bessel Optimizer: e Probability Based Incremental Learning (PBIL) Classifier: K-Ary Tree Hashing |
id12(2) [68] | Activities of Daily Living (ADLs) | Classify limited content | DL | CNN-based DTL-HID |
id12(2)e [69] | Recognition modes of locomotion and postures recognition of sporadic gestures | Classify limited content | DL | DeepConvLSTM |
id12(5)e [70] | Walking Upstairs Downstairs Sitting Standing Lying | Classify limited content | DL | CNN |
id12(7)e [71] | Lying down, sitting, standing, walking, running, cycling, Nordic walking, ascending stairs, descending stairs, vacuum cleaning, ironing clothes, jumping rope | Classify limited content | CML DL | ① KNN ② Rotation forest ③ Neural network |
id4b [31] | Gait assessment | Accurate relative position of components to space | N/A | N/A |
id7d [72] | The activities of the human body in different situations | Classify limited content | Statistical model | Optimization algorithm: Adam and AdaDelta Post-processing: MEMM |
id8d [73] | The activities of the human body in different situations | Classify limited content | Meta-heuristic search algorithm | Optimization algorithm: Binary Grey Wolf Optimization (BGWO) Classifier: Decision tree (DT) classifier |
id9d [74] | Different modes of human activity | Classify limited content | Genetic algorithm | Genetic algorithm, GA |
Category | Dataset | Performance | Computational Time Complexity |
---|---|---|---|
id2(1) [24] | MPII Human Pose Dataset human posture dataset and the common objects in context (COCO) key points challenge data | 99.7% (validation dataset) | N/A |
id2(2) [39] | 10 Korean young males | Xsens as reference ① The RMSE for all joint angles is 8.4° ② The RMSE for all joint angles is 13.4° | ③ 20 ms delay |
id2(3) [40] | Specific action pictures from 31 students | The model’s accuracy in detecting throwing, relying, lying, jumping actions, and wearing helmets was 90.14%, 89.19%, 97.18%, 97.22%, and 93.67%, respectively. | N/A |
id2(4) [41] | N/A | F-Score (%) ① 96.84 ② 94.33 ③ 94.70 ④ 91.77 | N/A |
id2(7) [42] | 10 participants | Accuracy (%) ① 97.99 ② 98.33 ③ 99.01 | Training time ① 31 min ② 56 min ③ 54 min |
id2(1)e [33] | ① MPII human multi-person dataset ② COCO 2016 key points challenge dataset | ① 75.6% mAP ② 60.5% AP The champion of that year | N/A |
id2(2)e [32] | ① MPII human multi-person dataset ② COCO 2016 key points challenge dataset | ① 75.6% mAP ② 65.3% | Nvidia 1080 Ti and CUDA 8 ① 73 ms ② 74 ms |
id2(3)e [43] | Ten samples from one subject | Mean Bias (°) ① Chest Ref Back 0.3 Waist 5.7 Head −0.5 Shoulder −3.7 Center–waist 4.6 ② Chest 1.2 Under armpit −1.2 | N/A |
id2(5)e [44] | Kinetics (Pre-training) ① HMDB-51 ② UCF-101 | ① 80.2% ② 97.9% | N/A |
id2(12)e [45] | ① NTU RGB + D ② UoL 3D social activity dataset ③ Collective Activity Dataset (CAD) | ① 93.5% (NTU RGB + D dataset) ② 92.2% (UoL dataset) ③ 89.6% (Collective Activity Dataset) | A Core i5-4300U CPU is used to compute the running time. For one frame, the computational time for recognition of human action was 0.11 s. |
id3(3) [46] | Dataset from collected nine elderly activities with 11,000 records | 0.964 | N/A |
id3(5) [47] | Approximately 23.7 h running and walking data | NB: 72 scores RF: 83 scores SVM: 93 scores LSTM: 94 scores GRU: 94 scores Conv1D: 96 scores | NB: 130 s RF: 139 s SVM: 134 s LSTM: 47 s GRU: 46 s Conv1D: 28 s |
id3(10) [48] | Contains 3,341,184 samples covering activities at various terrain and fatigue levels | 0.978 | May not be feasible for battery-powered field devices |
id3(11)e [49] | ① UCI-HAR ② WISDM ③ PAMAP2 | ① 96.20% ② 97.21% ③ 95.27% | N/A |
id4(2) [50] | From three subjects | ② Slightly better than ①; both errors in the range of 3 to 8 degrees for all the joint angles | N/A |
id4(3) [51] | 50 asymptomatic adults | All free and fixed AROM. This system demonstrated adequate reliability (ICC ≥ 0.7). | N/A |
id4(4) [52] | 45 healthy participants | The mean and standard deviation of the PoV entropy feature is 2.602. | N/A |
id4(5) [53] | 13 healthy male university students | shoulder and elbow flexion/extension waveforms with mean CMCs > 0.93 shoulder adduction/abduction, and internal/external rotation waveforms with mean CMCs > 0.8 | N/A |
id4(6) [54] | 4 workers | Macro F1 score 0.870 | On (Intel Core i7-7700 CPU@ 2.8 GHz, 16 GB RAM, NIVIDA GeForce GTX 1060 GPU@16 GB RAM system) recognizes 256 postures per second |
id4(1)e [35] | 12 healthy participants | ① long complex task mean ± SD RMSE on all joints was 2.8° ± 1.6° short simple tasks was 1.2° ± 0.7° ② as reference | N/A |
id4(2)e [55] | 30 healthy participants | ① Single leg test: ICC value range = 0.70 to 0.80 Double leg test: 0.44 to 0.47 ② as reference | N/A |
id4(3)e [56] | 2 participants | ① 62–95% ② 63–94% ③ 68–96% ④ 65–96% ⑤ 63–94% | N/A |
id4(4)e [57] | 17 healthy participants | RMSE of shoulder flexion and extension is less than 10 ° | N/A |
id4(6)e [58] | 20 participants | For upright standing posture, average error ① 76 mm ② 87 mm | N/A |
id4(7)e [59] | 10 dairy workers | ① The RMSD ranges from 4.1 to 6.6° for the torso and from 7.2 to 12.1° for the upper arm ② as reference | N/A |
id4(8)e [60] | 21 participants with different professional levels | ① scenario1 91.23% ② scenario2 92.04% | Processing time (seconds) ① 524 ② 13 |
id4(9)e [61] | ① MPI-INF-3DHP ② Human 3.6 m | Mean Per Joint Position Error (MPJPE) ① 80.5 mm ② 124.7 mm | Capable of working in real time at a frequency of 30 Hz on a single TitanX (Pascal architecture) system |
id4(11)e [62] | 6 surgical faculty members | The neck and torso flexion/extension angles are accurate to 2.9 ± 0.9 degrees and 1.6 ± 1.1 degrees, respectively. Shoulder elevation is accurate to 6.8 ± 2.7 degrees. Elbow flexion is accurate to 8.2 ± 2.8 degrees. | N/A |
id4(12)e [63] | 17 healthy participants | ① The range of anterior sagittal pelvic angle RMSE is 2.7 °−8.9 ° ② as reference | N/A |
id5(1)e [64] | ① IM-DailyDepthActivity ② MSRAction3D ③ MSRDailyActivity3D | ① 72.86% ② 93.3% ③ 94.1% | N/A |
id5(3)e [65] | ① SBU Kinect Interaction ② UoL3D Social Activity ③ UT-Interaction | ① 91.25% ② 90.4% ③ 87.4% | N/A |
id5(5)e [66] | ① UCF50 ② hmdb51 ③ Olympic sports dataset | Key points detection accuracy ① 80.9% ② 82.1% ③ 81.7% Event classification accuracy ① 90.48% ② 89.2% ③ 90.83% | N/A |
id5(10)e [67] | ① DALIAC ② PAMPA2 ③ IM-LifeLog | ① 94.23% ② 94.07% ③ 96.40% | N/A |
id12(2) [68] | Two real-world ADL datasets are gathered: “Opportunity” (OPPO) and HANDY | OPPO: Micro-averaged 0.707 HANDY: F-1 Score 0.984 | N/A |
id12(2)e [69] | ① OPPORTUNITY dataset ② Skoda | F1 score ① Modes of Locomotion 0.895 Gesture Recognition 0.915 ② 0.985 | N/A |
id12(5)e [70] | ① WISDM ② UCI-HAR | ① 93.32 ② 97.63 | On Nexus 5X smartphone (cpu only), the system was able to classify about 28 samples per second. On Xeon E5-2640 v3 8-Core CPU and NVIDIA Titan X GPU, 149,600 samples per second. |
id12(7)e [71] | PAMAP2 | ① 0.890 ② 0.941 ③ 0.900 | N/A |
id4b [31] | 21 healthy adults | ② As reference Gait speed, step length: r and rc values > 0.9 Foot swing velocity: r = 0.93 rc = 0.54 stride time: R2 = 0.991 | N/A |
id7d [72] | ① USC-HAD ② IMSB ③ Mhealth human dataset | ① 91.25% ② 93.66% ③ 90.91% | N/A |
id8d [73] | ① MOTIONSENSE ② MHEALTH ③ IM-AccGyro human–machine dataset | ① 88.25% ② 93.95% ③ 96.83% | N/A |
id9d [74] | ① IM-WSHA ② WISDM ③ IM-SB ④ SMotion | ① 81.92% ② 95.37% ③ 90.17% ④ 94.58% | N/A |
Method | AP @ 0.5:0.95 | AP @ 0.5 | AP @ 0.75 | AP Medium | AP Large |
---|---|---|---|---|---|
OpenPose (CMU-Pose) | 61.8 | 84.9 | 67.5 | 57.1 | 68.2 |
Detectron (Mask R-CNN) | 67.0 | 88.0 | 73.1 | 62.2 | 75.6 |
AlphaPose | 73.3 | 89.2 | 79.1 | 69.0 | 78.6 |
Visual Sensors | |||
---|---|---|---|
Category | RGB Camera | Depth Camera | TOF |
Subdivision category | Multi-camera system | Time-of-Flight (ToF) (mainly represented by Kinect) | |
Sensor characteristics | No need for structured light, TOF transmitters and receivers, relatively low hardware costs, relying on natural light, can be used indoors and outdoors | Without the need for structured light, ToF emitters, and receivers, the hardware cost is relatively low. It relies on ambient light and can be used both indoors and outdoors. Multi-camera systems, which rely on visual features for image matching, have complex matching algorithms. | The measurement range of ToF is relatively long and unaffected by surface grayscale and features, and the depth distance calculation remains stable at the centimeter level without varying with distance. |
The directness and indirectness of sensor measurement | Indirect | Indirect | Unlike multi-camera systems that require algorithmic processing to output three-dimensional data, ToF can directly output the three-dimensional data of the measured object. |
Sampling rate | Moderate sampling rate | The sampling rate is the lowest among all schemes, which is not suitable for high-speed movement. | Low to medium sampling rate |
Applicable environment | Strong and dim light conditions have a significant impact. It captures less information compared to multi-camera systems. | Strong and dim light conditions have a significant impact. | It is essentially unusable under strong outdoor light conditions. |
Continuous working hours | No need for attachment installation, and there is no deviation in the sensor attachment position due to prolonged operation. | ||
Layout position | Usually installed in front of the person | Need a specific angle to cover the part of the body | Usually installed in front of the person |
Influence on human body movement after use | None | ||
Highest recognition type | Accurate relative position of components to space |
Non-Visual Sensors | ||||
---|---|---|---|---|
Category | Pressure Sensor | Accelerometers | Gyroscopes | Magnetometers |
Sensor characteristics | The architecture exhibits a straightforward design, characterized by rapid responsiveness and elevated sensitivity. | The system is simple, responsive, and sensitive, but integrating and filtering data can lead to cumulative errors. | High sensitivity, fast response, but with significant zero drift, poor stability, and the need for filtering and integration that can lead to cumulative errors. | Not affected by gravity, no need for integration. |
The three are often used in combination with each other. | ||||
Amount of information provided | Low | In non-visual schemes, this sensor can provide the most human body posture information. | ||
Sampling rate | It generally has a high sampling rate, making it more suitable for high-speed motion detection. | |||
Applicable environment | It is susceptible to the influence of environmental factors such as temperature and humidity and requires calibration and compensation. | It is easily affected by gravity and noise. | ||
Continuous working hours | No need for attachment installation, and there is no deviation in the sensor attachment position due to prolonged operation. | Prolonged operation can lead to displacement of the wearing position and drift of the sensor itself, which can significantly affect the accuracy of sensors, making them unsuitable for continuous long-term work. | ||
Influence on human body movement after use | It will have little impact on users | Installed on specific parts and joints of the body | ||
Highest recognition type | Classify limited content | Accurate relative position of components to space |
Algorithm Type | Computational Complexity | Accuracy | Applicability | Advantages and Disadvantages |
---|---|---|---|---|
Deep learning (CNN, LSTM) | High | Highest | Best for large, complex datasets | High accuracy but resource-intensive, especially in devices with limited resources. |
Traditional machine learning (SVM, k-NN) | Moderate | Moderate | Small-scale datasets, real-time tasks | Efficient on small datasets but less accurate on high-dimensional or large data. |
Statistical models (HMM) | Moderate | Moderate | Time series processing, continuous motion recognition | Good for handling dynamic motion, but limited when facing complex datasets. |
Geometric Transformation | Low | High (for 3D data) | Preprocessing in 3D posture analysis | Effective in reducing error but relies on high-quality sensor data for dynamic scenarios. |
Meta-heuristic algorithms (GA, BGWO) | High | High | Complex optimization tasks, feature selection | Optimizes model parameters effectively but computationally expensive, especially in large-scale search spaces. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yan, L.; Du, Y. Exploring Trends and Clusters in Human Posture Recognition Research: An Analysis Using CiteSpace. Sensors 2025, 25, 632. https://doi.org/10.3390/s25030632
Yan L, Du Y. Exploring Trends and Clusters in Human Posture Recognition Research: An Analysis Using CiteSpace. Sensors. 2025; 25(3):632. https://doi.org/10.3390/s25030632
Chicago/Turabian StyleYan, Lichuan, and You Du. 2025. "Exploring Trends and Clusters in Human Posture Recognition Research: An Analysis Using CiteSpace" Sensors 25, no. 3: 632. https://doi.org/10.3390/s25030632
APA StyleYan, L., & Du, Y. (2025). Exploring Trends and Clusters in Human Posture Recognition Research: An Analysis Using CiteSpace. Sensors, 25(3), 632. https://doi.org/10.3390/s25030632