Comprehensive Review of Vision-Based Fall Detection Systems
Abstract
:1. Introduction
2. Materials and Methods
3. Results
4. Discussion
4.1. Preprocessing
4.2. Characterization
4.2.1. Global
Silhouette Segmentation
Space–Time Methods
Optical Flow
Feature Descriptors
4.2.2. Local
4.2.3. Depth
Depth Map Representation
Skeleton Representation
4.3. Classification
4.3.1. Discriminative Models
Feature-Threshold-Based
Multivariate Exponentially Weighted Moving Average
Support Vector Machines
K-Nearest Neighbor
Decision Tree
Boost Classifier
Sparse Representation Classifier
Logistic Regression
Deep Learning Models
4.3.2. Generative Models
Hidden Markov Model
4.4. Tracking
4.4.1. Moving Average Filter
4.4.2. PID Filter
4.4.3. Kalman Filter
4.4.4. Particle Filter
4.4.5. Fused Images
4.4.6. Camshift
4.4.7. Deep Learning Architectures
4.5. Classifying Algorithms Performances
4.6. Validation Datasets
- The first group is integrated by a single dataset. It collects falls and activities of daily life (ADL) executed by volunteers whose results are recorded using different sensors, included RGB and IR cameras. It is used by a single system for validation purposes;
- The second group, which includes three datasets, incorporates depth and accelerometric data. By its relevance and number of reviewed systems using it in their performance evaluation, one dataset is especially important, UR fall detection [29]. This dataset, employed by over a third of all studied systems, includes 30 falls and 40 ADLs recorded by two depth systems, one providing frontal images and a second camera recording vertical ones. This information is accompanied by accelerometric data and was released in 2015;
- The third group is composed of nine datasets. They all mix ADLs and falls recorded in different scenarios by RGB cameras, either conventional or fish eye ones, from different perspectives and at different heights. Two of them exceed the mark of 10% users, LE2I [23] and the Multicam Fall Dataset [10].LE2I, published in 2013, is a dataset that includes 143 different types of falls performed by actors and 48 ADLs. These events were recorded in environments simulating the ones that could be found in an elderly home.Multicam includes 24 scenarios recorded with 8 IP cameras, so events can be analyzed from multiple perspectives. Twenty-two of the scenarios contains falls, while the other 2 only include confounding actions. Events are simulated by volunteers, and this dataset was released in 2010;
- The fourth group includes six datasets. Different activities, falls included, are recorded by depth systems. The two most used ones are the Fall Detection Dataset [30] and SDUFall [12], though both of them fall below the 10% users mark.Fall Detection Dataset, used by almost 10% of the systems, was published in 2017. The images in this dataset are recorded in five different rooms from eight different view angles, and five different volunteers take part in it.SDUFall, published in 2014, is another dataset that gathers depth information associated with six types of actions, being a fall one of them. Actions are repeated 30 times by 10 volunteers and are recorded by a depth system;
- The fifth group, composed of a single dataset, collects synthetic information. CMU Graphics Lab—motion capture library [55] is a dataset that contains biomechanical information related to human body movement captured through the use of motion capture (MoCap) technology. To generate that information, a group of volunteers, wearing sensors in different parts of their bodies, execute diverse activities. The information collected by the sensors is integrated through a human body model and stored in the dataset, so it can be used for development purposes. This approach to system development and validation has numerous advantages over conventional methods, as it gives developers the possibility of training their systems under any possible image perspective or occlusion situation. However, clutter and noise, the other important problems for optimal system performance, are not included in the information recorded in this database.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- United Nations. World Population Ageing 2017: Highlights; Department of Economic and Social Affairs, United Nations: New York, NY, USA, 2017. [Google Scholar]
- Sterling, D.A.; O’connor, J.A.; Bonadies, J. Geriatric falls: Injury severity is high and disproportionate to mechanism. J. Trauma Inj. Infect. Crit. Care 2001, 50, 116–119. [Google Scholar] [CrossRef]
- Vallabh, P.; Malekian, R. Fall detection monitoring systems: A comprehensive review. J. Ambient. Intell. Humaniz. Comput. 2018, 9, 1809–1833. [Google Scholar] [CrossRef]
- Rucco, R.; Sorriso, A.; Liparoti, M.; Ferraioli, G.; Sorrentino, P.; Ambrosanio, M.; Baselice, F. Type and Location of Wearable Sensors for Monitoring Falls during Static and Dynamic Tasks in Healthy Elderly: A Review. Sensors 2018, 18, 1613. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Yajai, A.; Rodtook, A.; Chinnasarn, K.; Rasmequan, S.; Apichet, Y. Fall detection using directional bounding box. In Proceedings of the 2015 12th International Joint Conference on Computer Science and Software Engineering (JCSSE), Hatyai, Thailand, 22–24 July 2015; pp. 52–57. [Google Scholar]
- Chong, C.-J.; Tan, W.-H.; Chang, Y.C.; Batcha, M.F.N.; Karuppiah, E. Visual based fall detection with reduced complexity horprasert segmentation using superpixel. In Proceedings of the 2015 IEEE 12th International Conference on Networking, Sensing and Control, Taipei, Taiwan, 9–11 April 2015; pp. 462–467. [Google Scholar]
- Rajabi, H.; Nahvi, M. An intelligent video surveillance system for fall and anesthesia detection for elderly and patients. In Proceedings of the 2015 2nd International Conference on Pattern Recognition and Image Analysis (IPRIA), Rasht, Iran, 11–12 March 2015; pp. 1–6. [Google Scholar]
- Juang, L.H.; Wu, M.N. Fall Down Detection Under Smart Home System. J. Med. Syst. 2015, 39, 107–113. [Google Scholar] [CrossRef] [PubMed]
- Mousse, M.A.; Motamed, C.; Ezin, E.C. Video-Based People Fall Detection via Homography Mapping of Foreground Polygons from Overlapping Cameras. In Proceedings of the 2015 11th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Bangkok, Thailand, 23–27 November 2015; pp. 164–169. [Google Scholar]
- Auvinet, E.; Rougier, C.; Meunier, J.; St-Arnaud, A.; Rousseau, J. Multiple Cameras Fall Data Set; Technical Report Number 1350; University of Montreal: Montreal, QC, Canada, 8 July 2011. [Google Scholar]
- Aslan, M.; Sengur, A.; Xiao, Y.; Wang, H.; Ince, M.C.; Ma, X. Shape feature encoding via Fisher Vector for efficient fall detection in depth-videos. Appl. Soft Comput. 2015, 37, 1023–1028. [Google Scholar] [CrossRef]
- Ma, X.; Wang, H.; Xue, B.; Zhou, M.; Ji, B.; Li, Y. Depth-Based Human Fall Detection via Shape Features and Improved Extreme Learning Machine. IEEE J. Biomed. Health Inform. 2014, 18, 1915–1922. [Google Scholar] [CrossRef]
- Bian, Z.-P.; Hou, J.; Chau, L.-P.; Magnenat-Thalmann, N. Fall Detection Based on Body Part Tracking Using a Depth Camera. IEEE J. Biomed. Health Inform. 2015, 19, 430–439. [Google Scholar] [CrossRef]
- Lin, C.; Wang, S.-M.; Hong, J.-W.; Kang, L.-W.; Huang, C.-L. Vision-Based Fall Detection through Shape Features. In Proceedings of the 2016 IEEE Second International Conference on Multimedia Big Data (BigMM), Taipei, Taiwan, 20–22 April 2016; pp. 237–240. [Google Scholar]
- Merrouche, F.; Baha, N. Depth camera based fall detection using human shape and movement. In Proceedings of the 2016 IEEE International Conference on Signal and Image Processing (ICSIP), Beijing, China, 13–15 August 2016; pp. 586–590. [Google Scholar]
- Gunale, K.G.; Mukherji, P. Fall detection using k-nearest neighbor classification for patient monitoring. In Proceedings of the 2015 International Conference on Information Processing (ICIP), Pune, India, 16–19 December 2015; pp. 520–524. [Google Scholar]
- Bhavya, K.R.; Park, J.; Park, H.; Kim, H.; Paik, J. Fall detection using motion estimation and accumulated image map. In Proceedings of the 2016 IEEE International Conference on Consumer Electronics-Asia (ICCE-Asia), Seoul, Korea, 26–28 October 2016; pp. 1–2. [Google Scholar]
- Wang, K.; Cao, G.; Meng, D.; Chen, W.; Cao, W. Automatic fall detection of human in video using combination of features. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; pp. 1228–1233. [Google Scholar]
- Barnich, O.; Van Droogenbroeck, M. ViBe: A Universal Background Subtraction Algorithm for Video Sequences. IEEE Trans. Image Process. 2011, 20, 1709–1724. [Google Scholar] [CrossRef] [Green Version]
- Chua, J.-L.; Chang, Y.C.; Lim, W.K. A simple vision-based fall detection technique for indoor video surveillance. Signal Image Video Process. 2015, 9, 623–633. [Google Scholar] [CrossRef]
- Pratap, U.; Khan, M.A.; Jalai, A.S. Human fall detection for video surveillance by handling partial occlusion scenario. In Proceedings of the 2016 11th International Conference on Industrial and Information Systems (ICIIS), Roorkee, India, 3–4 December 2016; pp. 280–284. [Google Scholar]
- Wang, X.; Liu, H.; Liu, M. A novel multi-cue integration system for efficient human fall detection. In Proceedings of the 2016 IEEE International Conference on Robotics and Biomimetics (ROBIO), Uttarakhand, India, 3–4 December 2016; pp. 1319–1324. [Google Scholar]
- Charfi, I.; Miteran, J.; Dubois, J.; Atri, M.; Tourki, R. Optimized spatio-temporal descriptors for real-time fall detection: Comparison of support vector machine and Adaboost-based classification. J. Electron. Imaging 2013, 22, 041106. [Google Scholar] [CrossRef]
- Alaoui, A.Y.; El Hassouny, A.; Thami, R.O.H.; Tairi, H. Video based human fall detection using von Mises distribution of motion vectors. In Proceedings of the 2017 Intelligent Systems and Computer Vision (ISCV), Fez, Morocco, 17–19 April 2017; pp. 1–5. [Google Scholar]
- Charfi, I.; Miteran, J.; Dubois, J.; Atri, M.; Tourki, R. Definition and Performance Evaluation of a Robust SVM Based Fall Detection Solution. In Proceedings of the 2012 Eighth International Conference on Signal Image Technology and Internet Based Systems, Naples, Italy, 25–29 November 2012; pp. 218–224. [Google Scholar]
- Yajai, A.; Rasmequan, S. Adaptive directional bounding box from RGB-D information for improving fall detection. J. Vis. Commun. Image Represent. 2017, 49, 257–273. [Google Scholar] [CrossRef]
- Lewandowski, B.; Wengefeld, T.; Schmiedel, T.; Gross, H.-M. I see you lying on the ground–Can I help you? Fast fallen person detection in 3D with a mobile robot. In Proceedings of the 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), Lisbon, Portugal, 28 August–1 September 2017; pp. 74–80. [Google Scholar] [CrossRef]
- Harrou, F.; Zerrouki, N.; Sun, Y.; Houacine, A. Vision-based fall detection system for improving safety of elderly people. IEEE Instrum. Meas. Mag. 2017, 20, 49–55. [Google Scholar] [CrossRef] [Green Version]
- Kepski, M.; Kwolek, B. Embedded system for fall detection using body-worn accelerometer and depth sensor. In Proceedings of the 2015 IEEE 8th International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications (IDAACS), Warsaw, Poland, 24–26 September 2015; Volume 2, pp. 755–759. [Google Scholar]
- Adhikari, K.; Bouchachia, H.; Nait-Charif, H. Activity recognition for indoor fall detection using convolutional neural network. In Proceedings of the 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, Japan, 8–12 May 2017. [Google Scholar]
- Basavaraj, G.M.; Kusagur, A. Vision based surveillance system for detection of human fall. In Proceedings of the 2017 2nd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, 19–20 May 2017; pp. 1516–1520. [Google Scholar]
- De Miguel, K.; Brunete, A.; Hernando, M.; Gambao, E. Home Camera-Based Fall Detection System for the Elderly. Sensors 2017, 17, 2864. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Yao, L.; Min, W.; Lu, K. A New Approach to Fall Detection Based on the Human Torso Motion Model. Appl. Sci. 2017, 7, 993. [Google Scholar] [CrossRef] [Green Version]
- Antonello, M.; Carraro, M.; Pierobon, M.; Menegatti, E. Fast and robust detection of fallen people from a mobile robot. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4159–4166. [Google Scholar]
- IASLAB-RGBD Fallen Person Dataset. 2019. Available online: http://robotics.dei.unipd.it/reid/index.php/downloads (accessed on 27 January 2021).
- Mohd, M.N.H.; Nizam, Y.; Suhaila, S.; Jamil, M.M.A. An optimized low computational algorithm for human fall detection from depth images based on Support Vector Machine classification. In Proceedings of the 2017 IEEE International Conference on Signal and Image Processing Applications (ICSIPA), Kuching, Malaysia, 12–14 September 2017; pp. 407–412. [Google Scholar]
- Cippitelli, E.; Gambi, E.; Gasparrini, S.; Spinsante, S. TST Fall Detection Dataset v2. Available online: https://ieee-dataport.org/documents/tst-fall-detection-dataset-v2 (accessed on 27 January 2021).
- The Fall Detection Datase. Available online: https://falldataset.com/ (accessed on 27 January 2021).
- Joshi, N.B.; Nalbalwar, S. A fall detection and alert system for an elderly using computer vision and Internet of Things. In Proceedings of the 2017 2nd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, 19–20 May 2017; pp. 1276–1281. [Google Scholar]
- Otanasap, N.; Boonbrahm, P. Pre-impact fall detection approach using dynamic threshold based and center of gravity in multiple Kinect viewpoints. In Proceedings of the 2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE), NakhonSiThammarat, Thailand, 12–14 July 2017; pp. 1–6. [Google Scholar]
- Feng, Q.; Gao, C.; Wang, L.; Zhang, M.; Du, L.; Qin, S. Fall detection based on motion history image and histogram of oriented gradient feature. In Proceedings of the 2017 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Xiamen, China, 6–9 November 2017; pp. 341–346. [Google Scholar]
- Hernandez-Mendez, S.; Maldonado-Mendez, C.; Marin-Hernandez, A.; Rios-Figueroa, H.V. Detecting falling people by autonomous service robots: A ROS module integration approach. In Proceedings of the 2017 International Conference on Electronics, Communications and Computers (CONIELECOMP), Cholula, Mexico, 22–24 February 2017; pp. 1–7. [Google Scholar]
- Kwolek, B.; Kępski, M. Human fall detection on embedded platform using depth maps and wireless accelerometer. Comput. Methods Programs Biomed. 2014, 117, 489–501. [Google Scholar] [CrossRef]
- Kasturi, S.; Jo, K.-H. Human fall classification system for ceiling-mounted kinect depth images. In Proceedings of the 2017, 17th International Conference on Control, Automation and Systems (ICCAS), Jeju, Korea, 18–21 October 2017; pp. 1346–1349. [Google Scholar]
- Kasturi, S.; Jo, K.-H. Classification of human fall in top Viewed kinect depth images using binary support vector machine. In Proceedings of the 2017 10th International Conference on Human System Interactions (HSI), Ulsan, Korea, 17–19 July 2017; pp. 144–147. [Google Scholar]
- Pattamaset, S.; Charoenpong, T.; Charoenpong, P.; Chianrabutra, C. Human fall detection by using the body vector. In Proceedings of the 2017 9th International Conference on Knowledge and Smart Technology (KST), Chonburi, Thailand, 1–4 February 2017; pp. 162–165. [Google Scholar]
- Taghvaei, S.; Jahanandish, M.H.; Kosuge, K. Auto Regressive Moving Average Hidden Markov Model for Vision-based Fall Prediction-An Application for Walker Robot. Assist. Technol. 2016, 29, 19–27. [Google Scholar] [CrossRef] [PubMed]
- Galvao, Y.M.; Albuquerque, V.A.; Fernandes, B.J.T.; Valenca, M.J.S. Anomaly detection in smart houses: Monitoring elderly daily behavior for fall detecting. In Proceedings of the 2017 IEEE Latin American Conference on Computational Intelligence (LA-CCI), Arequipa, Peru, 8–10 November 2017; pp. 1–6. [Google Scholar]
- Tran, T.-H.; Le, T.-L.; Hoang, V.-N.; Vu, H. Continuous detection of human fall using multimodal features from Kinect sensors in scalable environment. Comput. Methods Programs Biomed. 2017, 146, 151–165. [Google Scholar] [CrossRef]
- Tran, T.-H.; Le, T.-L.; Pham, D.-T.; Hoang, V.-N.; Khong, V.-M.; Tran, Q.-T.; Nguyen, T.-S.; Pham, C. A multi-modal multi-view dataset for human fall analysis and preliminary investigation on modality. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 1947–1952. [Google Scholar]
- Li, X.; Pang, T.; Liu, W.; Wang, T. Fall detection for elderly person care using convolutional neural networks. In Proceedings of the 2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; pp. 1–6. [Google Scholar]
- Fan, Y.; Levine, M.D.; Wen, G.; Qiu, S. A deep neural network for real-time detection of falling humans in naturally occurring scenes. Neurocomputing 2017, 260, 43–58. [Google Scholar] [CrossRef]
- Baldewijns, G.; Debard, G.; Mertes, G.; Vanrumste, B.; Croonenborghs, T. Bridging the gap between real-life data and simulated data by providing a highly realistic fall dataset for evaluating camera-based fall detection algorithms. Heal. Technol. Lett. 2016, 3, 6–11. [Google Scholar] [CrossRef] [Green Version]
- Abobakr, A.; Hossny, M.; Nahavandi, S. A Skeleton-Free Fall Detection System From Depth Images Using Random Decision Forest. IEEE Syst. J. 2017, 12, 2994–3005. [Google Scholar] [CrossRef]
- CMU Graphics Lab.—Motion Capture Library. Available online: http://mocap.cs.cmu.edu/ (accessed on 27 January 2021).
- Dai, B.; Yang, D.; Ai, L.; Zhang, P. A Novel Video-Surveillance-Based Algorithm of Fall Detection. In Proceedings of the 2018 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Beijing, China, 13–15 October 2018; pp. 1–6. [Google Scholar]
- Mastorakis, G.; Ellis, T.; Makris, D. Fall detection without people: A simulation approach tackling video data scarcity. Expert Syst. Appl. 2018, 112, 125–137. [Google Scholar] [CrossRef] [Green Version]
- Sehairi, K.; Chouireb, F.; Meunier, J. Elderly fall detection system based on multiple shape features and motion analysis. In Proceedings of the 2018 International Conference on Intelligent Systems and Computer Vision (ISCV), Fez, Morocco, 2–4 April 2018; pp. 1–8. [Google Scholar]
- Lu, K.-L.; Chu, E.T.-H. An Image-Based Fall Detection System for the Elderly. Appl. Sci. 2018, 8, 1995. [Google Scholar] [CrossRef] [Green Version]
- Panahi, L.; Ghods, V. Human fall detection using machine vision techniques on RGB–D images. Biomed. Signal Process. Control. 2018, 44, 146–153. [Google Scholar] [CrossRef]
- Rahnemoonfar, M.; Alkittawi, H. Spatio-temporal convolutional neural network for elderly fall detection in depth video cameras. In Proceedings of the 2018 IEEE International Conference on Big Data (Big Data), Seattle, WA, USA, 10–13 December 2018; pp. 2868–2873. [Google Scholar]
- Ricciuti, M.; Spinsante, S.; Gambi, E. Accurate Fall Detection in a Top View Privacy Preserving Configuration. Sensors 2018, 18, 1754. [Google Scholar] [CrossRef] [Green Version]
- Ko, M.; Kim, S.; Kim, M.-G.; Kim, K. A Novel Approach for Outdoor Fall Detection Using Multidimensional Features from A Single Camera. Appl. Sci. 2018, 8, 984. [Google Scholar] [CrossRef] [Green Version]
- Ali, S.F.; Khan, R.; Mahmood, A.; Hassan, M.T.; Jeon, M. Using Temporal Covariance of Motion and Geometric Features via Boosting for Human Fall Detection. Sensors 2018, 18, 1918. [Google Scholar] [CrossRef] [Green Version]
- Min, W.; Yao, L.; Lin, Z.; Liu, L. Support vector machine approach to fall recognition based on simplified expression of human skeleton action and fast detection of start key frame using torso angle. IET Comput. Vis. 2018, 12, 1133–1140. [Google Scholar] [CrossRef]
- Min, W.; Cui, H.; Rao, H.; Li, Z.; Yao, L. Detection of Human Falls on Furniture Using Scene Analysis Based on Deep Learning and Activity Characteristics. IEEE Access 2018, 6, 9324–9335. [Google Scholar] [CrossRef]
- Shanshan, X.; Xi, C. Fall detection method based on semi-contour distances. In Proceedings of the 2018 14th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 12–16 August 2018; pp. 785–788. [Google Scholar]
- CENTRE FOR DIGITAL HOME—MMU. Available online: http://foe.mmu.edu.my/digitalhome/FallVideo.zip (accessed on 27 January 2021).
- El Kaid, A.; Baïna, K.; Baïna, J. Reduce False Positive Alerts for Elderly Person Fall Video-Detection Algorithm by convolutional neural network model. Procedia Comput. Sci. 2019, 148, 2–11. [Google Scholar] [CrossRef]
- Ma, C.; Shimada, A.; Uchiyama, H.; Nagahara, H.; Taniguchi, R.-I. Fall detection using optical level anonymous image sensing system. Opt. Laser Technol. 2019, 110, 44–61. [Google Scholar] [CrossRef]
- Kumar, D.; Ravikumar, A.K.; Dharmalingam, V.; Kafle, V.P. Elderly Health Monitoring System with Fall Detection Using Multi-Feature Based Person Tracking. In Proceedings of the 2019 ITU Kaleidoscope: ICT for Health: Networks, Standards and Innovation (ITU K), Atlanta, GA, USA, 4–6 December 2019. [Google Scholar] [CrossRef]
- MOT Dataset. Available online: https://motchallenge.net/ (accessed on 27 January 2021).
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Computer Vision-ECCV 2014, ECCV 2014. Lecture Notes in Computer Science; Springer: Cham, Swizerland, 2014; pp. 740–755. [Google Scholar]
- Harrou, F.; Zerrouki, N.; Sun, Y.; Houacine, A. An Integrated Vision-Based Approach for Efficient Human Fall Detection in a Home Environment. IEEE Access 2019, 7, 114966–114974. [Google Scholar] [CrossRef]
- Brieva, J.; Ponce, H.; Moya-Albor, E.; Martinez-Villasenor, L. An Intelligent Human Fall Detection System Using a Vision-Based Strategy. In Proceedings of the 2019 IEEE 14th International Symposium on Autonomous Decentralized System (ISADS), Utrecht, The Netherlands, 8–10 April 2019; pp. 1–5. [Google Scholar]
- Hua, M.; Nan, Y.; Lian, S. Falls Prediction Based on Body Keypoints and Seq2Seq Architecture. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27 October–3 November 2019; pp. 1251–1259. [Google Scholar]
- Hasan, M.; Islam, S.; Abdullah, S. Robust Pose-Based Human Fall Detection Using Recurrent Neural Network. In Proceedings of the 2019 IEEE International Conference on Robotics, Automation, Artificial-intelligence and Internet-of-Things (RAAICON), Dhaka, Bangladesh, 29 November–1 December 2019; pp. 48–51. [Google Scholar]
- Soni, P.K.; Choudhary, A. Automated Fall Detection From a Camera Using Support Vector Machine. In Proceedings of the 2019 Second International Conference on Advanced Computational and Communication Paradigms (ICACCP), Gangtok, India, 25–28 February 2019; pp. 1–6. [Google Scholar]
- Espinosa, R.; Ponce, H.; Gutiérrez, S.; Martínez-Villaseñor, L.; Brieva, J.; Moya-Albor, E. A vision-based approach for fall detection using multiple cameras and convolutional neural networks: A case study using the UP-Fall detection dataset. Comput. Biol. Med. 2019, 115, 103520. [Google Scholar] [CrossRef]
- Martínez-Villaseñor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Núñez-Martínez, J.; Peñafort-Asturiano, C. UP-Fall Detection Dataset: A Multimodal Approach. Sensors 2019, 19, 1988. [Google Scholar] [CrossRef] [Green Version]
- Kalita, S.; Karmakar, A.; Hazarika, S.M. Human Fall Detection during Activities of Daily Living using Extended CORE9. In Proceedings of the 2019 Second International Conference on Advanced Computational and Communication Paradigms (ICACCP), Gangtok, India, 25–28 February 2019; pp. 1–6. [Google Scholar]
- Maldonado-Bascón, S.; Iglesias-Iglesias, C.; Martín-Martín, P.; Lafuente-Arroyo, S. Fallen People Detection Capabilities Using Assistive Robot. Electron. 2019, 8, 915. [Google Scholar] [CrossRef] [Green Version]
- Cai, X.; Li, S.; Liu, X.; Han, G. A Novel Method Based on Optical Flow Combining with Wide Residual Network for Fall Detection. In Proceedings of the 2019 IEEE 19th International Conference on Communication Technology (ICCT), Xi’an, China, 10–19 October 2019; pp. 715–718. [Google Scholar]
- Kong, X.; Chen, L.; Wang, Z.; Meng, L.; Tomiyama, H.; Chen, Y. Robust Self-Adaptation Fall-Detection System Based on Camera Height. Sensors 2019, 19, 3768. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Kong, X.; Meng, Z.; Nojiri, N.; Iwahori, Y.; Meng, L.; Tomiyama, H. A HOG-SVM Based Fall Detection IoT System for Elderly Persons Using Deep Sensor. Procedia Comput. Sci. 2019, 147, 276–282. [Google Scholar] [CrossRef]
- Carlier, A.; Peyramaure, P.; Favre, K.; Pressigout, M. Fall Detector Adapted to Nursing Home Needs through an Optical-Flow based CNN. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; Volume 2020, pp. 5741–5744. [Google Scholar]
- Wang, B.-H.; Yu, J.; Wang, K.; Bao, X.-Y.; Mao, K.-M. Fall Detection Based on Dual-Channel Feature Integration. IEEE Access 2020, 8, 103443–103453. [Google Scholar] [CrossRef]
- Menacho, C.; Ordonez, J. Fall detection based on CNN models implemented on a mobile robot. In Proceedings of the 2020 17th International Conference on Ubiquitous Robots (UR), Kyoto, Japan, 22–26 June 2020; pp. 284–289. [Google Scholar]
- Zhong, C.; Ng, W.W.Y.; Zhang, S.; Nugent, C.; Shewell, C.; Medina-Quero, J. Multi-occupancy Fall Detection using Non-Invasive Thermal Vision Sensor. IEEE Sens. J. 2020, 21, 1. [Google Scholar] [CrossRef]
- Sun, G.; Wang, Z. Fall detection algorithm for the elderly based on human posture estimation. In Proceedings of the 2020 Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Busan, Korea, 13–16 October 2020; pp. 172–176. [Google Scholar]
- Liu, J.-X.; Tan, R.; Sun, N.; Han, G.; Li, X.-F. Fall Detection under Privacy Protection Using Multi-layer Compressed Sensing. In Proceedings of the 2020 3rd International Conference on Artificial Intelligence and Big Data (ICAIBD), Chengdu, China, 28–31 May 2020; pp. 247–251. [Google Scholar]
- Thummala, J.; Pumrin, S. Fall Detection using Motion History Image and Shape Deformation. In Proceedings of the 2020 8th International Electrical Engineering Congress (iEECON), Chiang Mai, Thailand, 4–6 March 2020; pp. 1–4. [Google Scholar] [CrossRef]
- Zhang, J.; Wu, C.; Wang, Y. Human Fall Detection Based on Body Posture Spatio-Temporal Evolution. Sensors 2020, 20, 946. [Google Scholar] [CrossRef] [Green Version]
- Kottari, K.N.; Delibasis, K.; Maglogiannis, I. Real-Time Fall Detection Using Uncalibrated Fisheye Cameras. IEEE Trans. Cogn. Dev. Syst. 2019, 12, 588–600. [Google Scholar] [CrossRef]
- Delibasis, K.; Goudas, T.; Maglogiannis, I. A novel robust approach for handling illumination changes in video segmentation. Eng. Appl. Artif. Intell. 2016, 49, 43–60. [Google Scholar] [CrossRef]
- PIROPO (People in Indoor ROoms with Perspective and Omnidirectional Cameras). Available online: https://www.gti.ssr.upm.es/research/gti-data/databases (accessed on 27 January 2020).
- Feng, Q.; Gao, C.; Wang, L.; Zhao, Y.; Song, T.; Li, Q. Spatio-temporal fall event detection in complex scenes using attention guided LSTM. Pattern Recognit. Lett. 2020, 130, 242–249. [Google Scholar] [CrossRef]
- Xu, Q.; Huang, G.; Yu, M.; Guo, Y.; Huang, G. Fall prediction based on key points of human bones. Phys. A Stat. Mech. its Appl. 2020, 540, 123205. [Google Scholar] [CrossRef]
- Shahroudy, A.; Liu, J.; Ng, T.-T.; Wang, G. NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar]
- Htun, S.N.; Zin, T.T.; Tin, P. Image Processing Technique and Hidden Markov Model for an Elderly Care Monitoring System. J. Imaging 2020, 6, 49. [Google Scholar] [CrossRef]
- Kalinga, T.; Sirithunge, C.; Buddhika, A.; Jayasekara, P.; Perera, I. A Fall Detection and Emergency Notification System for Elderly. In Proceedings of the 2020 6th International Conference on Control, Automation and Robotics (ICCAR), Singapore, 20–23 April 2020; pp. 706–712. [Google Scholar]
- Chen, W.; Jiang, Z.; Guo, H.; Ni, X. Fall Detection Based on Key Points of Human-Skeleton Using OpenPose. Symmetry 2020, 12, 744. [Google Scholar] [CrossRef]
- Cai, X.; Li, S.; Liu, X.; Han, G. Vision-Based Fall Detection With Multi-Task Hourglass Convolutional Auto-Encoder. IEEE Access 2020, 8, 44493–44502. [Google Scholar] [CrossRef]
- Chen, Y.; Li, W.; Wang, L.; Hu, J.; Ye, M. Vision-Based Fall Event Detection in Complex Background Using Attention Guided Bi-Directional LSTM. IEEE Access 2020, 8, 161337–161348. [Google Scholar] [CrossRef]
- Chen, Y.; Kong, X.; Meng, L.; Tomiyama, H. An Edge Computing Based Fall Detection System for Elderly Persons. Procedia Comput. Sci. 2020, 174, 9–14. [Google Scholar] [CrossRef]
- Wang, X.; Jia, K. Human Fall Detection Algorithm Based on YOLOv3. In Proceedings of the 2020 IEEE 5th International Conference on Image, Vision and Computing (ICIVC), Qingdao, China, 23–25 July 2020; pp. 50–54. [Google Scholar]
- Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
- Horprasert, T.; Harwood, D.; Davis, L.S. A Statistical Approach for Real-time Robust Background Subtraction and Shadow Detection. In Proceedings of the IEEE ICCV’99 FRAME-RATE Workshop, Kerkyra, Greece, 20 September 1999. [Google Scholar]
- Mittal, A.; Paragios, N. Motion-based background subtraction using adaptive kernel density estimation. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004. [Google Scholar]
- Mousse, M.A.; Motamed, C.; Ezin, E.C. Fast Moving Object Detection from Overlapping Cameras. In Proceedings of the 12th International Conference on Informatics in Control, Automation and Robotics, Colmar, France, 21–23 July 2015; pp. 296–303. [Google Scholar]
- Mario, I.; Chacon, M.; Sergio, G.D.; Javier, V.P. Simplified SOM-neural model for video segmentation of moving objects. In Proceedings of the 2009 International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 474–480. [Google Scholar]
- Nguyen, V.-T.; Le, T.-L.; Tran, T.-H.; Mullot, R.; Courboulay, V.; Van-Toi, N. A new hand representation based on kernels for hand posture recognition. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; Volume 1, pp. 1–6. [Google Scholar]
- Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the ICNN’95-International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar] [CrossRef]
- Bobick, A.F.; Davis, J.W. The recognition of human movement using temporal templates. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 257–267. [Google Scholar] [CrossRef] [Green Version]
- Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the Imaging Understanding Workshop, Vancouber, Canada, 24–28 August 1981; pp. 121–130. [Google Scholar]
- Shi, J.; Tomasi, C. Good Features to Track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, DC, USA, 21–23 June 1994; pp. 593–600. [Google Scholar]
- Candès, E.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–37. [Google Scholar] [CrossRef]
- Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
- Nizam, Y.; Haji Mohd, M.N.; Abdul Jamil, M.M. A Study on Human Fall Detection Systems: Daily Activity Classification and Sensing Techniques. Int. J. Integr. Eng. 2016, 8, 35–43. [Google Scholar]
- Kalita, S.; Karmakar, A.; Hazarika, S.M. Efficient extraction of spatial relations for extended objects vis-à-vis human activity recognition in video. Appl. Intell. 2018, 48, 204–219. [Google Scholar] [CrossRef]
- Rosenblatt, F. The perceptron: A probabilistic model for information storage and organization in the brain. Psychol. Rev. 1958, 65, 386–408. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
- Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. arXiv 2014, arXiv:1409.3215. [Google Scholar]
- Hubel, D.H.; Wiesel, T.N. Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J. Physiol. 1962, 160, 106–154. [Google Scholar] [CrossRef]
- Fukushima, K. Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol. Cybern. 1980, 36, 193–202. [Google Scholar] [CrossRef]
- LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
- Tygert, M.; Bruna, J.; Chintala, S.; LeCun, Y.; Piantino, S.; Szlam, A. A Mathematical Motivation for Complex-Valued Convolutional Networks. Neural Comput. 2016, 28, 815–825. [Google Scholar] [CrossRef]
- Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
- Junejo, I.N.; Foroosh, H. Euclidean path modeling for video surveillance. Image Vis. Comput. 2008, 26, 512–528. [Google Scholar] [CrossRef]
- Maldonado, C.; Rios-Figueroa, H.V.; Mezura-Montes, E.; Marin, A.; Marin-Hernandez, A. Feature selection to detect fallen pose using depth images. In Proceedings of the 2016 International Conference on Electronics, Communications and Computers (CONIELECOMP), Cholula, Mexico, 24–26 February 2016; pp. 94–100. [Google Scholar]
- Labayrade, R.; Aubert, D.; Tarel, J.-P. Real time obstacle detection in stereovision on non flat road geometry through "v-disparity" representation. In Proceedings of the Intelligent Vehicle Symposium, Versailles, France, 17–21 June 2002. [Google Scholar]
- Schmiedel, T.; Einhorn, E.; Gross, H.-M. IRON: A fast interest point descriptor for robust NDT-map matching and its application to robot localization. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 3144–3151. [Google Scholar]
- Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Read. Comput. Vis. 1987, 24, 726–740. [Google Scholar] [CrossRef]
- Solbach, M.D.; Tsotsos, J.K. Vision-Based Fallen Person Detection for the Elderly. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 1433–1442. [Google Scholar]
- Zhang, K.; Chen, S.-C.; Whitman, D.; Shyu, M.-L.; Yan, J.; Zhang, C. A progressive morphological filter for removing nonground measurements from airborne LIDAR data. IEEE Trans. Geosci. Remote Sens. 2003, 41, 872–882. [Google Scholar] [CrossRef] [Green Version]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Tax, D.M.J.; Duin, R.P.W. Support Vector Data Description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef] [Green Version]
- Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
- Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2011, 29, 1189–1232. [Google Scholar] [CrossRef]
- Rabiner, L.; Juang, B. An introduction to hidden Markov models. IEEE ASSP Mag. 1986, 3, 4–16. [Google Scholar] [CrossRef]
- Natarajan, P.; Nevatia, R. Online, Real-time Tracking and Recognition of Human Actions. In Proceedings of the 2008 IEEE Workshop on Motion and video Computing, Copper Mountain, CO, USA, 8–9 January 2008; pp. 1–8. [Google Scholar] [CrossRef]
- Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef] [Green Version]
- Gordon, N. Bayesian Methods for Tracking. Ph.D. Thesis, Mathematics Department Imperial College, London, UK, 1993. Available online: https://spiral.imperial.ac.uk/bitstream/10044/1/7783/1/NeilGordon-1994-PhD-Thesis.pdf (accessed on 27 January 2020).
- Kangas, M.; Vikman, I.; Nyberg, L.; Korpelainen, R.; Lindblom, J.; Jämsä, T. Comparison of real-life accidental falls in older people with experimental falls in middle-aged test subjects. Gait Posture 2012, 35, 500–505. [Google Scholar] [CrossRef] [PubMed]
- Klenk, J.; Becker, C.; Lieken, F.; Nicolai, S.; Maetzler, W.; Alt, W.; Zijlstra, W.; Hausdorff, J.M.; Van Lummel, R.C.; Chiari, L.; et al. Comparison of acceleration signals of simulated and real-world backward falls. Med. Eng. Phys. 2011, 33, 368–373. [Google Scholar] [CrossRef] [PubMed]
- Thilo, F.J.S.; Hahn, S.; Halfens, R.; Schols, J.M. Usability of a wearable fall detection prototype from the perspective of older people–A real field testing approach. J. Clin. Nurs. 2018, 28, 310–320. [Google Scholar] [CrossRef]
- Demiris, G.; Chaudhuri, S.; Thompson, H.J. Older Adults’ Experience with a Novel Fall Detection Device. Telemed. J. E. Health 2016, 22, 726–732. [Google Scholar] [CrossRef] [Green Version]
- Ren, L.; Peng, Y. Research of Fall Detection and Fall Prevention Technologies: A Systematic Review. IEEE Access 2019, 7, 77702–77722. [Google Scholar] [CrossRef]
Reference | Year | Characterization (Global/Local/Depth) | Classification | Input Signal | Used Datasets | Performance |
---|---|---|---|---|---|---|
A. Yajai et al. [5] | 2015 | Skeleton joint tracking model provided by MS Kinect® is used to track joints and build a 2D and 3D bounding box around the body/depth characterization | Feature-threshold-based.
| Depth | This system-specific video dataset—no public access at revision time | Accuracy 98.43% Specificity 98.75% Recall 98.12% |
C. -J. Chong et al. [6] | 2015 | Pixel clustering and background (Horprasert)/global characterization | Feature-threshold-based. Method 1:
| Red-green-blue (RGB) | Specific video dataset—no public access at revision time | Method 1 Sensitivity 66.7% Specificity 80% Method 2 Sensitivity 72.2% Specificity 90% |
H. Rajabi et al. [7] | 2015 | Foreground extraction through background subtraction (Gaussian mixed models—GMM) and Sobel filter application/ global characterization | Feature-threshold-based.
| RGB | This system-specific video dataset—no public access at revision time | Fall detection success rate 81% |
L. H. Juang et al. [8] | 2015 | Foreground extraction through background subtraction (optical flow-based) and human joints identified/global characterization | Support vector machine (SVM) | RGB | This system-specific video dataset—no public access at revision time | Accuracy up to 100% |
M. A. Mousse et al. [9] | 2015 | Foreground extraction through pixel color and brightness distortion determination and integration of foreground maps through homography/global characterization | Feature-threshold-based. Ratio observed silhouette area/silhouette area projected on the ground plane | RGB—2 ORTHOGONAL VIEWS | Multicam Fall Dataset [10] | Sensitivity 95.8% Specificity 100% |
Muzaffer Aslan et al. [11] | 2015 | Human silhouette is segmented using depth information, and curvature scale space (CSS) is calculated and encoded in a Fisher vector/depth characterization | SVM | Depth | SDUFall [12] | Average accuracy 88.01% |
Z. Bian et al. [13] | 2015 | Silhouette extraction by using depth information. Human body joints identified and tracked with torso rotation/depth characterization | SVM | Depth | This system-specific video dataset—no public access at revision time | Sensitivity 95.8% Specificity 100% |
C. Lin et al. [14] | 2016 | Foreground extraction through background subtraction (GMM)/global characterization | Feature-threshold-based.
| RGB | This system-specific video dataset—no public access at revision time | Not published |
F. Merrouche et al. [15] | 2016 | Foreground extraction by using the difference between depth frames and head tracking through particle filter/depth characterization | Feature-threshold-based.
| Depth | SDUFall [12] | Sensitivity 90.76% Specificity 93.52% Accuracy 92.98% |
K. G. Gunale et al. [16] | 2016 | Foreground extraction through background subtraction (direct comparison)/global characterization | K-nearest neighbor (KNN) | RGB | Chute dataset—no public access at revision time | Accuracy Fall 90% No fall 100% |
K. R. Bhavya et al. [17] | 2016 | Foreground extraction through background subtraction (direct comparison)/global characterization + optical flow (OF)/global characterization | KNN on MHI and OF features | RGB | This system-specific video dataset—no public access at revision time | Not published |
Kun Wang et al. [18] | 2016 | Segmentation through vibe [19] and histogram of oriented gradients (HOG) and local binary pattern (LBP)/global characterization + feature maps obtained through convolutional neural network (CNN)/ local characterization | SVM-linear kernel | RGB | Multicam Fall Dataset [10] and SIMPLE Fall Detection Dataset [20] and This system-specific video dataset—no public access at revision time | Sensitivity 93.7% Specificity 92% |
U. Pratap et al. [21] | 2016 | Foreground extraction through background subtraction (GMM)/global characterization | Feature-threshold-based.
| RGB | Specific video datasets—no public access at revision time | Fall detection rate 92% False alarm rate 6.25% |
X. Wang et al. [22] | 2016 | Segmentation through vibe [19] and upper body database populated and sparse OF determined/global characterization | Feature-threshold-based.
| RGB | LE2I [23] | Average precision 81.55% |
A. Y. Alaoui et al. [24] | 2017 | Foreground extraction through background subtraction (direct comparison)/global characterization + OF/global characterization | No classification algorithm reported | RGB | CHARFI2012 Dataset [25] | Precision 91% Sensitivity 86.66% |
Apichet Yajai et al. [26] | 2017 | Skeleton joint tracking model provided by MS Kinect®/depth characterization | Feature-threshold-based. Aspect ratios:
| Depth | This system-specific video dataset—no public access at revision time | Accuracy 98.15% Sensitivity 97.75% Specificity 98.25% |
B. Lewandowski et al. [27] | 2017 | voxels around the point cloud are calculated. The ones classified as human are clustered, and IRON features are calculated/local characterization | Feature-threshold-based.
| Depth | This system-specific video dataset—no public access at revision time | Sensitivity in operational environments 99% |
F. Harrou et al. [28] | 2017 | Foreground extraction through background subtraction (direct comparison)/depth characterization | Multivariate exponentially weighted moving average (MEWMA)-SVM KNN Artificial neural network (ANN) Naïve Bayes (NB) | RGB | UR Fall Detection [29] & Fall Detection Dataset [30] | Accuracy KNN 91.94% ANN 95.15% NB 93.55% NEWMA-SVM 96.66% |
G. M. Basavaraj et al. [31] | 2017 | Foreground extraction through background subtraction (median)/global characterization | Feature-threshold-based.
| RGB | This system-specific video dataset—no public access at revision time | Accuracy Fall 86.66% Non-fall 90% |
K. Adhikari et al. [30] | 2017 | Foreground extraction through background subtraction (direct comparison) using both RGB techniques and depth ones and Feature maps obtained through CNN/local and depth characterization | Softmax based on features vector from CNN | Depth | This system-specific video dataset—no public access at revision time | Overall, accuracy 74% System sensitivity to lying pose 99% |
Koldo De Miguel et al. [32] | 2017 | Foreground extraction through background subtraction (GMM) + Sparse OF determined/global characterization | KNN on silhouette and OF features | RGB | This system-specific video dataset—no public access at revision time | Accuracy 96.9% Sensitivity 96% Specificity 97.6% |
Leiyue Yao et al. [33] | 2017 | Skeleton joint tracking model provided by MS Kinect®/depth characterization | Feature-threshold-based
| Depth | This system-specific video dataset—no public access at revision time | Accuracy 97.5% True positive rate 98% True negative rate 97% |
M. Antonello et al. [34] | 2017 | voxels around the point cloud are calculated. Then they are segmented in homogeneous patches and the ones classified as human are gathered and classified or not as a human lying body/depth characterization | SVM—radial-based kernel | Depth | IASLAB-RGBD fallen person Dataset [35] | Set A Accuracy: single view (SV) 0.87/SV+map verification (MV) 0.92 Precision: SV 0.73/SV+MV 0.85 Recall: SV 0.85/SV+MV 0.85 Set B Accuracy: SV 0.88/SV+MV 0.9 Precision: SV 0.8/SV+MV 0.87 Recall: SV 0.86/SV+MV 0.81 |
M. N. H. Mohd et al. [36] | 2017 | Skeleton joint tracking model provided by MS Kinect® is used to determine joint positions and speeds/depth characterization | SVM based on joints speeds and rule-based decision-based on joints position in relation to knees | Depth | TST Fall Detection [37] and UR Fall Detection [29] and Falling Detection [38] | Accuracy 97.39% Specificity 96.61% Sensitivity 100% |
N. B. Joshi et al. [39] | 2017 | Foreground extraction through background subtraction (GMM)/global characterization | Feature-threshold-based.
| RGB | LE2I [23] | Specificity 92.98% Accuracy 91.89% |
N. Otanasap et al. [40] | 2017 | Skeleton joint tracking model provided by MS Kinect®/depth characterization | Feature-threshold-based.
| Depth | This system-specific video dataset—no public access at revision time | Sensitivity 97% Accuracy 100% |
Q. Feng et al. [41] | 2017 | CNN is used to detect and track people, and Sub-MHI are correlated to each person BB/local characterization | SVM | RGB | UR Fall Detection [29] | Precision 96.8% Recall 98.1% F1 97.4% |
S. Hernandez-Mendez et al. [42] | 2017 | Foreground extraction through background subtraction (direct comparison) and silhouette tracking. Then centroid and features are determined/depth characterization | Feature-threshold-based.
| Depth | Depth And Accelerometric Dataset [43] and this system-specific video dataset—no public access at revision time | The fallen pose is detected correctly on 100% of occasions. |
S. Kasturi et al. [44] | 2017 | Foreground extraction through background subtraction (direct comparison)/depth characterization | SVM | Depth | UR Fall Detection [29] | Sensitivity 100% Specificity 88.33% |
S. Kasturi et al. [45] | 2017 | Foreground extraction through background subtraction (direct comparison)/depth characterization | SVM | Depth | UR Fall Detection [29] | Accuracy Total testing accuracy 96.34% |
S. Pattamaset et al. [46] | 2017 | Body vector construction and CG identification taking as starting point 16 parts of the human body/depth characterization | Feature-threshold-based.
| Depth | This system-specific video dataset—no public access at revision time | Accuracy 100% |
Sajjad Taghvaei et al. [47] | 2017 | Foreground extraction through background subtraction/depth characterization | Hidden Markov model (HMM) | Depth | This system-specific video dataset—no public access at revision time | Accuracy 84.72% |
Y. M. Galvão et al. [48] | 2017 | Median square error (MSE) every 3 frames/global characterization | Multilayer perceptron (MLP) KNN SVM—polynomial kernel | RGB | UR Fall Detection [29] | F1 score: MLP 0.991 KNN 0.988 SVM—polynomial kernel 0.988 |
Thanh-Hai Tran et al. [49] | 2017 | Skeleton joint tracking model provided by MS Kinect®/depth characterization or Motion map extraction from RGB images and gradient kernel descriptor calculated/global characterization | Feature-threshold-based.
| Depth or RGB | UR Fall Detection [29] and LE2I [23] and Multimodal Multiview Dataset of Human Activities [50] | UR Dataset Sensitivity 100% Specificity 99.23% LE2I Dataset Sensitivity 97.95% Specificity 97.87% MULTIMODAL Dataset (Average) Sensitivity 92.62% Specificity 100% |
X. Li et al. [51] | 2017 | Foreground extraction through background subtraction (direct comparison) and feature maps obtained through CNN/ local characterization | Softmax based on features vector from CNN | RGB | UR Fall Detection [29] | Sensitivity 100% Specificity 99.98% Accuracy 99.98% |
Yaxiang Fan et al. [52] | 2017 | Feature maps obtained through CNN from dynamic images/local characterization | Classification made by fully connected last layers of CNNs | RGB | Multicam Fall Dataset [10] & LE2I [23] and High-Quality Dataset [53] and This system-specific video dataset—no public access at revision time | Sensitivity LE2I 98.43% Multicam 97.1% HIGH-QUALITY FALL SIM 74.2% SYSTEM Dataset 63.7% |
A. Abobakr et al. [54] | 2018 | Silhouette extraction by using depth information. A feature vector of different body pixels based on depth difference between pairs of points is created/depth characterization | Random decision forest for pose recognition and SVM for movement identification | Depth | UR Fall Detection [29] and CMU Graphics Lab—motion capture library [55] | Accuracy 96% Precision 91% Sensitivity 100% Specificity 93% |
B. Dai et al. [56] | 2018 | Foreground extraction through background subtraction (direct comparison)/global characterization | Feature-threshold-based.
| RGB | UR Fall Detection [29] and This system-specific video dataset—no public access at revision time | Sensitivity 95% Specificity 96.7% |
Georgios Mastorakis et al. [57] | 2018 | Depth images are used to determine head velocity profile/depth characterization | Feature-threshold-based.
| Depth | Specific video dataset developed for [43] (A) and [12] (B)– no public access at revision time | A Dataset Sensitivity 100% Specificity 100% B Dataset Sensitivity 90.88% Specificity 98.48% |
K. Sehairi et al. [58] | 2018 | Foreground extraction through background subtraction (self-organizing maps) and feature extraction associated with each silhouette/global characterization |
| RGB | LE2I [23] | Accuracy SVM-RBF 99.27% KNN 98.91% ANN 99.61% |
Kun-Lin Lu et al. [59] | 2018 | Person detection through CNN YoLOv3 and feature extraction of the generated bounding box/local characterization | Feature-threshold-based
| RGB | This system-specific video dataset—no public access at revision time | Recall 100% Precision 93.94% Accuracy 95.96% |
Leila Panahi et al. [60] | 2018 | Foreground extraction through background subtraction (depth information) and silhouette tracking. Then ellipse is established around the silhouette, and features are determined/depth characterization | SVM & Threshold-based decision
| Depth | Depth and Accelerometric Dataset [43] | Average results SVM Sensitivity 98.52% Specificity 97.35% Threshold-based decision Sensitivity 98.52% Specificity 97.35% |
M. Rahnemoonfar et al. [61] | 2018 | Feature maps obtained through CNN/depth characterization | Softmax based on features vector from CNN | Depth | SDUFall [12] | Accuracy 97.58% |
Manola Ricciuti et al. [62] | 2018 | Foreground extraction through background subtraction (direct comparison)/depth characterization | SVM | Depth | This system-specific video dataset—no public access at revision time | Accuracy 98.6% |
Myeongseob Ko et al. [63] | 2018 | Depth map from monocular images and silhouette detection through particle swarm optimization/global characterization | Feature-threshold-based
| RGB | This system-specific video dataset—no public access at revision time | Accuracy 97.7% Sensitivity 95.7% Specificity 98.7% |
Syed F. Ali et al. [64] | 2018 | Foreground extraction through background subtraction (GMM)/global characterization | Boosted J48 | RGB | UR Fall Detection [29] and Multicam Fall Dataset [10] | Accuracies Multicam (2 classes) 99.2% Multicam (2 classes) 99.25% UR FALL 99% |
W. Min et al. [65] | 2018 | Skeleton joint tracking model provided by MS Kinect® is used to estimate vertical/torso angle/depth characterization | SVM | Depth | TST Fall Detection [37] | Accuracy 92.05% |
W. Min et al. [66] | 2018 | Object recognition through CNN and features of human shape sorted out as well as their spatial relations with furniture in the image/local characterization | Automatic engine classifier based on similarities (minimum quadratic error) between real-time actions and activity class features | RGB | This system-specific video dataset—no public access at revision time and UR Fall Detection [29] | Precision 94.44% Recall 94.95% Accuracy 95.5% |
X. ShanShan et al. [67] | 2018 | Foreground extraction through background subtraction (GMM)/global characterization | SVM-radial kernel | RGB | Center For Digital Home Dataset– MMU [68] | Sensitivity 96.87% Accuracy 86.79% |
Amal El Kaid et al. [69] | 2019 | Feature maps obtained through convolutional layers of a CNN/local characterization | Softmax based on features vector from CNN | RGB | This system-specific video dataset—no public access at revision time | Reduces false positives of angel assistance system by 17% by discarding positives assigned to people in a wheelchair |
Chao Ma et al. [70] | 2019 | Face masking to preserve privacy and feature maps obtained through CNN/local characterization | Autoencoder SVM | RGB + IR | UR Fall Detection [29] and Multicam Fall Dataset [10] and Fall Detection Dataset [30] and This system-specific video Dataset—no public access at revision time | Autoencoder Sensitivity 93.3% Specificity 92.8% SVM Sensitivity 90.8% Specificity 89.6% |
D. Kumar et al. [71] | 2019 | Silhouette segmentation by edge detection through HOG/global characterization + silhouette center angular velocity determined by long short-term memory (LSTM) model/local characterization | feature-threshold-based.
| RGB | MOT Dataset [72] and UR Fall Detection [29] and COCO Dataset [73] | Accuracy 98.1% |
F. Harrou et al. [74] | 2019 | Foreground extraction through background subtraction (direct comparison)/global characterization | SVM
| RGB | UR Fall Detection [29] & Fall Detection Dataset [30] | Accuracy: Linear kernel 93.93% Polynomial kernel 94.35% Radial kernel 96.66% |
J. Brieva et al. [75] | 2019 | Feature maps obtained through CNN from OF/ local characterization | Softmax based on features vector from CNN | RGB | This system-specific video dataset—no public access at revision time | Precision 95.27% Recall 95.42% F1 95.34% |
M. Hua et al. [76] | 2019 | Human keypoints identified by OpenPose (convolutional pose machines and human body vector construction) and recurrent neural network (RNN)-LSTM ANN used for pose prediction/local characterization | Fully connected layer | RGB | LE2I [23] | Precision 90.8% Recall 98.3% F1 0.944 |
M. M. Hasan et al. [77] | 2019 | Human keypoints identified by OpenPose (convolutional pose machines and human body vector construction) and RNN-LSTM ANN/local characterization | Softmax based on features vector from RNN-LSTM | RGB | UR Fall Detection [29] & Fall Detection Dataset [30] & Multicam Fall Dataset [10] | URFD Sensitivity 99% Specificity 96% FDD Sensitivity 99% Specificity 97% Multicam Sensitivity 98% Specificity 96% |
P. K. Soni et al. [78] | 2019 | Foreground extraction through background subtraction (GMM)/global characterization | SVM | RGB | UR Fall Detection [29] | Specificity 97.1% Sensitivity 98.15% |
Ricardo Espinosa et al. [79] | 2019 | OF extracted from 1-s windows/global characterization + Feature maps obtained through CNN/local characterization |
| RGB | UPFALL [80] | Sensitivity Softmax 97.95% SVM 14.1% RF 14.3% MLP 11.03% KNN 14.35% |
S. Kalita et al. [81] | 2019 | BBs established in hands, head and legs through extended core9 framework/local characterization | SVM | RGB | UR Fall Detection [29] | Sensitivity 93.33% Specificity 95% Accuracy 94.28% |
Saturnino Maldonado-Bascón et al. [82] | 2019 | Person detection through CNN YoLOv3 and feature extraction of the generated BB /local characterization | SVM | RGB | IASLAB-RGBD fallen person dataset [35] and This system-specific video dataset—no public access at revision time | Average results Precision 88.75% Recall 77.7% |
X. Cai et al. [83] | 2019 | OF/global characterization + Wide residual network/local characterization | Softmax classifier implemented in the last layer of the ANN | RGB | UR Fall Detection [29] | accuracy 92.6% |
Xiangbo Kong et al. [84] | 2019 | Segmentation by model provided by MS Kinect® + depth map and CNN used for feature maps creation/depth characterization | Softmax based on features vector from CNN implemented in its last layer | Depth | This system-specific video dataset—no public access at revision time | Depending on the camera height accuracy, results between 80.1% and 100% are obtained |
Xiangbo Kong et al. [85] | 2019 | Foreground extraction through background subtraction (Depth information) and HOG is calculated as a classifying feature | SVM-linear kernel | Depth | This system-specific video Dataset—no public access at revision time | Sensitivity 97.6% Specificity 100% |
A. CARLIER et al. [86] | 2020 | Dense OF/global characterization + feature maps obtained through CNN/ local characterization | Fully connected layer | RGB | UR Fall Detection [29] and Multicam Fall Dataset [10] and LE2I [23] | Sensitivity 86.2% False discovery rate 11.6% |
B. Wang et al. [87] | 2020 | Human keypoints identified by OpenPose (convolutional pose machines and human body vector construction) and followed by DeepSORT (CNN able to track numerous objects simultaneously)/local characterization | Classifiers are used to sort out falling state and fallen state
| RGB | UR Fall Detection [29] & Fall Detection Dataset [30] & LE2I [23] | F1-score Falling state GDBT 95.69% DT 84.85% RF 95.92% SVM 96.1% KNN 93.78% MLP 97.41% Fallen state GDBT 95.27% DT 95.45% RF 96.8% SVM 95.22% KNN 94.22% MLP 94.46% |
C. Menacho et al. [88] | 2020 | Dense OF/global characterization and feature maps obtained through CNN/ local characterization | Fully connected layer | RGB | UR Fall Detection [29] | Accuracy 88.55% |
C. Zhong et al. [89] | 2020 | Binarization based on IR threshold + edge identification/global characterization + feature maps obtained through convolutional layers of an ANN/local characterization | Based on features maps from CNN:
| IR | This system-specific video dataset—no public access at revision time | Multi-occupancy scenarios F1 score: RBFNN 89.57 (+/−0.62) SVM 88.74% (+/−1.75) Softmax 87.37% (+/−1.4) DT 88.9% (+/−0.68) |
G. Sun et al. [90] | 2020 | pose estimation through OpenPose (convolutional pose machines and human body vector construction) and single-shot multibox detector-MobileNet (SSD-MobileNet)/local characterization |
| RGB | COCO Dataset [73] and a specific video dataset—no public access at revision time | Sensitivity SVM 92.5% KNN 93.8% SVDD 94.6% |
J. Liu et al. [91] | 2020 | Local binary pattern histograms from three orthogonal planes (LBP-TOP) applied over optical Flow after robust principal component analysis (RPCA) techniques have been applied over incoming video signals. | Sparse representations classification (SRC) | RGB | UR Fall Detection [29] & Fall Detection Dataset [30] | Accuracy: FDD dataset 98% URF dataset 99.2% |
J. Thummala et al. [92] | 2020 | Foreground extraction through background subtraction (GMM)/global characterization | Feature-threshold-based. Object height/width ratio, ratio change speed and MHI. | RGB | LE2I [23] | Accuracy 95.16% |
Jin Zhang et al. [93] | 2020 | Human keypoints identified by CNN (convolutional pose machines and human body vector construction)/local characterization | Logistic regression classifier based on:
| RGB | This system-specific video dataset—no public access at revision time | Fall detection rate 98.7% False alarm rate 1.05% |
K. N. Kottar et al. [94] | 2020 | Segmentation through vibe [19] and illumination change-resistant algorithm (ICA) [95] then main silhouette axis determination | Feature-threshold-based.
| RGB | This system-specific video dataset—no public access at revision time and PIROPO [96] | Specific database accuracy ICA—87%–96.34% VIBE—78.05%–86.5% PIROPO—ICA Walk accuracy 95% Seat accuracy 98.65% |
Qi Feng et al. [97] | 2020 | Feature maps obtained through convolutional layers of a CNN and LSTM/local characterization | Softmax based on features vector from ANN implemented in its last layer | RGB | Multicam Fall Dataset [10] and UR Fall Detection [29] and this system-specific video dataset—no public access at revision time | Multicam Dataset Sensitivity 91.6% Specificity 93.5% UR Dataset Precision 94.8% Recall 91.4% THIS SYSTEM Dataset Precision 89.8% Recall 83.5% |
Qingzhen Xu et al. [98] | 2020 | Human keypoints identified by OpenPose (convolutional pose machines and human body vector construction) and CNN used for feature maps creation/local characterization | Softmax based on features vector from CNN implemented in its last layer | RGB | UR Fall Detection [29] and Multicam Fall Dataset [10] and NTU RGB+D Dataset [99] | Accuracy rate 91.7% |
Swe N. Htun et al. [100] | 2020 | Foreground extraction through background subtraction (GMM)/global characterization | Hidden Markov model (HMM) based onObservable data:
| RGB | LE2I [23] | Precision 99.05% Recall 98.37% Accuracy 99.8% |
T. Kalinga et al. [101] | 2020 | Skeleton joint tracking model provided by MS Kinect® is used to determine joint speeds and angles of different body parts/depth characterization | Feature-threshold-based.
| Depth | This system-specific video dataset—no public access at revision time | Accuracy 92.5% Sensitivity 95.45% Specificity 88% |
Weiming Chen et al. [102] | 2020 | Human keypoints identified by OpenPose (convolutional pose machines and human body vector construction)/local characterization | Feature-threshold-based
| RGB | This system-specific video dataset—no public access at revision time | Accuracy 97% Sensitivity 98.3% Specificity 95% |
X. Cai et al. [103] | 2020 | Feature maps obtained through hourglass convolutional auto-encoder (HCAE) ANN/local characterization | Softmax based on features vector from HCAE | RGB | UR Fall Detection [29] | Sensitivity 100% Specificity 93% Accuracy 96.2% |
Y. Chen et al. [104] | 2020 | Foreground extraction through CNN and Bi-LSTM ANN/local characterization | Softmax based on features vector from RNN-Bi-LSTM | RGB | UR Fall Detection [29] and This system-specific video dataset—no public access at revision time | URFD Precision 0.897 Recall 0.813 F1 0.852 Specific dataset Precision 0.981 Recall 0.923 F1 0.948 |
Yuxi Chen et al. [105] | 2020 | Feature maps obtained through 3 different CNNs (LeNet, AlexNet y GoogLeNet)/depth characterization | Classification made by fully connected last layers of CNNs | Depth | Video dataset developed for the system in [84] | Average values Lenet Sensitivity 82.78% Specificity 98.07% AlexNet Sensitivity 86.84% Specificity 98.41% GoogLeNet Sensitivity 92.87% Specificity 99% |
X. Wang et al. [106] | 2020 | Feature maps obtained through convolutional layers of an ANN/local characterization | Logistic function to identify frame-by-frame two classes in the prediction layer (person and fallen) | RGB | UR Fall Detection [29] & Fall Detection Dataset [30] | Average precision (AP) for fallen 0.97 mean average precision (mAP) for both classes 0.83 |
Reference | Year | Input Signal | ANN/Classifiers and Performance | |||||
---|---|---|---|---|---|---|---|---|
C. -J. Chong et al. [6] | 2015 | RGB | Method 1 BB aspect ratio and CG position | |||||
Sensitivity 66.7% | ||||||||
Specificity 80% | ||||||||
Method 2 Ellipse orientation and aspect ratio + MHI | ||||||||
Sensitivity 72.2% | ||||||||
Specificity 90% | ||||||||
F. Harrou et al. [28] | 2017 | RGB | Accuracy | Sensitivity | Specificity | |||
KNN | 91.94% | 100% | 86.00% | |||||
ANN | 95.15% | 100% | 91.00% | |||||
NB | 93.55% | 100% | 88.60% | |||||
MEWMA-SVM | 96.66% | 100% | 94.93% | |||||
Y. M. Galvão et al. [48] | 2017 | RGB | F1 score | |||||
Multilayer perceptron (MLP) 0.991 | ||||||||
K-nearest neighbors (KNN) 0.988 | ||||||||
SVM—polynomial kernel 0.988 | ||||||||
Leila Panahi et al. [60] | 2018 | Depth | Average results | |||||
SVM | ||||||||
Sensitivity 98.52% | ||||||||
Specificity 97.35% | ||||||||
Threshold-based decision | ||||||||
Sensitivity 98.52% | ||||||||
Specificity 97.35% | ||||||||
K. Sehairi et al. [58] | 2018 | RGB | Accuracy | |||||
SVM-RBF 99.27% | ||||||||
KNN 98.91% | ||||||||
ANN 99.61% | ||||||||
Chao Ma et al. [70] | 2019 | RGB+IR | Autoencoder | |||||
Sensitivity 93.3% | ||||||||
Specificity 92.8% | ||||||||
SVM | ||||||||
Sensitivity 90.8% | ||||||||
Specificity 89.6% | ||||||||
F. Harrou et al. [74] | 2019 | RGB | Accuracy: | |||||
K-NN 91.94% | ||||||||
ANN 95.16% | ||||||||
Naïve Bayes 93.55% | ||||||||
Decision tree 90.48% | ||||||||
SVM 96.66% | ||||||||
Ricardo Espinosa et al. [79] | 2019 | RGB | Sensitivity | Specificity | ||||
Softmax | 97.95% | 83.08% | ||||||
SVM | 14.10% | 90.03% | ||||||
RF | 14.30% | 91.26% | ||||||
MLP | 11.03% | 93.65% | ||||||
KNN | 14.35% | 90.96% | ||||||
Xiangbo Kong et al. [84] | 2019 | Depth | HOG+SVM | LeNet | AlexNet | GoogLeNet | ETDA-Net | |
Average accuracy | 89.48% | 88.28% | 93.53% | 96.59% | 95.66% | |||
Average specificity | 95.43% | 97.18% | 97.56% | 98.76% | 99.35% | |||
Average sensitivity | 83.75% | 74.54% | 87.10% | 88.74% | 91.87% | |||
B. Wang et al. [87] | 2020 | RGB | F1 score | |||||
Falling state | ||||||||
GDBT 95.69% | ||||||||
DT 84.85% | ||||||||
RF 95.92% | ||||||||
SVM 96.1% | ||||||||
KNN 93.78% | ||||||||
MLP 97.41% | ||||||||
Fallen state | ||||||||
GDBT 95.27% | ||||||||
DT 95.45% | ||||||||
RF 96.8% | ||||||||
SVM 95.22% | ||||||||
KNN 94.22% | ||||||||
MLP 94.46% | ||||||||
C. Zhong et al. [89] | 2020 | IR | F1 score | |||||
RBFNN 89.57 (+/−0.62) | ||||||||
SVM 88.74% (+/−1.75) | ||||||||
Softmax 87.37% (+/−1.4) | ||||||||
DT 88.9% (+/−0.68) | ||||||||
C. Menacho et al. [88] | 2020 | RGB | Accuracy | |||||
VGG-16 87.81% | ||||||||
VGG-19 88.66% | ||||||||
Inception V3 92.57% | ||||||||
ResNet50 92.57% | ||||||||
Xception 92.57% | ||||||||
ANN proposed in this system 88.55% | ||||||||
G. Sun et al. [90] | 2020 | RGB | Sensitivity | Specificity | ||||
SVM | 92.50% | 93.70% | ||||||
KNN | 93.80% | 92.30% | ||||||
SVDD | 94.60% | 93.80% | ||||||
Yuxi Chen et al. [105] | 2020 | Depth | Average values | |||||
Lenet | ||||||||
Sensitivity 82.78% | ||||||||
Specificity 98.07% | ||||||||
AlexNet | ||||||||
Sensitivity 86.84% | ||||||||
Specificity 98.41% | ||||||||
GoogLeNet | ||||||||
Sensitivity 92.87% | ||||||||
Specificity 99% |
Signal Type | Dataset Name | Characteristics |
---|---|---|
Accelerometric and electroencephalogram (EEG) and RGB and passive infrared (IR) | Upfall [80] | 17 volunteers execute falls and activities of daily life (ADL) of different types recorded by an accelerometer, EEG, RGB and passive IR systems |
Depth and Accelerometric | Depth and accelerometric dataset [43] | Volunteers execute several activities, and falls are recorded by a depth system and accelerometers. |
TST fall detection [37] | 11 volunteers execute 4 fall types and 4 ADLs recorded by RGB-depth (RGB-D) and accelerometer systems | |
UR fall detection [29] | 30 falls and 40 ADLs recorded by RGB-D and accelerometer systems | |
RGB | Center for digital home data set—MMU [68] | 20 videos, including 31 falls and several ADLs |
LE2I [23] | 191 different activities, including ADLs and 143 falls | |
Charfi2012 dataset [25] | 250 video sequences in four different locations, 192 containing falls, and 57 containing ADLs. Actors, under different light conditions, move in environments where occlusion exits and cluttered and textured background is common | |
High-quality dataset [53] | It is a fall detection dataset that attempts to approach the quality of a real-life fall dataset. It has realistic settings and fall scenarios. In detail, 55 fall scenarios and 17 normal activity scenarios were filmed by five web-cameras in a room similar to one in a nursing home | |
Multicam fall dataset [10] | The video data set is composed of several simulated normal daily activities and falls viewed from 8 different cameras and performed by one subject in 24 scenarios | |
Simple fall detection dataset [20] | The dataset contains 30 daily activities such as walking, sitting down, squatting down, and 21 fall activities such as forward falls, backward falls and sideway falls | |
MO dataset [72] | MOT dataset intends to be a framework for the fair evaluation of multiple people tracking algorithms. In this framework, the designers provide:
| |
COCO dataset [73] | COCO is a large-scale object detection, segmentation, and captioning dataset designed to show common objects in context | |
Piropo [96] | Multiple activities recorded in two different scenarios with both conventional and fish eye cameras | |
Depth | IASLAB-RGB fallen person dataset [35] | It consists of several static and dynamic sequences with 15 different people and 2 different environments |
Multimodal multiview dataset of human activities [50] | It consists of 2 datasets recorded simultaneously by 2 Kinect systems including ADLs and falls in a living room equipped with a bed, a cupboard, a chair and surrounding office objects illuminated by neon lamps on the ceiling or by sunlight | |
Sdufall [12] | 10 volunteers develop 6 activities recorded by RGB-D systems | |
Falling detection [38] | 6 volunteers perform 26 falls and similar activities recorded by RGB-D systems. | |
Fall detection dataset [30] | 5 volunteers execute 5 different types of fall | |
NTU RGB+ dataset [99] | It is a large-scale dataset for human action recognition. It contains 56,880 action samples and includes 4 different modalities of data for each sample: RGB videos, depth map sequences, 3D skeletal data and IR videos | |
Synthetic Movement Databases | CMU Graphics Lab—motion capture library [55] | Library that captures synthetic movements through movement capture (MoCap) technology |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Gutiérrez, J.; Rodríguez, V.; Martin, S. Comprehensive Review of Vision-Based Fall Detection Systems. Sensors 2021, 21, 947. https://doi.org/10.3390/s21030947
Gutiérrez J, Rodríguez V, Martin S. Comprehensive Review of Vision-Based Fall Detection Systems. Sensors. 2021; 21(3):947. https://doi.org/10.3390/s21030947
Chicago/Turabian StyleGutiérrez, Jesús, Víctor Rodríguez, and Sergio Martin. 2021. "Comprehensive Review of Vision-Based Fall Detection Systems" Sensors 21, no. 3: 947. https://doi.org/10.3390/s21030947
APA StyleGutiérrez, J., Rodríguez, V., & Martin, S. (2021). Comprehensive Review of Vision-Based Fall Detection Systems. Sensors, 21(3), 947. https://doi.org/10.3390/s21030947