Computer Vision Tasks for Ambient Intelligence in Children’s Health
Abstract
:1. Introduction
- Neurocognitive impairment (e.g., based on Prechtl General Movement Assessment—GMA) or early signs of neurocognitive developmental disorders (e.g., Autism Spectrum Disorders—ASD or Attention Deficit Hyperactivity Disorders—ADHD).
- Dysmorphisms (e.g., cleft lip) or physical or motor impairments (e.g., gait and walking disorders) due to genetic disorders or surgery.
- The well-being and health status of newborns (e.g., vital signs and sleep monitoring in the nursery or in the Neonatal Intensive Care Unit—NICU) and children.
2. Face Analysis and Head Movements
- ScopusQUERY “TITLE-ABS-KEY ((newborn OR baby OR children OR toddler OR infant) AND (face OR facial) AND (analysis OR detection OR recognition OR tracking) AND “computer vision” AND PUBYEAR > 2014 AND PUBYEAR < 2024 that returned 158 documents;
- Web of Science Core Collection((((ALL=(children)) OR ALL=(infant)) OR ALL=(baby)) OR ALL=(newborn)) AND ((ALL=(face)) OR ALL=(facial)) AND ((ALL=(analysis)) OR ALL=(detection) OR ALL=(recognition)) AND ALL=(computer) AND ALL=(vision)), refined in the YEARS from 2015 to 2023, that returned 197 documents;
- Scholarallintitle: children OR newborn OR babies OR infants OR face OR facial OR analysis OR recognition OR detection OR tracking OR “computer vision”, refined in the YEARS from 2015 to 2023, that returned 158 documents.
2.1. Face Morphology Analysis
2.2. Head and Gaze Tracking and Analysis
Work (Year) | Method | Clinical Task | Metrics | Dataset Population/Age (h = hours, w = weeks, m = months, y = years) |
---|---|---|---|---|
[33] (2019) | OpenFace | Early detection of ASD signs | Qualitative | 6 children |
[27] (2021) | Computation of 51 facial landmark + computation of rotation parameters between the landmarks and a 3D canonical face model | ASD diagnosis | Qualitative | 104 toddlers (age: 16–31 m) |
[34] (2021) | Faster R-CNN algorithm to fine-tune a pre-trained ResNet-101 | Monitoring of paediatric patients in critical settings | acc = 84% | 59 paediatric patients |
2.3. Facial Expressions Analysis
- Handle meltdown crisis. Studies such as [47,48] consider the safety of autistic children during a meltdown crisis. Meltdown signals are not associated with a specific facial expression, but with a mixture of abnormal facial expressions related to complex emotions. Through the evaluation of a set of spatio-temporal geometric facial features of micro-expressions, the authors demonstrate that the proposed system can automatically distinguish a compound emotion of autistic children during a meltdown crisis from the normal state and timely notify caregivers.
- Support specialists in diagnosing and evaluating ASD children. In [41], the authors propose a CV module consisting of four main components aimed at face detection, facial landmark detection, multi-face tracking and facial action unit extraction. The authors highlight how the proposed system could provide a noninvasive framework to apply to pre-school children in order to understand the underlying mechanisms of the difficulties in the use, sharing and response to emotions typical of ASD.
- Computationally analyse how children with ASD produce facial expressions with respect to their typically developing peers. In [56,57,58], the authors propose a framework aimed at computationally assessing how ASD and typically developing children produce facial expressions. Such a framework, which works on a sequence of images captured by a webcam under unconstrained conditions, locates and tracks multiple landmarks to monitor facial muscle movements involved in the production of facial expressions (thus performing a type of virtual electromyography). The output from these virtual sensors is then fused to model the individual’s ability to produce facial expressions. The results correlate with psychologists’ ratings, demonstrating how the proposed framework can effectively quantify the emotional competence of children with ASD to produce facial expressions.
- Early detect symptoms of autism. Despite advances in the literature, it is still difficult to identify early markers that can effectively detect the manifestation of symptoms of ASD. Carpenter and colleagues [49] collected videos of 104 young children (22 with ASD) watching short movies on a tablet. They then used a CV approach to automatically detect and track specific facial landmarks in the recorded videos to estimate the children’s facial expressions (positive, neutral, all others) and differentiate between children with and without ASD. In these cases, children with ASD were more likely to show ’neutral’ facial expressions, while children without ASD were more likely to show ’all other’ facial expressions (raised eyebrows, open mouth, engaged, etc.).
2.4. Multimodal Analysis
2.5. Publicly Available Datasets
- COPE Database [61,62]: This database contains 204 photographs of 26 newborns (between 18–36 h old) who were photographed while experiencing the pain of a heel lance and a variety of stressors, including being moved from one cot to another (a stressor that produces crying that is not in response to pain), a puff of air on the nose (a stressor that produces eye squinting), and friction on the outer lateral surface of the heel (a stressor that produces facial expressions of distress similar to those of pain). In addition to these four facial displays, the database contains images of the newborns in a neutral resting state. All subjects were born in a large Midwestern hospital in the United States. All newborns involved in the study were Caucasian, evenly divided between the sexes (13 boys and 12 girls), and in good health.
- CAFE Database [75]: The CAFE set is a collection of 1192 photographs of 2- to 8-year-old children posing with the six basic emotions defined by Ekman [82]: sadness, happiness, surprise, anger, disgust and fear. It also includes a seventh neutral expression. Such a set is also racially and ethnically diverse, with 27 African American, 16 Asian, 77 Caucasian/European American, 23 Latino, and 11 South Asian children. Photographs include enough face variability to allow independent researchers to determine and study the natural variation in human facial expressions. The children were asked to pose with their mouths open and closed for each expression except surprise. Surprised faces were open-mouthed only. Open-mouthed, disgusted faces usually included a tongue protrusion.
- CLOCK Database [70]: This database was generated by a multi-site longitudinal project known as CLOCK (Craniofacial microsomia: Longitudinal Outcomes in Children pre-Kindergarten), which examined the neurodevelopmental and phenotypic outcomes of children with craniofacial microsomia (CFM) and demographically matched controls [83]. Two age-appropriate emotion induction tasks were used to elicit positive and negative facial expressions. In the positive emotion task, an experimenter blew bubbles at the infant. In the negative emotion task, an experimenter presented the infant with a toy car, allowed the infant to play, then removed the car and covered it with a clear plastic container. Each video was approximately 2 min long (745 K and 634 K recorded frames). The video resolution was 1920 × 1080. FACS coders manually annotated for nine action units: AU1 (inner brow raised), AU2 (outer brow raised), AU3 (inner brow pulled together), AU4 (lowered eyebrow), AU6 (raised cheek), AU9 (nose), AU10 (nose wrinkle), AU9 (nasal wrinkling), AU12 (corner of lips pulled back), AU20 (lip stretching) and (lip stretching) and AU28 (lip sucking).
- LIRIS-CSE Database [71]: It features video clips and dynamic images consisting of 26,000 frames depicting 12 children from diverse ethnic backgrounds. This database showcases children’s natural, unforced facial expressions across various scenarios, featuring six universal or prototypical emotional expressions: happiness, sadness, surprise, anger, disgust, and fear as defined by Ekman [73]. The recordings were made in unconstrained environments, enabling free head and hand movements while sitting freely. In contrast to other public databases, the authors assert that they were capable of gathering children’s natural expressions as they happened due to the unconstrained environment. The database has been validated by 22 human raters.
- GestATional Database [23]: It comprises 130 neonates recruited between October 2015 and October 2017. Clinical staff at Nottingham University NHS Trust Hospital, Nottingham, UK carried out recruitment and sorted the neonates into five groups based on their prematurity status. The data gathered included: (i) images of the neonates’ faces, feet, and ears; (ii) case report forms with important information such as the baby’s gestational age, days of life at the time of the visit, current weight, Ballard Score, the mother’s medical history, and information related to the delivery. It is important that technical term abbreviations are explained when they are first used, and that a logical flow of information is maintained with causal connections between statements.
- FF-NFS-MIAMI Database [68,69]: It is a database documenting spontaneous behaviour in 43 four-month-old infants. Infants’ interactions with their mothers were recorded during a Face-to-Face/Still-Face (FF/SF) protocol [84]. The FF/SF protocol elicits both positive and negative effects. It assesses infant responses to parent unresponsiveness, an age-appropriate stressor. AUs were manually annotated from the video by certified FACS coders for four action units: AU4 (brow lowering), AU6 (cheek raising), AU12 (lip corner pulling) and AU20 (lip stretching). The combination of AU6 and AU12 is associated with a positive effect; AU4 and AU20 are associated with a negative effect. The video resolution is 1288 × 964. There are 116,000 manually annotated frames in 129 videos of 43 infants.
- USF-MNPAD-I Database [65]: The University of South Florida Multimodal Neonatal Pain Assessment (USF-MNPAD-I) Dataset was collected from 58 neonates (27–41 weeks gestational age) while they were hospitalised in the NICU, undergoing procedural and postoperative procedures. It comprises video footage (face, head, and body), audio (crying sounds), vital signs (heart rate, blood pressure, oxygen saturation), and cortical activity. Additionally, it includes continuous pain scores, following the NIPS (Neonatal Infant Pain Scale) scale [85], for each pain indicator and medical notes for all neonates. This dataset was obtained as a component of a continuous project centred on creating avant garde automated approaches for tracking and evaluating neonatal pain and distress.
Dataset | Reference | Number of Subjects | Type of Data | Age of Subjects | Year | Publicly Available |
---|---|---|---|---|---|---|
COPE | [61,62] | 26 | Images | Neonates: (age: 18–36 h) | 2005 | Yes |
CAFE | [75] | 154 | Images | Children (age: 2–8 years) | 2014 | Yes |
CLOCK | [70] | 80 | Video | Children (age: 4–5 years) | 2017 | No |
LIRIS-CSE | [71] | 12 | Video | Children (age: 6–12 years) | 2019 | Yes |
GestATional | [23] | 130 | Images | Neonates (gestational age: 28–40 weeks) | 2019 | No |
FF-NFS-MIAMI | [68,69] | 43 | Video | Infants | 2020 | No |
USF-MNPAD-I | [65] | 58 | Video, audio, physiological, contextual, information | Neonates (age: 27–41 weeks) | 2021 | Yes |
2.6. New Computer Vision Perspectives for More Accurate Face Analyses
3. Body Analysis
- ScopusQUERY “TITLE-ABS-KEY ( ( children OR infants OR babies ) AND ( body OR limbs OR head ) AND ( motion OR movements ) AND computer AND vision ) AND PUBYEAR > 2014 AND PUBYEAR < 2024 that returned 105 documents;
- Web of Science Core Collection((((ALL=(children)) OR ALL=(infants)) OR ALL=(babies)) AND ALL=(computer ) AND ALL=(vision) AND ((ALL=(motion)) OR ALL=(movements)) AND ( (ALL=(body)) OR (ALL=(limbs)) OR (ALL=(head)) )), refined in the YEARS from 2015 to 2023, that returned 60 documents;and
- Scholarallintitle: children OR babies OR infants OR motion OR movements “computer vision”, refined in the YEARS from 2015 to 2023, that returned 132 documents.
3.1. Common Datasets and Tools for Human Pose Estimation
Dataset | Reference | Number of Subjects | Type of Data | Age of Subjects | Year | Publicly Available |
---|---|---|---|---|---|---|
SSBD | [124] | 75 | RGB videos (attributes of the behaviour) | Neonates: (age: 0–7 m) | 2013 | Yes |
MINI-RGBD | [115] | 12 | Synthetic videos (RGB, Depth, 2D-3D joint positions) | Neonates: (age: 0–7 m) | 2018 | Yes |
BHT | [117] | 20 | RGB images (body parts segmentation) | Neonates: (age: 0–6 m) | 2019 | No |
babyPose | [122,123] | 16 | depth Videos (limb-joint locations) | Neonates: (Gestation Period: 24–37 w) | 2019 | Yes |
Youtube-infant | [119] | 104 | Videos: (BINS score) | Neonates: (age: 6–9 w) | 2020 | Yes |
SyRip | [116] | 140 | Synthetic and Real Images: (fully 2D body joints) | Neonates: (age: 0–12 m) | 2021 | Yes |
AIMS | [118] | NA | Synthetic and Real Images: (AIMS pose label) | Neonates: (age: 0–6 m) | 2022 | No |
3.2. Monitoring of Lying Children
3.3. Posture/Gait Analysis
3.4. New Research Directions for More Accurate Infants Pose Estimation
4. Discussion
4.1. Gaps and Open Challenges
- Privacy and ethical concerns: Collecting data from children requires strict adherence to privacy laws and regulations, such as the General Data Protection Regulation (GDPR) and the Children’s Online Privacy Protection Act (COPPA) in the United States. These laws require obtaining explicit consent from parents or guardians and ensuring the anonymity and security of children’s personal information. Meeting these requirements can be complex and time-consuming.
- Parental consent: Obtaining parental consent for data collection can be difficult, especially if it involves sensitive information or requires active participation from children. Parents may be concerned about the potential risks of data misuse or the potential impact on their child’s privacy. Building trust and addressing these concerns is crucial, and it often involves clear communication and transparency about data handling practices.
- Limited accessibility: Children may have limited access to technology or may not be able to provide consistent or reliable data due to various factors like socioeconomic disparities, geographical location, or cultural norms. This can result in biased or incomplete datasets, which can negatively impact the performance and fairness of AI models.
- Dynamic and diverse nature of children’s behaviour: Children’s behaviour, cognition, and language skills undergo rapid development and change over time. Creating a dataset that adequately captures this dynamic nature requires extensive longitudinal studies, which can be resource-intensive and time-consuming.
- Ethical considerations in data collection: Collecting data from vulnerable populations, such as children, requires special care to ensure their well-being and protection. Researchers must consider the potential emotional or psychological impact on children and ensure that the data collection process is designed ethically and with sensitivity.
- Limited sample size: Children constitute a smaller population subset compared to adults, making it challenging to gather a sufficiently large and diverse dataset. Limited data can lead to overfitting, where the AI model performs well on the training data but fails to generalize to new examples.
- Consent withdrawal and data management: Children’s participation in data collection should be voluntary, and they or their parents should have the right to withdraw consent at any time. Managing and removing data associated with withdrawn consent can be challenging, especially if it has already been incorporated into AI training models.
4.2. Ethico-Legal Considerations
- Privacy: the privacy and confidentiality of children and their parents are treated with high standards, as already introduced in the previous section. This hinders the rapid development of technology to some extent but ensures that children’s dignity and respect are properly taken into account. It is worth noting that when ambient intelligence comes into play, privacy becomes an issue not only for patients and parents but also for clinicians and caregivers. Addressing this issue at the technical level requires the adoption of privacy-preservation approaches such as those based on privacy-preserving visual sensors (e.g., depth or thermal sensors) or those based on ad hoc techniques able to ensure context-aware visual privacy and retain all the information contained in RGB cameras [166]. This may help reduce the feeling of intrusion in parents and caregivers.
- Extensive validation: scientists are aware of the inherent limitations of data-inductive techniques, such as those CV methods that use machine learning approaches. The accuracy of these methods is closely related to the type and quality of data used to train and develop them. For this reason, it is very important to perform extensive technical and clinical validation of such methods to verify their ability to generalise and handle unknown conditions. Standardised external validation and multi-centre studies should be carefully planned, together with standardised evaluation metrics, to demonstrate the reliability of the methods developed, particularly in terms of generalisability, safety and clinical value.
- Transparency: the use of technology should be made clear and transparent, thus avoiding any grey areas and uncertainties in their adoption. This entails accounting for the relevant details about the data used, the actors involved, the choices and processes enacted during development along with the main scope and limitations of the CV and ambient intelligence tools. In addition, meaningful motivations behind their outputs should be provided, especially when they are used to support diagnostic and prognostic processes. Only this way, end-users and beneficiaries, mainly children, caregivers, clinicians, nurses and parents can really be aware and empowered by the CV- and AI-powered technologies and gather trust in them [167,168,169]. The final goal is actually to contribute to collaborative decision-making, by augmenting caregivers and recipients with powerful information-processing tools.
- Accountability: healthcare professionals are responsible for justifying their actions and decisions to patients and their families, and are liable for any potential positive or negative impact on the patient’s health. The use of decision support technologies, such as those based on CV and ambient intelligence, should be clearly modelled in the legal framework of medical liability to avoid any grey area when clinicians decide to use the results of a tool or follow a suggestion received. This is still a very controversial issue. On a technical level, CV applications can implement traceability tools that document their entire development lifecycle, making it easier to deal with cases where something goes wrong.
5. Conclusions
- The collection and availability of larger datasets, also covering longer periods of children monitoring;
- The improvement of current solutions thanks to more precise and advanced methods, also based on foundational vision models;
- The integration of different types of visual sensors, such as thermal cameras that might provide relevant information for instance about the development of the thermoregulatory system of newborns;
- The integrated processing of multimodal data, such as audio signals (e.g., to monitor children’s crying), IoT data (e.g., from smart mattresses) and videos, thereby allowing, for example, a comprehensive monitoring of the health and well-being status of newborns in nurseries or in NICUs;
- The optimization of computing and sensing facilities to enable technology diffusion in resource-limited and most needy countries.
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Leo, M.; Farinella, G.M. Computer Vision for Assistive Healthcare; Academic Press: Cambridge, MA, USA, 2018. [Google Scholar]
- Esteva, A.; Chou, K.; Yeung, S.; Naik, N.; Madani, A.; Mottaghi, A.; Liu, Y.; Topol, E.; Dean, J.; Socher, R. Deep learning-enabled medical computer vision. NPJ Digit. Med. 2021, 4, 5. [Google Scholar] [CrossRef]
- Aleksic, S.; Atanasov, M.; Agius, J.C.; Camilleri, K.; Cartolovni, A.; Climent-Peerez, P.; Colantonio, S.; Cristina, S.; Despotovic, V.; Ekenel, H.K.; et al. State of the Art of Audio- and Video-Based Solutions for AAL. arXiv 2022, arXiv:2207.01487. [Google Scholar]
- Haque, A.; Milstein, A.; Li, F.-F. Illuminating the dark spaces of healthcare with ambient intelligence. Nature 2020, 585, 202. [Google Scholar] [CrossRef] [PubMed]
- Andreu, Y.; Chiarugi, F.; Colantonio, S.; Giannakakis, G.; Giorgi, D.; Henriquez, P.; Kazantzaki, E.; Manousos, D.; Marias, K.; Matuszewski, B.J.; et al. Wize Mirror—A smart, multisensory cardio-metabolic risk monitoring system. Comput. Vis. Image Underst. 2016, 148, 3–22. [Google Scholar] [CrossRef]
- Chaddad, A.; Peng, J.; Xu, J.; Bouridane, A. Survey of Explainable AI Techniques in Healthcare. Sensors 2023, 23, 634. [Google Scholar] [CrossRef] [PubMed]
- Dunne, R.; Morris, T.; Harper, S. A survey of ambient intelligence. ACM Comput. Surv. 2021, 54, 1–27. [Google Scholar] [CrossRef]
- Leo, M.; Carcagnì, P.; Mazzeo, P.L.; Spagnolo, P.; Cazzato, D.; Distante, C. Analysis of facial information for healthcare applications: A survey on computer vision-based approaches. Information 2020, 11, 128. [Google Scholar] [CrossRef]
- Dimitri, P. Child health technology: Shaping the future of paediatrics and child health and improving NHS productivity. Arch. Dis. Child. 2019, 104, 184–188. [Google Scholar] [CrossRef]
- Sacks, L.; Kunkoski, E.; Noone, M. Digital Health Technologies in Pediatric Trials. Ther. Innov. Regul. Sci. 2022, 56, 929–933. [Google Scholar] [CrossRef]
- Senechal, E.; Jeanne, E.; Tao, L.; Kearney, R.; Shalish, W.; Sant’Anna, G. Wireless monitoring devices in hospitalized children: A scoping review. Eur. J. Pediatr. 2023, 182, 1991–2003. [Google Scholar] [CrossRef]
- Leo, M.; Bernava, G.M.; Carcagnì, P.; Distante, C. Video-Based Automatic Baby Motion Analysis for Early Neurological Disorder Diagnosis: State of the Art and Future Directions. Sensors 2022, 22, 866. [Google Scholar] [CrossRef]
- Silva, N.; Zhang, D.; Kulvicius, T.; Gail, A.; Barreiros, C.; Lindstaedt, S.; Kraft, M.; Bölte, S.; Poustka, L.; Nielsen-Saines, K.; et al. The future of General Movement Assessment: The role of computer vision and machine learning—A scoping review. Res. Dev. Disabil. 2021, 110, 103854. [Google Scholar] [CrossRef] [PubMed]
- Marcroft, C.; Khan, A.; Embleton, N.D.; Trenell, M.; Plötz, T. Movement recognition technology as a method of assessing spontaneous general movements in high risk infants. Front. Neurol. 2015, 5, 284. [Google Scholar] [CrossRef] [PubMed]
- Hallemans, A.; Van de Walle, P.; Wyers, L.; Verheyen, K.; Schoonjans, A.; Desloovere, K.; Ceulemans, B. Clinical usefulness and challenges of instrumented motion analysis in patients with intellectual disabilities. Gait Posture 2019, 71, 105–115. [Google Scholar] [CrossRef]
- Washington, P.; Park, N.; Srivastava, P.; Voss, C.; Kline, A.; Varma, M.; Tariq, Q.; Kalantarian, H.; Schwartz, J.; Patnaik, R.; et al. Data-driven diagnostics and the potential of mobile artificial intelligence for digital therapeutic phenotyping in computational psychiatry. Biol. Psychiatry Cogn. Neurosci. Neuroimaging 2020, 5, 759–769. [Google Scholar] [CrossRef] [PubMed]
- de Belen, R.A.J.; Bednarz, T.; Sowmya, A.; Del Favero, D. Computer vision in autism spectrum disorder research: A systematic review of published studies from 2009 to 2019. Transl. Psychiatry 2020, 10, 333. [Google Scholar] [CrossRef]
- Mercan, E.; Morrison, C.; Stuhaug, E.; Shapiro, L.; Tse, R. Novel computer vision analysis of nasal shape in children with unilateral cleft lip. J. Cranio-Maxillo Surg. Off. Publ. Eur. Assoc. Cranio-Maxillo Surg. 2018, 46, 35–43. [Google Scholar] [CrossRef]
- Wu, J.; Tse, R.; Shapiro, L. Automated face extraction and normalization of 3d mesh data. In Proceedings of the 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014; pp. 750–753. [Google Scholar]
- Wu, J.; Heike, C.; Birgfeld, C.; Evans, K.; Maga, M.; Morrison, C.; Saltzman, B.; Shapiro, L.; Tse, R. Measuring Symmetry in Children With Unrepaired Cleft Lip: Defining a Standard for the Three-Dimensional Midfacial Reference Plane. Cleft Palate-Craniofacial J. Off. Publ. Am. Cleft Palate-Craniofacial Assoc. 2016, 53, 695–704. [Google Scholar] [CrossRef]
- Wu, J.; Liang, S.; Shapiro, L.; Tse, R. Measuring Symmetry in Children With Cleft Lip. Part 2: Quantification of Nasolabial Symmetry Before and After Cleft Lip Repair. Cleft Palate-Craniofacial J. Off. Publ. Am. Cleft Palate-Craniofacial Assoc. 2016, 53, 705–713. [Google Scholar] [CrossRef]
- Narayanan, D.L.; Ranganath, P.; Aggarwal, S.; Dalal, A.; Phadke, S.; Mandal, K. Computer-aided Facial Analysis in Diagnosing Dysmorphic Syndromes in Indian Children. Indian Pediatr. 2019, 56, 1017–1019. [Google Scholar] [CrossRef]
- Torres Torres, M.; Valstar, M.; Henry, C.; Ward, C.; Sharkey, D. Postnatal gestational age estimation of newborns using Small Sample Deep Learning. Image Vis. Comput. 2019, 83–84, 87–99. [Google Scholar] [CrossRef] [PubMed]
- Winter, R. What’s in a face? Nat. Genet. 1996, 12, 124–129. [Google Scholar] [CrossRef] [PubMed]
- Gurovich, Y.; Hanani, Y.; Bar, O.; Fleischer, N.; Gelbman, D.; Basel-Salmon, L.; Krawitz, P.; Kamphausen, S.; Zenker, M.; Bird, L.; et al. DeepGestalt-identifying rare genetic syndromes using deep learning. arXiv 2018, arXiv:1801.07637. [Google Scholar]
- Hustinx, A.; Hellmann, F.; Sümer, Ö.; Javanmardi, B.; André, E.; Krawitz, P.; Hsieh, T.C. Improving Deep Facial Phenotyping for Ultra-rare Disorder Verification Using Model Ensembles. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5018–5028. [Google Scholar]
- Bovery, M.; Dawson, G.; Hashemi, J.; Sapiro, G. A Scalable Off-the-Shelf Framework for Measuring Patterns of Attention in Young Children and Its Application in Autism Spectrum Disorder. IEEE Trans. Affect. Comput. 2021, 12, 722–731. [Google Scholar] [CrossRef] [PubMed]
- Chang, Z.; Di Martino, M.; Aiello, R.; Baker, J.; Carpenter, K.; Compton, S.; Davis, N.; Eichner, B.; Espinosa, S.; Flowers, J.; et al. Computational Methods to Measure Patterns of Gaze in Toddlers With Autism Spectrum Disorder. JAMA Pediatr. 2021, 175, 827–836. [Google Scholar] [CrossRef]
- Varma, M.; Washington, P.; Chrisman, B.; Kline, A.; Leblanc, E.; Paskov, K.; Stockham, N.; Jung, J.Y.; Sun, M.W.; Wall, D. Identification of social engagement indicators associated with autism spectrum disorder using a game-based mobile application. J. Med. Internet Res. 2021, 24, e31830. [Google Scholar] [CrossRef]
- Hashemi, J.; Campbell, K.; Carpenter, K.; Harris, A.; Qiu, Q.; Tepper, M.; Espinosa, S.; Schaich Borg, J.; Marsan, S.; Calderbank, R.; et al. A scalable app for measuring autism risk behaviors in young children: A technical validity and feasibility study. In Proceedings of the 5th EAI International Conference on Wireless Mobile Communication and Healthcare, London, UK, 14–16 October 2015; pp. 23–27. [Google Scholar]
- Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
- Ramirez-Duque, A.; Frizera-Neto, A.; Bastos, T.F. Robot-Assisted Autism Spectrum Disorder Diagnostic Based on Artificial Reasoning. J. Intell. Robot Syst. 2019, 96, 267–281. [Google Scholar] [CrossRef]
- Prinsen, V.; Jouvet, P.; Al Omar, S.; Masson, G.; Bridier, A.; Noumeir, R. Automatic eye localization for hospitalized infants and children using convolutional neural networks. Int. J. Med. Inform. 2021, 146, 104344. [Google Scholar] [CrossRef]
- King, D.E. Dlib-ml: A machine learning toolkit. J. Mach. Learn. Res. 2009, 10, 1755–1758. [Google Scholar]
- Baltrusaitis, T.; Robinson, P.; Morency, L.P. OpenFace: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the Advances in Neural Information Processing Systems 28: Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Zaker, N.; Mahoor, M.H.; Mattson, W.I.; Messinger, D.S.; Cohn, J.F. A comparison of alternative classifiers for detecting occurrence and intensity in spontaneous facial expression of infants with their mothers. In Proceedings of the 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Shanghai, China, 22–26 April 2013; pp. 1–6. [Google Scholar]
- Zanette, S.; Gao, X.; Brunet, M.; Bartlett, M.S.; Lee, K. Automated decoding of facial expressions reveals marked differences in children when telling antisocial versus prosocial lies. J. Exp. Child Psychol. 2016, 150, 165–179. [Google Scholar] [CrossRef]
- Del Coco, M.; Leo, M.; Carcagnì, P.; Spagnolo, P.; Mazzeo, P.L.; Bernava, M.; Marino, F.; Pioggia, G.; Distante, C. A Computer Vision Based Approach for Understanding Emotional Involvements in Children with Autism Spectrum Disorders. In Proceedings of the 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), Venice, Italy, 22–29 October 2017; pp. 1401–1407. [Google Scholar]
- Sun, Y.; Shan, C.; Tan, T.; Long, X.; Pourtaherian, A.; Zinger, S.; de With, P. Video-based discomfort detection for infants. Mach. Vis. Appl. 2019, 30, 933–944. [Google Scholar] [CrossRef]
- Zamzmi, G.; Paul, R.; Salekin, M.S.; Goldgof, D.; Kasturi, R.; Ho, T.; Sun, Y. Convolutional Neural Networks for Neonatal Pain Assessment. IEEE Trans. Biom. Behav. Identity Sci. 2019, 1, 192–200. [Google Scholar] [CrossRef]
- Filntisis, P.P.; Efthymiou, N.; Koutras, P.; Potamianos, G.; Maragos, P. Fusing Body Posture With Facial Expressions for Joint Recognition of Affect in Child–Robot Interaction. IEEE Robot. Autom. Lett. 2019, 4, 4011–4018. [Google Scholar] [CrossRef]
- Nagpal, S.; Singh, M.; Vatsa, M.; Singh, R.; Noore, A. Expression Classification in Children Using Mean Supervised Deep Boltzmann Machine. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 236–245. [Google Scholar]
- Martinez, A.; Pujol, F.; Mora, H. Application of Texture Descriptors to Facial Emotion Recognition in Infants. Appl. Sci. 2020, 10, 1115. [Google Scholar] [CrossRef]
- Jarraya, S.K.; Masmoudi, M.; Hammami, M. A comparative study of Autistic Children Emotion recognition based on Spatio-Temporal and Deep analysis of facial expressions features during a Meltdown Crisis. Multimed. Tools Appl. 2021, 80, 83–125. [Google Scholar] [CrossRef]
- Jarraya, S.K.; Masmoudi, M.; Hammami, M. Compound Emotion Recognition of Autistic Children During Meltdown Crisis Based on Deep Spatio-Temporal Analysis of Facial Geometric Features. IEEE Access 2020, 8, 69311–69326. [Google Scholar] [CrossRef]
- Carpenter, K.; Hahemi, J.; Campbell, K.; Lippmann, S.; Baker, J.; Egger, H.; Espinosa, S.; Vermeer, S.; Sapiro, G.; Dawson, G. Digital Behavioral Phenotyping Detects Atypical Pattern of Facial Expression in Toddlers with Autism. Autism Res. Off. J. Int. Soc. Autism Res. 2021, 14, 488–499. [Google Scholar] [CrossRef]
- Hoti, K.; Chivers, P.T.; Hughes, J.D. Assessing procedural pain in infants: A feasibility study evaluating a point-of-care mobile solution based on automated facial analysis. Lancet Digit. Health 2021, 3, 623–634. [Google Scholar] [CrossRef]
- Salekin, S.; Zamzmi, G.; Goldgof, D.; Kasturi, R.; Ho, T.; Sun, Y. Multimodal spatio-temporal deep learning approach for neonatal postoperative pain assessment. Comput. Biol. Med. 2021, 129, 104150. [Google Scholar] [CrossRef]
- Washington, P.; Kalantarian, H.; Kent, J.; Husic, A.; Kline, A.; Leblanc, E.; Hou, C.; Mutlu, C.; Dunlap, K.; Penev, Y.; et al. Training affective computer vision models by crowdsourcing soft-target labels. Cogn. Comput. 2021, 13, 1363–1373. [Google Scholar] [CrossRef]
- Ertugrul, I.O.; Ahn, Y.A.; Bilalpur, M.; Messinger, D.S.; Speltz, M.L.; Cohn, J.F. Infant AFAR: Automated facial action recognition in infants. Behav. Res. Methods 2022, 55, 1024–1035. [Google Scholar] [CrossRef]
- Zamzmi, G.; Pai, C.Y.; Goldgof, D.; Kasturi, R.; Ashmeade, T.; Sun, Y. A Comprehensive and Context-Sensitive Neonatal Pain Assessment Using Computer Vision. IEEE Trans. Affect. Comput. 2022, 13, 28–45. [Google Scholar] [CrossRef]
- Qayyum, A.; Razzak, I.; Moustafa, N.; Mazher, M. Progressive ShallowNet for large scale dynamic and spontaneous facial behaviour analysis in children. Image Vis. Comput. 2022, 119, 104375. [Google Scholar] [CrossRef]
- Leo, M.; Carcagnì, P.; Distante, C.; Mazzeo, P.L.; Spagnolo, P.; Levante, A.; Petrocchi, S.; Lecciso, F. Computational Analysis of Deep Visual Data for Quantifying Facial Expression Production. Appl. Sci. 2019, 9, 4542. [Google Scholar] [CrossRef]
- Leo, M.; Carcagnì, P.; Distante, C.; Spagnolo, P.; Mazzeo, P.L.; Rosato, A.C.; Petrocchi, S.; Pellegrino, C.; Levante, A.; De Lumè, F.; et al. Computational Assessment of Facial Expression Production in ASD Children. Sensors 2018, 18, 3993. [Google Scholar] [CrossRef]
- Leo, M.; Carcagnì, P.; Coco, M.D.; Spagnolo, P.; Mazzeo, P.L.; Celeste, G.; Distante, C.; Lecciso, F.; Levante, A.; Rosato, A.C.; et al. Towards the automatic assessment of abilities to produce facial expressions: The case study of children with ASD. In Proceedings of the 20th Italian National Conference on Photonic Technologies (Fotonica 2018), Lecce, Italy, 23–25 May 2018; pp. 1–4. [Google Scholar]
- Oster, H. Baby Facs: Facial Action Coding System for Infants and Young Children; Unpublished Monograph and Coding Manual; New York University: New York, NY, USA, 2006. [Google Scholar]
- Li, C.; Pourtaherian, A.; van Onzenoort, L.; Ten, W.E.T.A.; de With, P.H.N. Infant Facial Expression Analysis: Towards a Real-Time Video Monitoring System Using R-CNN and HMM. IEEE J. Biomed. Health Inform. 2021, 25, 1429–1440. [Google Scholar] [CrossRef] [PubMed]
- Brahnam, S.; Chuang, C.F.; Shih, F.Y.; Slack, M.R. SVM classification of neonatal facial images of pain. In Proceedings of the International Workshop on Fuzzy Logic and Applications, Crema, Italy, 15–17 September 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 121–128. [Google Scholar]
- Brahnam, S.; Chuang, C.F.; Sexton, R.; Shih, F. Machine assessment of neonatal facial expressions of acute pain. Decis. Support Syst. 2007, 43, 1242–1254. [Google Scholar] [CrossRef]
- Redmon, J.; Farhadi, A. Yolov3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
- Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Salekin, M.S.; Zamzmi, G.; Hausmann, J.; Goldgof, D.; Kasturi, R.; Kneusel, M.; Ashmeade, T.; Ho, T.; Sun, Y. Multimodal neonatal procedural and postoperative pain assessment dataset. Data Brief 2021, 35, 106796. [Google Scholar] [CrossRef] [PubMed]
- Haines, N.; Bell, Z.; Crowell, S.; Hahn, H.; Kamara, D.; McDonough-Caplan, H.; Shader, T.; Beauchaine, T. Using automated computer vision and machine learning to code facial expressions of affect and arousal: Implications for emotion dysregulation research. Dev. Psychopathol. 2019, 31, 871–886. [Google Scholar] [CrossRef]
- Bruer, K.C.; Zanette, S.; Ding, X.P.; Lyon, T.D.; Lee, K. Identifying Liars Through Automatic Decoding of Children’s Facial Expressions. Child Dev. 2020, 91, e995–e1011. [Google Scholar] [CrossRef] [PubMed]
- Chen, M.; Chow, S.M.; Hammal, Z.; Messinger, D.S.; Cohn, J.F. A person-and time-varying vector autoregressive model to capture interactive infant-mother head movement dynamics. Multivar. Behav. Res. 2020, 56, 739–767. [Google Scholar] [CrossRef]
- Hammal, Z.; Cohn, J.F.; Messinger, D.S. Head movement dynamics during play and perturbed mother-infant interaction. IEEE Trans. Affect. Comput. 2015, 6, 361–370. [Google Scholar] [CrossRef] [PubMed]
- Hammal, Z.; Chu, W.S.; Cohn, J.F.; Heike, C.; Speltz, M.L. Automatic action unit detection in infants using convolutional neural network. In Proceedings of the 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII), San Antonio, TX, USA, 23–26 October 2017; pp. 216–221. [Google Scholar]
- Khan, R.A.; Crenn, A.; Meyer, A.; Bouakaz, S. A novel database of children’s spontaneous facial expressions (LIRIS-CSE). Image Vis. Comput. 2019, 83, 61–69. [Google Scholar] [CrossRef]
- Salakhutdinov, R.; Hinton, G. Deep boltzmann machines. In Proceedings of the 12th International Conference on Artificial Intelligence and Statistics, Clearwater Beach, FL, USA, 16–18 April 2009; pp. 448–455. [Google Scholar]
- Ekman, P. Facial Expressions of Emotion: New Findings, New Questions. Psychol. Sci. 1992, 3, 34–38. [Google Scholar] [CrossRef]
- Langner, O.; Dotsch, R.; Bijlstra, G.; Wigboldus, D.H.; Hawk, S.T.; Van Knippenberg, A. Presentation and validation of the Radboud Faces Database. Cogn. Emot. 2010, 24, 1377–1388. [Google Scholar] [CrossRef]
- LoBue, V.; Trasher, C. The Child Affective Facial Expression (CAFE) Set: Validity and Reliability from Untrained Adults. Front. Psychol. 2014, 5, 1532. [Google Scholar] [CrossRef]
- Belpaeme, T.; Baxter, P.; De Greeff, J.; Kennedy, J.; Read, R.; Looije, R.; Neerincx, M.; Baroni, I.; Zelati, M.C. Child-robot interaction: Perspectives and challenges. In Proceedings of the Social Robotics: 5th International Conference—ICSR 2013, Bristol, UK, 27–29 October 2013; Springer: Cham, Switzerland, 2013; pp. 452–459. [Google Scholar]
- Xu, K.; Ji, B.; Wang, Z.; Liu, J.; Liu, H. An Auxiliary Screening System for Autism Spectrum Disorder Based on Emotion and Attention Analysis. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 2299–2304. [Google Scholar]
- Zhang, Y.; Kong, M.; Zhao, T.; Hong, W.; Xie, D.; Wang, C.; Yang, R.; Li, R.; Zhu, Q. Auxiliary diagnostic system for ADHD in children based on AI technology. Front. Inf. Technol. Electron. Eng. 2021, 22, 400–414. [Google Scholar] [CrossRef]
- Xiong, X.; De la Torre, F. Supervised Descent Method and Its Applications to Face Alignment. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 532–539. [Google Scholar]
- Kanade, T.; Cohn, J.; Tian, Y. Comprehensive database for facial expression analysis. In Proceedings of the Fourth IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France, 28–30 March 2000; pp. 46–53. [Google Scholar]
- Nag, A.; Haber, N.; Voss, C.; Tamura, S.; Daniels, J.; Ma, J.; Chiang, B.; Ramachandran, S.; Schwartz, J.N.; Winograd, T.; et al. Toward Continuous Social Phenotyping: Analyzing Gaze Patterns in an Emotion Recognition Task for Children With Autism Through Wearable Smart Glasses. J. Med. Internet Res. 2020, 22, e13810. [Google Scholar] [CrossRef] [PubMed]
- Ekman, P.; Friesen, W. Pictures of Facial Affect; Consulting Psychologists Press: Palo Alto, CA, USA, 1976. [Google Scholar]
- Luquetti, D.V.; Speltz, M.L.; Wallace, E.R.; Siebold, B.; Collett, B.R.; Drake, A.F.; Johns, A.L.; Kapp-Simon, K.A.; Kinter, S.L.; Leroux, B.G.; et al. Methods and challenges in a cohort study of infants and toddlers with craniofacial microsomia: The CLOCK study. Cleft Palate-Craniofacial J. 2019, 56, 877–889. [Google Scholar] [CrossRef] [PubMed]
- Adamson, L.B.; Frick, J.E. The Still Face: A History of a Shared Experimental Paradigm. Infancy 2003, 4, 451–473. [Google Scholar] [CrossRef]
- Hudson-Barr, D.; Capper-Michel, B.; Lambert, S.; Palermo, T.M.; Morbeto, K.; Lombardo, S. Validation of the pain assessment in neonates (PAIN) scale with the neonatal infant pain scale (NIPS). Neonatal Netw. 2002, 21, 15–21. [Google Scholar] [CrossRef]
- Taigman, Y.; Yang, M.; Ranzato, M.; Wolf, L. DeepFace: Closing the Gap to Human-Level Performance in Face Verification. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar] [CrossRef]
- Schroff, F.; Kalenichenko, D.; Philbin, J. FaceNet: A Unified Embedding for Face Recognition and Clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- Ruiz, N.; Chong, E.; Rehg, J.M. Fine-Grained Head Pose Estimation Without Keypoints. In Proceedings of the The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Guo, J.; Zhu, X.; Yang, Y.; Yang, F.; Lei, Z.; Li, S.Z. Towards Fast, Accurate and Stable 3D Dense Face Alignment. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Prados-Torreblanca, A.; Buenaposada, J.M.; Baumela, L. Shape Preserving Facial Landmarks with Graph Attention Networks. In Proceedings of the 33rd British Machine Vision Conference 2022—BMVC 2022, London, UK, 21–24 November 2022; BMVA Press: Surrey, UK, 2022. [Google Scholar]
- Albiero, V.; Chen, X.; Yin, X.; Pang, G.; Hassner, T. img2pose: Face Alignment and Detection via 6DoF, Face Pose Estimation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
- Hempel, T.; Abdelrahman, A.A.; Al-Hamadi, A. 6d Rotation Representation For Unconstrained Head Pose Estimation. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 2496–2500. [Google Scholar] [CrossRef]
- Abate, A.F.; Bisogni, C.; Castiglione, A.; Nappi, M. Head Pose Estimation: An Extensive Survey on Recent Techniques and Applications. Pattern Recogn. 2022, 127, 108591. [Google Scholar] [CrossRef]
- Asperti, A.; Filippini, D. Deep Learning for Head Pose Estimation: A Survey. SN Comput. Sci. 2023, 4, 349. [Google Scholar] [CrossRef]
- Deng, J.; Guo, J.; Niannan, X.; Zafeiriou, S. ArcFace: Additive Angular Margin Loss for Deep Face Recognition. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
- Deng, J.; Guo, J.; Zhou, Y.; Yu, J.; Kotsia, I.; Zafeiriou, S. RetinaFace: Single-stage Dense Face Localisation in the Wild. arXiv 2019, arXiv:1905.00641. [Google Scholar]
- Qi, D.; Tan, W.; Yao, Q.; Liu, J. YOLO5Face: Why Reinventing a Face Detector. arXiv 2021, arXiv:2105.12931. [Google Scholar]
- Bharadwaj, S.; Bhatt, H.S.; Vatsa, M.; Singh, R. Domain Specific Learning for Newborn Face Recognition. IEEE Trans. Inf. Forensics Secur. 2016, 11, 1630–1641. [Google Scholar] [CrossRef]
- Wen, D.; Fang, C.; Ding, X.; Zhang, T. Development of Recognition Engine for Baby Faces. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 3408–3411. [Google Scholar]
- Wang, Z.; Zeng, F.; Liu, S.; Zeng, B. OAENet: Oriented attention ensemble for accurate facial expression recognition. Pattern Recognit. 2021, 112, 107694. [Google Scholar] [CrossRef]
- Zhuang, X.; Liu, F.; Hou, J.; Hao, J.; Cai, X. Transformer-Based Interactive Multi-Modal Attention Network for Video Sentiment Detection. Neural Process. Lett. 2022, 54, 1943–1960. [Google Scholar] [CrossRef]
- Yang, H.; Zhu, K.; Huang, D.; Li, H.; Wang, Y.; Chen, L. Intensity enhancement via GAN for multimodal face expression recognition. Neurocomputing 2021, 454, 124–134. [Google Scholar] [CrossRef]
- Zhang, T.; Tang, K. An Efficacious Method for Facial Expression Recognition: GAN Erased Facial Feature Network (GE2FN). In Proceedings of the 2021 13th International Conference on Machine Learning and Computing (ICMLC 2021), Shenzhen China, 26 February–1 March 2021; pp. 417–422. [Google Scholar] [CrossRef]
- Schoneveld, L.; Othmani, A.; Abdelkawy, H. Leveraging recent advances in deep learning for audio-Visual emotion recognition. Pattern Recognit. Lett. 2021, 146, 1–7. [Google Scholar] [CrossRef]
- Sciortino, G.; Farinella, G.M.; Battiato, S.; Leo, M.; Distante, C. On the estimation of children’s poses. In Proceedings of the International Conference on Image Analysis and Processing, Catania, Italy, 11–15 September 2017; pp. 410–421. [Google Scholar]
- Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.; Sheikh, Y. OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 172–186. [Google Scholar] [CrossRef] [PubMed]
- Zha, D.; Bhat, Z.P.; Chen, Y.W.; Wang, Y.; Ding, S.; Chen, J.; Lai, K.H.; Bhat, M.Q.; Jain, A.K.; Reyes, A.C.; et al. Autovideo: An automated video action recognition system. arXiv 2021, arXiv:2108.04212. [Google Scholar]
- Mathis, A.; Mamidanna, P.; Cury, K.M.; Abe, T.; Murthy, V.N.; Mathis, M.W.; Bethge, M. DeepLabCut: Markerless pose estimation of user-defined body parts with deep learning. Nat. Neurosci. 2018, 21, 1281–1289. [Google Scholar] [CrossRef]
- Torralba, A.; Russell, B.C.; Yuen, J. Labelme: Online image annotation and applications. Proc. IEEE 2010, 98, 1467–1484. [Google Scholar] [CrossRef]
- Liu, S.; Huang, X.; Fu, N.; Ostadabbas, S. Heuristic weakly supervised 3d human pose estimation in novel contexts without any 3d pose ground truth. arXiv 2021, arXiv:2105.10996. [Google Scholar]
- Bernava, G.M.; Leo, M.; Carcagnì, P.; Distante, C. An Advanced Tool for Semi-automatic Annotation for Early Screening of Neurodevelopmental Disorders. In Proceedings of the Image Analysis and Processing—ICIAP 2022 Workshops: ICIAP International Workshops, Lecce, Italy, 23–27 May 2022; Revised Selected Papers, Part II. Springer: Cham, Swizterland, 2022; pp. 154–164. [Google Scholar]
- Bazarevsky, V.; Grishchenko, I.; Raveendran, K.; Zhu, T.; Zhang, F.; Grundmann, M. Blazepose: On-device real-time body pose tracking. arXiv 2020, arXiv:2006.10204. [Google Scholar]
- Orlandi, S.; Guzzetta, A.; Bandini, A.; Belmonti, V.; Barbagallo, S.D.; Tealdi, G.; Mazzotti, S.; Scattoni, M.L.; Manfredi, C. AVIM—A contactless system for infant data acquisition and analysis: Software architecture and first results. Biomed. Signal Process. Control 2015, 20, 85–99. [Google Scholar] [CrossRef]
- Baccinelli, W.; Bulgheroni, M.; Simonetti, V.; Fulceri, F.; Caruso, A.; Gila, L.; Scattoni, M.L. Movidea: A software package for automatic video analysis of movements in infants at risk for neurodevelopmental disorders. Brain Sci. 2020, 10, 203. [Google Scholar] [CrossRef]
- Hesse, N.; Bodensteiner, C.; Arens, M.; Hofmann, U.G.; Weinberger, R.; Sebastian Schroeder, A. Computer vision for medical infant motion analysis: State of the art and rgb-d data set. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018; pp. 32–49. [Google Scholar]
- Huang, X.; Fu, N.; Liu, S.; Ostadabbas, S. Invariant representation learning for infant pose estimation with small data. In Proceedings of the 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), Jodhpur, India, 15–18 December 2021; pp. 1–8. [Google Scholar]
- Zhang, Q.; Xue, Y.; Huang, X. Online training for body part segmentation in infant movement videos. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 489–492. [Google Scholar]
- Yang, C.Y.; Jiang, Z.; Gu, S.Y.; Hwang, J.N.; Yoo, J.H. Unsupervised Domain Adaptation Learning for Hierarchical Infant Pose Recognition with Synthetic Data. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022; pp. 1–6. [Google Scholar]
- Chambers, C.; Seethapathi, N.; Saluja, R.; Loeb, H.; Pierce, S.R.; Bogen, D.K.; Prosser, L.; Johnson, M.J.; Kording, K.P. Computer vision to automatically assess infant neuromotor risk. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 2431–2442. [Google Scholar] [CrossRef]
- Aylward, G. Bayley Infant Neurodevelopmental Screener; Pearson: San Antonio, TX, USA, 1995. [Google Scholar]
- Piper, M.C.; Pinnell, L.E.; Darrah, J.; Maguire, T.; Byrne, P.J. Construction and validation of the Alberta Infant Motor Scale (AIMS). Can. J. Public Health Rev. Can. De Sante Publique 1992, 83, S46–S50. [Google Scholar]
- Moccia, S.; Migliorelli, L.; Carnielli, V.; Frontoni, E. Preterm infants’ pose estimation with spatio-temporal features. IEEE Trans. Biomed. Eng. 2019, 67, 2370–2380. [Google Scholar] [CrossRef]
- Migliorelli, L.; Moccia, S.; Pietrini, R.; Carnielli, V.P.; Frontoni, E. The babyPose dataset. Data Brief 2020, 33, 106329. [Google Scholar] [CrossRef]
- Rajagopalan, S.; Dhall, A.; Goecke, R. Self-stimulatory behaviours in the wild for autism diagnosis. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 755–761. [Google Scholar]
- Khanam, F.T.Z.; Perera, A.G.; Al-Naji, A.; Gibson, K.; Chahl, J. Non-contact automatic vital signs monitoring of infants in a neonatal intensive care unit based on neural networks. J. Imaging 2021, 7, 122. [Google Scholar] [CrossRef]
- Einspieler, C.; Prechtl, H.F. Prechtl’s assessment of general movements: A diagnostic tool for the functional assessment of the young nervous system. Ment. Retard. Dev. Disabil. Res. Rev. 2005, 11, 61–67. [Google Scholar] [CrossRef] [PubMed]
- Sun, Y.; Kommers, D.; Wang, W.; Joshi, R.; Shan, C.; Tan, T.; Aarts, R.M.; van Pul, C.; Andriessen, P.; de With, P.H. Automatic and continuous discomfort detection for premature infants in a NICU using video-based motion analysis. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 5995–5999. [Google Scholar]
- Gibson, K.; Al-Naji, A.; Fleet, J.; Steen, M.; Esterman, A.; Chahl, J.; Huynh, J.; Morris, S. Non-contact heart and respiratory rate monitoring of preterm infants based on a computer vision system: A method comparison study. Pediatr. Res. 2019, 86, 738–741. [Google Scholar] [CrossRef] [PubMed]
- Hussain, T.; Muhammad, K.; Khan, S.; Ullah, A.; Lee, M.Y.; Baik, S.W. Intelligent baby behavior monitoring using embedded vision in IoT for smart healthcare centers. J. Artif. Intell. Syst. 2019, 1, 110–124. [Google Scholar] [CrossRef]
- Sahin, I.; Modi, A.; Kokkoni, E. Evaluation of OpenPose for Quantifying Infant Reaching Motion. Arch. Phys. Med. Rehabil. 2021, 102, e86. [Google Scholar] [CrossRef]
- Balta, D.; Kuo, H.; Wang, J.; Porco, I.G.; Morozova, O.; Schladen, M.M.; Cereatti, A.; Lum, P.S.; Della Croce, U. Characterization of Infants’ General Movements Using a Commercial RGB-Depth Sensor and a Deep Neural Network Tracking Processing Tool: An Exploratory Study. Sensors 2022, 22, 7426. [Google Scholar] [CrossRef] [PubMed]
- Moro, M.; Pastore, V.P.; Tacchino, C.; Durand, P.; Blanchi, I.; Moretti, P.; Odone, F.; Casadio, M. A markerless pipeline to analyze spontaneous movements of preterm infants. Comput. Methods Programs Biomed. 2022, 226, 107119. [Google Scholar] [CrossRef]
- Ni, H.; Xue, Y.; Ma, L.; Zhang, Q.; Li, X.; Huang, S.X. Semi-supervised body parsing and pose estimation for enhancing infant general movement assessment. Med. Image Anal. 2023, 83, 102654. [Google Scholar] [CrossRef] [PubMed]
- Dechemi, A.; Bhakri, V.; Sahin, I.; Modi, A.; Mestas, J.; Peiris, P.; Barrundia, D.E.; Kokkoni, E.; Karydis, K. Babynet: A lightweight network for infant reaching action recognition in unconstrained environments to support future pediatric rehabilitation applications. In Proceedings of the 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN), Vancouver, BC, Canada, 8–12 August 2021; pp. 461–467. [Google Scholar]
- Huang, X.; Wan, M.; Luan, L.; Tunik, B.; Ostadabbas, S. Computer Vision to the Rescue: Infant Postural Symmetry Estimation from Incongruent Annotations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 1909–1917. [Google Scholar]
- Rehouma, H.; Noumeir, R.; Jouvet, P.; Bouachir, W.; Essouri, S. A computer vision method for respiratory monitoring in intensive care environment using RGB-D cameras. In Proceedings of the 2017 Seventh International Conference on Image Processing Theory, Tools and Applications (IPTA), Montreal, QC, Canada, 28 November–1 December 2017; pp. 1–6. [Google Scholar]
- Kidziński, Ł.; Yang, B.; Hicks, J.L.; Rajagopal, A.; Delp, S.L.; Schwartz, M.H. Deep neural networks enable quantitative movement analysis using single-camera videos. Nat. Commun. 2020, 11, 4054. [Google Scholar] [CrossRef] [PubMed]
- Ossmy, O.; Adolph, K.E. Real-time assembly of coordination patterns in human infants. Curr. Biol. 2020, 30, 4553–4562. [Google Scholar] [CrossRef] [PubMed]
- Ali, A.; Negin, F.F.; Bremond, F.F.; Thümmler, S. Video-based behavior understanding of children for objective diagnosis of autism. In Proceedings of the VISAPP 2022—17th International Conference on Computer Vision Theory and Applications, Online. 6–8 February 2022. [Google Scholar]
- Ferrer-Mallol, E.; Matthews, C.; Stoodley, M.; Gaeta, A.; George, E.; Reuben, E.; Johnson, A.; Davies, E.H. Patient-led development of digital endpoints and the use of computer vision analysis in assessment of motor function in rare diseases. Front. Pharmacol. 2022, 13, 916714. [Google Scholar] [CrossRef]
- Jocher, G.; Stoken, A.; Borovec, J.; Changyu, L.; Hogan, A.; Diaconu, L.; Ingham, F.; Poznanski, J.; Fang, J.; Yu, L.; et al. ultralytics/yolov5: V3. 1-bug fixes and performance improvements. Zenodo 2020, 1. [Google Scholar] [CrossRef]
- Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
- Teed, Z.; Deng, J. Raft: Recurrent all-pairs field transforms for optical flow. In Proceedings of the ECCV 2020: 16th European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Proceedings, Part II 16. Springer: Cham, Switzerland, 2020; pp. 402–419. [Google Scholar]
- Zheng, C.; Zhu, S.; Mendieta, M.; Yang, T.; Chen, C.; Ding, Z. 3d human pose estimation with spatial and temporal transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 11656–11665. [Google Scholar]
- Zhang, J.; Gong, K.; Wang, X.; Feng, J. Learning to Augment Poses for 3D Human Pose Estimation in Images and Videos. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10012–10026. [Google Scholar] [CrossRef]
- Tang, Z.; Qiu, Z.; Hao, Y.; Hong, R.; Yao, T. 3D Human Pose Estimation With Spatio-Temporal Criss-Cross Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 4790–4799. [Google Scholar]
- Einfalt, M.; Ludwig, K.; Lienhart, R. Uplift and Upsample: Efficient 3D Human Pose Estimation with Uplifting Transformers. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 2903–2913. [Google Scholar]
- Garau, N.; Conci, N. CapsulePose: A variational CapsNet for real-time end-to-end 3D human pose estimation. Neurocomputing 2023, 523, 81–91. [Google Scholar] [CrossRef]
- Clark, H.; Coll-Seck, A.M.; Banerjee, A.; Peterson, S.; Ameratunga, S.; Balabanova, D.; Bhan, M.K.; Bhutta, Z.A.; Borrazzo, J.; Doherty, T.; et al. A future for the world’s children? A WHO-UNICEF-Lancet Commission. Lancet 2020, 395, 605–658. [Google Scholar] [CrossRef]
- Hanson, M.A.; Cooper, C.; Aihie Sayer, A.; Eendebak, R.J.; Clough, G.F.; Beard, J.R. Developmental aspects of a life course approach to healthy ageing. J. Physiol. 2016, 594, 2147–2160. [Google Scholar] [CrossRef]
- van Heerden, A.; Leppanen, J.; Rotheram-Borus, M.J.; Worthman, C.M.; Kohrt, B.A.; Skeen, S.; Giese, S.; Hughes, R.; Bohmer, L.; Tomlinson, M. Emerging Opportunities Provided by Technology to Advance Research in Child Health Globally. Glob. Pediatr. Health 2020, 7, 1–9. [Google Scholar] [CrossRef]
- Magrini, M.; Salvetti, O.; Carboni, A.; Curzio, O. An Interactive Multimedia System for Treating Autism Spectrum Disorder. In Proceedings of the ECCV 2016 Workshops: European Conference on Computer Vision, Amsterdam, The Netherlands, 8–10 and 15–16 October 2016; Hua, G., Jégou, H., Eds.; Springer: Cham, Switzerland, 2016; pp. 331–342. [Google Scholar]
- Magrini, M.; Curzio, O.; Carboni, A.; Moroni, D.; Salvetti, O.; Melani, A. Augmented Interaction Systems for Supporting Autistic Children. Evolution of a Multichannel Expressive Tool: The SEMI Project Feasibility Study. Appl. Sci. 2019, 9, 3081. [Google Scholar] [CrossRef]
- Moor, M.; Banerjee, O.; Abad, Z.S.H.; Krumholz, H.M.; Leskovec, J.; Topol, E.J.; Rajpurkar, P. Foundation models for generalist medical artificial intelligence. Nature 2023, 616, 259–265. [Google Scholar] [CrossRef]
- Lantos, J.D.; Meadow, W.L. Neonatal Bioethics: The Moral Challenges Of Medical Innovation; The Johns Hopkins University Press: Baltimore, MD, USA, 2008. [Google Scholar]
- Liu, J.; Chen, X.X.; Wang, X.L. Ethical issues in neonatal intensive care units. J. Matern.-Fetal Neonatal Med. 2016, 29, 2322–2326. [Google Scholar] [CrossRef]
- Botkin, J.R. Ethical issues in pediatric genetic testing and screening. Curr. Opin. Pediatr. 2016, 28, 700–704. [Google Scholar] [CrossRef]
- Ake-Kob, A.; Blazeviciene, A.; Colonna, L.; Cartolovni, A.; Dantas, C.; Fedosov, A.; Florez-Revuelta, F.; Fosch-Villaronga, E.; He, Z.; Klimczuk, A.; et al. State of the Art on Ethical, Legal, and Social Issues Linked to Audio- and Video-Based AAL Solutions. 2022. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4075151 (accessed on 30 July 2023). [CrossRef]
- Walsh, V.; Oddie, S.; McGuire, W. Ethical Issues in Perinatal Clinical Research. Neonatology 2019, 116, 52–57. [Google Scholar] [CrossRef] [PubMed]
- Alexander, D.; Quirke, M.; Doyle, C.; Hill, K.; Masterson, K.; Brenner, M. Technology solutionism in paediatric intensive care: Clinicians’ perspectives of bioethical considerations. BMC Med. Ethics 2023, 24, 55. [Google Scholar]
- Adams, C.; Pente, P.; Lemermeyer, G.; Rockwell, G. Artificial Intelligence Ethics Guidelines for K-12 Education: A Review of the Global Landscape. In Proceedings of the International Conference on Artificial Intelligence in Education, Utrecht, The Netherlands, 14–18 June 2021; Roll, I., McNamara, D., Sosnovsky, S., Luckin, R., Dimitrova, V., Eds.; Springer: Cham, Switzerland, 2021; pp. 24–28. [Google Scholar]
- Adams, C.; Pente, P.; Lemermeyer, G.; Rockwell, G. Ethical principles for artificial intelligence in K-12 education. Comput. Educ. Artif. Intell. 2023, 4, 100131. [Google Scholar] [CrossRef]
- McStay, A.; Rosner, G. Emotional artificial intelligence in children’s toys and devices: Ethics, governance and practical remedies. Big Data Soc. 2021, 8, 1–16. [Google Scholar] [CrossRef]
- Boch, S.; Sezgin, E.; Lin Linwood, S. Ethical artificial intelligence in paediatrics. Lancet Child Adolesc. Health 2022, 6, 833–835. [Google Scholar] [CrossRef] [PubMed]
- Thai, K.; Tsiandoulas, K.H.; Stephenson, E.A.; Menna-Dack, D.; Zlotnik Shaul, R.; Anderson, J.A.; Shinewald, A.R.; Ampofo, A.; McCradden, M.D. Perspectives of Youths on the Ethical Use of Artificial Intelligence in Health Care Research and Clinical Care. JAMA Netw. Open 2023, 6, e2310659. [Google Scholar] [CrossRef] [PubMed]
- Ravi, S.; Climent-Pérez, P.; Florez-Revuelta, F. A review on visual privacy preservation techniques for active and assisted living. Multimed. Tools Appl. 2023. [Google Scholar] [CrossRef]
- Jovanovic, M.; Mitrov, G.; Zdravevski, E.; Lameski, P.; Colantonio, S.; Kampel, M.; Tellioglu, H.; Florez-Revuelta, F. Ambient Assisted Living: Scoping Review of Artificial Intelligence Models, Domains, Technology, and Concerns. J. Med. Internet Res. 2022, 24, e36553. [Google Scholar] [CrossRef]
- Colantonio, S.; Jovanovic, M.; Zdravevski, E.; Lameski, P.; Tellioglu, H.; Kampel, M.; Florez-Revuelta, F. Are Active and Assisted Living applications addressing the main acceptance concerns of their beneficiaries? Preliminary insights from a scoping review. In Proceedings of the 15th International Conference on PErvasive Technologies Related to Assistive Environments, Corfu Island, Greece, 29 June–1 July 2022; pp. 414–421. [Google Scholar]
- Lekadir, K.; Osuala, R.; Gallin, C.; Lazrak, N.; Kushibar, K.; Tsakou, G.; Aussó, S.; Alberich, L.C.; Marias, K.; Tsiknakis, M.; et al. FUTURE-AI: Guiding Principles and Consensus Recommendations for Trustworthy Artificial Intelligence in Medical Imaging. arXiv 2021, arXiv:2109.09658. [Google Scholar]
Work (Year) | Method | Clinical Task | Metrics | Dataset Population/Age (h = hours, w = weeks, m = months, y = years) |
---|---|---|---|---|
[18,19] (2014, 2018) | Extraction of features related to nasolabial symmetry | Quantification of facial asymmetry in children pre- and post- primary cleft lip repair | Qualitative | 50 infants and 50 children (age: 8–10 y) |
[20,21] (2016) | Geometrical approach + landmarks identified by computer-based template mesh deformation | Quantification of facial asymmetry in children with unilateral cleft lip nasal deformity | Qualitative + Symmetry Scores | 49 infants (age 4–10 m) |
[22] (2019) | Face2Gene CLINIC app (based on a CNN) | Recognition of facial dysmorphisms due to genetic disorders | acc = 72.5% | 51 children |
[23] (2019) | CNN + SVR | Estimation of postnatal gestational age | 7.98 days RMSE | 130 newborns (gestational age: 28–40 w) |
Work (Year) | Method | Clinical Task | Metrics | Dataset Population/Age (h = hours, w = weeks, m = months, y = years) |
---|---|---|---|---|
[39] (2013) | AAM + HOG features; comparison PCA-LMNN vs. Laplacian Eigenmap and SVM vs. K-Nearest Neighbour | Assessment of the dynamics of face-to-face interactions with the mother | ICC | 12 infants (age: 1 m–1 y) |
[40] (2016) | CERT | Lie detection | Qualitative | Children (age: 6–11 y) |
[41] (2017) | HOG computation + Landmark Detection by CLNF + Facial AU intensities computation | Diagnosis and evaluation of ASD children | Entropy score + Similarity metrics | children (age: 5–6 y) |
[42] (2019) | Geometric and appearance features/facial landmark-based template matching + SVM | Pain assessment | AUC = 0.87/0.97 | 22 infants (age: 1 m–1 y) |
[43] (2019) | Neonatal CNN | Pain assessment | acc = 97% | 84 neonates (age: 18 h–41 w) |
[44] (2019) | CNN + ResNet | Robot assisted therapy | acc = 72% | children (age: 6–12 y) |
[45] (2019) | Mean Supervised Deep Boltzmann Machine | Emotion detection and recognition | acc = 75% | 154 children (age: 2–8 y) |
[46] (2020) | Texture and geometric descriptors: LBP, LTP and RB + SVM | Pain assessment | acc = 95% | 26 neonates (age: 18–36 h) |
[47,48] (2020, 2021) | Deep spatiotemporal geometric facial features + Recurrent Neural Network | ASD meltdown crisis management | acc = 85.8% | children (age: 4–11 y) |
[49] (2021) | Facial landmark detection and tracking | Early detection of ASD symptoms | AUC from 0.62 to 0.73 | 104 toddlers (age: 1–2 y) |
[50] (2021) | PainCheck Infant | Pain assessment | Correlation with standard scores: r = 0.82–0.88; p < 0.0001 | 40 infants (age: 2–9 m) |
[51] (2021) | YOLO face detector + VGG-16 for facial features extraction + LSTM | Pain assessment | acc = 79% | 58 neonates (age: 27–41 w) |
[52] (2021) | ResNet-152 | Emotion detection and recognition | Balanced acc = 79.1% | 154 children (age: 2–8 y) |
[53] (2022) | VGG-16 network | Emotion detection and recognition | AUC = 0.82 | 123 children (age: 1 m–5 y) |
[54] (2022) | Strain-based, geometric-based, texture-based and gradient-based facial features | Pain assessment | acc = 95.56% | 31 neonates |
[55] (2022) | Progressive lightweight shallow learning (ShallowNet) | Emotion detection and recognition | Acc = 99.06% | 12 children (age: 6–12 y) |
Work (Year) | Method | Clinical Task | Metrics | Datase Population/Age (h = hours, w = weeks, m = months, y = years) |
---|---|---|---|---|
[30] (2015) | Facal Expressions + Head Pose by IntraFace Software 2015 | Detection of early indicators of ASD | ICC | 20 toddlers (age: 16–30 m) |
[77] (2020) | facial Expression An. + Gaze Tracking by Classical computer vision methods | Detection of early indicators of ASD | acc = 97.12% | 10 children (age: 6–11 y) |
[78] (2021) | facial Expression An. + Gaze Tracking + 3D Body Pose by Classical computer vision methods | ADHD diagnosis | acc = 80.25% | children (age: 6–12 y) |
Work (Year) | Method | Clinical Task | Metrics | Dataset Population/Age (w = weeks, y = years) |
---|---|---|---|---|
[127] (2019) | Optical flow + SVM | Discomfort Moments | acc = 0.86 AUC of 0.94 | 11 (34 w) |
[128] (2019) | Skin detection + Motion magnification | Vital Sign | Limit of agreement = b.p.m., r.p.m. | 10 (23 w–40 w) |
[129] (2019) | Motion Detection | Vital Sign | acc = 87% | 1/unknown |
[119] (2020) | OpenPose + kinematic features + Naïve Bayesian Class | Neuromotor Risk | RMSE = 0.92 | 19 (10 w) |
[130] (2021) | OpenPose | Reaching Trajectories | 95% confidence in hands tracking | 12 (48 w) |
[131] (2022) | basic tracking primitives | GMA | qualitative | 8 (3 m–5 m) |
[132] (2022) | DeepLabCut + kinematic features + RF | Neuromotor Risk | acc = 0.62 | 142 (40 w) |
[133] (2023) | FVGAN +SiamParseNet | GMA | 161 (49 w–60 w) |
Work (Year) | Method | Clinical Task | Metrics | Dataset Population/Age (in years) |
---|---|---|---|---|
[137] (2020) | OpenPose + Motion Param + CNN | Gait Analysis to predict surgery | AUC = 0.71 | 1026 (5–11) |
[138] (2020) | AutoViDev + arms and legs time series distance | assessing coordination | qualitative | 24 (1) |
[139] (2022) | Optical flow + RGB + 3D CNN | ASD/Healthy | acc = 86.04% | 60 (3–6) |
[140] (2022) | OpenPose + motion parameters | Evaluating dystrophy | qualitative | 11 (13) |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Germanese, D.; Colantonio, S.; Del Coco, M.; Carcagnì, P.; Leo, M. Computer Vision Tasks for Ambient Intelligence in Children’s Health. Information 2023, 14, 548. https://doi.org/10.3390/info14100548
Germanese D, Colantonio S, Del Coco M, Carcagnì P, Leo M. Computer Vision Tasks for Ambient Intelligence in Children’s Health. Information. 2023; 14(10):548. https://doi.org/10.3390/info14100548
Chicago/Turabian StyleGermanese, Danila, Sara Colantonio, Marco Del Coco, Pierluigi Carcagnì, and Marco Leo. 2023. "Computer Vision Tasks for Ambient Intelligence in Children’s Health" Information 14, no. 10: 548. https://doi.org/10.3390/info14100548
APA StyleGermanese, D., Colantonio, S., Del Coco, M., Carcagnì, P., & Leo, M. (2023). Computer Vision Tasks for Ambient Intelligence in Children’s Health. Information, 14(10), 548. https://doi.org/10.3390/info14100548