Next Article in Journal
A Smart Camera Trap for Detection of Endotherms and Ectotherms
Next Article in Special Issue
Printed and Flexible ECG Electrodes Attached to the Steering Wheel for Continuous Health Monitoring during Driving
Previous Article in Journal
Removal of Non-Specifically Bound Proteins Using Rayleigh Waves Generated on ST-Quartz Substrates
Previous Article in Special Issue
A Wearable, Multi-Frequency Device to Measure Muscle Activity Combining Simultaneous Electromyography and Electrical Impedance Myography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review

by
Vinothini Selvaraju
1,2,
Nicolai Spicher
1,
Ju Wang
1,
Nagarajan Ganapathy
1,
Joana M. Warnecke
1,
Steffen Leonhardt
3,
Ramakrishnan Swaminathan
2 and
Thomas M. Deserno
1,*
1
Peter L. Reichertz Institute for Medical Informatics of TU Braunschweig and Hannover Medical School, D-38106 Braunschweig, Germany
2
Non-Invasive Imaging and Diagnostic Laboratory, Biomedical Engineering, Department of Applied Mechanics, Indian Institute of Technology Madras, Chennai 600036, India
3
Chair for Medical Information Technology, Helmholtz-Institute for Biomedical Engineering, RWTH Aachen University, D-52074 Aachen, Germany
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(11), 4097; https://doi.org/10.3390/s22114097
Submission received: 6 April 2022 / Revised: 18 May 2022 / Accepted: 18 May 2022 / Published: 28 May 2022
(This article belongs to the Special Issue Sensors toward Unobtrusive Health Monitoring II)

Abstract

:
In recent years, noncontact measurements of vital signs using cameras received a great amount of interest. However, some questions are unanswered: (i) Which vital sign is monitored using what type of camera? (ii) What is the performance and which factors affect it? (iii) Which health issues are addressed by camera-based techniques? Following the preferred reporting items for systematic reviews and meta-analyses (PRISMA) statement, we conduct a systematic review of continuous camera-based vital sign monitoring using Scopus, PubMed, and the Association for Computing Machinery (ACM) databases. We consider articles that were published between January 2018 and April 2021 in the English language. We include five vital signs: heart rate (HR), respiratory rate (RR), blood pressure (BP), body skin temperature (BST), and oxygen saturation (SpO2). In total, we retrieve 905 articles and screened them regarding title, abstract, and full text. One hundred and four articles remained: 60, 20, 6, 2, and 1 of the articles focus on HR, RR, BP, BST, and SpO2, respectively, and 15 on multiple vital signs. HR and RR can be measured using red, green, and blue (RGB) and near-infrared (NIR) as well as far-infrared (FIR) cameras. So far, BP and SpO2 are monitored with RGB cameras only, whereas BST is derived from FIR cameras only. Under ideal conditions, the root mean squared error is around 2.60 bpm, 2.22 cpm, 6.91 mm Hg, 4.88 mm Hg, and 0.86 °C for HR, RR, systolic BP, diastolic BP, and BST, respectively. The estimated error for SpO2 is less than 1%, but it increases with movements of the subject and the camera-subject distance. Camera-based remote monitoring mainly explores intensive care, post-anaesthesia care, and sleep monitoring, but also explores special diseases such as heart failure. The monitored targets are newborn and pediatric patients, geriatric patients, athletes (e.g., exercising, cycling), and vehicle drivers. Camera-based techniques monitor HR, RR, and BST in static conditions within acceptable ranges for certain applications. The research gaps are large and heterogeneous populations, real-time scenarios, moving subjects, and accuracy of BP and SpO2 monitoring.

1. Introduction

Physiological processes of the human body are associated with electrical, chemical, kinematic, and acoustic changes [1]. Although their measurement may not provide direct physiological evaluation or diagnosis, they yield valuable information regarding human activities and physiology [2]. This information is particularly useful in monitoring the cardiovascular system [3]. Several authors such as Lu et al., have comprehensively investigated electrophysiological signals such as electrocardiography (ECG) or optical signals such as photoplethysmography (PPG) to record vital signs [4]. Commonly, the four primary vital signs are heart rate (HR), respiratory rate (RR), blood pressure (BP), and body temperature (BT), and oxygen saturation (SpO2) is usually referred to as the fifth vital sign [5,6].
Vital signs offer information on the cardiovascular and respiratory systems [7]. The HR measures the number of heart contractions per minute and is usually given in beats per minute (bpm). The respiratory cycle includes inhalation and exhalation, and the RR is typically measured in cycles per minute (cpm) [8]. The maximum pressure exerted in the arteries during cardiac contraction is known as systolic BP, whereas the minimum pressure during cardiac relaxation is described as diastolic BP [9]. BT is a measurement of the body’s capacity to produce and release heat and is an early indicator of illness [10]. SpO2 measures the relative concentration of oxygenated hemoglobin with respect to the total amount of hemoglobin, and it is an essential physiological parameter to assess the oxygen supply to the human body [11,12]. These vital signs are monitored routinely in clinical practice using contact-based technology such as ECG recorders, respiratory belts, finger pulse oximeters, sphygmomanometers, and thermometers [5,13,14]. However, contact-based measures are obtrusive and may result in skin irritations. For instance, wet ECG electrodes need replacement after 48 h. Therefore, noncontact and unobtrusive vital sign monitoring is an emerging field of research [14]. For example, capacitive coupled ECG can be used to measure the electrical activity of the heart in an indirect way. It is recorded without physical contact to the skin through clothing and can be integrated into chairs, beds, or driver seats [15]. However, capacitive ECG is strongly affected by movement artifacts, coupling impedance fluctuation, and electromagnetic interference [16]. On the other hand, camera-based techniques are the focus of current research [13,17].
Image recording systems are inexpensive, broadly available, and easily combined with the increasing processing power of microcomputer systems. Utilizing a camera, Wu et al., firstly obtained dermal perfusion changes from the skin surface [18]. The underlying idea of image recording is similar to that of detecting the amplitude of finger pulse. Due to the contraction and relaxation phases of the human heart, the blood travels throughout the vascular system. It can be detected in every part of the body to different degrees. These pulsatile blood flow causes subtle changes in the skin color that is detected by image recording devices [18,19,20]. It is also referred as the color-based approach. Another approach is to measure the cyclic motions of the body caused by cardiorespiratory activity. It is known as the motion-based approach or ballistocardiography (BCG) [21,22]. The use of image or video data might offer noninvasive and unobtrusive measurement of vital signs [7,13].
Camera-based monitoring is applicable in clinical and nonclinical settings. In clinical settings, target groups include geriatric patients [23], newborns [24,25,26,27,28], patients in intensive care units (ICU) [29], patients undergoing magnetic resonance imaging [30], or patients before, during, and after surgery [31]. Noncontact measurement reduces the risk of infection and increases the patient comfort [23] and does not disturb the patient, which is important not only for detecting sleep disorders (e.g., apnea) but also for hypertension, heart failure, and arrhythmia [32,33,34]. Nonclinical applications are in-home [33] or in-vehicle monitoring (e.g., driver) [35,36], activity monitoring (e.g., fitness) [37], biometric authentication [38], and stress detection [39,40]. Such applications are integrated conveniently, affordably, and safely into daily life. In addition to physiological monitoring, psychological conditions can also be evaluated, for example in the car, to prevent accidents [35,36,41]. In addition, camera-based vital sign monitoring with mobile phones supports telehealth in rural regions, where clinics and specialized medical equipment are scarce [42,43].
The majority of work focus imaging PPG (iPPG, PPGi, PPGI), which is also named camera-based PPG, remote PPG (rPPG), distance PPG, noncontact PPG, or video PPG [15,19,23,44,45,46,47]. Classically, rPPG also stands for reflective PPG [15,48]. In the following, we therefore use iPPG for all image-based PPG measures. The HR and the HR variability (HRV) as well as the pulse rate (PR) and its variability (PRV) derive from ECG-based as well as PPG-based measures, respectively. Several authors such as Bánhalmi et al., confirmed the good correlations between HRV and PRV [49]. Therefore, we use the term HR and HRV to address both of the techniques.
iPPG measures subtle skin color variations or cyclic movements of the body to estimate HR, RR, BP, BST, or SpO2. However, the field of camera-based vital sign measurement comprises not only red–green–blue (RGB, wavelength: 350–740 nm) receptors [7,50] but also infrared cameras, usually in near-infrared (NIR, 740–1000 nm) [51,52,53], and far-infrared (FIR, >1 μm) [15,25,29,34]. They measure the absorbance, reflectance, or transmission of light [43]. For instance, RGB cameras capture the reemitted light after passing through various skin tissues, while hemoglobin and melanin change the optical skin properties. The captured light may reflect the cardiac-related parameters that can be derived from variations of the intensity pixels in images. In addition, the respiratory-based component is based on the detection of periodic chest movements caused by the lungs volume changes while breathing. This method is reliable, safe, and affordable as it does not require special equipment, and it is used for long-term monitoring [21,54]. However, it is inapplicable in dark or low-light conditions [55].
FIR is also known as thermal imaging or infrared thermography [15,56,57]. It detects the temperature fluctuations around the nostril area as the temperature during inhalation and exhalation is similar to that of the external environment and the human body, respectively [15,32]. The cyclical ejection of blood from the heart to the head, through the carotid arteries and thoracic aorta, which results in periodic vertical motions of the head, may be used to monitor HR through utilization of FIR [57]. In addition, body skin temperature (BST) is a noncontact and noninvasive method of measuring the temperature continuously by the detection of infrared radiation emitted from the body. It can be monitored using FIR. FIR measures this infrared radiation, which results in estimating BST [58]. Therefore, it is suitable during the night in applications such as sleep monitoring, neonates, or ICU monitoring [25,59]. In addition, it has an advantage under varying illumination conditions and can help with privacy concerns [32]. However, most of the FIR cameras are high-end products and expensive [27], whereas consumer-accessible FIR cameras are challenged by low pixel resolution and sampling rates [60]. The NIR camera overcomes such issues.
NIR cameras work similar to RGB cameras, but they can also operate at night and in dark conditions if infrared (IR) lamps illuminate the scene [53,61]. Operation without visible illumination is advantageous for applications such as driver monitoring [36] and sleep studies [52,53,62]. The drawback of NIR cameras is the lower absorption by hemoglobin and a lower signal-to-noise ratio (SNR) [36,41]. However, camera-based vital sign monitoring faces many challenges associated with the hardware, motion artifacts, variability of natural illumination, skin variance, and the camera-subject distance. These factors negatively affect the accuracy of vital sign measurements [59].
In recent years, several iPPG review articles focus on HR measurement [50,63,64,65]. Zaunseder et al., focused on HR and HRV, as well as other derivable parameters including pulse transit time (PTT) and pulse wave velocity to remotely examine the peripheral vascular system [66]. Addison et al., explored RR measurements using depth cameras [67]. Maurya et al., analyzed noncontact RR monitoring of newborn using RGB-, IR-, and radar-based modalities [28]. Rehouma et al., comprehensively surveyed RR measurements using noncontact sensors, including cameras [68]. Khanam et al., reviewed motion- and color-based methods and focus on motion artefacts, illumination and distance variations, different sensors and acquisition settings, and demographic parameters [69]. Recently, Steinman et al., highlighted BP measurements utilizing smartphone technologies and video cameras [70]. Haford et al., presented an interesting review of demographic parameters, camera parameters, study design, and obtained accuracy on the reported vital signs such as HR, RR, SpO2, and BP [71]. Antink et al., provided an overview of camera-based measurements of vital signs using RGB and IR cameras [13]. However, existing reviews disregard how the vital signs are extracted from the camera frames. Most review articles focused on either HR or RR measurement using RGB cameras. Only a few considered BP, BST, and SpO2.
To fill this gap, we present a complete review of recent advances in camera-based techniques for all five vital signs. We further aim to identify challenges of existing works, get insight in how to address these challenges, and identify future research prospects. In particular, we want to answer the following research questions:
  • Which vital sign is monitored using what type of camera?
  • What is the performance and which factors affect it?
  • Which health issues are addressed by camera-based techniques?

2. Methods

We performed the literature review according to the preferred reporting items for systematic reviews and meta-analyses (PRISMA) workflow [72].

2.1. Selecting Databases

In previous work, we have identified PubMed, Scopus, and the Association for Computing Machinery (ACM) digital libraries as most relevant [73]. We searched these databases for articles published in the English language between January 2018 and April 2021.

2.2. Composing the Search Query

Our main purpose is to retrieve the relevant literature on measuring the vital signs continuously, without contact, and remotely using different camera modalities. We developed the search string using relevant terms (Appendix A.1, Appendix A.2 and Appendix A.3 for PubMed, Scopus, and ACM, respectively). Relevant terms address the
  • Sensor technology: camera, video, image, RGB, infrared, thermal, thermography;
  • Sensor setting: remote, noncontact, contactless, contact-free, touchless, vision-based;
  • Biosignal: biosignal, biomedical signal, physiological signal, cardiorespiratory signal, heart rate, heartbeat, pulse rate, respiratory rate, breathing rate, blood pressure, body temperature, oxygen saturation; and the
  • Modality: photoplethysmogra* (to include photoplethysmogram, photoplethysmograph, and photoplethysmography).

2.3. Inclusion and Exclusion Criteria

We include and exclude articles according to:
  • Inclusion criteria
    a.
    English language
    b.
    Human research
    c.
    Journal, Conference
  • Exclusion criteria
    a.
    Review, systematic review, survey
From the retrieved articles, we first eliminated duplicates. The selection process consists of three stages: title, abstract, and full-text screening. In title screening stage, we removed work focusing on aspects not associated with vital sign measurement (e.g., face detection without vital signs monitoring and facial expression analysis) and work using other sensors (e.g., spectroscopy, radars). Subsequently, we performed the same selection process to the abstracts and to the full text. After literature analysis, we identified population parameters (e.g., age, number of subjects, subjects’ skin tone), camera parameters (e.g., camera, frame rate, resolution), applied algorithms, performance metrics, and predicted vital signs.

3. Results

Figure 1 depicts the PRISMA workflow and the results of the literature screening. A total of 1031 articles were found; from those, 124 duplicates were removed. The titles and abstract of the remaining 905 articles were screened and 781 were considered irrelevant. Subsequently, the full texts were examined with 19 being found to be irrelevant for this review and one full-text was inaccessible. To that end, we include 104 articles, all of which met the following eligibility criteria:
  • Optical image (RGB, NIR, or FIR)
  • Extraction of vital sign (HR, RR, BP, BST, or SpO2)
  • Comparison to ground truth
Each year delivered 20 to 35 records. Researchers mainly focused on HR and RR, followed by BP, BST, and SpO2 (Table 1). The following subsections presents the results of the literature review. They are organized in the following order: data acquisition, image processing, signal processing, and performance metrics (Figure 2). At first, video or image data from smartphones, webcams, or digital cameras is acquired with synchronously obtained ground truth (Section 3.1). Existing datasets are also used (Section 3.2). As shown in Section 3.3, image processing techniques are then employed to detect the region of interest (ROI) and the raw signal is extracted from the ROI. In Section 3.4, we discuss signal processing techniques, mainly deep learning (DL) approaches, to remove motion and illumination artifacts. After optional data fusion (Section 3.5), the vital sign is extracted from the signal (Section 3.6) and compared to the ground truth using performance metrics (Section 3.7).

3.1. Data Acquisition

The data used in the analyzed articles was acquired in indoor (e.g., laboratory, hospital, home), or outdoor environments (e.g., in-car). Twenty articles used videos from existing datasets [52,59,74,75,76,77,78,79,80,81,82,83,84,85,86,87]. In the following, we describe the main findings regarding study design, hardware setup, and ground truth (Figure 3) that is presented in Table S1.

3.1.1. Study Design

The study design covers aspects from study population, namely number of subjects, age, gender, skin color, and the various types of settings. Researchers mainly recruited volunteers. Frequent populations were healthy [12,37,39,46,51,60,88,89,90,91,92,93,94,95,96,97], ICU patients [29,98,99,100], neonates or infants [24,25,60,98,99,101,102]. The majority of the articles obtained the data on their own (Figure 4). Number of subjects was mostly between 10 and 30.
Regarding the age, study subjects were between 18 and 85 years old [53,60,90,103,104,105,106,107], while newborns aged from a few days up to 40 weeks [24,25,26,98,99,102,108,109] with different gender. Thirty-two articles reported their skin color. The Fitzpatrick scale (I to VI) was most commonly used to determine the skin color of subjects (8 articles) [43,60,92,101,110,111]. Visually resemble colors like white, pale, yellow, brown, or black skin reported in six papers [20,87,95,112,113]. In 18 papers, the subjects’ country of origin or geographical locations were indirectly used as skin color indicators [24,39,51,59,75,85,90,98,99,104,109,114].

3.1.2. Hardware Setup

In the reviewed articles, RGB, NIR, FIR, or a combination of RGB with NIR or FIR camera were used (Figure 5). RGB cameras such as smartphone cameras [43,49,104,115], webcams [20,95,116,117], or digital cameras [56,78,102,108] were utilized often, due to their low costs and easy availability. Further, most researchers use single RGB camera, followed by combination of cameras.
Several parameters affect the image quality of cameras, such as spectrum range, sensor hardware, frame rate, frame resolution, or camera–subject distance. Frame rate ranges from 1 frame per second (fps) to 500 fps. For fast moving subjects, frame rates of 100 fps to 500 fps were employed [45,49,118,119], however 10–30 fps was used often. Rapczynski et al., reported that higher frame rates do not contribute more information but increase demands for storage space, transmission, and computation instead [59,80]. Applied resolutions vary greatly, ranging from 2080 px, down to 47 px. Although different ranges of resolutions are observed in articles, a resolution of 320 pixels (px) × 240 px is sufficient to recover a signal with an adequate SNR [120]. However, the 480 px to 720 px range widely used.
In addition, environmental (e.g., illumination, temperature) and behavioral (e.g., body movement) parameters impact the image quality [117]. For charge coupled devices (CCD) cameras, illumination is one of the key parameters that is provided by sunlight through windows [59,95], ambient illumination from ceiling fluorescent lights [46,87,101,102,107,110,121], phone flash lights [12,43,49], uniform white light [37,122,123] or combination of ambient light and sunlight through windows [81,86,109].
Figure 6 shows the conceptual semantic relationship between camera images and vital signs. The majority of articles was estimated the HR by utilizing RGB images. Similar to HR, RR measurements were calculated using both RGB, NIR and FIR images. Measurements of BP and SpO2 were extracted from RGB images. FIR images were used to calculate BST only. None of the research used RGB images to estimate BST based on the author’s knowledge.

3.1.3. Ground Truth

In addition to the camera-based modalities, all research employed typically assessable or commonly available measurement devices to record the vital sign’s ground truth (Figure 7). Fifty-five articles used manual palpation (e.g., radial artery), ECG, or pulse oximeters to record the HR [39,49,86,95,102,107,110,112]. Sixteen articles utilized manually counting (e.g., visual inspection), respiratory belt, pressure sensors, or polysomnography to calculate the RR [32,51,60,89,94,105,106,124,125]. Six articles measured BP by utilizing manual sphygmomanometers or a blood pressure cuff [37,113,119,123,126]. Two and one articles used clinical thermometers and pulse oximeter that delivers the reference for BT and SpO2, respectively [61,127]. Nine articles utilized patient monitor, or wrist-based wearable devices to track vital signs (e.g., Philips patient monitor) [20,25,93,99,102,108,128,129,130] and fifteen articles delivered multiple vital signs measurements [43,51,53,57,94,106,122,131,132].

3.2. Existing Datasets

Publicly available datasets (Table 2) have been recorded while the subjects are watching video stimuli [133], watching music videos [134], playing a stressful game [135], moving their face (e.g., speaking, slow and fast rotation) [136], using various illuminations [137,138] and during driving [41]. Different cameras captured the ROI in the VIPL_HR [137], TokyoTech Remote PPG [118] and MERL-Rice near infrared pulse dataset (MR-NIRP) [41]. All experiments were conducted in an indoor setting except MR-NIRP. The frame rate varies between 20 fps and 60 fps. However, the majority of the databases captures video with RGB cameras at a frame rate of 30 fps. Along with camera-based modalities, typical clinical parameters such as HR, RR, BT, and SpO2 were obtained. All databases are available upon the approval of the end user license agreement except publicly accessible MR-NIRP database.

3.3. Image Processing

Once the video data is available, it is fed into a typical pipeline for image-based vital sign measurements (Figure 2). Usually, image processing contains ROI detection, tracking, enhancement, color decomposition, and raw signal extraction.

3.3.1. ROI Detection

Depending on the parameter of interest, the ROIs differ. For instance, the ROI for HR measurements is the location of the facial artery (e.g., cheeks) [46,59,79,151], while RR often is extracted from the nasal or torso region [29,31,91,128,152]. The orbital region and the full frame are the preferred ROIs to calculate the BST [61,100,127,132]. Forehead, forearm, palm, and cheeks are included to measure BP [37,113,119,126].
Figure 8 depicts the workflow of the HR vital signs measurement from different camera images. These articles considered the face [83,85,87], forehead [20,43,46,102], cheeks [75,79,132,153], nose [40,94,109], chin [40], and palm [37]. Furthermore, four articles partitioned the entire frame or face into sub-ROIs and select the most reliable sub-ROI [27,77,91,119]. Having more capillaries and being unaffected by facial expression, the forehead and the cheeks are most commonly extracted as ROI [38,95,110,130,151,154].
Different algorithms identify the ROI either manually [29,57,101,102,105,111,125,128,152] or automatically, where the Viola–Jones algorithm is used most frequently (Table 3). In addition to the listed articles, four articles employ the discriminative response map fitting algorithm to locate the face landmark points [77,87,107,153]. Addison et al., detected a seed point on the forehead and the skin around it and performed a flood fill for each frame that recursively aggregates all adjacent skin pixels within the preset tolerances [110]. Recently, neural-network-based ROI detection has gained increased attention and yielded higher accuracy compared to conventional approaches [78,155].

3.3.2. ROI Tracking

ROI extraction can be done by detecting the ROI in all frames or by identifying ROI in the first frame and tracking over all frames [20]. Sixteen research teams used the Kanade–Lucas–Tomasi algorithm [29,31,36,59,60,76,85,89,107,122]. It is also effectively tracks the ROI in NIR images [156]. Alternate tracking using kernel correlation filters show false detections in out of sight targets [75,109,157,158]. Liu et al., employed the compressive tracking algorithm to detect multiple subjects [114]. Hochhausen et al., and Pavlidis et al., utilized the particle filter-based object tracking algorithm in thermal images, but this method requires a multicore computer system [132,152,159].
Table 3. Automatic detection of ROI by utilizing various classical and DL methods.
Table 3. Automatic detection of ROI by utilizing various classical and DL methods.
AlgorithmsDescriptionAdvantageDisadvantage
Viola–Jones
[18,75,81,83,87,93,106,115,129,153]
It utilizes Haar-like features and Adaboost algorithm to construct a cascade classifier. It works well on full, frontal, and well-lit facts. It suffers from faces in a crowd, face rotation, inclined or angled faces, expression variations, and low image resolution.
Histogram of oriented Gradients
[86,151,160]
It constructs the feature by calculating the gradient direction histogram on the local area of the image. Fast running speed and identifies 68 facial landmark points.It may be influenced by light intensity and detection and inaccurate location of feature points on profile.
Multitask cascaded convolutional neural network [154]It is a convolutional-neural-network (CNN)-based framework, which consists of three stages for joint face detection and alignment. Accurate face detection, less affected by light intensity and direction.It may provide sophisticated models and calculation, which may result in a slow running speed; only five feature points can be tracked.
Single shot multibox detector
[78]
It is a fast convolutional neural network to detect faces using a single neural network.Fast processing speed and multiscale feature map is adopted.The robustness of the network to small object detection may not too high.
You look only once
[100,161]
It is one stage detector based on object detection. Fastest object detection algorithm. It utilizes full image as context information which is possible to achieve real-time requirements.It requires a graphics-processing-based computational machine. It may be relatively sensitive to the scale of the object.
Template matching
[51,61,106]
It matches the image by providing a base template which to compare.Relatively easier to implement and use.Not suitable for complex templates, no face in the frame, or occlusion of face.

3.3.3. Image Enhancement

Enhancement of low-resolution images improves the resulting vital signs. For instance, subtle color changes in the ROI are difficult to measure reliably. The temporal magnification approach relies basically on two main types: Lagrange and Euler [98]. The Lagrange technique tracks the mobility of each pixel, whereas the Eulerian approach splits the image into grid elements and tracks the movement of pixels [98,162]. So far, eight research teams utilized Eulerian video magnification to magnify the subtle movements [24,125]. Moya-Albor et al., applied the Hermite transform-based Eulerian video magnification technique with improved accuracy [97]. Furthermore, Kwasniewska et al., and Yu et al., introduced the super resolution-based deep neural network to improve the performance of vital sign estimation of hallucinated thermal image [125] and the spatiotemporal video enhancement network, respectively [163].

3.3.4. Color Channel Decomposition

Depending on the color space, channel decomposition separates colors or intensity and illumination. Most commonly, researchers use RGB; hue saturation value (HSV); and luminance, chroma blue, chroma red (YCbCr). Sixty-five articles explored RGB channels [37,95,104,110]. Green channel contains more information related to cardiac pulse [38,104,130]. Pai et al., mentioned the intensity of red and blue channels to eliminate illumination variations [164]. Only two articles regarded the red channel [115,151].
The HSV color space has been explored by three research teams [43,114,130]. HSV does not require color normalization and is robust to illumination variations [154]. Sanyal & Nundy explored the hue channel to measure the fluctuations associated with skin color changes and used for iPPG signal extraction [43,114]. Bánhalmi et al., examined luma component from YCbCr [49]. Ryu et al., combined chroma blue and chroma red channels providing a high SNR [142]. YCbCr and HSV-based skin-representing pixels have been simply compared with a threshold [81,98,116,154]. In addition, Li et al., Rapczynsk et al., and Barbieri et al., employed the Conaire’s method [85], skin probability-based approach [80], and a decision tree classifier [92] to segment the skin pixels, respectively.

3.3.5. Raw Signal Extraction

Averaging the ROI in each frame extract the one-dimensional iPPG signal that results in reduction of quantization noise [54,142,164]. The majority of research (94 articles) articles spatially averaged to obtain one-dimensional signal [29,46,59,60,79,107,109,115,152]. On the other hand, several studies such as Zhu et al., utilized feature points to track the cyclic motions of the ROI [165].

3.4. Signal Processing

Signal processing usually conducts the preprocessing of the raw signal and vital sign extraction.

3.4.1. Preprocessing

Obtained from spatial averaging of ROI pixels, the vital signs’ raw signals comprise noise from subject movement, facial expression, illumination variation, and dynamic environments. Hence, before extracting vital signs, the raw signals need appropriate pre-processing (Table 4).
The critical parameter of a band-pass is its bandwidth. According to the Association for the Advancement of Medical Instrumentation (AAMI), the bandwidth should be in between 30 bpm to 200 bpm for HR studies in human adults [168]. However, HR varies depending on a variety of factors including age, exercise, medications, and medical conditions [6]. Therefore, different filter bandwidth settings have been applied such as, 0.3–6 Hz [92], 0.6–2.8 Hz [93], 0.6–3 Hz [114], 0.75–3 Hz [59], 0.5–4 Hz [110], 0.7–4 Hz [79,130,160], 0.65–4 Hz [116], 0.83–3.17 Hz [122], and 0.8–2.2 Hz [43]. Since infants under the age of one year have a higher HR than adults, ranging from 110 to 160 bpm [98,169], research on infants reported frequencies of 1.83–2.67 Hz [108], 1.3–5 Hz [109], and 1.67–3.33 Hz [101].
The RR of adult lies between 12 cpm and 20 cpm [170,171]. The RR varies with factors such as age, exercise, and respiratory problems. The researchers have considered the frequency ranges of 0.05–1.5 Hz [32], 0.18–0.5 Hz [43], 0.05–2 Hz [105], and 0.1–5 Hz [89]. Similar to HR, infants have a higher RR than adults, ranging from 40–60 cpm [27]. The bandwidths of RR were in the range of 0.1–3 Hz [27], 0.5–1 Hz [108], and 0.2–2 Hz [102]. Band-pass filters in BP studies have the bandwidths of 0.7–2 Hz [37] and 0–6 Hz [172]. With respect to SpO2, researchers have used a bandwidth of 1.67–3.33 Hz [101] and 0.5–5 Hz [12].
Although motion artifacts are a key problem while sleeping as well as in infant studies, only a few articles address this problem. Benedetto et al., removed motion artifacts manually [39]. Eight studies identified body motion threshold-based [32,37,90,95,115,153,173]. Lorato et al., applied a gross motion detector to remove torso and chest motions [26]. In neonates, Rossol et al., used a motion history image algorithm [24].

3.4.2. Vital Sign Extraction

According to the various vital signs, researchers conduct several techniques to extract the vital signs from the raw signals.

Heart Rate

The video-extracted signal is composed of a component representing the blood volume pulse as well as a noise component. The dimensionality reduction is commonly categorized as blind source separation (BSS), such as independent component analysis (ICA) and principal component analysis (PCA), model-based methods, and DL, that is employed to extract pulse related information [168] (Table 5). BSS extracts the source signals without previous knowledge. Further, ICA may require more computation to separate the components [174]. To overcome this problem, the variants of ICA, namely FastICA [154], viterbi decoding-based ICA [83], project_ICA [174], and joint BSS along with ensemble empirical mode decomposition [157] are used by Zheng et al., Raseena & Ghosh, Qi et al., and Lee et al., respectively. On the other hand, Lewandowska et al., employed a PCA in 2012 [175]. Barbieri et al., reported result on the zero-phase component analysis, a variant of PCA, to find the peaks with more precision [92,162]. However, according to Huang et al., BSS-based methods leave the skin color information unutilized [168]. Chrominance (CHROM)-based [162] and plane-orthogonal-to-skin (POS) are traditional model-based methods. It provides more stable under complicated and nonstationary circumstances [168].
Recently, DL techniques gained a great amount of interest which are categorized as non-end-to-end and end-to-end. Non-end-to-end techniques apply DL either in image or in signal processing [181]. Hsu et al., employed CNN to extract HR by applying two-dimensional time–frequency representation [78,145]. Qiu et al., utilized spatial decomposition and temporal filtering to extract feature images and cascaded them with a CNN [75]. Moya-Albor et al., trained CNN by feeding magnifications from a Hermite transform [97]. On the other hand, end-to-end learn features on their own to generate the vital sign that is employed by a few works and described as follows: Yu et al., used spatiotemporal video enhancement networks and the rPPGNet to implement end-to-end HR extraction without utilizing classical algorithms [163]. Huang et al., demonstrated two dimensional convolutional along with long short-term memory to extract HR [182]. Further, Perepelkina et al., utilized HeartTrack that is constructed by a CNN [121].
In addition to HR, other cardiac-based metrics, such as HRV and interbeat interval (IBI), are used to assess physical and mental health in daily activities [40,164]. The sympathetic nervous system, which regulates blood vessel diameter, is inversely correlated to HRV [172]. HRV calculates the variations between the heart beats, whereas IBI measures the time interval between the two consecutive heart beats [56,104]. According to the joint Task Force of the European Society of Cardiology and the North American Society of Pacing and Electrophysiology, the HRV may be evaluated either in the time or in the frequency domain [104,183]. Typical parameters in the time domain are standard deviation of interbeat intervals, root mean square of successive differences in interbeat intervals, proportion of successive normal-to-normal that differ by more than 50 ms, and mean of PP interval [40,93,104,118]. Frequency domain features are very low frequency components, low frequency components, high frequency components, and the ratio of low frequency to high frequency [92,104]. These features are analyzed statistically. Typically, researchers perform filtering or decomposition of the signal before extracting these features in the time or frequency domains [85,92,95,117,118].

Respiratory Rate

RGB cameras detect the RR from breathing motion, whereas IR cameras identify both temperature fluctuations caused by breaths and motion [57,131]. FIR camera resolution varies strongly (Table 6). In RGB images, dimensionality reduction methods, namely PCA, zero-phase component analysis and POS, were applied by Zhu et al. [53], Iozza et al. [89], and Chen et al. [56], respectively. Recently, Schrumpf et al., utilized empirical mode decomposition to extract the respiratory related component from intensity-based signal. It decomposes the signal into numerous intrinsic mode components and identifies the relevant component using power spectral density [91].
RR was monitored using thermal camera in fifteen articles. In this, eleven articles showed that inhaling and exhaling at regular intervals causes a proportional decrease and increase of the temperature around the nasal cavities, respectively, which can be converted into RR [29,31,60,94,125,128,152]. Three articles investigated periodic motions of chest in the full frame [25,26,27].

Blood Pressure

A common approach of calculating BP is to extract waveform-related features such as temporal, areas, amplitude, and derivative features by utilizing systolic and diastolic peaks [46,88]. Sugita et al., extracted the fluctuations of BP waveform from palm ROI [37,186]. However, such indirect methods cannot observe the BP absolutely. The PTT is a valuable metric for tracking variations in BP throughout regular activities [186]. It is possible to obtain by calculating the time taken for an arterial pulse to travel along an arterial segment by monitoring at two different sites on a blood vessel [111,119]. It can be seen from the difference between vessels near and far from the heart. Shirbani et al., measured the PTT between the forehead and left-hand ROI of the iPPG signal [111]. The time difference from two ROIs codes the blood pressure [186]. On the other hand, studies employed the signal processing techniques to extract hemoglobin related signal features that are used to estimate BP. Zhou et al., applied ICA and peak detection to identify the peak and valley of the wave in order to estimate the systolic BP and diastolic BP respectively [113].
Stimulus (e.g., food, exercise activities) has an impact on the sympathetic nervous system and elevates the BP [122,126]. Zhou et al., conducted the test during rest condition [113]. Sugita et al., and Takahashi et al., conducted the experiment during exercise activities (e.g., squat exercise) [37,119]. Vasoconstriction is the narrowing (constriction) of blood vessels by small muscles in their walls, which was stimulated by Oiwa et al., via cold temperature [126]. These experiments conducted in laboratory settings, used facial or hand/palm ROIs. We have not identified any study in a real-world setting.

Body Skin Temperature

Six articles acquired thermal videos to measure BST by obtaining the highest temperature of forehead [61,132], or the entire frame [94,100,127,131]. Dagdanpurev et al., employed the K-nearest neighbor algorithm to differ healthy and infected patients [127]. We did not find any method to derive BST from RGB.

Oxygen Saturation

Forehead and hand backside are common ROIs to measure SpO2 from RGB images. Classically, the SpO2 is calculated from the ratio of absorbance at red and blue channels (ratio-to-ratio). It is based on the effective blood volume and the hemoglobin concentration [12]. However effective blood volume can be influenced by peripheral vasoconstriction, local temperature, or cardiac index. To overcome this problem, Sun et al., explored multiple linear regression models [12]. Wieler et al., extracted the oxygen desaturation from the facial region of the forehead in neonates [101]. From this region, red and blue channels obtained to calculate the peak-to-trough ratio of two wavelengths. It corresponds to the maximum and minimum absorption of oxygen in blood and is an indirect measure of peripheral SpO2 [101].

3.5. Data Fusion

To overcome limitations by a singular data source, some articles fuse data from different sources. We differ image-based and signal-based fusion that uses several video sources and combines the data already on a frame-based level and after the raw signals have been extracted separately from the different videos, respectively. For example, Lorato et al., merged multiple thermal camera images into a single image before computing RR and automatically detect the core pixel that corresponds to the respiration [25]. Villarroel et al., combined patient detection and skin segmentation from an optical flow network [99]. Kurihara et al., combined RGB and NIR face patches to estimate the HR, which is helpful in dark environments and eliminate background illumination [187]. McDuff et al., delivered an intriguing investigation that involved the use of nine distinct cameras with diverse angles for HR estimation [120].
On the other hand, Kado et al., used histogram fusion to select suitable face patch pairs in HR measurements, based on thresholding [55]. Pereira et al., explored three different sensor fusion strategies: median frequency of all valid ROIs, highest signal quality index, and Bayesian fusion [27]. Scebba et al., introduced a signal quality-based fusion approach to calculate the RR by determining the signal quality of various respiratory signals such as thermal airflow and respiratory induced motion. They also used S2 fusion, a combination of signal quality-based fusion with apnea detection to obtain the RR [60]. Fujita et al., implemented a CNN to select the spectrum from a set of multiple obtained ROI from a single frame [159].
In summary, fusion-based techniques can significantly improve the robustness of vital sign monitoring, despite the motion and illumination problems in real-time situations [15]. For instance, sensor fusion techniques that employ multiple cameras positioned at different angles can reliably track the ROI of the subject and compensate motion artifacts (e.g., head turn). A combination of RGB and NIR cameras for day and night environments is helpful in low light applications [187]. Furthermore, signal fusion might allow single camera approaches to assess multiple ROIs simultaneously.

3.6. Vital Sign Estimation

The conversion of the signal into a vital sign can be done in the time (34 studies) or in the frequency domain (59 studies). In the time domain, peak detection has diffusely been applied to obtain the IBI from which vital signs are computed. Peak detection may reduce the error by averaging all the IBI over the video duration [32,40,77,105,115,116,118,151]. In the frequency domain, a global or windowed (short-time) Fourier transform may identify the peaks with maximum power, which then are simply turned into a rate of a vital sign [59,60,78,82,94,117,122,130].

3.7. Performance Assessment

3.7.1. Performance Metrics

The mean error and its standard deviation, MAE, mean relative error, mean of error rate percentage, limits of agreement, RMSE, CC, and correlation of determination assess the performance of the various algorithms.

Heart Rate

Figure 9 depicts the various conditions investigated in HR measurements. Studies conducted in static conditions resulted in low RMSE [49,95,107,111,114,130]. Dynamic conditions such as subject movement, speaking, facial expressions, and exercise activities provided less reliable results [82,87,107,130,156] (Figure 9a). HR measurements perform at various distances 0.15–3 m throughout the studies (Figure 9b). Less than 1 m yielded satisfactory results but the RMSE increased with larger distances between the subject and the camera [59,87]. Outliers are due to movements [94]. Five research teams particularly investigated on distance variance [59,83,87,156,188]. For instance, Zhang et al., studied 40–80 cm and reported that 60 cm provides the optimal performance [156].
Similarly, varying the intensity of the illumination causes a different result. At static conditions, light intensities of 150 lux and 300 lux yielded RMSEs of 5.56 bpm and 5.58 bpm, respectively. Dynamic conditions, such as rotation, mixed action, four feet walk, reached RMSEs of 8.09 bpm, 9.96 bpm, 8.60 bpm at 150 lux and 7.36 bpm, 9.88 bpm, and 7.72 bpm at 300 lux respectively [130]. Hassan et al., varied intensity levels such as 340–380 lux, 430–470 lux, and 510–550 lux to check the impact on HR estimation, which resulted in 20.27 bpm, 16.86 bpm, and 13.98 bpm, respectively [87]. Hence, it shows that increasing the illumination reduces the error in RGB. Low-light (1 lux) and bright illumination (184 lux) conditions by utilizing NIR light resulted in RMSE of 1.06 bpm and 0.86 bpm [51]. Zhang et al., varied the intensity levels for the driving condition between 0 lux and 80 lux and demonstrated with webcam and NIR video. The performance deteriorated if the intensity level is less than 20 lux in webcam video, whereas NIR video provides better result even at 2 lux. Zheng et al., also reported the optimal illumination at 40 lux for normal driving situations [156]. However, drastic illumination changes might occur during driving due to the streetlamps and headlights [36]. In summary, HR measurement using RGB cameras suffer from darkness, whereas NIR cameras yield better results.
Variation of skin tone (e.g., fair, brown, black) also influences the performance. Under static condition, the RMSE for fair skin was 0.07 bpm [95]. On the contrary, the RMSE for fair skin with limited subject movement condition is 17.75 bpm [87]. Song et al., investigate Asian subjects at 1 m distance and 2704 px × 1520 px, reaching 1.50 bpm in static condition [59]. The Chinese subject group attains an RMSE of 2.77 bpm with minor mobility [114]. The RMSE of 17.10 bpm reaches with brown skin and minimal physical activity [87]. For dark skin, participants in stationary state yielded an RMSE of 16.49 bpm [87]. Addison et al., reports the RMSE of 1.74 bpm, 2.32 bpm, 1.88 bpm, 2.04 bpm, 2.89 bpm, and 2.43 bpm on the Fitzpatrick scales I to VI, respectively [110]. From these studies, we can conclude that increased melanin concentration lowers the performance of HR estimation.

Respiratory Rate

Researchers measured the RR in different conditions: normal and simulated breathing with controlled frequency [27,60,89,91], different clothing [105], and different camera-subject distances [29,52]. Studies conducted in stay-still situations yielded better results (Table 6) [27,29,105,124,128,152]. In RGB images, Chen et al. [56] and Sanyal & Nundy [43] performed a steady state with minimal movement and yielded an RMSE of 2.16 cpm, and 3.88 cpm, respectively. In NIR images of sleeping subjects, Zhu et al. achieved an RMSE of 2.10 cpm [53]. Chen et al., utilized NIR in a still state with varying illumination and reached an RMSE of 0.70 cpm [51]. In a regular setting, Pereira et al. [27], Negishi et al., [106], Kwasniewska et al. [125], and Cosar et al. [94] used thermal images and obtained an RMSE of 0.31 cpm, 1.44 cpm, 3.40 cpm, and 3.81 cpm, respectively. In contrast, movements lower the performance down to 2.52 cpm and 6.20 cpm as shown by Negishi et al. [106] and Cosar et al. [94], respectively. From all these findings, the median RMSE is 2.1 cpm and 3.9 cpm for distances d ≤ 1m and d > 1 m, respectively (Figure 10). Hence, RMSE grows with increasing distance in the same way as HR does [29].
Real time applications for neonatal care involve both HR and RR measurements [24,25,26,27]. HR measurements of inactive infants achieved a RMSE of 1.1 bpm [109]. However, infant studies provided larger limits of agreement due to motion and unconsciously swinging hands in HR and RR measurements [24,98,102,108].

Blood Pressure

Zhou et al., reported RMSE of systolic and diastolic BP without movement in a distance of 60 cm as 6.91 mmHg and 4.88 mmHg, respectively [113]. Additionally, they varied distances (30–90 cm) and illumination (75 to 500 lux). Substantial inaccuracies arise when the distance is large and the ambient light is bright or faint. Zhou et al., reported best results with a light intensity of 150–200 lux and a camera-subject distance of 50–60 cm [113]. Sugita et al., achieved RMSE of 25.7 mm Hg during cycling exercise [37]. Studies with food stimuli (e.g., chocolates) [122] and cold stimuli [88,126] further affected BP.

Body Skin Temperature

Dagdanpurev et al., reported significant results between healthy subjects and infected patients and a correlation of 0.82 during steady state [127]. Cosar et al., obtained an RMSE of 0.86 °C and 0.88 °C in steady state and moving conditions, respectively [94].

Oxygen Saturation

Sun et al., explored the peripheral SpO2 in healthy adults using smartphone cameras and a ROI from the back of hand [12]. They reported estimated errors less than 1%. Wieler et al., investigated the oxygen desaturation on healthy newborns with an RGB-based approach and reported 75% sensitivity and 20% positive predictive value [101].

3.7.2. Factors Affecting the Performance

Facial hair and skin tone variation have a strong effect on vital sign estimation (Figure 11) [128]. In this, darker skin tones have a higher melanin concentration [87,111], which absorbs a large amount of incoming light, causes the signal to attenuate, and weaken the measured iPPG signals [104,164]. In addition, makeup weakens the pulse signal [96,120].
Disregarding the camera type, camera parameters such as resolution and frame rate always have a tradeoff [120]. If the image resolution is too low, quantization noise interferes the iPPG signal, that is difficult to reduce by spatial averaging in video frames. Although resolution is high, pixels within ROI may be insufficient at far recording distance [59]. Hence, increased camera-subject distance reduces the performance (Figure 9b). According to Nyquist theorem, a frame rate of 8 fps is required to acquire the data with a HR of 240 bpm and a RR of 90 cpm. However, studies most commonly used the frame rate of 30 fps for vital sign estimation.
Meanwhile, as the amount of light received by the camera sensor decreases, the signal strength of the iPPG pulse also drops. As a result, the impact of video quality, particularly the seldom used ultrahigh definition must be studied at different camera-subject distances with diverse angles [59]. Motion artifacts are an important factor, too. Several techniques (e.g., ICA, CHROM) significantly reduce such artifacts but it may suffer from head orientation, fast moving, partially covered ROI, and emotions in real-time [41,82,87,168].
Last but not least, environmental parameters such as illumination and outside temperature influence the accuracy of iPPG. Sunlight or artificial light are the most prevalent sources. Sunlight provides almost white light and reliable measurements. In contrast, artificial light results in noisy measures [117,189]. High illuminations yield better performance [87,130]. However, fluctuations in illumination such as obtained in a driving vehicle are most difficult to handle and may also occur in indoor environments [45,117].

3.8. Applications

iPPG is used in a variety of environments including clinical and nonclinical settings. Clinically, iPPG performs general health examinations on a regular or continuous basis [35,99,104]. Additionally, it is applied to patients in critical care units, the post-anesthesia care units, during surgery [29,31,128], and to newborns [25,27,98,99,101,102,108,109] as well as geriatric patients [23].
Nonclinical settings include in-home [33] and in-vehicle health monitoring [35], exercise tracking [190], sleep monitoring [53], and stress monitoring [40]. Home-based vital signs monitoring is extremely helpful in the current pandemic situation due to limited hospital space and facilities [33]. It is also useful for monitoring individuals of all ages, particularly geriatric patients [23] and newborns [102,108,109]. In the automotive area, physiological and psychological condition of the driver are monitored to avoid accidents [35,156,168]. Huang et al., additionally monitor vital signs of vehicle passengers [168].
Although fitness is beneficial to health, too excessive activities might have serious consequences. Hence, it is critical to monitor the physiological status during exercises [49], such as stationary cycling [37]. Sleep monitoring addresses the vital signs as well as its abnormalities, such as sleep apnea and heart disease [32,60,93]. In addition, stress and anxiety monitoring avoids panic attack and accidents during driving [40].

4. Discussion

We perceive that work done in this field within the short period of this review (January 2018–April 2021) is significant and it may open up new research directions. We also believe that noncontact vital sign monitoring has gained increased attention among researchers and physicians during the current pandemic situation. We limited to image-based techniques for the scope of this review and excluded non-image-based noncontact techniques. This section discusses on the review’s limitations and responses to the research questions. Further, we comprehensively explain the research gaps in these fields as well as future research directions.

4.1. Limitations

In this paper, we have reviewed the state-of-the-art in vital sign monitoring using cameras. We searched PubMed, Scopus, and ACM databases for relevant articles that have been published during the last three years. We have screened 1031 articles and included 104. There might be other sources that deliver further relevant articles. Although we carefully selected our search terms (Appendix A), we meanwhile observed that our search cannot be reproduced anymore. The number of returns today differs from that of April 2021, as in particular Scopus has obviously updated its entries including more articles. This may be due to late record submissions to the databases or electronic first publication styles. We might have missed relevant articles due to the selection of our search strings. So, this review may not cover the complete literature.

4.2. Research Questions

We now answer our research questions:
  • Which vital sign is monitored using what type of camera? Using RGB cameras, the iPPG signal is extracted from a ROI in the video frames to monitor HR, SpO2, and BP based on color intensity changes, whereas RR is monitored based on body motion. NIR cameras measure the HR and RR as similar to RGB. In addition, thermal cameras extract HR, and RR based on periodic motions of the torso area, breathe airflow, or vertical movement of the face, whereas BST is obtained from the highest temperature of the ROI (Figure 6).
  • What is the performance and which factor affects it? Static conditions avoid physical movement in well-controlled environments. Here, error rates are less than 5 bpm and 3 cpm for HR and RR, respectively (Figure 9a). The performance suffers considerably from body motion and face expressions and is further lowered if the camera-subject distance is more than 1 m (Figure 9b and Figure 10). In addition, changes in illumination considerably impacts the performance. Low light illumination decreases accuracy. These effects are quantified using several metrics (e.g., MAE, RMSE, CC).
  • Which health issues are addressed by camera-based techniques? The working horses’ applications are ICU patients, neonatal, and geriatric monitoring as well as daily monitoring at home and hospital, during driving, and while performing exercises. Further, since March 2020, the COVID-19 pandemic is ongoing globally. As the virus affects the lungs, noncontact SpO2 monitoring is a valuable alternative to contact-based methods, as it avoids transmission of the virus. HR, RR, and BST can also be monitored to keep the best track of subject’s health. It is also further useful for monitoring vital signs in a noncontact way during other pandemic situations.

4.3. Research Gaps and Future Research Directions

In the following, we discuss the future of camera-based health monitoring and if camera-based approaches could eventually replace contact-based modalities in the clinical environment. First, the number of vital signs is limited with the vast majority of studies focusing on HR, RR, and BP [28,50,66,191]. Other measures are arterial stiffness and distal pulse reflection [88]. So far, BST cannot be measured using RGB cameras. Additionally, there is difference between body core temperature (BCT), BT, and BST. BCT refers to the temperature of the internal tissues of the body (e.g., pulmonary artery). Due to the balance of heat generation, absorption and loss, BCT maintains at constant temperature [58]. However, it can only be measured invasively by utilizing catheters. BT is a contact-based technique that utilizes sensors at various locations of the body (e.g., oral, axillary, rectal), which makes it difficult to use for infants or children. Body mass index, age, and the metabolic rate impact BT [192]. BST is an indicator of skin blood flow that may reflects the temperature of the body [58]. Current studies focus on BST estimation to measure core temperature from facial regions utilizing FIR imaging. However, physical activities and environmental factors differ the measure from the medically more relevant BCT. Obtaining BCT from the facial region using FIR is still under research, and machine learning might help here.
Although the research on iPPG is relatively recent, it has rapidly spread from the laboratory to everyday environments research. As a substitute of conventional contact-based techniques, iPPG will be able to assess almost all physiological measurements that contact PPG provides. However, it remains under research due to various reasons that is discussed below in detail. Until now, only stationary studies in well-controlled settings yield reliable vital signs. The performance degrades significantly if the illumination varies, the subject has darker skin tone or is in motion, or if camera–subject distance is too large. In this, RGB images with proper illumination provide better signal performance, whereas, NIR cameras have a lower SNR due to the lower absorption of hemoglobin [41]. However, a few studies recently concentrated on NIR camera-based vital sign monitoring owing to the advantages of functioning in dark conditions (e.g., sleep monitoring, driver monitoring) [34,41,93]. Even though, it is not explored widely. Future research on NIR cameras might yield a path for camera-based vital sign monitoring in all light conditions.
We further are concerned that most studies and existing databases use a small number of participants (n = 10–30) and homogeneous groups (e.g., same skin group or particular geographical location). In addition, there might be bias in age and gender due to changes in the skin properties. There is no work focusing exclusively on healthy elder subjects or children above age of two. Hence, further studies are required to effectively evaluate the influence of age and gender. Most of the studies disregard skin colors although the melanin concentration has a strong impact on the measurements. Developing robust algorithms using DL might help here. Furthermore, the majority of the studies was conducted in indoor environments and obtained reliable results. Outdoor environments, on the other hand, can result in large amounts of noise, making it difficult to monitor vital signs [36,168]. This limits the translation into clinical practice and real-time environments. Though the techniques need further improvement before being used in clinical medicine, home-based indoor applications can be explored already [193]. Future research is required to focus on resolving the issues by tuning these parameters, namely camera–subject distance, camera parameters, and illumination, in various indoor and outdoor environments to obtain the best possible results that might improve the robustness in practical applications (Figure 12). As of today, subject and camera parameters are not mentioned in most of the article. Researchers need to improve and follow existing guidelines such as STARE-HI [194].
On the other hand, a few studies have shown to yield stable results even under nonstationary conditions [75,76,115,121,157,190]. However, most of the studies suffer from motion artifacts due to the dynamic situations. Although a band-pass filter extracts an interested frequency region, utilization of motion artifact-based algorithms will help to reduce the error. But only a few studies employed them so far [24,26,37,49,89,190]. In addition, the conventional signal processing methods, such as parametric-based BSS method and model-based methods, reported in several articles to recover the pulse signal, may fail due to lack of previous knowledge and assumption of skin tone vectors, respectively. However, nonparametric-based fast kernel density-ICA called semi-BSS, presented by Song et al., is reported to have promising results with varied resolution and camera–subject distance [59]. Although DL-based approaches have significant potential, they are still in the early stages of research for iPPG. However, end-to-end DL networks as demonstrated by Perepelkina et al., and Yu et al., could be a potential approach which have been shown to outperform classical signal processing techniques [121,163]. In the future, semi-BSS-based techniques and DL networks could be employed to overcome the classical signal processing methods.
The majority of the articles relied on a single camera-based physiological monitoring, either RGB or IR. However, single cameras cannot persistently acquire the frames due to head or body posture variations or presence of obstacles. This might lead to missing observations or loss of data. Thereby, multi-camera-based researches could be able to overcome this challenge with the help of fusion techniques [120]. Studies such as Liu et al., detect HR from multiple people by employing a single camera frame and yield promising results [114]. Hence, combining these two methods, multi--camera and multi-person-based vital sign monitoring could be extended with the help of surveillance camera in hospital, in-home, and workspace environments. Our group recently explored these measurements for continuous health monitoring in indoor with the help of convolutional neural network [195]. In addition, camera-based monitoring can extract multiple vital signs (Table 1). It could be included in multi-camera and multi-person measurements. Future research might focus on these techniques in real-world scenarios to develop more practical applications with sufficient reliability. This can also be developed into optical-sensors-based devices and smartphone-based or webcam-based applications that might results in continuous vital signs monitoring or periodic measurement [196,197]. In addition, in the last 20 years, the world has frequently faced a great number of pandemic situations such as SARS in 2002, swine flu in 2009, Ebola in 2014, and now the coronavirus in 2020. So, the current situation urgently requires developing towards noncontact and unobtrusive vital signs monitoring devices without virus transmission. It might be helpful in the pandemic situation for noncontact monitoring of vital signs and for people under home quarantine.
Our review demonstrated a wide variety of clinical and nonclinical applications, including smart homes and smart cars, which allow for continuous monitoring of vital signs in everyday life [33,35]. In the smart home, camera-based monitoring of vital signs can be combined with gait analysis, fall detection, and sleep monitoring. This could minimize costs and improve out-of-hospital treatments. To reach these application areas, further developments are required. The technology needs robustness in dynamic settings and be able to cope with movement, facial expression, varying illumination, and different camera-subject distance. Furthermore, it requires the development of robust algorithms [87]. Nonetheless, future research should be directed toward mobile health (mHealth) applications to provide users a timely and consistent feedback of their health status [158]. mHealth must perform without hardware attachment and provide user-friendly interfaces. Due to the increasing number of smartphone users and continuously improved smartphone sensor technology (e.g., camera), we expect that many studies of this review transform into smartphone applications and serve in telemedicine.

5. Conclusions

Camera-based techniques monitor vital signs unobtrusively. This review provides an overview of data acquisition technology (hardware), image and signal processing (software), accuracy, and application areas. As of today, HR and RR are reliably monitored using RGB cameras in controlled settings only, but other vital signs are still lacking robust and sufficiently precise systems. Subject, camera, setting, and environmental parameters have a significant impact on the accuracy. To overcome these effects, robust algorithms based on advanced signal processing or DL are urgently needed. Additionally, fusion-based approaches (e.g., multiple ROI or multiple cameras) bear the potential of enhancing reliability. Different ranges of hardware and software parameter can be investigated to obtain the best possible results for various environments. With respect to the COVID-19 pandemic, a potential application could be to deploy cameras for smart-home-based health monitoring of subjects undergoing quarantine. Recently, the U.S. Food and Drug Administration (FDA) approved smart phone applications for therapy e.g., for tinnitus [198]. In light of this development, smartphone camera-based vital sign monitoring could also be part of digital health. We call for collaboration across the world to collect publicly available dataset of large diversity and size as a basis to make camera-based vital signs monitoring sufficiently robust to be translated into smartphone apps and regularly used in future mHealth and telemedicine.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22114097/s1, Table S1. Extracted information from all articles.

Author Contributions

Conceptualization, V.S., N.S. and T.M.D.; methodology, V.S., N.S. and T.M.D.; literature review, V.S., N.S., J.W., N.G., J.M.W., S.L. and T.M.D.; writing—original draft preparation, V.S.; writing—review and editing, V.S., N.S., J.W., N.G., J.M.W., S.L., R.S. and T.M.D.; visualization, V.S., N.S., J.W. and T.M.D.; supervision, T.M.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Source data for all figures are provided with the paper and Supplementary File.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in the manuscript:
ECGElectrocardiography
PPG Photoplethysmography
HR Heart rate
RR Respiratory rate
BP Blood pressure
BTBody temperature
SpO2Oxygen saturation
bpmbeats per minute
cpm cycles per minute
BCGballistocardiography
ICUIntensive care units
iPPGImaging PPG
rPPGRemote PPG
HRVHR variability
PRPulse rate
PRVPulse rate variability
RGBRed–green–blue
NIRNear-infrared
FIRFar-infrared
BSTBody skin temperature
IRInfrared
pxpixels
SNRSignal-to-noise ratio
PRISMAPreferred reporting items for systematic reviews and meta-analyses
ACMAssociation for Computing Machinery
ROIregion of interest
fps frame per second
MR-NIRPMERL-Rice near infrared pulse dataset
CCD Charge coupled devices
EULAEnd user license agreement
HSVHue saturation value
YCbCrluminance, chroma blue, chroma red
AAMIAssociation for the Advancement of Medical Instrumentation
BSS Blind source separation
ICAIndependent component analysis
PCAPrincipal component analysis
DLDeep learning
CHROMChrominance model
POSPlane-orthogonal-to-skin
IBIInterbeat interval
PTTPulse transit time
MAEMean absolute error
RMSERoot mean square error
CCCorrelation coefficient
BCTBody core temperature
mHealth Mobile health
FDAFood and Drugs Administration

Appendix A. Searching String

Appendix A.1. PubMed

(“RGB”[Title] OR “infrared”[Title] OR “thermal”[Title] OR “thermography”[Title] OR “camera”[Title] OR “video”[Title] OR “image”[Title] OR “imaging”[Title] OR “remote”[Title] OR “non contact”[Title] OR “non-contact”[Title] OR “touchless”[Title] OR “contactless”[Title] OR “contact-free”[Title] OR “vision-based”[Title])
AND
  • (“biosignal”[Title] OR “bio signal”[Title] OR “biomedical signal”[Title] OR “physiological”[Title] OR “cardiorespiratory signal”[Title] OR “cardiorespiratory signals”[Title] OR “ppg”[Title] OR “photoplethysmograph”[Title] OR “photoplethysmography”[Title] OR “photoplethysmogram”[Title] OR “vital sign”[Title] OR “vital signs”[Title] OR “respiration”[Title] OR “respiratory”[Title] OR “breathing”[Title] OR “heart rate”[Title] OR “pulse rate”[Title] OR “heart beat”[Title] OR “body temperature”[Title] OR “oxygen saturation”[Title] OR “blood pressure”[Title] OR “face”[Title] OR “facial”[Title])
AND
  • (“processing”[Title] OR “analytics”[Title] OR “analysis”[Title] OR “analyse”[Title] OR “analysing”[Title] OR “analyze”[Title] OR “analyzing”[Title] OR “measure”[Title] OR “measurement”[Title] OR “measuring”[Title] OR “sensing”[Title] OR “monitor”[Title] OR “monitoring”[Title] OR “estimation”[Title] OR “estimate”[Title] OR “estimating”[Title] OR “quantification”[Title] OR “identification”[Title] OR “recognition”[Title] OR “detect”[Title] OR “detection”[Title] OR “detecting”[Title] OR “tracking”[Title] OR “feature map”[Title] OR “feature maps”[Title])
NOT
  • (review[pt] OR systematic review[pt] OR survey[pt])
AND
  • (english[la])
AND
  • (“2018/01/01”[DP]: “2021/04/30” [DP])

Appendix A.2. Scopus

TITLE((“RGB” OR “infrared” OR “thermal” OR “thermography” OR “camera” OR “video” OR “image” OR “imaging” OR “remote” OR “non contact” OR “non-contact” OR “touchless” OR “contactless” OR “contact-free” OR “vision-based”)
AND
  • (“biosignal” OR “bio signal” OR “biomedical signal” OR “physiological” OR “cardiorespiratory signal” OR “cardiorespiratory signals” OR “ppg” OR “photoplethysmograph” OR “photoplethysmography” OR “photoplethysmogram” OR “vital sign” OR “vital signs” OR “respiration” OR “respiratory” OR “breathing” OR “heart rate” OR “pulse rate” OR “heart beat” OR “body temperature” OR “oxygen saturation” OR “blood pressure” OR “face” OR “facial”)
AND
  • (“processing” OR “analytics” OR “analysis” OR “analyse” OR “analysing” OR “analyze” OR “analyzing” OR “measure” OR “measurement” OR “measuring” OR “sensing” OR “monitor” OR “monitoring” OR “estimation” OR “estimate” OR “estimating” OR “quantification” OR “identification” OR “recognition” OR “detect” OR “detection” OR “detecting” OR “tracking” OR “feature map” OR “feature maps”))
AND
  • PUBDATETXT(“January 2021” OR “February 2021” OR “March 2021” OR “April 2021” OR “January 2020” OR “February 2020” OR “March 2020” OR “April 2020” OR “May 2020” OR “June 2020” OR “July 2020” OR “August 2020” OR “September 2020” OR “October 2020” OR “November 2020” OR “December 2020” OR “January 2019” OR “February 2019” OR “March 2019” OR “April 2019” OR “May 2019” OR “June 2019” OR “July 2019” OR “August 2019” OR “September 2019” OR “October 2019” OR “November 2019” OR “December 2019” OR “January 2018” OR “February 2018” OR “March 2018” OR “April 2018” OR “May 2018” OR “June 2018” OR “July 2018” OR “August 2018” OR “September 2018” OR “October 2018” OR “November 2018” OR “December 2018”)
AND NOT
  • (“review” OR “survey”)
AND
  • (LIMIT-TO (LANGUAGE, “English”))
AND
  • (LIMIT-TO (PUBSTAGE, “final”))
AND
  • (LIMIT-TO (SUBJAREA, “MEDI”) OR LIMIT-TO (SUBJAREA, “ENGI”) OR LIMIT-TO (SUBJAREA, “NEUR”) OR LIMIT-TO (SUBJAREA, “COMP”) OR LIMIT-TO (SUBJAREA, “HEAL”) OR LIMIT-TO (SUBJAREA, “MULT”))

Appendix A.3. ACM

Title: ((“RGB” OR “infrared” OR “thermal” OR “thermography” OR “camera” OR “video” OR “image” OR “imaging” OR “remote” OR “non contact” OR “non-contact” OR “touchless” OR “contactless” OR “contact-free” OR “vision-based”)
AND
  • (“biosignal” OR “bio signal” OR “biomedical signal” OR “physiological” OR “cardiorespiratory signal” OR “cardiorespiratory signals” OR “ppg” OR “photoplethysmograph” OR “photoplethysmography” OR “photoplethysmogram” OR “vital sign” OR “vital signs” OR “respiration” OR “respiratory” OR “breathing” OR “heart rate” OR “pulse rate” OR “heart beat” OR “body temperature” OR “oxygen saturation” OR “blood pressure” OR “face” OR “facial”)
AND
  • (“processing” OR “analytics” OR “analysis” OR “analyse” OR “analysing” OR “analyze” OR “analyzing” OR “measure” OR “measurement” OR “measuring” OR “sensing” OR “monitor” OR “monitoring” OR “estimation” OR “estimate” OR “estimating” OR “quantification” OR “identification” OR “recognition” OR “detect” OR “detection” OR “detecting” OR “tracking” OR “feature map” OR “feature maps”))
NOT
  • “review” NOT “survey”
AND
  • 2018:2021

References

  1. Sörnmo, L.; Laguna, P. Bioelectrical Signal Processing in Cardiac and Neurological Applications; Academic Press: Cambridge, MA, USA, 2005; Volume 8. [Google Scholar]
  2. Zhou, G.; Wang, Y.; Cui, L. Biomedical Sensor, Device and Measurement Systems. In Advances in Bioengineering; Serra, P.A., Ed.; InTech: London, UK, 2015; ISBN 978-953-51-2141-1. [Google Scholar]
  3. Levick, J.R. An Introduction to Cardiovascular Physiology; Butterworths: London, UK; Boston, MA, USA, 1991; ISBN 978-0-7506-1028-5. [Google Scholar]
  4. Lu, G.; Yang, F.; Taylor, J.A.; Stein, J.F. A Comparison of Photoplethysmography and ECG Recording to Analyse Heart Rate Variability in Healthy Subjects. J. Med. Eng. Technol. 2009, 33, 634–641. [Google Scholar] [CrossRef] [PubMed]
  5. Evans, D.; Hodgkinson, B.; Berry, J. Vital Signs in Hospital Patients: A Systematic Review. Int. J. Nurs. Stud. 2001, 38, 643–650. [Google Scholar] [CrossRef]
  6. Rohmetra, H.; Raghunath, N.; Narang, P.; Chamola, V.; Guizani, M.; Lakkaniga, N.R. AI-Enabled Remote Monitoring of Vital Signs for COVID-19: Methods, Prospects and Challenges. Computing 2021, 1–27. [Google Scholar] [CrossRef]
  7. Wang, W.; Wang, X. Contactless Vital Signs Monitoring; Academic Press: Cambridge, MA, USA, 2022. [Google Scholar]
  8. Hlastala, M.P.; Berger, A.J. Physiology of Respiration; Oxford University Press: Oxford, UK, 2001. [Google Scholar]
  9. Sherwood, L. Human Physiology: From Cells to Systems, 9th ed.; Brooks/Cole, Cengage Learning: Boston, MA, USA, 2016; ISBN 978-1-285-86693-2. [Google Scholar]
  10. Ruiz, M.; Garcia, J.; Fernández, B. Body Temperature and Its Importance as a Vital Constant. Rev. Enferm. 2009, 32, 44–52. [Google Scholar]
  11. Hafen, B.B.; Sharma, S. Oxygen Saturation; StatPearls Publishing: Treasure Island, FL, USA, 2021. [Google Scholar]
  12. Sun, Z.; He, Q.; Li, Y.; Wang, W.; Wang, R.K. Robust Non-Contact Peripheral Oxygenation Saturation Measurement Using Smartphone-Enabled Imaging Photoplethysmography. Biomed. Opt. Express 2021, 12, 1746. [Google Scholar] [CrossRef]
  13. Antink, C.H.; Lyra, S.; Paul, M.; Yu, X.; Leonhardt, S. A Broader Look: Camera-Based Vital Sign Estimation across the Spectrum. Yearb. Med. Inf. 2019, 28, 102–114. [Google Scholar] [CrossRef] [Green Version]
  14. Zheng, Y.-L.; Ding, X.-R.; Poon, C.C.Y.; Lo, B.P.L.; Zhang, H.; Zhou, X.-L.; Yang, G.-Z.; Zhao, N.; Zhang, Y.-T. Unobtrusive Sensing and Wearable Devices for Health Informatics. IEEE Trans. Biomed. Eng. 2014, 61, 1538–1554. [Google Scholar] [CrossRef]
  15. Leonhardt, S.; Leicht, L.; Teichmann, D. Unobtrusive Vital Sign Monitoring in Automotive Environments—A Review. Sensors 2018, 18, 3080. [Google Scholar] [CrossRef] [Green Version]
  16. Magjarevic, R.; Ottenbacher, J.; Heuer, S. Motion Artefacts in Capacitively Coupled ECG Electrodes. In IFMBE Proceedings, Proceedings of the World Congress on Medical Physics and Biomedical Engineering, Munich, Germany, 7–12 September 2009; Dössel, O., Schlegel, W.C., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; Volume 25/4, pp. 1059–1062. ISBN 978-3-64203-8-815. [Google Scholar]
  17. Kumar, M.; Veeraraghavan, A.; Sabharwal, A. DistancePPG: Robust Non-Contact Vital Signs Monitoring Using a Camera. Biomed. Opt. Express 2015, 6, 1565. [Google Scholar] [CrossRef] [Green Version]
  18. Wu, T.; Blazek, V.; Schmitt, H.J. Photoplethysmography Imaging: A New Noninvasive and Noncontact Method for Mapping of the Dermal Perfusion Changes. In Proceedings of the SPIE 4163, Optical Techniques and Instrumentation for the Measurement of Blood Composition, Structure, and Dynamics, Amsterdam, The Netherlands, 4–8 July 2000. [Google Scholar]
  19. Huelsbusch, M.; Blazek, V. Contactless Mapping of Rhythmical Phenomena in Tissue Perfusion Using PPGI. In Proceedings of the SPIE 4683, Medical Imaging 2002: Physiology and Function from Multidimensional Images, San Diego, CA, USA, 23–28 February 2002; pp. 110–117. [Google Scholar]
  20. Nadrag, C.; Poenaru, V.; Suciu, G. Heart Rate Measurement Using Face Detection in Video. In Proceedings of the 2018 International Conference on Communications (COMM), Bucharest, Romania, 14–16 June 2018. [Google Scholar]
  21. Chaichulee, S.; Villarroel, M.; Jorge, J.; Arteta, C.; McCormick, K.; Zisserman, A.; Tarassenko, L. Cardio-Respiratory Signal Extraction from Video Camera Data for Continuous Non-Contact Vital Sign Monitoring Using Deep Learning. Physiol. Meas. 2019, 40, 115001. [Google Scholar] [CrossRef] [Green Version]
  22. Lee, H.; Cho, A.; Whang, M. Vision-Based Measurement of Heart Rate from Ballistocardiographic Head Movements Using Unsupervised Clustering. Sensors 2019, 19, 3263. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Yu, X.; Laurentius, T.; Bollheimer, C.; Leonhardt, S.; Antink, C.H. Noncontact Monitoring of Heart Rate and Heart Rate Variability in Geriatric Patients Using Photoplethysmography Imaging. IEEE J. Biomed. Health Inform. 2021, 25, 1781–1792. [Google Scholar] [CrossRef] [PubMed]
  24. Rossol, S.L.; Yang, J.K.; Toney-Noland, C.; Bergin, J.; Basavaraju, C.; Kumar, P.; Lee, H.C. Non-Contact Video-Based Neonatal Respiratory Monitoring. Children 2020, 7, 171. [Google Scholar] [CrossRef] [PubMed]
  25. Lorato, I.; Stuijk, S.; Meftah, M.; Kommers, D.; Andriessen, P.; van Pul, C.; de Haan, G. Multi-Camera Infrared Thermography for Infant Respiration Monitoring. Biomed. Opt. Express 2020, 11, 4848–4861. [Google Scholar] [CrossRef]
  26. Lorato, I.; Stuijk, S.; Meftah, M.; Kommers, D.; Andriessen, P.; van Pul, C.; de Haan, G. Towards Continuous Camera-Based Respiration Monitoring in Infants. Sensors 2021, 21, 2268. [Google Scholar] [CrossRef]
  27. Pereira, C.B.; Yu, X.; Goos, T.; Reiss, I.; Orlikowsky, T.; Heimann, K.; Venema, B.; Blazek, V.; Leonhardt, S.; Teichmann, D. Noncontact Monitoring of Respiratory Rate in Newborn Infants Using Thermal Imaging. IEEE Trans. Biomed. Eng. 2019, 66, 1105–1114. [Google Scholar] [CrossRef]
  28. Maurya, L.; Kaur, P.; Chawla, D.; Mahapatra, P. Non-Contact Breathing Rate Monitoring in Newborns: A Review. Comput. Biol. Med. 2021, 132, 104321. [Google Scholar] [CrossRef]
  29. Chan, P.; Wong, G.; Dinh Nguyen, T.; Nguyen, T.; McNeil, J.; Hopper, I. Estimation of Respiratory Rate Using Infrared Video in an Inpatient Population: An Observational Study. J. Clin. Monit. Comput. 2020, 34, 1275–1284. [Google Scholar] [CrossRef]
  30. Spicher, N.; Kukuk, M.; Maderwald, S.; Ladd, M.E. Initial Evaluation of Prospective Cardiac Triggering Using Photoplethysmography Signals Recorded with a Video Camera Compared to Pulse Oximetry and Electrocardiography at 7T MRI. BioMed. Eng. OnLine 2016, 15, 126. [Google Scholar] [CrossRef] [Green Version]
  31. Kim, J.; Kwon, J.H.; Kim, E.; Yoo, S.K.; Shin, C. Respiratory Measurement Using Infrared Thermography and Respiratory Volume Monitor during Sedation in Patients Undergoing Endoscopic Urologic Procedures under Spinal Anesthesia. J. Clin. Monit. Comput. 2019, 33, 647–656. [Google Scholar] [CrossRef]
  32. Jakkaew, P.; Onoye, T. Non-Contact Respiration Monitoring and Body Movements Detection for Sleep Using Thermal Imaging. Sensors 2020, 20, 6307. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, J.; Spicher, N.; Warnecke, J.M.; Haghi, M.; Schwartze, J.; Deserno, T.M. Unobtrusive Health Monitoring in Private Spaces: The Smart Home. Sensors 2021, 21, 864. [Google Scholar] [CrossRef] [PubMed]
  34. Hu, M.; Zhai, G.; Li, D.; Fan, Y.; Duan, H.; Zhu, W.; Yang, X. Combination of Near-Infrared and Thermal Imaging Techniques for the Remote and Simultaneous Measurements of Breathing and Heart Rates under Sleep Situation. PLoS ONE 2018, 13, e0190466. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Wang, J.; Warnecke, J.; Haghi, M.; Deserno, T. Unobtrusive Health Monitoring in Private Spaces: The Smart Vehicle. Sensors 2020, 20, 2442. [Google Scholar] [CrossRef]
  36. Nowara, E.M.; Marks, T.K.; Mansour, H.; Veeraraghavan, A. SparsePPG: Towards Driver Monitoring Using Camera-Based Vital Signs Estimation in Near-Infrared. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1272–1281. [Google Scholar]
  37. Sugita, N.; Noro, T.; Yoshizawa, M.; Ichiji, K.; Yamaki, S.; Homma, N. Estimation of Absolute Blood Pressure Using Video Images Captured at Different Heights from the Heart. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 4458–4461. [Google Scholar]
  38. Kim, C.-I.; Lee, J.-H. The Non-Contact Biometric Identified Bio Signal Measurement Sensor and Algorithms. THC 2018, 26, 215–228. [Google Scholar] [CrossRef] [Green Version]
  39. Benedetto, S.; Caldato, C.; Greenwood, D.C.; Bartoli, N.; Pensabene, V.; Actis, P. Remote Heart Rate Monitoring—Assessment of the Facereader RPPg by Noldus. PLoS ONE 2019, 14, e0225592. [Google Scholar] [CrossRef] [Green Version]
  40. Wei, J.; Luo, H.; Wu, S.J.; Zheng, P.P.; Fu, G.; Lee, K. Transdermal Optical Imaging Reveal Basal Stress via Heart Rate Variability Analysis: A Novel Methodology Comparable to Electrocardiography. Front. Psychol. 2018, 9, 98. [Google Scholar] [CrossRef] [Green Version]
  41. Nowara, E.M.; Marks, T.K.; Mansour, H.; Veeraraghavan, A. Near-Infrared Imaging Photoplethysmography During Driving. IEEE Trans. Intell. Transport. Syst. 2020, 23, 3589–3600. [Google Scholar] [CrossRef]
  42. Ding, X.; Nassehi, D.; Larson, E.C. Measuring Oxygen Saturation With Smartphone Cameras Using Convolutional Neural Networks. IEEE J. Biomed. Health Inform. 2019, 23, 2603–2610. [Google Scholar] [CrossRef]
  43. Sanyal, S.; Nundy, K.K. Algorithms for Monitoring Heart Rate and Respiratory Rate From the Video of a User’s Face. IEEE J. Transl. Eng. Health Med. 2018, 6, 1–11. [Google Scholar] [CrossRef]
  44. Liu, H.; Allen, J.; Zheng, D.; Chen, F. Recent Development of Respiratory Rate Measurement Technologies. Physiol. Meas. 2019, 40, 07TR01. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Cerina, L.; Iozzia, L.; Mainardi, L. Influence of Acquisition Frame-Rate and Video Compression Techniques on Pulse-Rate Variability Estimation from VPPG Signal. Biomed. Eng. Biomed. Tech. 2017, 64, 53–65. [Google Scholar] [CrossRef] [PubMed]
  46. Fleischhauer, V.; Ruprecht, N.; Sorelli, M.; Bocchi, L.; Zaunseder, S. Pulse Decomposition Analysis in Photoplethysmography Imaging. Physiol. Meas. 2020, 41, 095009. [Google Scholar] [CrossRef] [PubMed]
  47. Gudi, A.; Bittner, M.; van Gemert, J. Real-Time Webcam Heart-Rate and Variability Estimation with Clean Ground Truth for Evaluation. Appl. Sci. 2020, 10, 8630. [Google Scholar] [CrossRef]
  48. Wang, W.; den Brinker, A.C.; de Haan, G. Single-Element Remote-PPG. IEEE Trans. Biomed. Eng. 2019, 66, 2032–2043. [Google Scholar] [CrossRef] [PubMed]
  49. Bánhalmi, A.; Borbás, J.; Fidrich, M.; Bilicki, V.; Gingl, Z.; Rudas, L. Analysis of a Pulse Rate Variability Measurement Using a Smartphone Camera. J. Healthc. Eng. 2018, 2018, 1–15. [Google Scholar] [CrossRef]
  50. Hassan, M.A.; Malik, A.S.; Fofi, D.; Saad, N.; Karasfi, B.; Ali, Y.S.; Meriaudeau, F. Heart Rate Estimation Using Facial Video: A Review. Biomed. Signal Process. Control 2017, 38, 346–360. [Google Scholar] [CrossRef]
  51. Chen, W.; Hernandez, J.; Picard, R.W. Estimating Carotid Pulse and Breathing Rate from Near-Infrared Video of the Neck. Physiol. Meas. 2018, 39, 10NT01. [Google Scholar] [CrossRef] [Green Version]
  52. Queiroz, L.; Oliveira, H.; Yanushkevich, S.; Ferber, R. Video-Based Breathing Rate Monitoring in Sleeping Subjects. In Proceedings of the 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11 October 2020; pp. 2458–2464. [Google Scholar]
  53. Zhu, K.; Li, M.; Akbarian, S.; Hafezi, M.; Yadollahi, A.; Taati, B. Vision-Based Heart and Respiratory Rate Monitoring During Sleep—A Validation Study for the Population at Risk of Sleep Apnea. IEEE J. Transl. Eng. Health Med. 2019, 7, 1–8. [Google Scholar] [CrossRef]
  54. Nowara, E.; McDuff, D. Combating the Impact of Video Compression on Non-Contact Vital Sign Measurement Using Supervised Learning. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27–28 October 2019; pp. 1706–1712. [Google Scholar]
  55. Kado, S.; Monno, Y.; Moriwaki, K.; Yoshizaki, K.; Tanaka, M.; Okutomi, M. Remote Heart Rate Measurement from RGB-NIR Video Based on Spatial and Spectral Face Patch Selection. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5676–5680. [Google Scholar]
  56. Chen, M.; Zhu, Q.; Zhang, H.; Wu, M.; Wang, Q. Respiratory Rate Estimation from Face Videos. In Proceedings of the 2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Chicago, IL, USA, 19–22 May 2019; pp. 1–4. [Google Scholar]
  57. Barbosa Pereira, C.; Czaplik, M.; Blazek, V.; Leonhardt, S.; Teichmann, D. Monitoring of Cardiorespiratory Signals Using Thermal Imaging: A Pilot Study on Healthy Human Subjects. Sensors 2018, 18, 1541. [Google Scholar] [CrossRef] [Green Version]
  58. Wartzek, T.; Mühlsteff, J.; Imhoff, M. Temperature Measurement. Biomed. Tech. Biomed. Eng. 2011, 56, 241–257. [Google Scholar] [CrossRef] [PubMed]
  59. Song, R.; Zhang, S.; Cheng, J.; Li, C.; Chen, X. New Insights on Super-High Resolution for Video-Based Heart Rate Estimation with a Semi-Blind Source Separation Method. Comput. Biol. Med. 2020, 116, 103535. [Google Scholar] [CrossRef] [PubMed]
  60. Scebba, G.; Da Poian, G.; Karlen, W. Multispectral Video Fusion for Non-Contact Monitoring of Respiratory Rate and Apnea. IEEE Trans. Biomed. Eng. 2021, 68, 350–359. [Google Scholar] [CrossRef] [PubMed]
  61. Zhou, Y.; Ghassemi, P.; Chen, M.; McBride, D.; Casamento, J.P.; Pfefer, T.J.; Wang, Q. Clinical Evaluation of Fever-Screening Thermography: Impact of Consensus Guidelines and Facial Measurement Location. J. Biomed. Opt. 2020, 25, 097002. [Google Scholar] [CrossRef]
  62. Li, M.H.; Yadollahi, A.; Taati, B. Noncontact Vision-Based Cardiopulmonary Monitoring in Different Sleeping Positions. IEEE J. Biomed. Health Inform. 2017, 21, 1367–1375. [Google Scholar] [CrossRef]
  63. Sikdar, A.; Behera, S.K.; Dogra, D.P. Computer-Vision-Guided Human Pulse Rate Estimation: A Review. IEEE Rev. Biomed. Eng. 2016, 9, 91–105. [Google Scholar] [CrossRef]
  64. Rouast, P.V.; Adam, M.T.P.; Chiong, R.; Cornforth, D.; Lux, E. Remote Heart Rate Measurement Using Low-Cost RGB Face Video: A Technical Literature Review. Front. Comput. Sci. 2018, 12, 858–872. [Google Scholar] [CrossRef]
  65. Wang, C.; Pun, T.; Chanel, G. A Comparative Survey of Methods for Remote Heart Rate Detection from Frontal Face Videos. Front. Bioeng. Biotechnol. 2018, 6, 33. [Google Scholar] [CrossRef] [Green Version]
  66. Zaunseder, S.; Trumpp, A.; Wedekind, D.; Malberg, H. Cardiovascular Assessment by Imaging Photoplethysmography—A Review. Biomed. Eng. Biomed. Tech. 2018, 63, 617–634. [Google Scholar] [CrossRef]
  67. Addison, A.P.; Addison, P.S.; Smit, P.; Jacquel, D.; Borg, U.R. Noncontact Respiratory Monitoring Using Depth Sensing Cameras: A Review of Current Literature. Sensors 2021, 21, 1135. [Google Scholar] [CrossRef]
  68. Rehouma, H.; Noumeir, R.; Essouri, S.; Jouvet, P. Advancements in Methods and Camera-Based Sensors for the Quantification of Respiration. Sensors 2020, 20, 7252. [Google Scholar] [CrossRef] [PubMed]
  69. Khanam, F.-T.-Z.; Al-Naji, A.; Chahl, J. Remote Monitoring of Vital Signs in Diverse Non-Clinical and Clinical Scenarios Using Computer Vision Systems: A Review. Appl. Sci. 2019, 9, 4474. [Google Scholar] [CrossRef] [Green Version]
  70. Steinman, J.; Barszczyk, A.; Sun, H.-S.; Lee, K.; Feng, Z.-P. Smartphones and Video Cameras: Future Methods for Blood Pressure Measurement. Front. Digit. Health 2021, 3, 770096. [Google Scholar] [CrossRef] [PubMed]
  71. Harford, M.; Catherall, J.; Gerry, S.; Young, J.; Watkinson, P. Availability and Performance of Image-Based, Non-Contact Methods of Monitoring Heart Rate, Blood Pressure, Respiratory Rate, and Oxygen Saturation: A Systematic Review. Physiol. Meas. 2019, 40, 06TR01. [Google Scholar] [CrossRef] [PubMed]
  72. Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G. PRISMA Group Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement. PLoS Med. 2009, 6, e1000097. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  73. Hsu, W.; Baumgartner, C.; Deserno, T.M.; Section Editors of the IMIA Yearbook Section on Sensors, Signals, and Imaging Informatics. Notable Papers and New Directions in Sensors, Signals, and Imaging Informatics. Yearb. Med. Inf. 2021, 30, 150–158. [Google Scholar] [CrossRef]
  74. Zhang, S.; Song, R.; Cheng, J.; Zhang, Y.; Chen, X. A Feasibility Study of a Video-Based Heart Rate Estimation Method with Convolutional Neural Networks. In Proceedings of the 2019 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications (CIVEMSA), Tianjin, China, 14–16 June 2019; pp. 1–5. [Google Scholar]
  75. Qiu, Y.; Liu, Y.; Arteaga-Falconi, J.; Dong, H.; El Saddik, A. EVM-CNN: Real-Time Contactless Heart Rate Estimation from Facial Video. IEEE Trans. Multimed. 2019, 21, 1778–1787. [Google Scholar] [CrossRef]
  76. Jaiswal, K.B.; Meenpal, T. Continuous Pulse Rate Monitoring from Facial Video Using RPPG. In Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, India, 1–3 July 2020; pp. 1–5. [Google Scholar]
  77. Liu, X.; Yang, X.; Jin, J.; Wong, A. Detecting Pulse Wave from Unstable Facial Videos Recorded From Consumer-Level Cameras: A Disturbance-Adaptive Orthogonal Matching Pursuit. IEEE Trans. Biomed. Eng. 2020, 67, 3352–3362. [Google Scholar] [CrossRef]
  78. Hsu, G.-S.J.; Xie, R.-C.; Ambikapathi, A.; Chou, K.-J. A Deep Learning Framework for Heart Rate Estimation from Facial Videos. Neurocomputing 2020, 417, 155–166. [Google Scholar] [CrossRef]
  79. Maki, Y.; Monno, Y.; Tanaka, M.; Okutomi, M. Remote Heart Rate Estimation Based on 3D Facial Landmarks. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 2634–2637. [Google Scholar]
  80. Rapczynski, M.; Werner, P.; Al-Hamadi, A. Effects of Video Encoding on Camera-Based Heart Rate Estimation. IEEE Trans. Biomed. Eng. 2019, 66, 3360–3370. [Google Scholar] [CrossRef]
  81. Gupta, P.; Bhowmick, B.; Pal, A. MOMBAT: Heart Rate Monitoring from Face Video Using Pulse Modeling and Bayesian Tracking. Comput. Biol. Med. 2020, 121, 103813. [Google Scholar] [CrossRef] [PubMed]
  82. Umematsu, T.; Tsujikawa, M. Head-Motion Robust Video-Based Heart Rate Estimation Using Facial Feature Point Fluctuations. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 1–4. [Google Scholar]
  83. Raseena, K.T.; Ghosh, P.K. A Maximum Likelihood Formulation to Exploit Heart Rate Variability for Robust Heart Rate Estimation From Facial Video. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5191–5194. [Google Scholar]
  84. Kurihara, K.; Sugimura, D.; Hamamoto, T. Non-Contact Heart Rate Estimation via Adaptive RGB/NIR Signal Fusion. IEEE Trans. Image Process. 2021, 30, 6528–6543. [Google Scholar] [CrossRef] [PubMed]
  85. Li, P.; Benezeth, Y.; Macwan, R.; Nakamura, K.; Gomez, R.; Li, C.; Yang, F. Video-Based Pulse Rate Variability Measurement Using Periodic Variance Maximization and Adaptive Two-Window Peak Detection. Sensors 2020, 20, 2752. [Google Scholar] [CrossRef] [PubMed]
  86. Tsou, Y.-Y.; Lee, Y.-A.; Hsu, C.-T.; Chang, S.-H. Siamese-RPPG Network: Remote Photoplethysmography Signal Estimation from Face Videos. In Proceedings of the 35th Annual ACM Symposium on Applied Computing, Brno, Czech Republic, 30 March 2020; pp. 2066–2073. [Google Scholar]
  87. Hassan, M.A.; Malik, A.S.; Fofi, D.; Karasfi, B.; Meriaudeau, F. Towards Health Monitoring Using Remote Heart Rate Measurement Using Digital Camera: A Feasibility Study. Measurement 2020, 149, 106804. [Google Scholar] [CrossRef] [PubMed]
  88. Sorelli, M.; Kopietz, C.; Zaunseder, S.; Bocchi, L. Pulse Decomposition Analysis in Camera-Based Photoplethysmography. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 3179–3182. [Google Scholar]
  89. Iozza, L.; Lázaro, J.; Cerina, L.; Silvestri, D.; Mainardi, L.; Laguna, P.; Gil, E. Monitoring Breathing Rate by Fusing the Physiological Impact of Respiration on Video-Photoplethysmogram with Head Movements. Physiol. Meas. 2019, 40, 094002. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  90. Paul, M.; Yu, X.; Wu, B.; Weiss, C.; Antink, C.H.; Blazek, V.; Leonhardt, S. Waveform Analysis for Camera-Based Photoplethysmography Imaging. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 2713–2718. [Google Scholar]
  91. Schrumpf, F.; Monch, C.; Bausch, G.; Fuchs, M. Exploiting Weak Head Movements for Camera-Based Respiration Detection. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 6059–6062. [Google Scholar]
  92. Barbieri, R.; Ficarelli, L.; Levi, R.; Negro, M.; Cerina, L.; Mainardi, L. Identification and Tracking of Physiological Parameters from Skin Using Video Photoplethysmography. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 6822–6825. [Google Scholar]
  93. Yu, X.; Paul, M.; Antink, C.H.; Venema, B.; Blazek, V.; Bollheimer, C.; Leonhardt, S.; Teichmann, D. Non-Contact Remote Measurement of Heart Rate Variability Using Near-Infrared Photoplethysmography Imaging. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 846–849. [Google Scholar]
  94. Cosar, S.; Yan, Z.; Zhao, F.; Lambrou, T.; Yue, S.; Bellotto, N. Thermal Camera Based Physiological Monitoring with an Assistive Robot. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5010–5013. [Google Scholar]
  95. Melchor Rodríguez, A.; Ramos-Castro, J. Video Pulse Rate Variability Analysis in Stationary and Motion Conditions. BioMed. Eng. OnLine 2018, 17, 11. [Google Scholar] [CrossRef] [Green Version]
  96. Wang, W.; Shan, C. Impact of Makeup on Remote-PPG Monitoring. Biomed. Phys. Eng. Express 2020, 6, 035004. [Google Scholar] [CrossRef]
  97. Moya-Albor, E.; Brieva, J.; Ponce, H.; Rivas-Scott, O.; Gomez-Pena, C. Heart Rate Estimation Using Hermite Transform Video Magnification and Deep Learning. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 2595–2598. [Google Scholar]
  98. Chen, Q.; Jiang, X.; Liu, X.; Lu, C.; Wang, L.; Chen, W. Non-Contact Heart Rate Monitoring in Neonatal Intensive Care Unit Using RGB Camera. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 5822–5825. [Google Scholar]
  99. Villarroel, M.; Chaichulee, S.; Jorge, J.; Davis, S.; Green, G.; Arteta, C.; Zisserman, A.; McCormick, K.; Watkinson, P.; Tarassenko, L. Non-Contact Physiological Monitoring of Preterm Infants in the Neonatal Intensive Care Unit. NPJ Digit. Med. 2019, 2, 128. [Google Scholar] [CrossRef] [Green Version]
  100. Lyra, S.; Mayer, L.; Ou, L.; Chen, D.; Timms, P.; Tay, A.; Chan, P.Y.; Ganse, B.; Leonhardt, S.; Hoog Antink, C. A Deep Learning-Based Camera Approach for Vital Sign Monitoring Using Thermography Images for ICU Patients. Sensors 2021, 21, 1495. [Google Scholar] [CrossRef]
  101. Wieler, M.E.; Murphy, T.G.; Blecherman, M.; Mehta, H.; Bender, G.J. Infant Heart-Rate Measurement and Oxygen Desaturation Detection with a Digital Video Camera Using Imaging Photoplethysmography. J. Perinatol. 2021, 41, 1725–1731. [Google Scholar] [CrossRef]
  102. Gibson, K.; Al-Naji, A.; Fleet, J.; Steen, M.; Esterman, A.; Chahl, J.; Huynh, J.; Morris, S. Non-Contact Heart and Respiratory Rate Monitoring of Preterm Infants Based on a Computer Vision System: A Method Comparison Study. Pediatr. Res. 2019, 86, 738–741. [Google Scholar] [CrossRef] [PubMed]
  103. Pavlidis, I.; Garza, I.; Tsiamyrtzis, P.; Dcosta, M.; Swanson, J.W.; Krouskop, T.; Levine, J.A. Dynamic Quantification of Migrainous Thermal Facial Patterns—A Pilot Study. IEEE J. Biomed. Health Inform. 2019, 23, 1225–1233. [Google Scholar] [CrossRef] [PubMed]
  104. Huynh, S.; Balan, R.K.; Ko, J.; Lee, Y. VitaMon: Measuring Heart Rate Variability Using Smartphone Front Camera. In Proceedings of the 17th Conference on Embedded Networked Sensor Systems, New York, NY, USA, 10–13 November 2019; pp. 1–14. [Google Scholar]
  105. Massaroni, C.; Lo Presti, D.; Formica, D.; Silvestri, S.; Schena, E. Non-Contact Monitoring of Breathing Pattern and Respiratory Rate via RGB Signal Measurement. Sensors 2019, 19, 2758. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  106. Negishi, T.; Sun, G.; Liu, H.; Sato, S.; Matsui, T.; Kirimoto, T. Stable Contactless Sensing of Vital Signs Using RGB-Thermal Image Fusion System with Facial Tracking for Infection Screening. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 4371–4374. [Google Scholar]
  107. Liu, X.; Yang, X.; Jin, J.; Li, J. Self-Adaptive Signal Separation for Non-Contact Heart Rate Estimation from Facial Video in Realistic Environments. Physiol. Meas. 2018, 39, 06NT01. [Google Scholar] [CrossRef]
  108. Cobos-Torres, J.-C.; Abderrahim, M.; Martínez-Orgado, J. Non-Contact, Simple Neonatal Monitoring by Photoplethysmography. Sensors 2018, 18, 4362. [Google Scholar] [CrossRef] [Green Version]
  109. Paul, M.; Karthik, S.; Joseph, J.; Sivaprakasam, M.; Kumutha, J.; Leonhardt, S.; Hoog Antink, C. Non-Contact Sensing of Neonatal Pulse Rate Using Camera-Based Imaging: A Clinical Feasibility Study. Physiol. Meas. 2020, 41, 024001. [Google Scholar] [CrossRef]
  110. Addison, P.S.; Jacquel, D.; Foo, D.M.H.; Borg, U.R. Video-Based Heart Rate Monitoring across a Range of Skin Pigmentations during an Acute Hypoxic Challenge. J. Clin. Monit. Comput. 2018, 32, 871–880. [Google Scholar] [CrossRef] [Green Version]
  111. Shirbani, F.; Hui, N.; Tan, I.; Butlin, M.; Avolio, A.P. Effect of Ambient Lighting and Skin Tone on Estimation of Heart Rate and Pulse Transit Time from Video Plethysmography. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Montreal, QC, Canada, 20–24 July 2020; pp. 2642–2645. [Google Scholar]
  112. Tang, C.; Lu, J.; Liu, J. Non-Contact Heart Rate Monitoring by Combining Convolutional Neural Network Skin Detection and Remote Photoplethysmography via a Low-Cost Camera. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1309–1315. [Google Scholar]
  113. Zhou, Y.; Ni, H.; Zhang, Q.; Wu, Q. The Noninvasive Blood Pressure Measurement Based on Facial Images Processing. IEEE Sens. J. 2019, 19, 10624–10634. [Google Scholar] [CrossRef]
  114. Liu, L.; Zhao, Y.; Kong, L.; Liu, M.; Dong, L.; Ma, F.; Pang, Z. Robust Real-Time Heart Rate Prediction for Multiple Subjects from Facial Video Using Compressive Tracking and Support Vector Machine. J. Med. Imag. 2018, 5, 024503. [Google Scholar] [CrossRef]
  115. McGinnis, R.S.; McGinnis, E.W.; Petrillo, C.; Ferri, J.; Scism, J.; Price, M. Validation of Smartphone Based Heart Rate Tracking for Remote Treatment of Panic Attacks. IEEE J. Biomed. Health Inform. 2021, 25, 656–662. [Google Scholar] [CrossRef]
  116. Shoushan, M.M.; Reyes, B.A.; Rodriguez, A.M.; Chong, J.W. Non-Contact HR Monitoring via Smartphone and Webcam During Different Respiratory Maneuvers and Body Movements. IEEE J. Biomed. Health Inform. 2021, 25, 602–612. [Google Scholar] [CrossRef] [PubMed]
  117. Bieber, G.; Antony, N.; Haescher, M. Touchless Heart Rate Recognition by Robots to Support Natural Human-Robot Communication. In Proceedings of the 11th PErvasive Technologies Related to Assistive Environments Conference, Corfu, Greece, 26 June 2018; pp. 415–420. [Google Scholar]
  118. Maki, Y.; Monno, Y.; Yoshizaki, K.; Tanaka, M.; Okutomi, M. Inter-Beat Interval Estimation from Facial Video Based on Reliability of BVP Signals. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 6525–6528. [Google Scholar]
  119. Takahashi, R.; Ogawa-Ochiai, K.; Tsumura, N. Non-Contact Method of Blood Pressure Estimation Using Only Facial Video. Artif. Life Robot. 2020, 25, 343–350. [Google Scholar] [CrossRef]
  120. McDuff, D.J.; Blackford, E.B.; Estepp, J.R. Fusing Partial Camera Signals for Noncontact Pulse Rate Variability Measurement. IEEE Trans. Biomed. Eng. 2018, 65, 1725–1739. [Google Scholar] [CrossRef] [PubMed]
  121. Perepelkina, O.; Artemyev, M.; Churikova, M.; Grinenko, M. HeartTrack: Convolutional Neural Network for Remote Video-Based Heart Rate Monitoring. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1163–1171. [Google Scholar]
  122. Gonzalez Viejo, C.; Fuentes, S.; Torrico, D.; Dunshea, F. Non-Contact Heart Rate and Blood Pressure Estimations from Video Analysis and Machine Learning Modelling Applied to Food Sensory Responses: A Case Study for Chocolate. Sensors 2018, 18, 1802. [Google Scholar] [CrossRef] [Green Version]
  123. Yang, D.; Xiao, G.; Wei, J.; Luo, H. Preliminary Assessment of Video-Based Blood Pressure Measurement According to ANSI/AAMI/ISO81060-2:2013 Guideline Accuracy Criteria: Anura Smartphone App with Transdermal Optimal Imaging Technology. Blood Press. Monit. 2020, 25, 295–298. [Google Scholar] [CrossRef]
  124. Laurie, J.; Higgins, N.; Peynot, T.; Fawcett, L.; Roberts, J. An Evaluation of a Video Magnification-Based System for Respiratory Rate Monitoring in an Acute Mental Health Setting. Int. J. Med. Inform. 2021, 148, 104378. [Google Scholar] [CrossRef]
  125. Kwasniewska, A.; Szankin, M.; Ruminski, J.; Kaczmarek, M. Evaluating Accuracy of Respiratory Rate Estimation from Super Resolved Thermal Imagery. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 2744–2747. [Google Scholar]
  126. Oiwa, K.; Bando, S.; Nozawa, A. Contactless Blood Pressure Sensing Using Facial Visible and Thermal Images. Artif. Life Robot. 2018, 23, 387–394. [Google Scholar] [CrossRef]
  127. Dagdanpurev, S.; Sun, G.; Choimaa, L.; Abe, S.; Matsui, T. Clinical Application of Multiple Vital Signs-Based Infection Screening System in a Mongolian Hospital: Optimization of Facial Temperature Measurement by Thermography at Various Ambient Temperature Conditions Using Linear Regression Analysis. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5313–5316. [Google Scholar]
  128. Kwon, H.-M.; Ikeda, K.; Kim, S.-H.; Thiele, R.H. Non-Contact Thermography-Based Respiratory Rate Monitoring in a Post-Anesthetic Care Unit. J. Clin. Monit. Comput. 2020, 35, 1291–1297. [Google Scholar] [CrossRef]
  129. Xie, K.; Fu, C.-H.; Liang, H.; Hong, H.; Zhu, X. Non-Contact Heart Rate Monitoring for Intensive Exercise Based on Singular Spectrum Analysis. In Proceedings of the 2019 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), San Jose, CA, USA, 28–30 March 2019; pp. 228–233. [Google Scholar]
  130. Prakash, S.K.A.; Tucker, C.S. Bounded Kalman Filter Method for Motion-Robust, Non-Contact Heart Rate Estimation. Biomed. Opt. Express 2018, 9, 873. [Google Scholar] [CrossRef]
  131. Negishi, T.; Abe, S.; Matsui, T.; Liu, H.; Kurosawa, M.; Kirimoto, T.; Sun, G. Contactless Vital Signs Measurement System Using RGB-Thermal Image Sensors and Its Clinical Screening Test on Patients with Seasonal Influenza. Sensors 2020, 20, 2171. [Google Scholar] [CrossRef] [Green Version]
  132. Chuang, Y.T.; Lai, S.J.; Chang, T.F.; Lin, Y.H. Image-Based Skin Temperature and Pulse Rate Measuring System. In Proceedings of the 2020 IEEE International Conference on Consumer Electronics—Taiwan (ICCE-Taiwan), Taoyuan, Taiwan, 28–30 September 2020; pp. 1–2. [Google Scholar]
  133. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A Multimodal Database for Affect Recognition and Implicit Tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef] [Green Version]
  134. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  135. Bobbia, S.; Macwan, R.; Benezeth, Y.; Mansouri, A.; Dubois, J. Unsupervised Skin Tissue Segmentation for Remote Photoplethysmography. Pattern Recognit. Lett. 2019, 124, 82–90. [Google Scholar] [CrossRef]
  136. Stricker, R.; Muller, S.; Gross, H.-M. Non-Contact Video-Based Pulse Rate Measurement on a Mobile Service Robot. In Proceedings of the the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 1056–1062. [Google Scholar]
  137. Niu, X.; Han, H.; Shan, S.; Chen, X. VIPL-HR: A Multi-Modal Database for Pulse Estimation from Less-Constrained Face Video. arXiv 2018, arXiv:1810.04927. [Google Scholar]
  138. Heusch, G.; Anjos, A.; Marcel, S. A Reproducible Study on Remote Heart Rate Measurement. arXiv 2017, arXiv:1709.00962. [Google Scholar]
  139. HCI Tagging Database—Home. Available online: https://mahnob-db.eu/hci-tagging (accessed on 8 May 2022).
  140. DEAP: A Dataset for Emotion Analysis Using Physiological and Audiovisual Signals. Available online: https://www.eecs.qmul.ac.uk/mmv/datasets/deap/ (accessed on 8 May 2022).
  141. FAVIP Database. Available online: https://spire.ee.iisc.ac.in/spire/database.php (accessed on 23 May 2022).
  142. Ryu, J.; Hong, S.; Liang, S.; Pak, S.; Chen, Q.; Yan, S. Research on the Combination of Color Channels in Heart Rate Measurement Based on Photoplethysmography Imaging. J. Biomed. Opt. 2021, 26, 025003. [Google Scholar] [CrossRef]
  143. Benezeth, Y. UBFCrPPG. Available online: https://sites.google.com/view/ybenezeth/ubfcrppg (accessed on 8 May 2022).
  144. Pulse Rate Detection Dataset-PURE. Available online: https://www.tu-ilmenau.de/en/university/departments/department-of-computer-science-and-automation/profile/institutes-and-groups/institute-of-computer-and-systems-engineering/group-for-neuroinformatics-and-cognitive-robotics/data-sets-code/pulse-rate-detection-dataset-pure (accessed on 8 May 2022).
  145. Hsu, G.-S.; Ambikapathi, A.; Chen, M.-S. Deep Learning with Time-Frequency Representation for Pulse Estimation from Facial Videos. In Proceedings of the 2017 IEEE International Joint Conference on Biometrics (IJCB), Denver, CO, USA, 1–4 October 2017; pp. 383–389. [Google Scholar]
  146. VIPL-HR Database. Available online: https://vipl.ict.ac.cn/zygx/sjk/201811/t20181129_32716.html (accessed on 8 May 2022).
  147. COHFACE | Zenodo. Available online: https://zenodo.org/record/4081054#.YosWt1RBy3A (accessed on 23 May 2022).
  148. Technology-MMSE-HR Dataset (Multimodal Spontaneous Expression-Heart Rate Dataset). Available online: https://binghamton.technologypublisher.com/tech/MMSE-HR_dataset_(Multimodal_Spontaneous_Expression-Heart_Rate_dataset) (accessed on 8 May 2022).
  149. TokyoTech Remote PPG Dataset. Available online: http://www.ok.sc.e.titech.ac.jp/res/VitalSensing/remoteIBI/Dataset.html (accessed on 8 May 2022).
  150. MR-NIRP Database. Available online: https://computationalimaging.rice.edu/mr-nirp-dataset/ (accessed on 8 May 2022).
  151. Ronca, V.; Giorgi, A.; Rossi, D.; Di Florio, A.; Di Flumeri, G.; Aricò, P.; Sciaraffa, N.; Vozzi, A.; Tamborra, L.; Simonetti, I.; et al. A Video-Based Technique for Heart Rate and Eye Blinks Rate Estimation: A Potential Solution for Telemonitoring and Remote Healthcare. Sensors 2021, 21, 1607. [Google Scholar] [CrossRef]
  152. Hochhausen, N.; Barbosa Pereira, C.; Leonhardt, S.; Rossaint, R.; Czaplik, M. Estimating Respiratory Rate in Post-Anesthesia Care Unit Patients Using Infrared Thermography: An Observational Study. Sensors 2018, 18, 1618. [Google Scholar] [CrossRef] [Green Version]
  153. Lee, K.; Park, D.; Cho, C.; Ko, H. Robust Remote Heart Rate Estimation in Car Driving Environment. In Proceedings of the 2018 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 12–14 January 2018; pp. 1–2. [Google Scholar]
  154. Zheng, K.; Ci, K.; Cui, J.; Kong, J.; Zhou, J. Non-Contact Heart Rate Detection When Face Information Is Missing during Online Learning. Sensors 2020, 20, 7021. [Google Scholar] [CrossRef]
  155. Selvaraju, V.; Spicher, N.; Swaminathan, R.; Deserno, T.M. Face Detection from In-Car Video for Continuous Health Monitoring. In Proceedings of the SPIE Medical Imaging: Imaging Informatics for Healthcare, Research, and Applications, San Diego, CA, USA, 4 April 2022; Volume 12037, pp. 49–56. [Google Scholar]
  156. Zhang, Q.; Zhou, Y.; Song, S.; Liang, G.; Ni, H. Heart Rate Extraction Based on Near-Infrared Camera: Towards Driver State Monitoring. IEEE Access 2018, 6, 33076–33087. [Google Scholar] [CrossRef]
  157. Lee, K.; Lee, J.; Ha, C.; Han, M.; Ko, H. Video-Based Contactless Heart-Rate Detection and Counting via Joint Blind Source Separation with Adaptive Noise Canceller. Appl. Sci. 2019, 9, 4349. [Google Scholar] [CrossRef] [Green Version]
  158. Cai, K.; Chen, J.; Zhang, L.; Lin, Q.; Ji, H. A Passive Heart Rate Measurement Method Using Camera. In Proceedings of the 2018 International Conference on Computing and Pattern Recognition—ICCPR ’18, Shenzhen, China, 23–25 June 2018; pp. 68–72. [Google Scholar]
  159. Fujita, Y.; Hiromoto, M.; Sato, T. Fast And Robust Heart Rate Estimation From Videos Through Dynamic Region Selection. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 3024–3027. [Google Scholar]
  160. Le, D.Q.; Lie, W.-N.; Nhu, Q.N.Q.; Nguyen, T.T.A. Heart Rate Estimation Based on Facial Image Sequence. In Proceedings of the 2020 5th International Conference on Green Technology and Sustainable Development (GTSD), Ho Chi Minh City, Vietnam, 27–28 November 2020; pp. 449–453. [Google Scholar]
  161. Zhu, M.; He, Y.; He, Q. A Review of Researches on Deep Learning in Remote Sensing Application. IJG 2019, 10, 1–11. [Google Scholar] [CrossRef] [Green Version]
  162. Gutstein, S. Remote Photoplethysmography for Cardiovascular Monitoring; AD1080252; Defense Technical Information Center: Fort Belvoir, VA, USA, 2019; pp. 1–37. [Google Scholar]
  163. Yu, Z.; Peng, W.; Li, X.; Hong, X.; Zhao, G. Remote Heart Rate Measurement from Highly Compressed Facial Videos: An End-to-End Deep Learning Solution with Video Enhancement. arXiv 2019, arXiv:1907.11921. [Google Scholar]
  164. Pai, A.; Veeraraghavan, A.; Sabharwal, A. HRVCam: Robust Camera-Based Measurement of Heart Rate Variability. J. Biomed. Opt. 2021, 26, 022707. [Google Scholar] [CrossRef] [PubMed]
  165. Li, F.; Zhao, Y.; Kong, L.; Dong, L.; Liu, M.; Hui, M.; Liu, X. A Camera-Based Ballistocardiogram Heart Rate Measurement Method. Rev. Sci. Instrum. 2020, 91, 054105. [Google Scholar] [CrossRef] [PubMed]
  166. Chen, L.; Liu, N.; Hu, M.; Zhai, G. RGB-Thermal Imaging System Collaborated with Marker Tracking for Remote Breathing Rate Measurement. In Proceedings of the 2019 IEEE Visual Communications and Image Processing (VCIP), Sydney, Australia, 1–4 December 2019; pp. 1–4. [Google Scholar]
  167. Wu, H.; Wang, T.; Dai, T.; Lin, Y.; Wang, Y. A Real-Time Vision-Based Heart Rate Measurement Framework for Home Nursing Assistance. In Proceedings of the 3rd International Conference on Crowd Science and Engineering, Singapore, 28–31 July 2018; pp. 1–7. [Google Scholar]
  168. Huang, P.-W.; Wu, B.-J.; Wu, B.-F. A Heart Rate Monitoring Framework for Real-World Drivers Using Remote Photoplethysmography. IEEE J. Biomed. Health Inform. 2021, 25, 1397–1408. [Google Scholar] [CrossRef] [PubMed]
  169. Nageotte, M.P. Fetal Heart Rate Monitoring. Semin. Fetal Neonatal Med. 2015, 20, 144–148. [Google Scholar] [CrossRef]
  170. Hill, B.; Annesley, S.H. Monitoring Respiratory Rate in Adults. Br. J. Nurs. 2020, 29, 12–16. [Google Scholar] [CrossRef]
  171. Sylvia S, M. Understanding Human Anatomy & Physiology, 5th ed.; WCB/McGraw-Hill: New York, NY, USA, 2004. [Google Scholar]
  172. Luo, H.; Yang, D.; Barszczyk, A.; Vempala, N.; Wei, J.; Wu, S.J.; Zheng, P.P.; Fu, G.; Lee, K.; Feng, Z.-P. Smartphone-Based Blood Pressure Measurement Using Transdermal Optical Imaging Technology. Circ. Cardiovasc. Imaging 2019, 12, e008857. [Google Scholar] [CrossRef] [Green Version]
  173. Okada, G.; Mitsuhashi, R.; Kagawa, K.; Kawahito, S.; Tsumura, N. Noncontact Heart Rate Measurement Using a High-Sensitivity Camera in a Low-Light Environment. Artif. Life Robot. 2019, 24, 6–11. [Google Scholar] [CrossRef]
  174. Qi, L.; Yu, H.; Xu, L.; Mpanda, R.S.; Greenwald, S.E. Robust Heart-Rate Estimation from Facial Videos Using Project_ICA. Physiol. Meas. 2019, 40, 085007. [Google Scholar] [CrossRef] [PubMed]
  175. Lewandowska, M.; Nowak, J. Measuring Pulse Rate with a Webcam. J. Med. Imaging Health Inform. 2012, 2, 87–92. [Google Scholar] [CrossRef]
  176. Poh, M.-Z.; McDuff, D.J.; Picard, R.W. Advancements in Noncontact, Multiparameter Physiological Measurements Using a Webcam. IEEE Trans. Biomed. Eng. 2011, 58, 7–11. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  177. Poh, M.-Z.; McDuff, D.J.; Picard, R.W. Non-Contact, Automated Cardiac Pulse Measurements Using Video Imaging and Blind Source Separation. Opt. Express 2010, 18, 10762. [Google Scholar] [CrossRef]
  178. Gauci, L.; Falzon, O.; Camilleri, K.P. PCA-Driven Detection and Enhancement of Microchanges in Video Data Associated with Heart Rate. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 3892–3895. [Google Scholar]
  179. Boccignone, G.; Conte, D.; Cuculo, V.; D’Amelio, A.; Grossi, G.; Lanzarotti, R. An Open Framework for Remote-PPG Methods and Their Assessment. IEEE Access 2020, 8, 216083–216103. [Google Scholar] [CrossRef]
  180. Wang, W.; Stuijk, S.; de Haan, G. A Novel Algorithm for Remote Photoplethysmography: Spatial Subspace Rotation. IEEE Trans. Biomed. Eng. 2016, 63, 1974–1984. [Google Scholar] [CrossRef]
  181. Zheng, K.; Ci, K.; Li, H.; Shao, L.; Sun, G.; Liu, J.; Cui, J. Heart Rate Prediction from Facial Video with Masks Using Eye Location and Corrected by Convolutional Neural Networks. Biomed. Signal Processing Control 2022, 75, 103609. [Google Scholar] [CrossRef]
  182. Huang, B.; Chang, C.-M.; Lin, C.-L.; Chen, W.; Juang, C.-F.; Wu, X. Visual Heart Rate Estimation from Facial Video Based on CNN. In Proceedings of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November 2020; pp. 1658–1662. [Google Scholar]
  183. Task Force of the European Society of Cardiology the North American Society of Pacing Electrophysiology. Heart Rate Variability: Standards of Measurement, Physiological Interpretation, and Clinical Use. Circulation 1996, 93, 1043–1065. [Google Scholar] [CrossRef] [Green Version]
  184. Becker, C.; Achermann, S.; Rocque, M.; Kirenko, I.; Schlack, A.; Dreher-Hummel, T.; Zumbrunn, T.; Bingisser, R.; Nickel, C.H. Camera-Based Measurement of Respiratory Rates Is Reliable. Eur. J. Emerg. Med. 2018, 25, 416–422. [Google Scholar] [CrossRef]
  185. Jakkaew, P.; Onoye, T. An Approach to Non-Contact Monitoring of Respiratory Rate and Breathing Pattern Based on Slow Motion Images. In Proceedings of the 2019 IEEE International Conference on Consumer Electronics—Asia (ICCE-Asia), Bangkok, Thailand, 12–14 June 2019; pp. 47–51. [Google Scholar]
  186. Sugita, N.; Yoshizawa, M.; Abe, M.; Tanaka, A.; Homma, N.; Yambe, T. Contactless Technique for Measuring Blood-Pressure Variability from One Region in Video Plethysmography. J. Med. Biol. Eng. 2019, 39, 76–85. [Google Scholar] [CrossRef]
  187. Kurihara, K.; Sugimura, D.; Hamamoto, T. Adaptive Fusion of RGB/NIR Signals Based on Face/Background Cross-Spectral Analysis for Heart Rate Estimation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 4534–4538. [Google Scholar]
  188. Zhou, L.; Yin, M.; Xu, X.; Yuan, X.; Liu, X. Non-Contact Detection of Human Heart Rate with Kinect. Clust. Comput 2019, 22, 8199–8206. [Google Scholar] [CrossRef]
  189. Alonso-Cuevas, J.L.; Sanchez-Yanez, R.E. Simplified Heart Rate Measurement from Video. In Proceedings of the 3rd International Workshop on Interactive and Spatial Computing, Richardson, TX, USA, 12–13 April 2018; pp. 23–28. [Google Scholar]
  190. Zhu, Q.; Wong, C.-W.; Fu, C.-H.; Wu, M. Fitness Heart Rate Measurement Using Face Videos. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 2000–2004. [Google Scholar]
  191. Pursche, T.; Claus, R.; Tibken, B.; Moller, R. Using the Hilbert-Huang Transform to Increase the Robustness of Video Based Remote Heart-Rate Measurement from Human Faces. In Proceedings of the 2018 IEEE 8th International Conference on Consumer Electronics—Berlin (ICCE-Berlin), Berlin, Germany, 2–5 September 2018; pp. 1–4. [Google Scholar]
  192. Bastardot, F.; Marques-Vidal, P.; Vollenweider, P. Association of Body Temperature with Obesity. The CoLaus Study. Int. J. Obes. 2019, 43, 1026–1033. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  193. Maharatna, K.; Mazomenos, E.B.; Morgan, J.; Bonfiglio, S. Towards the Development of Next-Generation Remote Healthcare System: Some Practical Considerations. In Proceedings of the 2012 IEEE International Symposium on Circuits and Systems, Seoul, Korea, 20–23 May 2012; pp. 1–4. [Google Scholar]
  194. Talmon, J.; Ammenwerth, E.; Brender, J.; Dekeizer, N.; Nykanen, P.; Rigby, M. STARE-HI—Statement on Reporting of Evaluation Studies in Health Informatics. Int. J. Med. Inform. 2009, 78, 1–9. [Google Scholar] [CrossRef] [PubMed]
  195. Meratwal, M.; Spicher, N.; Deserno, T.M. Multi-Camera and Multi-Person Indoor Activity Recognition for Continuous Health Monitoring Using Long Short Term Memory. In Proceedings of the Medical Imaging 2022: Imaging Informatics for Healthcare, Research, and Applications, San Diego, CA, USA, 4 April 2022; Volume 12037, pp. 64–71. [Google Scholar]
  196. Xu, K. Silicon Electro-Optic Micro-Modulator Fabricated in Standard CMOS Technology as Components for All Silicon Monolithic Integrated Optoelectronic Systems. J. Micromech. Microeng. 2021, 31, 054001. [Google Scholar] [CrossRef]
  197. McDuff, D.; Hurter, C. InPhysible: Camouflage Against Video-Based Physiological Measurement. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 17–21 July 2018; pp. 5784–5789. [Google Scholar]
  198. Kalmeda. Die Mobile Tinnitus-Therapie Auf Rezept. Available online: https://www.kalmeda.de (accessed on 10 May 2022).
Figure 1. PRISMA diagram of literature screening.
Figure 1. PRISMA diagram of literature screening.
Sensors 22 04097 g001
Figure 2. General flow diagram of vital sign measurements from video/image data.
Figure 2. General flow diagram of vital sign measurements from video/image data.
Sensors 22 04097 g002
Figure 3. Data acquisition according to study design, hardware, and ground truth.
Figure 3. Data acquisition according to study design, hardware, and ground truth.
Sensors 22 04097 g003
Figure 4. Distribution of subject information including obtained data, number of subjects, participant type and available (A) and not available (NA) info of skin tone.
Figure 4. Distribution of subject information including obtained data, number of subjects, participant type and available (A) and not available (NA) info of skin tone.
Sensors 22 04097 g004
Figure 5. Distribution of hardware parameters, namely camera type, frame rate, resolution (r), and camera-subject distance.
Figure 5. Distribution of hardware parameters, namely camera type, frame rate, resolution (r), and camera-subject distance.
Sensors 22 04097 g005
Figure 6. Chord diagram of conceptual semantics between camera and vital sign.
Figure 6. Chord diagram of conceptual semantics between camera and vital sign.
Sensors 22 04097 g006
Figure 7. Distribution of ground truth.
Figure 7. Distribution of ground truth.
Sensors 22 04097 g007
Figure 8. Comprehensive work flow from various camera images to HR vital sign (Note: FF: full frame, VJ: Viola Jones algorithm, HoG: Histogram of oriented gradients; CNN: convolutional neural network; ManualD: manual ROI detection, NoD: no ROI detection, OtherD: Other ROI detection, ManualT: manual ROI tracking, KLT: Kanade–Lucas–Tomasi; KCF: kernel correlation filter; featT: feature tracking, OtherT: other ROI tracking, NoT: no ROI tracking, BSS: blind source separation algorithms, otherSP: other signal processing, ML: Machine learning FD: frequency domain, TD: time domain).
Figure 8. Comprehensive work flow from various camera images to HR vital sign (Note: FF: full frame, VJ: Viola Jones algorithm, HoG: Histogram of oriented gradients; CNN: convolutional neural network; ManualD: manual ROI detection, NoD: no ROI detection, OtherD: Other ROI detection, ManualT: manual ROI tracking, KLT: Kanade–Lucas–Tomasi; KCF: kernel correlation filter; featT: feature tracking, OtherT: other ROI tracking, NoT: no ROI tracking, BSS: blind source separation algorithms, otherSP: other signal processing, ML: Machine learning FD: frequency domain, TD: time domain).
Sensors 22 04097 g008
Figure 9. (a) RMSE of HR in static and dynamic conditions and, (b) varying distance between the subject and camera.
Figure 9. (a) RMSE of HR in static and dynamic conditions and, (b) varying distance between the subject and camera.
Sensors 22 04097 g009
Figure 10. RMSE of RR estimation with varying distance.
Figure 10. RMSE of RR estimation with varying distance.
Sensors 22 04097 g010
Figure 11. Factors affecting the vital signs performance.
Figure 11. Factors affecting the vital signs performance.
Sensors 22 04097 g011
Figure 12. The framework of main research gaps and future directions. In this, red color boxes represent the research gaps, and green color boxes refer to future research directions.
Figure 12. The framework of main research gaps and future directions. In this, red color boxes represent the research gaps, and green color boxes refer to future research directions.
Sensors 22 04097 g012
Table 1. Number n of articles from the considered years: January 2018 to April 2021.
Table 1. Number n of articles from the considered years: January 2018 to April 2021.
Article YearHRRRBPBSTSpO2MultipleTotal
2018252110736
2019149300329
2020165210327
202154001212
Total602062115104
Table 2. Existing databases.
Table 2. Existing databases.
DatabaseNo. of SubjectsCamera TypeCamera DetailFrame Rate
(fps)
Resolution
(px × px)
Ground Truth
MAHNOB-HCI
[133,139]
27RGBAllied Vision Stingray
F-046C; F-046B
60780 × 580ECG
DEAP
[134,140]
22RGBSony DCR-HC27E50720 × 576PPG, Respiration, EEG, EOG, EMG, GSR, BT
FAVIP
[83,141]
15RGBSamsung galaxy S3 and
iPhone 3GS
301280 × 720Pulse oximeter
UBFC-RPPG [135,142,143]42RGBLogitech C920 HD pro30640 × 480Pulse oximeter
PURE [136,144]10RGBevo274CVGE30640 × 480Finger pulse oximeter
Pulse from face
[78,145]
13RGBNikon D5300 camera501280 × 720Two Mio Alpha II wrist heart rate monitors
VIPL_HR
[137,146]
107RGBLogitech C31025960 × 720CONTEC CMS60C blood volume pulse recorder
NIRRealsense F20030640 × 480
RGB301920 × 1080
RGBHUAWEI P9 smart phone301920 × 1080
COHFACE
[138,147]
40RGBLogitech HD C52520640 × 480Blood volume pulse sensor, respiratory belt
MMSE-HR
[80,148]
40RGB, IRDi3D dynamic imaging system, FLIR A655sc251040 × 1392Biopac MP150 system—BP, HR
50640 × 480
TokyoTech Remote PPG [118,149]9RGB, NIRPrototype RGB-NIR camera300640 × 480Contact PPG sensor
MR-NIRP
[41,150]
18RGB, NIRFLIR Grasshopper3, Point Grey Grasshopper30640 × 640Finger pulse oximeter
Table 4. Preprocessing using different filtering methods.
Table 4. Preprocessing using different filtering methods.
FilteringDescription
Detrending filter
[89,92,99,113,130,153,154,160]
It removes the trend in signal
Moving average filter [29,55,116,125,126,132,154,166]It smooths a signal and suppresses high frequency noise
Band-pass filter [52,75,80,95,98,109,116,122,126,132,152,167]It eliminates the frequency components outside the bandwidth range
Table 5. Various algorithms for HR extraction.
Table 5. Various algorithms for HR extraction.
Signal Processing TechniquesCharacterization
ICA [176,177]It decomposes the signal and extracts independent components of pulse information from temporal RGB traces.
PCA [51,53,175,178]It utilizes a statistical technique to obtain uncorrelated components from RGB traces [151].
GREEN [38,104,130]In blood, hemoglobin and oxyhemoglobin absorb light of 520–580 nm, which is in the range of the camera’s green filter [38,104,130]. Hemoglobin absorbs green light in sufficient depth [81]. Therefore, the green channel is reported to have more information compared to blue or red channels [179].
CHROM [162]A linear combination of chrominance signals with the assumption of skin color necessitates a priori knowledge and eliminates motion artifacts but it may fail if pulse and specular signals are same [59].
POS [116]It projects the RGB-derived signals onto a plane orthogonal to the temporally normalized skin tone component.
Spatial subspace rotation [162,180]It utilizes the subspace of skin pixels and rotation measurements for extracting cardiac pulse information but it may require complete continuous sequence of camera frames to recover the pulse wave [120].
Kernel density ICA [59]It does not require a prior assumption of probability distributions of hidden sources and so-called semi-BSS method.
Table 6. Studies using RR by utilizing camera-based approaches.
Table 6. Studies using RR by utilizing camera-based approaches.
Ref.Camera TypeCamera DetailsFrame Rate (fps)Resolution (px × px)ROIGround TruthResults
[152]FIRInfratec VarioCAM HD head301024 × 768NosePhilips IntelliVue MP30 monitorCorrelation coefficient (CC): 0.607 upon arrival, 0.849 upon discharge
[29]FIROptrics PI 45080382 × 288NoseManual countingCC: near distance: 0.960; far distance: 0.508;
[27]FIRInfraTec VarioCAM HD head301024 × 768Full frame and split into sub-ROIAdults: Respiratory belt, Infants: Dräger M540 patient monitorRoot mean square error (RMSE): healthy adults: (sit still: 0.31 ± 0.09 cpm, stimulated breathing: 3.27 ± 0.72 cpm), infants: 4.15 ± 1.44 cpm
[125]FIRFLIR SC300030320 × 240Nostril area and mouth, nose and cheeks enclosedSubject finger flexion (upward–inhalation, downward–exhalation)RMSE: 3.40 cpm
[94]FIROptrics PI-45027382 × 288NoseManualRMSE: stay still: 3.81 cpm, moving: 6.20 cpm
[25]FIRFLIR Lepton 2.5,
FLIR lepton 3.5
8.760 × 80;
120 × 160
Full framePhilips MX700 patient monitorMean absolute error (MAE): 2.07 cpm
[32]FIRSeek Thermal Compact PRO for iPhone17640 × 480Highest temperature point and around itRespiration beltRMSE: 1.82 ± 0.75 cpm
[128]FIRFLIR T450sc30-NoseGE healthcare patient monitor, visual inspectionCC: 0.95
[57]FIRInfratec ImageIR 9300501024 × 768Nosepiezo plethysmography,
IntelliVue MP70 patient monitor
RMSE: 0.71 ± 0.30 cpm
[31]FIRFLIR T-42010320 × 240NostrilRespiratory volume monitorCC: 0.86 before sedation
[131]RGB;
FIR
Dual camera DFK23U618;
FLIR A315
15640 × 480;
320 × 240
Nose and mouthRespiration effort beltCC: 0.87; RMSE: 1.73 cpm
[106]RGB;
FIR
Dual camera DFK23U618;
FLIR A315
15640 × 480;
320 × 240
Nasal areaRespiratory effort beltRMSE: standing: 1.44 cpm, seated position with body movement: 2.52 cpm
[166]RGB, FIRMAG62 thermal imager10640 × 480Nostril regionSleep respiratory monitorCoefficient of determination: 0.905
[60]NIR;
FIR
NIR: see3cam_CU40,
FIR: FLIR lepton version 3.5
15; 8.7336 × 190, 160 × 120Chest, NostrilRespiratory beltRMSE: 4.44 cpm
[26]RGB;
FIR
RGB: IDS UI-2220SE;
FIR: FLIR Lepton 2.5
20; 8.7576 × 768;
60 × 80
Full framePhilips patient monitorMAE: 5.36 cpm
[124]RGBPoint Grey Flea 3 GigE-648 × 488ChestPolysomnographyMean error: non magnified: 0.874 cpm; magnified: 0.67 cpm
[24]RGBIP camera10320 × 180Full frameECG impedance pneumographyCC: 0.948; RMSE: 6.36 cpm
[184]RGBIDS uEye-222020--TorsoCapnographyAll clothing styles and respiratory patters (CC: 0.90–1.00) except winter coat-slow-deep scenario (CC:0.84)
[108]RGBDigital camera24/301920 × 1080 /1280 × 720Abdominal areaDräguer NICU monitorCC: 0.86
[91]RGBIDS UI-3160CP1201920 × 1080FaceUpper chest signalError: −0.25 to 0.5 cpm
[53]NIRPoint Grey Firefly MV USB 2.030640 × 480Full framePSG, ECG, Inductance plethysmographyCC: 0.80; RMSE: 2.10 ± 1.64 cpm; MAE: 0.82 ± 0.89 cpm
[102]RGBNikon D610, D5300301920 × 1080Abdominal areaPhilips intellivue monitorLimits of agreement: −22 to 23.6 cpm
[105]RGBCCD camera301280 × 720Jugular notchDifferential digital pressure sensorMAE: 0.39 cpm; Limits of agreements: (slim fit: ±0.98 cpm, loose fit: ±1.07 cpm)
[43]RGBSmartphone LG G230-ForeheadVisual inspection, Pulse oximeter, Heart rate monitorRMSE: hue: 3.88 cpm; Green: 5.68 cpm
[89]RGBLogitech C922/
GigE Sony XCG-C30C
601280 × 720/ 659 × 494Forehead, nose, cheeksRespiratory beltRelative error < 2% and inter quartile range < 5%
[51]NIRMonochromatic infrared camera62640 × 240Neck area with chin and upper chestChest beltCC: 0.99; RMSE: 0.70 cpm
[99]RGBJAI 3-CCD AT-200CL201620 × 1236SkinPhilips patient monitorMAE: 3.5 cpm
[52]NIRThermal imager MAG62, Avigilon H4 HD Dome-640 × 480ChestManualCoefficient of determination: dataset 1: 0.92, dataset 2: 0.87
[185]RGBSmartphone Galaxy S9+2401920 × 1080Abdomen or waist areaManual countingAccuracy 99.09%
[56]RGBCanon camera--Cheeks enclosedPPG sensor, Respiratory beltRMSE: 2.16 cpm
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Selvaraju, V.; Spicher, N.; Wang, J.; Ganapathy, N.; Warnecke, J.M.; Leonhardt, S.; Swaminathan, R.; Deserno, T.M. Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review. Sensors 2022, 22, 4097. https://doi.org/10.3390/s22114097

AMA Style

Selvaraju V, Spicher N, Wang J, Ganapathy N, Warnecke JM, Leonhardt S, Swaminathan R, Deserno TM. Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review. Sensors. 2022; 22(11):4097. https://doi.org/10.3390/s22114097

Chicago/Turabian Style

Selvaraju, Vinothini, Nicolai Spicher, Ju Wang, Nagarajan Ganapathy, Joana M. Warnecke, Steffen Leonhardt, Ramakrishnan Swaminathan, and Thomas M. Deserno. 2022. "Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review" Sensors 22, no. 11: 4097. https://doi.org/10.3390/s22114097

APA Style

Selvaraju, V., Spicher, N., Wang, J., Ganapathy, N., Warnecke, J. M., Leonhardt, S., Swaminathan, R., & Deserno, T. M. (2022). Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review. Sensors, 22(11), 4097. https://doi.org/10.3390/s22114097

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop