Next Article in Journal
Comparison of Different Cure Monitoring Techniques
Next Article in Special Issue
Mediator Microbial Biosensor Analyzers for Rapid Determination of Surface Water Toxicity
Previous Article in Journal
Exploiting an Intermediate Latent Space between Photo and Sketch for Face Photo-Sketch Recognition
Previous Article in Special Issue
Using Deep Learning to Predict Minimum Foot–Ground Clearance Event from Toe-Off Kinematics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Drivers’ Mental Workload Levels: Comparison of Machine Learning Methods Based on ECG and Infrared Thermal Signals

1
Department of Engineering and Geology, University G. d’Annunzio of Chieti-Pescara, 65127 Pescara, Italy
2
Department of Neurosciences, Imaging and Clinical Sciences, University G. d’Annunzio of Chieti-Pescara, 66100 Chieti, Italy
3
Next2U s.r.l., 65127 Pescara, Italy
4
Physiology Section, Department of Biomedical and Biotechnological Sciences, University of Catania, 95123 Catania, Italy
5
National Interuniversity Consortium of Science and Technology of Materials (INSTM), University of Messina, 98122 Messina, Italy
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(19), 7300; https://doi.org/10.3390/s22197300
Submission received: 31 August 2022 / Revised: 19 September 2022 / Accepted: 22 September 2022 / Published: 26 September 2022
(This article belongs to the Special Issue Feature Papers in Biosensors Section 2022)

Abstract

:
Mental workload (MW) represents the amount of brain resources required to perform concurrent tasks. The evaluation of MW is of paramount importance for Advanced Driver-Assistance Systems, given its correlation with traffic accidents risk. In the present research, two cognitive tests (Digit Span Test—DST and Ray Auditory Verbal Learning Test—RAVLT) were administered to participants while driving in a simulated environment. The tests were chosen to investigate the drivers’ response to predefined levels of cognitive load to categorize the classes of MW. Infrared (IR) thermal imaging concurrently with heart rate variability (HRV) were used to obtain features related to the psychophysiology of the subjects, in order to feed machine learning (ML) classifiers. Six categories of models have been compared basing on unimodal IR/unimodal HRV/multimodal IR + HRV features. The best classifier performances were reached by the multimodal IR + HRV features-based classifiers (DST: accuracy = 73.1%, sensitivity = 0.71, specificity = 0.69; RAVLT: accuracy = 75.0%, average sensitivity = 0.75, average specificity = 0.87). The unimodal IR features based classifiers revealed high performances as well (DST: accuracy = 73.1%, sensitivity = 0.73, specificity = 0.73; RAVLT: accuracy = 71.1%, average sensitivity = 0.71, average specificity = 0.85). These results demonstrated the possibility to assess drivers’ MW levels with high accuracy, also using a completely non-contact and non-invasive technique alone, representing a key advancement with respect to the state of the art in traffic accident prevention.

1. Introduction

Road accidents, indicated as one of the main causes of injury and death, are frequently related to the underestimation of drivers’ mental workload (MW) and fatigue [1]. The world of research is consistent in assuming that crash risks are strongly related to driver mental workload [2,3]. Hence, predicting cognitive states, such as mental overload, could be fundamental to prevent traffic accidents. The quantitative assessment of MW can be performed by means of neuroimaging and neurophysiological techniques and methods [4,5]. Indeed, several studies reported the use of behavioral measurement, such as eye blinking [6,7], and physiological measurement, such as Electrocardiogram (ECG) [8], Electroencephalogram (EEG) [9,10,11], and functional Near Infrared Spectroscopy (fNIRS) [6,12] to estimate MW. In a very recent review on the assessment of MW relying on physiological parameters, Tao et al. stated that cardiovascular, eye movement and EEG measures were the most frequently used across various research fields, reporting 76%, 66%, and 71% of the times a significant association with MW, respectively [13]. Relative to the ECG signal, among all the other physiological parameters, it can be considered one of the most suitable signals in the automotive research domain, since its detection ensures comfort and not excessive invasiveness for the driver if compared, for instance, with EEG measurements. Furthermore, ECG-derived parameters can also be established through cutting-edge technologies, which are now available to a large part of the population (i.e., smart devices) with a good level of reliability [14]. In a recent study, Tjolleng et al., developed a three-level classifier based on artificial neural networks relying on six ECG-derived features extracted in time and frequency domains [15]. In this study, drivers were asked to perform N-back tasks while driving on a static simulator. The developed model reached an accuracy of 82%. However, in general, there is a wide scientific literature about the relationship between physiological parameters (especially HRV) and MW. We recommend a very exhaustive review by Dias et al. [16].
However, the limitations in the assessment of behavioral/physiological parameters in real life driving through the above-mentioned techniques (contact probes, high sensitivity to driver’s motion, specific lighting conditions) prevent their large application in Advanced Driver Assistance Systems (ADAS), in which the use of non-contact sensors would be specifically desirable, and the main aim of the current study is to overtake these limitations due to contact and invasive measurement techniques by the use of a non-invasive and contactless methodology, the thermal infrared (IR) imaging, that has been proposed as a suitable alternative tool to estimate MW, just because of its contactless modality. IR imaging is a non-invasive technology that is able to infer the autonomic modulation of the superficial skin temperature [17]. Importantly, compared with visible cameras used to infer behavioral parameters, IR is not affected by illumination and can work in a completely dark environment. The use of IR imaging permits the estimation of the peripheral autonomic activity relying on the modulation of the skin temperature, which is a known expression of the psycho-physiological state of the subject [18,19,20]. Accordingly, experienced emotions, including stress or fatigue, can produce measurable changes in skin temperature [21,22].

2. Related Work

There is great attention in the research field on MW monitoring using thermal IR imaging. Kang et al. assessed affective training times by monitoring the cognitive load relying on facial temperature changes. Significant correlations (i.e., r 0.88 , 0.96 ) were found between the nose tip temperature and response time, accuracy, and the Modified Cooper Harper Scale ratings [23]. Stemberger et al. proposed a system for the evaluation of MW levels of aviators relying on the assessment of facial skin temperature. The method also relied on head pose estimation, measurement of the temperature variation over different facial regions, and an artificial neural network classifier. The system classified with good accuracy the MW into high, medium, and low levels 81% of the time [24].
Given the advantages of the use of IR imaging in psycho-physiological state monitoring, a relevant number of scientific works on the automotive research field are available. Most of these publications concern driver drowsiness/fatigue monitoring and emotional state detection [25,26,27,28,29,30]. Relative to drivers’ MW monitoring using thermal IR imaging, the literature is instead scarce. Or and Duffy used thermography to assess the relationship between MW and thermal patterns of facial regions of interest (ROIs). They found a significant correlation between the nose skin temperature change and the subjective workload score in both simulated and real-vehicle driving [31]. Pavlidis et al. [32], investigated the effects of cognitive, sensorimotor, emotional, and mixed stressors on driver arousal and performance during a driving simulator experiment. Perinasal perspiration, inferred by IR imaging, together with the measurement of steering angle and the lane departures on the left and right side of the road, revealed a more dangerous driving condition for both sensorimotor and mixed stressors compared to the baseline situation [32]. In a more recent work by Wang et al. [33], the correlation of facial skin temperature and its variation with the EEG-measured MW was examined in three different thermal environments (slightly cool, neutral, and slightly warm). They found that the absolute facial temperature had stronger correlations with MW than facial temperature variation and that the correlations were higher in the neutral thermal environment if compared with the other two thermal conditions [33]. Finally, Perpetuini et al. [12], developed a two-levels Support Vector Machine (SVM) classifier to predict the level of MW from IR imaging features. The Sample Entropy of the fNIRS signal was assumed to indicate MW and was used as output data for the model. The classifier showed a sensitivity of 77% and specificity of 69% [12].
Table 1 summarizes the approaches used in the related work cited above, reporting information for each one about the employed methodology and performance.
In the present work, the driver MW was established by means of IR imaging and supervised machine learning (ML) methods. Supervised ML approaches are part of Artificial Intelligence (AI) algorithms, able to automatically learn functions that map an input to an output based on known input–output pairs (training dataset). The function is inferred from labeled training data and can be used for mapping new datasets (test data), thus permitting to evaluate the accuracy of the learned function and estimate the level of generalization of the applied model [35].
Based on key features of thermal signals extracted from peculiar ROIs indicative of the psycho-physiological state and ECG derived parameters, ML-based classification models of MW were performed with the aim to distinguish among different levels of MW. Two cognitive tests, with their own subcategories, were chosen to investigate the drivers’ response to different and predefined levels of cognitive load in order to categorize the classes of MW. To develop an accurate and automated MW classification system, ML multimodal (based on both IR imaging and ECG derived features) and unimodal (IR imaging or ECG derived features) models were developed and compared. The principal innovation of the present study consists in the capability of distinguishing different MW levels based on the only monitoring of IR signals and/or ECG derived features. Of note, this work describes a novel approach for a contactless methodology dedicated to driver MW classification, constituting a significant improvement to actual ADAS technology and, in general, to road security level. Furthermore, the developed systems can be completely inherited from any other field of application in which it is desirable to accurately define the level of human MW, thus opening new opportunities and perspectives in the domains of ergonomics and human-machine interaction.

3. Materials and Methods

3.1. Participants

The experimental session involved 26 adults (17 males, age range 18–42, mean 30.89, standard deviation 6.08). Prior to the experimental sessions, the participants were adequately informed about the purpose and protocol of the study, and they signed an informed consent form resuming the methods and the purposes of the experimentation in accordance with the Declaration of Helsinki [36].
Participants were selected according to the following inclusion criteria:
-
possession of a driver’s license;
-
aged 18 years old or over;
Participants were excluded if they did not fall under the inclusion conditions and if they were diagnosed with mental/cognitive impairment.
A survey conducted through the administration of questionnaires revealed that, on average, participants had an experience of driving for (16.54 ± 5.84) years, they were used to driving (53.64 ± 18.45) hours per day and (6.29 ± 1.13) days per week. Furthermore, 72.73% declared that they drive only in an urban context, and 9.09% mainly on highways, whereas the 18.18% declared mixed context driving.

3.2. Experimental Protocol

Prior to testing, each subject was left in the experimental room for fifteen minutes to allow their baseline skin temperature to stabilize. The environmental conditions of the experimental room were set at a standardized temperature (23 °C) and humidity (50–60%) by the use of a thermostat.
The experimental sessions were performed using a static driver simulator (Figure 1a). Three 27 inch monitors were used to display the scenario, with a total video resolution of 5760 × 1080 pixels. The distance between the driver and monitors was 1.5 m. Drivers’ horizontal view angle was 150 degrees. Participants sat comfortably on the driver’ seat during both acclimatization and experimental periods.
The software used for the driving simulation was City Car Driving, Home Edition software-version 1.5 [37] (Figure 1b). The experimental protocol consisted in performing a 45 min driving simulation in an urban context. The experimental conditions were set a priori to ensure adverse driving conditions and deliver a reproducible experimental protocol to all study participants (i.e., Traffic density: 60%; Traffic behavior: Intense traffic; Pedestrian crossing the road in a wrong place: Often; Dangerous change of traffic: Often; Emergency braking of the car ahead: Often).
These conditions represented the baseline (BL) situation for the drivers and were selected to guarantee a non-monotonous environment. In particular, the settings associated to emergency situations and traffic guaranteed uncomfortable driving, since the participants were often driving in non-monotonous situations. After a BL period of fifteen minutes, two cognitive tests (i.e., Digit Span and Rey Auditory Verbal Learning tests) were administered to drivers. The administration of cognitive tasks allowed to manipulate the MW with respect to the baseline driving.
In detail, the Digit Span test (DST) is a cognitive test composed of two different tasks able to assess the abilities of short-term memory and working memory, the latter referring to the skill to retain information for a short time to manipulate them mentally [38]. For this test, the participant was asked to repeat sequences of digits verbally presented with a pace of one digit per second. The test started with a two-digit series and each time the sequence was repeated correctly a new set was presented with one more digit. If the participant could not remember a series, another one of the same length was proposed. If the participant was not able to repeat two sequences of the same length, the test ended. The digit span score consisted in the length of the longest correctly recalled sequence. For the purposes of the present study, the DST was composed of two parts:
  • Forward DST (repetition of digits in the same order to their presentation)
  • Backward DST (repetition of digits in the reverse order to their presentation)
The Rey Auditory Verbal Learning test (RAVLT) is a cognitive task able to evaluate verbal learning and long-term memory [39,40]. It was administered by reading the participant a list of 15 words at the pace of one word per second. At the end of the reading, the subject was asked to immediately repeat as many words as possible, in any order. This procedure was repeated with the same word list five consecutive times, recording different elements each time. This was the first part of the test and consisted in the Immediate Recall (ImmR).
After a 15-min time interval, during which the subject continued to drive, he/she was asked to remember (without the list being re-proposed by the examiner) as many words as possible from the list. This was the second part of the RAVLT and consisted in the Delayed Recall (DelR).
Subsequently, the last part of the test consisted in the Recognition (Rec) of the 15 words among the other words not present in the original list. The total amount of items was 46. This test allows for a qualitative evaluation of the memory performance in terms of facilitated recovery.
At the end of the test, the subject drove for 10 min without further test administration. The whole pipeline of the experimental procedure is summarized in Figure 2:

3.3. Data Acquisition and Analysis

During the execution of the experimental protocol, ECG signals and visible and thermal IR videos were concurrently acquired.
The ECG signals were recorded by means of Encephalan Mini (Medicom MTD system, Taganrog, Russia) using the lead configuration determined by the Standard Limb Leads (i.e., electrodes positioned at the right arm (RA), left arm (LA), and left leg (LL)) [41]. The ECG signals were acquired at a frequency rate of 256 Hz and band-pass filtered in the frequency band of [0.05–150] Hz. Furthermore a notch filter was used to eliminate the artifact due to the mains power supply (fnotch = 50 Hz).
An Intel RealSense D415 depth camera (Intel Corporation©, Santa Clara, California, USA) and a FLIR Boson 320LW IR thermal camera (FLIR corporation©, Wilsonville, Oregon, USA) were used to acquire visible and thermal IR videos, respectively. In detail, the visible camera is Full HD 1080p (1920 × 1080 pixel), whereas the spatial resolution of the thermal camera is 320 × 256 pixels. Relative to the thermal camera, FLIR Boson 320 relies on uncooled VOx microbolometer technology and it is featured with a thermal sensitivity of 50 mK. Of note, the output of FLIR Boson 320 is a 16 bit signal, linear with input flux (i.e., target irradiance) and independent from the camera’s temperature. This means that the output is not translated to absolute temperature (i.e., K/°C), and it ranges from 0 to 216.
For the purposes of this study, the two imaging devices were held together and aligned horizontally by means of a specifically designed frame made by Next2U® (Figure 3a,b).
Both visible and IR videos were recorded at a frequency rate of 10 Hz. The distance between the participant and the imaging system was 0.6 m (Figure 1a).
Concerning IR imaging data analysis, visible imagery were used to track facial landmarks (i.e., 68 points) through the software OpenFace [42], and, successively, they were co-registered to the thermal imagery by the estimation of the geometrical transformation between the visible and the IR optics, following the same procedure described in [43] (Figure 4a). Two ROIs were automatically determined on facial areas of physiological importance (i.e., nose tip and glabella) (Figure 4b). For each ROI, the average value of the pixels was extracted over time and representative features were computed over each experimental phase. To remove possible artifacts from thermal signals, the Hampel function (MATLAB 2021b©, Mathworks Natick, MA, USA) was employed [44]. The Hampel filter is a robust outlier detector relying on Median Absolute Deviation. For each sample of the signal, median and standard deviation are calculated using all neighboring values within a window of size SampWin. If the point of interest lies nSD standard deviations from the median it is identified as an outlier and is replaced by the median value. In this work, we chose SampWin = 15 s and nSD = 2. To take into account the initial values of the thermal signals, the average values of baseline (evaluated across a period of one minute before the DST phase) were subtracted from the raw thermal signals for each experimental phase (Figure 4c).
Subsequently, the following features were extracted from the thermal signals:
  • Mean value (MeanTemp);
  • Standard deviation (STD);
  • Kurtosis (K);
  • Skewness (S);
  • 90th percentile (90th P);
  • Sample Entropy (SampEn);
  • Ratio of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz) and in the high-frequency band (HF = [0.15–0.4] Hz) (LF/HF);
  • Mean value of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz);
  • Mean value of the power spectral density evaluated in the high-frequency band (HF = [0.15–0.4] Hz).
For the ECG data analysis, the elapsed time periods between the two successive R-peaks of the ECGs (RR signals) were extracted by means of a home-made MATLAB 2016b© script. The script was based on a peak detection procedure, in which parts of the signal exceeding two standard deviations were considered as R peaks. On the obtained RR signals (i.e., Heart Rate Variability (HRV) signal), six features were computed over the experimental phases:
  • Mean value (RRmean);
  • Standard deviation (SDNN);
  • Root mean square of successive differences (RMSSD);
  • Ratio of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz) and in the high-frequency band (HF = [0.15–0.4] Hz) (LF/HF);
  • Mean value of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz);
  • Mean value of the power spectral density evaluated in the high-frequency band (HF = [0.15–0.4] Hz).
To take into account the initial values of the HRV signals, each one of the features was normalized with respect to their baseline value. In particular, the ratio between each HRV feature during the experimental phases and the same feature evaluated during the baseline was computed and considered as input for the models.
Each one of the IR and ECG features were extracted relatively to the specific subcategory of test, which were used to define the label classes for the ML based models. For each one of these classes, features have been computed working on the IR or ECG signal acquired during the specific subtest. This aspect guarantees a balanced class numerosity because all the features for all the subjects were considered relative to each of the classes. A detailed description of the features computation is reported in the Appendix A Section.

3.4. Application of Supervised Machine Learning for Classification

Supervised ML is the process of learning a set of rules from instances with the aim of automatically find functions that map an input to an output. The function is inferred from labeled training data and can be used for mapping new dataset (test data) thus allowing to evaluate the accuracy of the learned function and estimate the level of generalization of the applied model [45]. In the present study the performances of six categories of classifiers were compared: Decision Trees (DT) [46], Discriminant Analysis (DA) [47], Logistic Regression (LR) [48], Support Vector Machines (SVM) [49], Nearest Neighbor (kNN) [50], and Ensemble Classifiers [51].
Linear, quadratic and cubic SVM classification models were considered in the present work.
Coarse, medium and fine kNN classification models were considered in the present work.
Relatively to the Ensemble classifiers, bagged trees, subspace discriminant, subspace kNN, and Random Under-Sampling (RUS) boosted trees were considered in the present work [52].
A k-fold cross validation (with k = 5) was used to protect against overfitting [53]. The procedure relies on partitioning the dataset into folds, each one with a training and validation dataset, and estimating the accuracy on each fold, guaranteeing the generalization of the model. To ensure that the samples of the same subject would have not been considered in both training and validation procedures, the folds were created so as to ensure that each subject was seen by the model only in the training or in the validation phases, and not in both of them. For the sake of clarity, for each set of features and each classifier model, since the subjects were 26, a set of 21 subjects were employed for training and a set of five drivers were used for testing. The procedure was iteratively repeated, randomizing the subjects involved as training and validation sets.
The machine learning-based analysis of data were performed by means of the Classification Learner App, MATLAB 2021b© [54]. For the purpose of this work, all the classification models available were considered.

4. Results

4.1. Drivers’ Performances on Cognitive Tasks

The DST score was calculated as the length of the longest correctly recalled sequence in both Forward and Backward phases. The RAVLT scores were calculated counting the total number of words repeated in the five repetitions over the ImmR phase, counting the total number of words recalled by the participants during the DelR phase and the total amount of correctly recognized world during the Rec phase.
Since the label classes of the developed models (Figure 12) were based on the specific subcategories of the DST and RAVLT, it was necessary to objectively assess if the effect on the performances of the test on the cohort of subjects was adherent with the one reported in the literature. Indeed, paired t-test analyses were performed on the scores obtained by the participants during the execution of the two cognitive tests. Significant differences were observed between Forward and Backward DST (t = 2.69, p < 0.01, degrees of freedom (dof) = 25), between ImmR and DelR (t = 21.90, p << 0.01, dof = 25), and between DelR and Rec scores (t = −44.82, p << 0.01, dof = 25). No significance was found in the comparison between ImmR and Rec. Participants’ scores are reported in Figure 5 (i.e., whiskers plot).
The mean values and standard deviations of the participants’ scores are reported in Table 2.

4.2. IR-Visible Video Processing

The method for combined visible and IR video processing has been validated in [43]. For the present study, on average 95.20% of the video frames were correctly processed. This percentage value referred to the number of frames with correctly identified facial landmarks with reference to the total number of frames.
Regarding the computational load, the average execution time of the developed algorithm was 0.09 s/frames with MATLAB 2016b© (64-bitWindows 7 Pro, Service Pack 1; Intel (R) Core (TM) i5 CPU; 8.00 GB RAM).

4.3. Performances of Supervised Machine Learning Approaches

Thermal and HRV features were first investigated and statistical analysis were performed to assess the most informative features among them. Student’s t-tests were performed among all the features over the experimental phases. The results are reported in Figure 6, Figure 7 and Figure 8 for DST and Figure 9, Figure 10 and Figure 11 for RAVLT.
Relative to DST, the skewness and the 90th percentile were the most informative IR features for the nosetip, whereas the LF/HF feature gave an important contribution for both nosetip and glabella (Figure 6 and Figure 7). No significant difference was revealed by HRV derived features between the two experimental phases (Figure 8).
Referring to RAVLT, the skewness, the 90th percentile and LF/HF were the most informative IR features relative to nosetip, showing significant differences in the comparison of ImmR—DelR (Figure 9). Instead, standard deviation and LF/HF were the most informative features relative to glabella for every comparison among the experimental phases, sampEn for ImmR vs. DelR comparison and LF and HF features for the comparison ImmR vs. Rec (Figure 10). For HRV derived features, HF and LF features were the most informative, both showing significant differences in the comparison DelR vs. Rec (Figure 11). HF features also showed a significant difference in the comparison ImmR vs. Rec (Figure 11).
In the present work, unimodal and multimodal ML-based approaches were developed, each one of them relying on six categories of classifiers: Decision Trees (DT), Discriminant Analysis (DA), Logistic Regression (LR), Support Vector Machines (SVM), Nearest Neighbor (kNN), and Ensemble Classifiers. In the unimodal approach, features extracted from IR signals or HRV signals were separately used as input signals. In the multimodal approach, instead, features extracted from both IR and HRV signals were used together as input data for the classification models. Two-level and three-level classification models were adopted for DST and RAVLT data, respectively. The scheme of the classification model is reported in Figure 12.
Notably, not all the features were used as input to the classifier models, but they were selected through a wrapper method [55,56]. This feature selection approach allows to consider only the minimal set of features that are relevant for the classification purpose. In particular, the random subset of features are evaluated as input features of the specific model and the subset of features that reach the best performance are chosen as input data. In this study, a number of 50 random combinations of features was chosen. After this procedure, for DST and RAVLT, the best feature sets were available and constituted the effective input data for the classifiers. The set of features after the wrapper procedure are summarized in Table 3.
The results of the classifications, in terms of accuracy, are reported in Table 4.
Relative to the classifiers with the best performances (highlighted in bold in Table 4), the Receiver Operating Characteristics (ROC) curves and confusion matrices are reported in Figure 13 for DST and Figure 14 and Figure 15 for RAVLT. ROC curves represent sensitivity (i.e., true positive rate) versus specificity (i.e., 1-false positive rate) across a range of values to evaluate the ability of the classifier to predict an outcome. An important parameter is the Area Under Curve (AUC), which summarizes the classifier performances. A model whose predictions are 100% wrong has an AUC of 0, whereas a model whose predictions are 100% correct has an AUC of 1.
In Figure 13, the performances of the linear SVM classifiers for both unimodal IR features-based models and the multimodal IR + HRV features-based classifier are reported. The unimodal IR features based model showed good performance (accuracy = 73.1%; AUC = 0.76; sensitivity = 0.73; specificity = 0.73; precision = 0.73; F1-score = 0.73) as well as the multimodal IR + HRV features based classifier (accuracy = 73.1%; AUC = 0.80; sensitivity = 0.69; specificity = 0.71; precision = 0.69; F1-score = 0.72).
Figure 14 shows the performances of the three-level unimodal IR features based classifier relying on the ensemble bagged tree model. Dealing with a three-level classifier, three ROC curve are presented, each one of them representing the comparison of one class with the cumulative class of the other two. The average performances are (AUC = 0.85; sensitivity = 0.71; specificity = 0.85; precision = 0.71; F1-score = 0.70).
Figure 15 shows the performances of the three-level multimodal IR + HRV features based classifier relying on the ensemble bagged tree model. The average performances are (AUC = 0.85; sensitivity = 0.75; specificity = 0.87; precision = 0.75; F1-score = 0.74).

5. Discussion

Monitoring the MW during driving situations is of paramount importance given its close relationship with the risk of road accidents. The main aim of the present study was to develop models, based on drivers’ psychophysiological features, that are able to discriminate the level of drivers’ MW. To this specific aim, two different cognitive tests (DST and RAVLT) were administered to twenty-six participants while driving in a simulated environment under non-monotonous situations. The statistical analyses on cognitive tests scores reveled significant differences between Forward and Backward DST and between ImmR and DelR and between ImmR and Rec in RAVLT, thus revealing different performances of subjects over the experimental phases.
DST and RAVLT were specifically chosen to study two different types of cognitive load, the former related to short-term memory and the latter related to long-term memory and verbal learning. DST and RAVLT are indicative also of the working memory capacity (WMC) of the subjects [57]. In particular, the Backward DST and the DelR-RAVLT have been reported as the most demanding in terms of WMC [57]. Estimating the WMC is of crucial importance in the research domain of the automotive sector since it has been demonstrated as a predictor of distracted driving [58]. In this study, the authors demonstrated that the levels of WMC affect the driving performances of individuals while engaged in cognitive distraction. Furthermore, they reported a mediation of WMC on the effect of distraction on braking response time.
In the current study, the possibility of recognizing different levels of MW through non-invasive techniques was investigated, and ML-based models relying on drivers’ psychophysiological features were developed and compared. HRV and IR thermal features were extracted over the experimental phases. In this context, it has to be underlined that the novelty of the present study consists in validating models able to classify different kinds of MW relying on the IR imaging technique, which is a completely non-invasive and contactless methodology. Relative to DST, the most informative IR features were the skewness and the 90th percentile for the nosetip and LF/HF for both nosetip and glabella (Figure 6 and Figure 7). HRV derived features revealed no significant difference (Figure 8). Regarding RAVLT, the most informative IR features relative to nosetip were the skewness, the 90th percentile, and LF/HF, which globally showed significant differences in the comparison of ImmR vs DelR (Figure 9). Instead, the most informative features relative to the glabella were the standard deviation and LF/HF for every comparison among the experimental phases, sampEn for ImmR vs. DelR comparison and LF and HF features for the comparison of ImmR vs. Rec (Figure 10). For HRV derived features, HF and LF features were the most informative, both showing significant differences in the comparison of DelR vs. Rec (Figure 11). HF features also showed a significant difference in the comparison of ImmR vs. Rec (Figure 11).
The relevance of IR features in discriminating different levels of MW is appreciable from the results mentioned above. Specifically, significant features relative to nosetip are commonly involved in both DST and RAVLT, whereas the features related to the glabella region are mostly involved during RAVLT. This result can be due to the major cognitive involvement in RAVLT since it has been demonstrated that thermal signals from the glabella/forehead are directly linked with cognitive load [59,60,61]. An important role is also played by the nosetip thermal features, and this result is in accordance with the literature [1,59,61]. Generally, the nosetip region has been reported as the most responsive during cognitive tasks, reflected by a drop in the nosetip temperature during cognitive task executions with respect to baseline conditions [62]. Of note, LF/HF for both the nosetip and glabella regions was demonstrated to be relevant for assessing the MW level. In fact, this feature accounts for the balance between the sympathetic nervous system (SNS) and the parasympathetic nervous system (PNS) activity. It has been inherited by HRV metrics and it is based on the assumption that LF power is generated by the SNS, while HF power is produced by the PNS [63]. The LF/HF feature has already been used for thermal imaging data analysis, showing good contribution in the psychophysiological state assessment of individuals [25,61].
Regarding the HRV derived features, it has been observed that they revealed statistical significance only in RAVLT, and, in particular, LF and HF features were the most informative. As mentioned above, they are relevant for the activation of SNS and PNS, respectively. Between them, the most informative was the LF feature, especially in the comparison between DelR and Rec. A recent review from Forte et al. reported HF and LF features from HRV as significant features indicative of cognitive performances [64]. However, the scientific community is not commonly in accordance on the fact that HRV can be a reliable indicator of MW, especially in the field of automotive. In fact, Paxion et al., in a review on mental workload and driving, highlighted some limits of HRV indicators [4]. Specifically, they argued that HRV is not exclusively sensitive to changes in MW, but is also related to energetic, thermoregulatory, respiratory, and emotional processes and physical activity. Furthermore, they reported that HRV is not always able to discriminate the level of difficulty, thus being an insufficient indicator to assess the MW.
This kind of finding is indeed reflected from the results of the present study, with reference not only to the most informative features but also in regard to the developed ML models. In this study, several models based on ML approaches have been compared based on unimodal IR/unimodal HRV/multimodal IR + HRV features. As shown in Table 4, the best classifier performances were reached by the unimodal IR features-based classifier and also from multimodal IR + HRV features-based models. Unimodal HRV- based models showed the worst performances with respect to the other approaches (i.e., the overall best accuracies reached were 59.6% and 47.4% for DST and RAVLT, respectively). Of note and with reference to the state of the art in ML and deep learning, in this work a specific typology of multimodal approach has been used. Indeed, referring to the recent work of Guarino et al., together with the unimodal approach defined by the authors also single-view learning, a single typology of multimodal procedure was adopted in the present study, referred to by the authors as the intermediate integration multi-view approach [65]. This particular multi-view approach has been chosen since there was the necessity of having a feature selection step (i.e., wrapper method) in the analysis pipeline, prior to the concatenation of IR and HRV features. Hence, the early integration multi-view approach was not considered, since the number of input features was similar to the number of participants. Further studies, instead, could be done to implement the late integration multi-view approach. In this regard, it is necessary to enlarge the sample size to benefit from more reliable single classifiers. For DST, the best performing models were based on two-classes of SVM classifier with linear kernel with unimodal IR features (accuracy = 73.1%; AUC = 0.76; sensitivity = 0.73; specificity = 0.73) and multimodal IR + HRV features (accuracy = 73.1%; AUC = 0.80; sensitivity = 0.69; specificity = 0.71) (Figure 13). For RAVLT, the best performing models were based on three-classes: ensemble bagged trees classifier with unimodal IR features (accuracy = 71.1%; average AUC = 0.85; average sensitivity = 0.71; average specificity = 0.85) and multimodal IR + HRV features (accuracy = 75.0%; average AUC = 0.85; average sensitivity = 0.75; average specificity = 0.87) (Figure 14). Of note, for RAVLT, three-class SVM classifiers with linear kernels also reported good performances with accuracies of 65.8% and 63.2% for unimodal IR and multimodal IR + HRV features based models (Table 4).
The high performances obtained from unimodal IR features-based classifiers are of paramount importance given the possibility of determining the level of drivers’ MW based on features collected by a non-contact device, i.e., the thermal camera. Thermal IR imaging outperformed the HRV measurements as well, constituting a reliable mean for assessing the level of MW in a ubiquitous and non-contact manner. This is an important result, given that in the automotive domain, especially in ADAS, one of the most important aims is to determine the psychophysiological state of the driver without interfering with him/her to avoid/prevent traffic accidents. The impact of such a result is interesting also in terms on ergonomics applied in the automotive field. In fact, the developed ML model could communicate the cognitive state of the driver and alert him/her in case of moderate/high MW. Furthermore, the results are obtained relying on a small-sized thermal camera (i.e., FLIR Boson 320), highly suitable for applications in a restricted environment, such as the cockpit of a vehicle.
However, some limitations have to be mentioned. First, further studies should be performed to increase the sample numerosity. The ML approaches used in this study relied on supervised learning, which is inherently a data-driven analysis; data-driven analyses are highly affected by the sample numerosity, and the performance of the model could indeed improve, reducing a possible overfitting effect driven by the limited sample size. Moreover, increasing the sample size could open the way to more sophisticated and powerful approaches based on deep learning modeling, which is the state of the art in data analysis in several areas of research.
Second, the current study focused on drivers with a limited age range (i.e., 18–42 years old), involving only young and middle-aged adults. The most important improvement of the method could be obtained, including in the study sample individuals with a wider age range. Furthermore, beyond increasing the sample size and age range, other factors, such as thermal comfort, gender and weather conditions during simulated driving sessions will be considered [66,67,68,69]. In fact, accounting for these factors could be of primary valence in automotive research, leading to a broad overview of all aspects concerning the object of the study.
Moreover, the present results refer to simulated driving conditions in which determinant variables for IR measurements, such as sunlight or forced ventilation, were not considered. Therefore, it would be desirable to also apply the developed methodology on real-driving situations in order to generalize the applicability of the technique.
As for being state-of-the-art, this is an original and novel study concerning drivers’ MW evaluation by means of thermal imaging, employing supervised ML algorithms. The present study, although addressed to limited and specific experimental conditions, underlines the feasibility of the method to be verified under wider operating situations. The present work represents a step forward in the perspective of the prevention of road accidents and, above all, it can constitute a turning point in the identification of various levels of mental workload, with benefits in several research domains, from ergonomics to human machine interaction.

6. Conclusions

In the present work, a novel method for drivers’ MW evaluation is presented. In particular, MW levels of the subjects while driving in a simulated environment were estimated with a high level of accuracy through ML algorithms applied to IR and HRV data. The presented work constitutes a step towards the establishment of a reliable detection of the MW levels in a non-invasive and contactless manner, ensuring the maintenance of an ecologic condition of driving, possibly contributing to the prevention of traffic accidents. Further directions for development will include the validation of the developed method directly on-board, with live evaluation of the cognitive workload level of the drivers. This will be particularly useful for all of the categories of long-time drivers, such as truck drivers and bus drivers, in order to prevent traffic accidents due to an excessive cognitive workload during driving activity.

Author Contributions

Conceptualization, D.C., D.P., C.F., A.M.; methodology, D.C., D.P., C.F., S.R.; software, D.C., M.T., S.N.; validation, D.P., C.F.; formal analysis, D.C.; investigation, D.C., D.P., C.F., L.M., S.R., A.G., G.F., F.R.; resources, S.R., A.G., G.F.; data curation, D.C., D.P., C.F., L.M.; writing—original draft preparation, D.C.; writing—review and editing, D.P., C.F., L.M., M.T., S.N., S.R., A.G., G.F., F.R., S.G.; supervision, S.G., A.M.; project administration, A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by PON FESR MIUR R&I 2014-2020-ADAS+ [grant number ARS01_00459], ECSEL Joint Undertaking (JU) European Union’s Horizon 2020 Heliaus [grant number 826131] and MISE ITINERE [grant number n. 1654, Accordi per l’innovazione—D. M. 2 August 2019].

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of University “G d’Annunzio” of Chieti-Pescara, Italy (protocol number 2077, date of approval: 14 May 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy issues. The codes developed for the purpose of the study are available on request to the corresponding author.

Conflicts of Interest

The authors declare that they have no conflict of interest.

Appendix A

This section describes the features used as input data for the ML models developed for the purpose of the study. IR and HRV features are illustrated.

Appendix A.1. IR Features

IR features have been directly extracted from the thermal signal relative to two ROIs: nosetip and glabella. An example of the thermal signals extracted over the experimental phases is reported in Figure A1.
Figure A1. Thermal signals extracted from nosetip (in blue) and glabella (in orange) ROIs over the experimental phases.
Figure A1. Thermal signals extracted from nosetip (in blue) and glabella (in orange) ROIs over the experimental phases.
Sensors 22 07300 g0a1
For each one of the experimental phases the following features have been computed for each ROI:
  • Mean value (MeanTemp)—average value of the thermal signal T over time (i.e., N samples) defined as:
M e a n T e m p = 1 N i = 1 N T i
2.
Standard deviation (STD)—standard deviation of the thermal signal T overtime (i.e., N samples) defined as:
S T D = 1 N 1 i = 1 N ( T i M e a n T e m p ) 2
3.
Kurtosis (K): fourth standardized moment, and it is the ratio between the fourth cental moment and the standard deviation. It is evaluated as follows:
K = 1 N i = 1 N ( T i M e a n T e m p ) 4 S T D 4
4.
Skewness (S)—third standardized moment, and it is the ratio between the third cental moment and the standard deviation. It is evaluated as follows:
S = 1 N i = 1 N ( T i M e a n T e m p ) 3 S T D 3
5.
90th percentile (90th P): is the temperature value below which the 90% of all temperature frequency distribution falls;
6.
Sample Entropy (SampEn): is defined as the negative natural logarithm of the conditional probability that signals that the subseries of length m (pattern length) that match pointwise within a tolerance r (similarity factor) also match at the m + 1 point. SampEn of a time series {t1,...,tN} of length N is computed employing the following set of equations:
S a m p E n m , r , N = ln U m + 1 r U m r U m r = N m τ 1 i = 1 N m τ C i m r C i m r = B i N m + 1 τ B i = n u m b e r   o f   j   w h e r e   d T i , T j r T i = t i , t i + τ , , t i + m 1 τ T j = t j , t j + τ , , t j + m 1 τ i j N m τ ,   j i
In this study, it has been considered that m = 2 and r = 0.2 ∙ SD of the signal. These parameters are commonly employed for complexity analysis of biological signals and they were chosen in accordance with [70].
7.
Mean value of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz)
8.
Mean value of the power spectral density evaluated in the high-frequency band (HF = [0.15–0.4] Hz)
9.
Ratio of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz) and in the high-frequency band (HF = [0.15–0.4] Hz) (LF/HF).
For features 7, 8, and 9, we report here the power spectral density of the thermal signal T(i) evaluated over a time window Tc:
P S D f = T f 2 T c
where T(f) is the Fourier transform of the signal T(i). For each one of the features 7 and 8, the average of the PSD has been evaluated in the reported frequency band.

Appendix A.2. HRV Features

Heart rate variability (HRV) refers to the fluctuations between consecutive heartbeatcycles. It is usually represented by the variation in the heart rate’s beat-to-beat temporal changes (RR intervals). Figure A2 represents an exemplificative ECG signal, in which some R peaks and RR intervals are highlighted.
Figure A2. Exemplificative ECG signal. Example of R peaks and RR intervals are highlighted.
Figure A2. Exemplificative ECG signal. Example of R peaks and RR intervals are highlighted.
Sensors 22 07300 g0a2
For each one of the experimental phases, the following features have been computed:
  • Mean value (RRmean)—average value of the RR intervals (RRi), evaluated as follows:
RRmean = 1 N i = 1 N R R i
2.
Standard deviation (SDNN)—Standard deviation of normal-to-normal interval, calculated as follows:
S D N N = 1 N 1 i = 1 N ( R R i R R m e a n ) 2
3.
Root mean square of successive differences (RMSSD): root-mean-square of successive RR interval differences, evaluated as follows:
R M S S D = 1 N 1 i = 1 N 1 ( R R i + 1 R R i ) 2
4.
Mean value of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz);
5.
Mean value of the power spectral density evaluated in the high-frequency band (HF = [0.15–0.4] Hz).
6.
Ratio of the power spectral density evaluated in the low-frequency band (LF = [0.04–0.15] Hz) and in the high-frequency band (HF = [0.15–0.4] Hz) (LF/HF).
For HRV features 4, 5, 6 refer to the computation described for IR features 7, 8, and 9.

References

  1. Kajiwara, S. Evaluation of Driver’s Mental Workload by Facial Temperature and Electrodermal Activity under Simulated Driving Conditions. Int. J. Automot. Technol. 2014, 15, 65–70. [Google Scholar] [CrossRef]
  2. Kantowitz, B.H.; Simsek, O. Secondary-task measures of driver workload. In Stress, Workload, and Fatigue; CRC Press: Boca Raton, FL, USA, 2000; ISBN 978-1-4106-0044-8. [Google Scholar]
  3. Da Silva, F.P. Mental Workload, Task Demand and Driving Performance: What Relation? Procedia Soc. Behav. Sci. 2014, 162, 310–319. [Google Scholar] [CrossRef]
  4. Paxion, J.; Galy, E.; Berthelon, C. Mental Workload and Driving. Front. Psychol. 2014, 5, 1344. [Google Scholar] [CrossRef] [PubMed]
  5. Charles, R.L.; Nixon, J. Measuring Mental Workload Using Physiological Measures: A Systematic Review. Appl. Ergon. 2019, 74, 221–232. [Google Scholar] [CrossRef] [PubMed]
  6. Foy, H.J.; Chapman, P. Mental Workload Is Reflected in Driver Behaviour, Physiology, Eye Movements and Prefrontal Cortex Activation. Appl. Ergon. 2018, 73, 90–99. [Google Scholar] [CrossRef] [PubMed]
  7. Marquart, G.; Cabrall, C.; de Winter, J. Review of Eye-Related Measures of Drivers’ Mental Workload. Procedia Manuf. 2015, 3, 2854–2861. [Google Scholar] [CrossRef]
  8. Heine, T.; Lenis, G.; Reichensperger, P.; Beran, T.; Doessel, O.; Deml, B. Electrocardiographic Features for the Measurement of Drivers’ Mental Workload. Appl. Ergon. 2017, 61, 31–43. [Google Scholar] [CrossRef]
  9. Borghini, G.; Astolfi, L.; Vecchiato, G.; Mattia, D.; Babiloni, F. Measuring Neurophysiological Signals in Aircraft Pilots and Car Drivers for the Assessment of Mental Workload, Fatigue and Drowsiness. Neurosci. Biobehav. Rev. 2014, 44, 58–75. [Google Scholar] [CrossRef]
  10. Di Flumeri, G.; Borghini, G.; Aricò, P.; Sciaraffa, N.; Lanzi, P.; Pozzi, S.; Vignali, V.; Lantieri, C.; Bichicchi, A.; Simone, A.; et al. EEG-Based Mental Workload Neurometric to Evaluate the Impact of Different Traffic and Road Conditions in Real Driving Settings. Front. Hum. Neurosci. 2018, 12, 509. [Google Scholar] [CrossRef]
  11. Di Flumeri, G.; Borghini, G.; Aricò, P.; Sciaraffa, N.; Lanzi, P.; Pozzi, S.; Vignali, V.; Lantieri, C.; Bichicchi, A.; Simone, A.; et al. Chapter 20–EEG-Based Mental Workload Assessment During Real Driving: A Taxonomic Tool for Neuroergonomics in Highly Automated Environments. In Neuroergonomics; Ayaz, H., Dehais, F., Eds.; Academic Press: Cambridge, MA, USA, 2019; pp. 121–126. ISBN 978-0-12-811926-6. [Google Scholar]
  12. Perpetuini, D.; Cardone, D.; Filippini, C.; Spadolini, E.; Mancini, L.; Chiarelli, A.M.; Merla, A. Can Functional Infrared Thermal Imaging Estimate Mental Workload in Drivers as Evaluated by Sample Entropy of the FNIRS Signal? In Proceedings of the 8th European Medical and Biological Engineering Conference, Portorož, Slovenia, 29 November–3 December 3 2020; Jarm, T., Cvetkoska, A., Mahnič-Kalamiza, S., Miklavcic, D., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 223–232. [Google Scholar]
  13. Tao, D.; Tan, H.; Wang, H.; Zhang, X.; Qu, X.; Zhang, T. A Systematic Review of Physiological Measures of Mental Workload. Int. J. Environ. Res. Public Health 2019, 16, 2716. [Google Scholar] [CrossRef] [Green Version]
  14. Georgiou, K.; Larentzakis, A.V.; Khamis, N.N.; Alsuhaibani, G.I.; Alaska, Y.A.; Giallafos, E.J. Can Wearable Devices Accurately Measure Heart Rate Variability? A Systematic Review. Folia Med. 2018, 60, 7–20. [Google Scholar] [CrossRef] [PubMed]
  15. Tjolleng, A.; Jung, K.; Hong, W.; Lee, W.; Lee, B.; You, H.; Son, J.; Park, S. Classification of a Driver’s Cognitive Workload Levels Using Artificial Neural Network on ECG Signals. Appl. Ergon. 2017, 59, 326–332. [Google Scholar] [CrossRef] [PubMed]
  16. Dias, R.D.; Ngo-Howard, M.C.; Boskovski, M.T.; Zenati, M.A.; Yule, S.J. Systematic Review of Measurement Tools to Assess Surgeons’ Intraoperative Cognitive Workload. Br. J. Surg. 2018, 105, 491–501. [Google Scholar] [CrossRef] [PubMed]
  17. Ammer, K.; Ring, F. The Thermal Human Body: A Practical Guide to Thermal Imaging; Jenny Stanford Publishing: New Delhi, India, 2019; ISBN 0-429-01998-X. [Google Scholar]
  18. Filippini, C.; Perpetuini, D.; Cardone, D.; Chiarelli, A.M.; Merla, A. Thermal Infrared Imaging-Based Affective Computing and Its Application to Facilitate Human Robot Interaction: A Review. Appl. Sci. 2020, 10, 2924. [Google Scholar] [CrossRef]
  19. Kosonogov, V.; Zorzi, L.D.; Honoré, J.; Martínez-Velázquez, E.S.; Nandrino, J.-L.; Martinez-Selva, J.M.; Sequeira, H. Facial Thermal Variations: A New Marker of Emotional Arousal. PLoS ONE 2017, 12, e0183592. [Google Scholar] [CrossRef]
  20. Perpetuini, D.; Cardone, D.; Filippini, C.; Chiarelli, A.M.; Merla, A. Modelling Impulse Response Function of Functional Infrared Imaging for General Linear Model Analysis of Autonomic Activity. Sensors 2019, 19, 849. [Google Scholar] [CrossRef]
  21. Cruz-Albarrán, I.A.; Benítez-Rangel, J.P.; Osornio-Ríos, R.A.; Morales-Hernández, L.A. Human Emotions Detection Based on a Smart-Thermal System of Thermographic Images. Infrared Phys. Technol. 2017, 81, 250–261. [Google Scholar] [CrossRef]
  22. Engert, V.; Merla, A.; Grant, J.A.; Cardone, D.; Tusche, A.; Singer, T. Exploring the Use of Thermal Infrared Imaging in Human Stress Research. PLoS ONE 2014, 9, e90782. [Google Scholar] [CrossRef]
  23. Kang, J.; McGinley, J.A.; McFadyen, G.; Babski-Reeves, K. Determining Learning Level and Effective Training Times Using Thermography. In Proceedings of the Army Science Conference, Orlando, FL, USA, 29 November–2 December 2006. [Google Scholar]
  24. Stemberger, J.; Allison, R.S.; Schnell, T. Thermal Imaging as a Way to Classify Cognitive Workload. In Proceedings of the 2010 Canadian Conference on Computer and Robot Vision, Ottawa, ON, Canada, 31 May–2 June 2010; pp. 231–238. [Google Scholar]
  25. Cardone, D.; Perpetuini, D.; Filippini, C.; Spadolini, E.; Mancini, L.; Chiarelli, A.M.; Merla, A. Driver Stress State Evaluation by Means of Thermal Imaging: A Supervised Machine Learning Approach Based on ECG Signal. Appl. Sci. 2020, 10, 5673. [Google Scholar] [CrossRef]
  26. Cardone, D.; Filippini, C.; Mancini, L.; Pomante, A.; Tritto, M.; Nocco, S.; Perpetuini, D.; Merla, A. Driver Drowsiness Evaluation by Means of Thermal Infrared Imaging: Preliminary Results. In Proceedings of the Infrared Sensors, Devices, and Applications XI. Int. Soc. Opt. Photonics 2021, 11831, 118310. [Google Scholar]
  27. Ebrahimian-Hadikiashari, S.; Nahvi, A.; Homayounfard, A.; Bakhoda, H. Monitoring the Variation in Driver Respiration Rate from Wakefulness to Drowsiness: A Non-Intrusive Method for Drowsiness Detection Using Thermal Imaging. J. Sleep Sci. 2018, 3, 1–9. [Google Scholar]
  28. Knapik, M.; Cyganek, B. Driver’s Fatigue Recognition Based on Yawn Detection in Thermal Images. Neurocomputing 2019, 338, 274–292. [Google Scholar] [CrossRef]
  29. Yamakoshi, T.; Yamakoshi, K.; Tanaka, S.; Nogawa, M.; Park, S.B.; Shibata, M.; Sawada, Y.; Rolfe, P.; Hirose, Y. Feasibility Study on Driver’s Stress Detection from Differential Skin Temperature Measurement. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–24 August 2008; pp. 1076–1079. [Google Scholar]
  30. Zhang, M.; Ihme, K.; Drewitz, U. Discriminating Drivers’ Emotions through the Dimension of Power: Evidence from Facial Infrared Thermography and Peripheral Physiological Measurements. Transp. Res. Part F Traffic Psychol. Behav. 2019, 63, 135–143. [Google Scholar] [CrossRef]
  31. Or, C.K.L.; Duffy, V.G. Development of a Facial Skin Temperature-Based Methodology for Non-Intrusive Mental Workload Measurement. Occup. Ergon. 2007, 7, 83–94. [Google Scholar] [CrossRef]
  32. Pavlidis, I.; Dcosta, M.; Taamneh, S.; Manser, M.; Ferris, T.; Wunderlich, R.; Akleman, E.; Tsiamyrtzis, P. Dissecting Driver Behaviors under Cognitive, Emotional, Sensorimotor, and Mixed Stressors. Sci. Rep. 2016, 6, 25651. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, X.; Li, D.; Menassa, C.C.; Kamat, V.R. Can Infrared Facial Thermography Disclose Mental Workload in Indoor Thermal Environments? In Proceedings of the 1st ACM International Workshop on Urban Building Energy Sensing, Controls, Big Data Analysis, and Visualization, New York, NY, USA, 13–14 November 2019; Association for Computing Machinery: New York, NY, USA; pp. 87–96. [Google Scholar]
  34. Praveena, M.; Jaiganesh, V. A Literature Review on Supervised Machine Learning Algorithms and Boosting Process. Int. J. Comput. Appl. 2017, 169, 32–35. [Google Scholar] [CrossRef]
  35. PMC, E. World Medical Association Declaration of Helsinki: Ethical Principles for Medical Research Involving Human Subjects. JAMA 2000, 284, 3043–3045. [Google Scholar]
  36. City Car Driving–Car Driving Simulator, PC Game. Available online: https://citycardriving.com/ (accessed on 26 June 2020).
  37. Orsini, A.; Grossi, D.; Capitani, E.; Laiacona, M.; Papagno, C.; Vallar, G. Verbal and Spatial Immediate Memory Span: Normative Data from 1355 Adults and 1112 Children. Ital. J. Neuro. Sci. 1987, 8, 537–548. [Google Scholar] [CrossRef]
  38. Carlesimo, G.A.; Caltagirone, C.; Gainotti, G.; Fadda, L.; Gallassi, R.; Lorusso, S.; Marfia, G.; Marra, C.; Nocentini, U.; Parnetti, L. The Mental Deterioration Battery: Normative Data, Diagnostic Reliability and Qualitative Analyses of Cognitive Impairment. ENE 1996, 36, 378–384. [Google Scholar] [CrossRef]
  39. Rey, A. L’examen Clinique En Psychologie [The Clinical Examination in Psychology]; Presses Universitaries De France: Oxford, UK, 1958; p. 222. [Google Scholar]
  40. Conover, M.B. Understanding Electrocardiography; Elsevier Health Sciences: Amsterdam, The Netherlands, 2002; ISBN 978-0-323-01905-7. [Google Scholar]
  41. Baltrušaitis, T.; Robinson, P.; Morency, L.-P. OpenFace: An Open Source Facial Behavior Analysis Toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
  42. Cardone, D.; Spadolini, E.; Perpetuini, D.; Filippini, C.; Chiarelli, A.M.; Merla, A. Automated Warping Procedure for Facial Thermal Imaging Based on Features Identification in the Visible Domain. Infrared Phys. Technol. 2021, 112, 103595. [Google Scholar] [CrossRef]
  43. Liu, H.; Shah, S.; Jiang, W. On-Line Outlier Detection and Data Cleaning. Comput. Chem. Eng. 2004, 28, 1635–1647. [Google Scholar] [CrossRef]
  44. Osisanwo, F.Y.; Akinsola, J.E.T.; Awodele, O.; Hinmikaiye, J.O.; Olakanmi, O.; Akinjobi, J. Supervised Machine Learning Algorithms: Classification and Comparison. Int. J. Comput. Trends Technol. 2017, 48, 128–138. [Google Scholar]
  45. Shalev-Shwartz, S.; Ben-David, S. Understanding Machine Learning: From Theory to Algorithms; Cambridge University Press: Cambridge, UK, 2014; ISBN 1-139-95274-9. [Google Scholar]
  46. Tharwat, A. Linear vs. Quadratic Discriminant Analysis Classifier: A Tutorial. Int. J. Appl. Pattern Recognit. 2016, 3, 145–180. [Google Scholar] [CrossRef]
  47. Dreiseitl, S.; Ohno-Machado, L. Logistic Regression and Artificial Neural Network Classification Models: A Methodology Review. J. Biomed. Inform. 2002, 35, 352–359. [Google Scholar] [CrossRef] [Green Version]
  48. Cervantes, J.; Garcia-Lamont, F.; Rodríguez-Mazahua, L.; Lopez, A. A Comprehensive Survey on Support Vector Machine Classification: Applications, Challenges and Trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
  49. Zhang, S.; Li, X.; Zong, M.; Zhu, X.; Cheng, D. Learning k for KNN Classification. ACM Trans. Intell. Syst. Technol. 2017, 8, 1–19. [Google Scholar] [CrossRef]
  50. Rokach, L. Ensemble-Based Classifiers. Artif. Intell. Rev. 2010, 33, 1–39. [Google Scholar] [CrossRef]
  51. Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A Survey on Ensemble Learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
  52. Refaeilzadeh, P.; Tang, L.; Liu, H. Cross-Validation. In Encyclopedia of Database Systems; Liu, L., Özsu, M.T., Eds.; Springer: New York, NY, USA, 2016; pp. 1–7. ISBN 978-1-4899-7993-3. [Google Scholar]
  53. Peck, G. Data Science with Matlab. Classification Techniques; CreateSpace Independent Publishing Platform: Scotts Valley, CA, USA, 2017; ISBN 978-1-979472-28-9. [Google Scholar]
  54. Di Credico, A.; Perpetuini, D.; Izzicupo, P.; Gaggi, G.; Cardone, D.; Filippini, C.; Merla, A.; Ghinassi, B.; Di Baldassarre, A. Estimation of Heart Rate Variability Parameters by Machine Learning Approaches Applied to Facial Infrared Thermal Imaging. Front. Cardiovasc. Med. 2022, 9, 893374. [Google Scholar] [CrossRef]
  55. Jović, A.; Brkić, K.; Bogunović, N. A Review of Feature Selection Methods with Applications. In Proceedings of the 2015 38th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), Opatija, Croatia, 25–29 May 2015; pp. 1200–1205. [Google Scholar]
  56. Fard, E.K.; Keelor, J.L.; Bagheban, A.A.; Keith, W.R. Comparison of the Rey Auditory Verbal Learning Test (RAVLT) and Digit Test among Typically Achieving and Gifted Students. Iran. J. Child Neurol. 2016, 10, 26–37. [Google Scholar]
  57. Louie, J.F.; Mouloua, M. Predicting Distracted Driving: The Role of Individual Differences in Working Memory. Appl. Ergon. 2019, 74, 154–161. [Google Scholar] [CrossRef] [PubMed]
  58. Abdelrahman, Y.; Velloso, E.; Dingler, T.; Schmidt, A.; Vetere, F. Cognitive Heat: Exploring the Usage of Thermal Imaging to Unobtrusively Estimate Cognitive Load. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2017, 1, 1–20. [Google Scholar] [CrossRef]
  59. Gaoua, N.; Racinais, S.; Grantham, J.; El Massioui, F. Alterations in Cognitive Performance during Passive Hyperthermia Are Task Dependent. Int. J. Hyperth. 2011, 27, 1–9. [Google Scholar] [CrossRef] [PubMed]
  60. Perpetuini, D.; Cardone, D.; Bucco, R.; Zito, M.; Merla, A. Assessment of the Autonomic Response in Alzheimer’s Patients During the Execution of Memory Tasks: A Functional Thermal Imaging Study. Curr. Alzheimer Res. 2018, 15, 951–958. [Google Scholar] [CrossRef]
  61. Itoh, M. Individual Differences in Effects of Secondary Cognitive Activity during Driving on Temperature at the Nose Tip. In Proceedings of the 2009 International Conference on Mechatronics and Automation, Changchun, China, 9–12 August 2009; pp. 7–11. [Google Scholar]
  62. Wang, L.; Duffy, V.G.; Du, Y. A Composite Measure for the Evaluation of Mental Workload. In Proceedings of the Digital Human Modeling; Duffy, V.G., Ed.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 460–466. [Google Scholar]
  63. Shaffer, F.; Ginsberg, J.P. An Overview of Heart Rate Variability Metrics and Norms. Front. Public Health 2017, 5, 258. [Google Scholar] [CrossRef] [Green Version]
  64. Forte, G.; Favieri, F.; Casagrande, M. Heart Rate Variability and Cognitive Function: A Systematic Review. Front. Neurosci. 2019, 13, 1–11. [Google Scholar] [CrossRef]
  65. Guarino, A.; Lettieri, N.; Malandrino, D.; Zaccagnino, R.; Capo, C. Adam or Eve? Automatic Users’ Gender Classification via Gestures Analysis on Touch Devices. Neural Comput. Appl. 2022, 1–23. [Google Scholar] [CrossRef]
  66. Daanen, H.A.; Van De Vliert, E.; Huang, X. Driving Performance in Cold, Warm, and Thermoneutral Environments. Appl. Ergon. 2003, 34, 597–602. [Google Scholar] [CrossRef]
  67. Mehler, B.; Reimer, B.; Coughlin, J.F. Physiological Reactivity to Graded Levels of Cognitive Workload across Three Age Groups: An on-Road Evaluation. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting, San Francisco, California, USA, 27 September–1 October 2010; SAGE Publications: Los Angeles, CA, USA, 2010; Volume 54, pp. 2062–2066. [Google Scholar]
  68. Son, J.; Lee, Y.; Kim, M.-H. Impact of Traffic Environment and Cognitive Workload on Older Drivers’ Behavior in Simulated Driving. Int. J. Precis. Eng. Manuf. 2011, 12, 135–141. [Google Scholar] [CrossRef]
  69. Son, J.; Reimer, B.; Mehler, B.; Pohlmeyer, A.E.; Godfrey, K.M.; Orszulak, J.; Long, J.; Kim, M.H.; Lee, Y.T.; Coughlin, J.F. Age and Cross-Cultural Comparison of Drivers’ Cognitive Workload and Performance in Simulated Urban Driving. Int. J. Automot. Technol. 2010, 11, 533–539. [Google Scholar] [CrossRef]
  70. Richman, J.S.; Moorman, J.R. Physiological Time-Series Analysis Using Approximate Entropy and Sample Entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Experimental setting: (a) static driver simulator; (b) screenshot of the driving simulation software (i.e., City Car Driving, Home Edition software-version 1.5, Forward Development, Ltd., Verona (WI), USA [37]).
Figure 1. Experimental setting: (a) static driver simulator; (b) screenshot of the driving simulation software (i.e., City Car Driving, Home Edition software-version 1.5, Forward Development, Ltd., Verona (WI), USA [37]).
Sensors 22 07300 g001
Figure 2. Pipeline of the experimental protocol.
Figure 2. Pipeline of the experimental protocol.
Sensors 22 07300 g002
Figure 3. Imaging system device settings: (a) position of the imaging system in the driving simulator; (b) detail of the imaging system device (visible and thermal camera horizontally aligned and held together by means of a 3d-printed support).
Figure 3. Imaging system device settings: (a) position of the imaging system in the driving simulator; (b) detail of the imaging system device (visible and thermal camera horizontally aligned and held together by means of a 3d-printed support).
Sensors 22 07300 g003
Figure 4. Processing of thermal and visible videos. (a) Software interface for the acquisition and processing of visible and thermal IR videos. (b) Thermal image with ROI drawn in red colors (Nosetip and Glabella); (c) thermal signal extracted from the two ROIs during the experimental phases. The values are obtained subtracting the mean value of the signals during the baseline phase.
Figure 4. Processing of thermal and visible videos. (a) Software interface for the acquisition and processing of visible and thermal IR videos. (b) Thermal image with ROI drawn in red colors (Nosetip and Glabella); (c) thermal signal extracted from the two ROIs during the experimental phases. The values are obtained subtracting the mean value of the signals during the baseline phase.
Sensors 22 07300 g004
Figure 5. Whiskers plot of the participants’ scores in DST (a) and RAVLT (b). Outliers are represented with red crosses.
Figure 5. Whiskers plot of the participants’ scores in DST (a) and RAVLT (b). Outliers are represented with red crosses.
Sensors 22 07300 g005
Figure 6. Thermal features relative to Nosetip ROI extracted during DST (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 6. Thermal features relative to Nosetip ROI extracted during DST (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g006
Figure 7. Thermal features relative to Glabella ROI extracted during DST (** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 7. Thermal features relative to Glabella ROI extracted during DST (** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g007
Figure 8. HRV features extracted during DST. Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 8. HRV features extracted during DST. Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g008
Figure 9. Thermal features relative to Nosetip ROI extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 9. Thermal features relative to Nosetip ROI extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g009
Figure 10. Thermal features relative to Glabella ROI extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 10. Thermal features relative to Glabella ROI extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g010
Figure 11. HRV features extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Figure 11. HRV features extracted during RAVLT (* p < 0.05; ** p < 0.01). Outliers are represented with red crosses. The titles of the single plots refer to abbreviations of features described in Section 3.3.
Sensors 22 07300 g011
Figure 12. Scheme of classification adopted in the present work: (a) scheme of the two-level classification model for DST; (b) scheme of the three-level classification model for RAVLT.
Figure 12. Scheme of classification adopted in the present work: (a) scheme of the two-level classification model for DST; (b) scheme of the three-level classification model for RAVLT.
Sensors 22 07300 g012
Figure 13. Performances of the SVM classifiers for DST: (a) ROC curve for unimodal IR features-based classifier; (b) confusion matrix unimodal IR features-based classifier; (c) ROC curve for multimodal IR + HRV features-based classifier; (d) confusion matrix multimodal IR + HRV features-based classifier.
Figure 13. Performances of the SVM classifiers for DST: (a) ROC curve for unimodal IR features-based classifier; (b) confusion matrix unimodal IR features-based classifier; (c) ROC curve for multimodal IR + HRV features-based classifier; (d) confusion matrix multimodal IR + HRV features-based classifier.
Sensors 22 07300 g013
Figure 14. Performances of the Ensemble bagged trees for unimodal IR features-based classifiers for RAVLT: (a) ROC curve for the classifier of class ImmR (i.e., class 0) vs. cumulative class (DelR + Rec) (i.e., class 1 + 2); (b) ROC curve for the classifier of class DelR (i.e., class 1) vs. cumulative class (ImmR + Rec) (i.e., class 0 + 2); (c) ROC curve for the classifier of class Rec (i.e., class 2) vs. cumulative class (ImmR + DelR) (i.e., class 0 + 1); (d) confusion matrix for the unimodal IR features-based classifier.
Figure 14. Performances of the Ensemble bagged trees for unimodal IR features-based classifiers for RAVLT: (a) ROC curve for the classifier of class ImmR (i.e., class 0) vs. cumulative class (DelR + Rec) (i.e., class 1 + 2); (b) ROC curve for the classifier of class DelR (i.e., class 1) vs. cumulative class (ImmR + Rec) (i.e., class 0 + 2); (c) ROC curve for the classifier of class Rec (i.e., class 2) vs. cumulative class (ImmR + DelR) (i.e., class 0 + 1); (d) confusion matrix for the unimodal IR features-based classifier.
Sensors 22 07300 g014
Figure 15. Performances of the Ensemble bagged trees for multimodal IR + hrv features-based classifiers for RAVLT: (a) ROC curve for the classifier of class ImmR (i.e., class 0) vs. cumulative class (DelR + Rec) (i.e., class 1 + 2); (b) ROC curve for the classifier of class DelR (i.e., class 1) vs. cumulative class (ImmR + Rec) (i.e., class 0 + 2); (c) ROC curve for the classifier of class Rec (i.e., class 2) vs. cumulative class (ImmR + DelR) (i.e., class 0 + 1); (d) confusion matrix for the multimodal IR features-based classifier.
Figure 15. Performances of the Ensemble bagged trees for multimodal IR + hrv features-based classifiers for RAVLT: (a) ROC curve for the classifier of class ImmR (i.e., class 0) vs. cumulative class (DelR + Rec) (i.e., class 1 + 2); (b) ROC curve for the classifier of class DelR (i.e., class 1) vs. cumulative class (ImmR + Rec) (i.e., class 0 + 2); (c) ROC curve for the classifier of class Rec (i.e., class 2) vs. cumulative class (ImmR + DelR) (i.e., class 0 + 1); (d) confusion matrix for the multimodal IR features-based classifier.
Sensors 22 07300 g015
Table 1. Summary of the most related work with research field, measured variables, methodological approach and performances reported.
Table 1. Summary of the most related work with research field, measured variables, methodological approach and performances reported.
AuthorsResearch FieldMeasured VariablesMethodological ApproachPerformance
Kang et al. [23] Military training monitoring
  • thermal IR imaging
  • Modified Cooper Harper Scale ratings
  • Reaction Time
  • Repeated measure ANOVA
  • Correlation
  • Nose temperature differs among experimental phases;
  • Significant correlation among all the measured variables
r 0.88 , 0.96
Stemberger et al. [24]Aviator training monitoring
  • thermal IR imaging
  • cognitive stress test
  • Repeated measure ANOVA
  • Correlation
  • Artificial Neural Network
  • Significant change in reaction time as a function of workload level (F(2) = 25.659, p < 0.001)
  • Negative relationship between task difficulty andpercentage of correct responses (r(33) = −0.64, p < 0.001)
  • 81% correct classification rate
Wang et al. [33]Thermal comfort and workload indoor
  • thermal IR imaging
  • EEG
  • environmental thermostat
  • Repeated measure ANOVA
  • Correlation
  • Random Forest classifier
  • Average prediction accuracy for all subjects under the slightly cool, neutral, and slightly warm environment is 45% ± 9%, 57% ± 9%, and 44% ± 9%, respectively (prediction of IR features on EEG features)
  • Stronger correlations between absolute facial skin temperature and mental workload are found in the neutral environment, compared to the slightly cool and slightly warm environments.
Or and Duffy [34]Car driver monitoring
  • thermal IR imaging
  • Modified Cooper-Harper scale rating
  • Repeated measure ANOVA
  • Correlation
  • The workload tasks had no significant effect on forehead temperature
  • Nose temperature showed a significant change after completing tasks for all conditions
  • Significant correlation between the nose skin temperature change and the subjective workload score (r = 0.32, p = 0.009)
Pavlidis et al. [32]Car driver monitoring
  • thermal IR imaging (perinasal signal to evaluate sympathetic activity)
  • NASA Task Load Index (TLX)
  • steering angle and maximum right-side/left-side lane departure
paired t-testsMean sympathetic arousal and mean steering performance during cognitive workload had significant deterioration with respect to no-stressor driving (p << 0.01)
Perpetuini et al. [12] Car driver monitoring
  • thermal IR imaging
  • fNIRS
SVM classifierSensitivity of 77% and specificity of 69%
Table 2. Participants’ scores statistics.
Table 2. Participants’ scores statistics.
DSTRAVLT
Forward DSTBackward DSTImmRDelRRec
Mean7.196.1541.698.9243.08
Standard Deviation2.002.139.713.033.74
Table 3. Selected features after wrapping method for each of the developed models.
Table 3. Selected features after wrapping method for each of the developed models.
Unimodal IR FeaturesUnimodal HRV FeaturesMultimodal IR + HRV Features
DSTNosetip:
STD; S; SampEn
Glabella:
K; S; LF; HF
RRmean
SDNN
HF
Nosetip:
STD; S; SampEn
Glabella:
K; S; LF; HF
HRV:
RRmean; SDNN; HF
RAVLTNosetip:
STD; K; S; 90thP; SampEn; LF/HF; LF; HF;
Glabella:
MeanTemp; K; S; SampEn; LF/HF; LF; HF;
RRmean; RMSSD; LF/HFNosetip:
STD; K; S; 90thP; SampEn; LF/HF; LF; HF;
Glabella:
MeanTemp; K; S; SampEn; LF/HF; LF; HF;
HRV:
RRmean; RMSSD; LF/HF
Table 4. Accuracy of the ML classifier for unimodal IR, unimodal HRV and multimodal IR + HRV features-based models. The models with the best accuracy are highlighted in bold.
Table 4. Accuracy of the ML classifier for unimodal IR, unimodal HRV and multimodal IR + HRV features-based models. The models with the best accuracy are highlighted in bold.
Unimodal IR FeaturesUnimodal HRV FeaturesMultimodal IR + HRV Features
DSTRAVLTDSTRAVLTDSTRAVLT
Decision Tree
Simple46.259.246.244.748.156.6
Medium50.059.250.039.548.153.9
Complex50.059.250.39.548.153.9
Discriminant Analysis
Linear69.256.659.632.963.555.3
Quadratic63.553.944.228.959.657.9
Logistic Regression69.2-50.0-73.1-
Support Vector Machine
Linear73.165.850.028.973.163.2
Quadratic65.453.942.326.361.551.3
Cubic51.956.651.935.561.551.3
K Nearest Neighbor
Coarse48.134.248.134.241.834.2
Medium55.847.450.031.657.744.7
Fine59.655.350.044.755.850.0
Ensemble
Bagged trees53.871.153.844.759.675.0
Subspace discriminant63.556.65032.959.656.6
Subspace kNN55.856.648.144.757.756.6
RUSboosted trees55.847.448.147.446.235.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cardone, D.; Perpetuini, D.; Filippini, C.; Mancini, L.; Nocco, S.; Tritto, M.; Rinella, S.; Giacobbe, A.; Fallica, G.; Ricci, F.; et al. Classification of Drivers’ Mental Workload Levels: Comparison of Machine Learning Methods Based on ECG and Infrared Thermal Signals. Sensors 2022, 22, 7300. https://doi.org/10.3390/s22197300

AMA Style

Cardone D, Perpetuini D, Filippini C, Mancini L, Nocco S, Tritto M, Rinella S, Giacobbe A, Fallica G, Ricci F, et al. Classification of Drivers’ Mental Workload Levels: Comparison of Machine Learning Methods Based on ECG and Infrared Thermal Signals. Sensors. 2022; 22(19):7300. https://doi.org/10.3390/s22197300

Chicago/Turabian Style

Cardone, Daniela, David Perpetuini, Chiara Filippini, Lorenza Mancini, Sergio Nocco, Michele Tritto, Sergio Rinella, Alberto Giacobbe, Giorgio Fallica, Fabrizio Ricci, and et al. 2022. "Classification of Drivers’ Mental Workload Levels: Comparison of Machine Learning Methods Based on ECG and Infrared Thermal Signals" Sensors 22, no. 19: 7300. https://doi.org/10.3390/s22197300

APA Style

Cardone, D., Perpetuini, D., Filippini, C., Mancini, L., Nocco, S., Tritto, M., Rinella, S., Giacobbe, A., Fallica, G., Ricci, F., Gallina, S., & Merla, A. (2022). Classification of Drivers’ Mental Workload Levels: Comparison of Machine Learning Methods Based on ECG and Infrared Thermal Signals. Sensors, 22(19), 7300. https://doi.org/10.3390/s22197300

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop