Next Article in Journal
Optimal Path Routing Protocol for Warning Messages Dissemination for Highway VANET
Next Article in Special Issue
A Meta-Analysis on Remote HRI and In-Person HRI: What Is a Socially Assistive Robot to Do?
Previous Article in Journal
A Privacy-Preserved ID-Based Secure Communication Scheme in 5G-IoT Telemedicine Systems
Previous Article in Special Issue
A New Robotic Endoscope Holder for Ear and Sinus Surgery with an Integrated Safety Device
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Investigating Methods for Cognitive Workload Estimation for Assistive Robots

1
Department of Computer Science, Tufts University, Medford, MA 02155, USA
2
Department of Electrical and Computer Engineering, Tufts University, Medford, MA 02155, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(18), 6834; https://doi.org/10.3390/s22186834
Submission received: 28 July 2022 / Revised: 29 August 2022 / Accepted: 5 September 2022 / Published: 9 September 2022
(This article belongs to the Special Issue Assistive Robots for Healthcare and Human-Robot Interaction)

Abstract

:
Robots interacting with humans in assistive contexts have to be sensitive to human cognitive states to be able to provide help when it is needed and not overburden the human when the human is busy. Yet, it is currently still unclear which sensing modality might allow robots to derive the best evidence of human workload. In this work, we analyzed and modeled data from a multi-modal simulated driving study specifically designed to evaluate different levels of cognitive workload induced by various secondary tasks such as dialogue interactions and braking events in addition to the primary driving task. Specifically, we performed statistical analyses of various physiological signals including eye gaze, electroencephalography, and arterial blood pressure from the healthy volunteers and utilized several machine learning methodologies including k-nearest neighbor, naive Bayes, random forest, support-vector machines, and neural network-based models to infer human cognitive workload levels. Our analyses provide evidence for eye gaze being the best physiological indicator of human cognitive workload, even when multiple signals are combined. Specifically, the highest accuracy (in %) of binary workload classification based on eye gaze signals is 80.45 ∓ 3.15 achieved by using support-vector machines, while the highest accuracy combining eye gaze and electroencephalography is only 77.08 ∓ 3.22 achieved by a neural network-based model. Our findings are important for future efforts of real-time workload estimation in the multimodal human-robot interactive systems given that eye gaze is easy to collect and process and less susceptible to noise artifacts compared to other physiological signal modalities.

1. Introduction

Assistive robots are envisioned to be deployed in many societal contexts, ranging from co-robots in warehouses and factories to socially assistive robots in health and eldercare settings. While the robots’ tasks will differ depending on the applications, what all of these contexts have in common is the need for the robot to understand the human interactants’ overall cognitive state in order to be able to provide effective help when it is needed. A person who is already busy with an important task and thus potentially overloaded, for example, should not be further stressed by having to answer questions from a robot or having to intervene in a robot’s task, especially if the robot’s task is not all that urgent. Higher cognitive workloads lead to worse human performance [1], and robots causing decreased human performance could quickly be rejected by humans. Thus, understanding when to best communicate or act so as to not overburden the human is a critical component of effective human-robot interaction.
However, there are several challenges involved in assessing human workload. For one, subjective measures such as the NASA-TLX that require humans to answer questions are typically not applicable (as they would exacerbate workload, although they could still be useful after an interaction for an artificial agent to learn about what contexts typically cause high workload). Objective measures utilizing physiological sensors and signal modalities (e.g., [2,3,4,5,6,7]), on the other hand, are not applicable in contexts where humans cannot be instrumented with sensors or they might not yield useful results (e.g., when physiological signals are subjected to significant noise due to motion artifacts). Moreover, even when recording conditions are ideal, it is still unclear which sensing modality and machine learning model to use and whether combining multiple modalities will provide better workload estimates compared to single modalities.
The goal of this paper was to evaluate different machine learning models using multiple physiological signals to determine the combination that yields the best workload estimates. For this purpose, we utilized eye gaze, EEG, arterial blood pressure (heart rate variability (HRV) and blood pressure variability (BPV)) data and their combinations from a multi-tasking simulated driving experiment ([8]; see also our work [9] for some initial analyses that are included here as well). In this experiment, three levels of cognitive workload were experimentally generated by requiring the participants (drivers) to perform secondary tasks such as braking and dialogue interactions at the same time as the primary driving task which induces higher workload levels (that have been verified by delayed reactions to secondary tasks). In addition to performing standard statistical analyses including ANOVAs, Tukey’s honestly significant difference (HSD) and Benjamini–Hochberg multiple pairwise tests, we employed five state-of-the-art machine learning models: (a) k-nearest neighbor (k-NN), (b) naive Bayes (NB), (c) random forest (RF), (d) support-vector machines (SVM), and (e) neural network-based models (NNMs) to evaluate their ability to detect the different workload levels. Our numerical results together with statistical analysis indicated that eye gaze, particularly pupil diameter, is the most reliable physiological indicator among all considered physiological parameters in assessing human cognitive workload. Moreover, we found out that fusing multiple physiological signals does not enhance the workload classification performance over just using the eye gaze signal alone.
We thus made the following contributions in this paper:
  • We analyzed data from a multi-modal, multi-task controlled driving simulation environment that involves multimodal interaction of several physiological signal modalities including eye gaze (pupillometry), EEG, and arterial blood pressure.
  • Our numerical results and statistical analysis demonstrated the effectiveness of pupil diameter in assessing cognitive workload. Particularly, our empirical results indicated that pupil diameter outperforms other physiological signals in cognitive workload prediction tasks regardless of the learning models.
  • Additionally, we showed that combining several physiological signal modalities does not provide a substantial increase in workload classification performance compared to using just eye gaze alone. This suggested that eye gaze is a sufficient modality for assessing cognitive workload in interactive, multi-modal, multi-task settings.
The remainder of this paper is organized as follows. Section 2 provides a summary of recent works on workload estimation/classification based on various types of physiological signals. Section 3 describes the details of how the driving simulation scenario is designed and how the signals are collected and processed. The workload evaluation, the details of physiological signal pre-processing steps, and feature extraction are summarized in Section 4. Five state-of-the-art machine learning models and their hyper-parameters are described in Section 5. Finally, we provide the numerical result in Section 6 and conclude in Section 7.

2. Background and Related Work

Cognitive workload is defined as the load imposed on working memory during a cognitive process and is one of the most important factors in designing and implementing robust human-machine systems [10,11]. Assessment of cognitive workload is mainly based on two methods: (a) subjective measurements and (b) objective measurements [2,12]. Subjective measurements use the perceived feeling and self-rating of operators to determine the cognitive workload. For subjective measurements, it is common to predict the cognitive workload by leveraging questions such as the proposed methods in the Administration-Task Load Index [13], Subjective Workload Assessment Technique [14], and Workload Profile [15]. On the other hand, objective measurements are mainly based on the recordings of task performance and physiological signals, such as EEG, heart rate, and eye movement [16,17,18], which we will review below.

2.1. EEG

EEG is considered a common and reliable indicator for assessing the human cognitive workload [2,3,4,19]. Berka et al. [3] recorded EEG signal streams from eighty healthy participants to investigate the association between task engagement and cognitive workload while performing learning and memory tasks. Another study first applied independent component analysis (ICA) to obtain the independent features from EEG, then predicted the cognitive workload based on these features [19]. So et al. [4] used short-term frontal EEG signals collected from twenty healthy volunteers performing four cognitive and motor tasks to evaluate the dynamic changes in mental workload. Although EEG signal can be assessed directly from the electrical activity of cortical and sub-cortical neurons with a very high temporal resolution in milliseconds, it is non-stationary and vulnerable to several types of noise such as blinking, frequency interference, and motion-related and sensor-based artifacts [20]. Several methods have been proposed to handle various types of noise appearing in EEG, including wavelet transform-based techniques, ICA-based methods, and adaptive filtering [21]. Usually, applying multiple denoising methods at the same time is necessary to handle multiple types of noise involved in EEG [21]. However, simultaneously applying several denoising algorithms not only increases the complexity of the denoising process but also leads to uncontrolled interactions between used algorithms [22]. In addition, the heavily used denoising algorithms potentially lead to the loss of relevant information in data. Therefore, denoising EEG signals usually requires practical experiences and sometimes needs manual processing steps [23].

2.2. HRV and BPV

There have been various works that leverage HRV parameters to predict cognitive workload. One study investigated real-time mental workload estimation via HRV parameters collected from 45 subjects [24]. Another study suggested that mental workload causes variations in non-linear HRV dynamics, which leads to increased risk of cardiac disorders [25]. Shakouri et al. [26] examined the correlation between HRV parameters and subjective workload during simulated driving in a NASA-TLX setup. There have been relatively few studies that link the cognitive workload with blood pressure variability (BPV). Stuiver et al. [27] investigated the effects of mental workload change in the BPV parameters collected from fifteen subjects during driving simulation. Hjortskov et al. [28] analyzed the impacts of mental stress with HRV and BPV and proposed that HRV is more sensitive to mental stress than BPV.

2.3. Eye Gaze

Eye gaze is another well-known indicator to assess human workload variations within a short time interval, making it a practical tool for predicting cognitive workload in real-time applications [29]. Considering that eye gaze is easier to collect in daily life situations and is less sensitive to undesired interference and motion artifacts than EEG, it is commonly used in workload estimation tasks [30,31]. Several studies used pupil diameter as a reliable indicator for assessing cognitive workload [30,32,33,34,35,36]. Pfleging et al. examined the correlation between pupil diameter and cognitive effort by adjusting the lighting conditions [33]. In another study, Palinko et al. suggested that eye tracking is one of the best approaches to predict cognitive workload in simulated driving, and pupil diameter is a reliable parameter to infer workload [30]. Bitkina et al. [35] investigated the efficiency of multiple eye tracking parameters including gaze fixation, pointing, and pupil diameter in workload estimation during simulated driving. Pang et al. [36] recorded several eye gaze parameters including fixation duration, blink duration, and pupil diameter to estimate different cognitive workload levels by adjusting the task difficulties. Additionally, some recent works evaluated eye gaze parameters for workload prediction in real-time applications [37,38]. Specifically, Kosch et al. [37] investigated real-time assessment of mental effort in the context of task complexity adaptation, while Appel et al. [38] used the normalized values of several eye gaze parameters such as pupil diameter, blink duration, and the number of blinks for online differentiation of cognitive workload.

2.4. Combination of Multiple Signal Modalities

There are also numerous works that aim to fuse multiple physiological signal types to boost the performance of mental workload estimation. For instance, the studies in [39,40,41] combined eye gaze and EEG signals to jointly predict the cognitive workload. Specifically, Khedher et al. [39] recorded both eye gaze and EEG signals from fifteen students in a virtual learning environment to classify two groups of learners: students who successfully resolved the tasks and students who did not. Additionally, their numerical results showed that the k-nearest neighbor classifier achieves the best classification efficiency over the other considered methods. Rozado and Dunser [40] combined the extracted features from EEG and pupil diameter to assess the cognitive workload [40]. Their results suggested that integrating EEG parameters with pupil diameter provides higher accuracy in workload classification compared to using a single signal type.
Even though combining multiple physiological signals may boost the accuracy of the trained model under some particular scenarios [40,42,43,44], it is not clear whether multi-modality learning, i.e., combining multiple physiological signals, can outperform single-modality learning in predicting the mental workload states. Indeed, some counter-examples appeared in the literature, practically pointing out that using multi-modality learning in workload prediction tasks may be harmful [41,45,46]. For example, Borys et al. [45] studied several combinations of EEG and pupillometry features to conclude that single model learning using eye-tracking data alone could achieve higher accuracy than other fusion-based models in cognitive workload classification tasks. Another study reviewed several research works on workload prediction from single modality and fusion of multiple signal modalities and concluded that most of the research works propose better performance in workload estimation by leveraging single modality compared to the fusion of multiple signal modalities [47]. They reported four studies within six that suggest that combining EEG and eye gaze does not enhance the overall performance in assessing workload over single usage of EEG or pupillometry, which is compatible with our findings.

2.5. Machine Learning for Detecting and Assessing Cognitive Workload

Machine learning models have been extensively used in literature for detecting and assessing cognitive workload. For instance, k-nearest neighbor was used in [39,48,49,50], naive Bayes was used in [51,52], random forest was used in [50,53], support-vector machines were used in [4,54,55,56,57,58], and neural network-based models were used in [49,50,54,59,60,61,62,63,64]. Since the performance of workload classification/prediction depends on learning models, previous works usually applied multiple learning models for detecting and assessing cognitive workload [39,65].
Finally, we refer the readers to recent surveys on using pupillometry, EEG, HRV, BPV, or their combinations for categorizing the mental workload in [16,47,66].

3. Experimental Data

We utilized pupillometry, EEG, and arterial blood pressure data from 80 subjects performing a multi-modal interactive driving task in a fixed-base driving simulator with an open-cab mockup (see [8,9] for details about the experimental setup). Ten braking events were generated, during which a vehicle appeared 200 m in front of the driver, causing the driver to brake quickly to avoid a collision. In addition, drivers had to answer 20 questions roughly every 30 to 60 s. Dialogue and braking events were arranged such that the time interval between the onset of the braking event and the end of the question, i.e., the “stimulus onset asynchrony” (SOA), was between −1 and +1 s with a step size of 0.5 s, for a total of five SOA conditions: SOA values of −1 and −0.5 represent cases where the braking event occurs 1 and 0.5 s after the completion of the question, respectively, while SOA values of 1 and 0.5 represent the cases where the braking event takes place 1 and 0.5 s before the completion of the question, respectively, and an SOA of 0 indicates that the braking event begins at the same time as the question ends.

3.1. Eye Gaze Recording

Eye gaze data was recorded with a Pupil Core (Pupil Labs, Berlin, Germany) eye tracker to obtain the pupillometry signal with a sampling rate of 400 Hz, which contains a 200 Hz binocular camera and a 120 Hz world camera. We performed three-step pre-processing to remove any out-of-band noise, sensory interference, and blink artifacts. First, we removed the signal segments with an amplitude of lower than 0.8 mm or greater than 10 mm by considering that the signal parts with an amplitude of lower than 0.8 mm are potential blink artifacts [67] and the quantitative pupil dilation widens up to 10 mm [68]. Second, we applied linear interpolation to fix the extracted parts [67]. Finally, we leveraged a fifth-order Butterworth low-pass filter with a cutoff frequency of 10 Hz to eliminate baseline wander [69]. Figure 1 illustrates the pre-processing steps of the pupillometry signal.

3.2. EEG Recording

EEG data was collected with an Enobio (Neuroelectrics, Cambridge, MA, USA) system with a 24-bit resolution and a sampling rate of 500 Hz with 3.14 cm 2 silver/silver chloride electrodes located corresponding to the international 10-10 system (FC1, FC2, FC5, FC6, CP1, CP2, CP5, and CP6). To pre-process the EEG signals, we applied a three-stage denoising method. First, we applied a sixth-order Butterworth band-pass filter between 0.1 Hz and 30 Hz to eliminate the out-of-band noise [70,71]. Second, we used independent component analysis (ICA) to decompose the mixture of signal streams to its hidden components. ICA is implemented with the presumption that the latent components are non-Gaussian and statistically independent, and it is an effective tool for extracting the blink artifacts from EEG signals [72]. We identified the blink artifacts with the instantaneous spikes in the amplitude and manually removed the ICA components, which are correlated with blink artifacts. Third, we applied Kalman smoothing, which is a prominent method to predict the state of dynamic linear structures in the presence of noise [73]. To achieve this, we utilized the Python library “Tsmoothie” [74] to smooth the EEG signals.

3.3. Arterial Blood Pressure Recording

A NIBP100D non-invasive fingertip plethysmography system was attached to participants’ left hands to collect the arterial blood pressure streams with a sampling rate of 20 Hz (BIOPAC Systems, Goleta, CA, USA). We pre-processed the arterial blood pressure (ABP) signal with three steps described in Figure 2. First, the ABP signal was upsampled to 250 Hz. Then, a Butterworth low-pass filter with a cutoff frequency of 5 Hz and an order of 5 was applied to eliminate the out-of-band noise. Finally, the systolic pressures (SP) and diastolic pressures (DP) were identified by applying both an amplitude threshold and a beat-to-beat duration threshold based on the behavior of cardiac dynamics. To achieve these, the candidates of SP and DP were obtained by taking the local maxima and local minima points from the ABP signal, respectively. The candidate SP points were evaluated as the potential heartbeats by considering that SP points taken from arterial blood pressure signals are reliable indicators to calculate heart rate and interbeat interval (IBI) [75,76,77]. Here, IBI represents the time interval between two consecutive heartbeats. Then, the amplitude threshold, which represents the threshold for the difference between the sequential DP and SP points, was determined as 10 mmHg. In other words, the subsequent DP and SP points whose amplitudes differ more than 10 mmHg were identified as possible SP and DP points. Next, minimum and maximum IBI values were determined to be 30 and 120 beats per minute (bpm), respectively. The intuition behind this selection was that the resting heart rate (HR) values of a normal adult remain within 60 bpm and 90 bpm [78]. We used a wider range of HR to eliminate the false detection of HR values based on subject-based variations.

4. Methods

4.1. Workload Evaluation

We generated three cognitive workload levels by combining the single driving event with additional tasks such as braking and dialogue: “Level 0” represents the baseline (i.e., the lowest level of cognitive workload while only driving without any additional task); “Level 1” represents the combination of the driving task and the dialogue task, where subjects had to answer questions like “What is your favorite color?”, “What pets do you have?”, “What sports do you watch?”, or “What type of movies do you like?”; and “Level 2” represent driving with simultaneous dialogue and braking events. To construct a reliable dataset for cognitive workload estimation/classification, we discarded the samples that were too noisy, leading to a balanced dataset of 348 samples, with each workload level containing 116 samples (from 43 healthy volunteers).
We used three SOA types within five to generate our dataset. Specifically, we did not include the SOA types of −1 and −0.5, which represent the case that the dialogue event ends before the onset of the braking event. The assumption behind this selection was that in the case of SOA events −1 and −0.5, the braking event starts after the completion of the question when the participant already thinks about the question and possibly generates the response in mind. Thus, the potential increase in cognitive workload based on the combination of braking and dialogue tasks would reduce. We, therefore, only leveraged the SOA types of 0, 0.5, and 1 to generate Level 2 in this study.

4.2. Pupillometry

We used the left pupillometry signal to obtain pupil dilation, the fixation count, and the fixation duration, assuming that the left and right pupil dilations are synchronous. We calculated percentage change in pupil size (PCPS) to avoid subject-based variations in pupil diameter by considering the range of pupil diameter fluctuations from person to person. The PCPS is calculated using the following equation [79]:
PCPS = Current value of pupil diameter Baseline value of pupil diameter Baseline value of pupil diameter × 100 % ,
Here, we determined the baseline measure of the pupil diameter by calculating the average value of a 1 s signal segment before the onset of the stimulus. We took 2.5 s time frames from pre-processed pupillometry signals for each workload level to determine PCPS values. Next, we calculated the average PCPS (APCPS) as follows [79]:
APCPS = 1 M i = 1 M P C P S i ,
where P C P S i represents the percentage change in pupil size for the ith sample, and M denotes the total number of samples in the time domain. In addition to PCPS and APCPS, we also obtained the number of fixations and calculated the mean of fixation duration within 2.5 s time windows.

4.3. Electroencephalography

To perform the statistical analysis of EEG, we calculated the power spectral density (PSD) of EEG streams to obtain the power distribution of the EEG signals in the frequency domain. To calculate the estimate of the PSD, we used Welch’s method, which partitions the data into segments to compute the estimation of the PSD by averaging the periodograms over segments [80,81]. To achieve this, we used non-overlapping 2 s windows with a sampling rate of 500 Hz. Then, we computed the absolute power corresponding to different sub-bands of EEG by defining the frequency ranges of each frequency band and approximating the area under the PSD curve. Based on that, we measured the average power of EEG signal from four frequency bands: 1 Hz to 4 Hz ( δ ), 4 Hz to 8 Hz ( θ ), 8 Hz to 12 Hz ( α ), and 12 Hz to 30 Hz ( β ), respectively. Here, based on the band-pass filter between 0.1 Hz and 30 Hz, which is suggested to eliminate the out-of-band noise from EEG signals [70,71] and given that within five frequency bands of EEG, δ , θ , and β are the most commonly used sub-bands for workload estimation [82,83,84], we excluded the gamma sub-band, which contains the frequencies higher than 30 Hz.

4.4. Arterial Blood Pressure

In this study, we considered the systolic pressures (SP) points as the heartbeats and used them to obtain the interbeat interval (IBI). We calculated the heart rate from the estimated IBI with the following equation:
HR = 6 × 10 4 IBI ( ms ) ( bpm ) .
Based on the estimated HR and IBI values, we acquired two types of measurements: (1) heart rate variability (HRV) and (2) blood pressure variability (BPV) from 30 s time windows for each cognitive workload level.

4.4.1. HRV

HRV is the variation in sequential IBIs over time. In this study, five time domain and six frequency domain HRV features were calculated from a 30 s ABP signal to investigate the performance of HRV in the assessment of cognitive workload.
Time Domain HRV Features. The time domain HRV features included the average of IBIs (meanIBI (ms)), the standard deviation of IBIs (SDIBI (ms)), the average of HR values (meanHR (bpm)), the standard deviation of HR values (SDHR (bpm)), and the root-mean-square of successive IBI differences (RMSSD). RMSSD was calculated with the following equation [85]:
RMSSD = 1 M 1 t = 1 M 1 ( I B I t + 1 I B I t ) 2 ( ms ) .
where M and I B I t represent the number of heartbeats over time and the interbeat interval between time t and t + 1 , respectively.
Frequency Domain HRV Features. First, the cubic spline interpolation was used to interpolate the sequential IBI values. Then, the fast Fourier transform (FFT) was applied to obtain the sample frequencies and the frequency components were calculated from the one-side frequency range of FFT. Six frequency domain features were taken, which include very low frequency (VLF) power (in ms 2 ), low frequency (LF) power (in ms 2 ), high frequency (HF) power (in ms 2 ), total power (in ms 2 ), HF normalized unit (HFnu) (%), and LF normalized unit (LFnu) (%). VLF, LF, and HF power represent the power of the frequency components of IBI variation lower than 0.04 Hz, between 0.04–0.15 Hz, and between 0.15–0.4 Hz, respectively [86]. The total power is the sum of the powers from VLF, LF, and HF.
LFnu and HFnu were formally defined as follows:
LFnu = LF LF + HF × 100 ( % )
HFnu = HF LF + HF × 100 ( % )

4.4.2. BPV

BPV is defined as the constant and dynamic variations in both systolic and diastolic blood pressures and was examined with different categories, such as long-term BPV (visit-to-visit), mid-term BPV (day-to-day), short-term BPV (over 24 h), and very short-term BPV (beat-to-beat) [87]. In this study, we took a 30 s ABP signal and calculated eight time domain BPV features together with twelve frequency domain BPV features to assess the very short-term BPV.
Time Domain BPV Features. We calculated four time domain BPV features, including the standard deviation (SD), successive variation (SV), coefficient of variation (CV), and average real variability (ARV). The definitions of time domain features were expressed with the following equations:
SD = 1 N 1 t = 1 N ( x t x ¯ ) ( mmHg ) ,
SV = 1 N 1 t = 1 N 1 ( x t + 1 x t ) 2 ( mmHg ) ,
CV = SD x ¯ × 100 ( % ) ,
ARV = 1 N 1 t = 1 N | x t + 1 x t | ( mmHg ) ,
where N, x t , and x ¯ represent the number of blood pressure values in time domain, the blood pressure value at time t, and the mean blood pressure, respectively [88].
Frequency Domain BPV Features. To determine the frequency domain BPV features, the cubic spline interpolation was applied to interpolate the SP and DP values. Next, the sample frequencies were determined via FFT, and a one-side frequency range of FFT was taken to calculate the frequency components. We calculated six frequency domain features of BPV, which contain very low frequency (VLF) power (in mmHg 2 ), low frequency (LF) power (in mmHg 2 ), high frequency (HF) power (in mmHg 2 ), total power (in mmHg 2 ), LFnu (%), and HFnu (%). Each feature was calculated for both SP and DP (12 features in total). VLF power, LF power, and HF power represent the power of the frequency components of blood pressure variation lower than 0.04 Hz, between 0.04–0.15 Hz, and between 0.15–0.4 Hz, respectively [89]. The total power is the sum of VLF, LF, and HF power [89]. LFnu and HFnu represent the LF and HF of blood pressure fluctuation in normalized units (both for SP and DP) [90], respectively. We used Equations (5) and (6) to calculate LFnu and HFnu for BPV.

4.5. Extracted Features of Pupillometry and EEG

4.5.1. Power Spectral Density of EEG

One of the most widely used extracted features from EEG is power spectral density (PSD) [91,92,93,94]. PSD quantifies the amount of power distributed in each frequency band of EEG signal [95]. From the pre-processed EEG data, we extracted the PSD by measuring the power in four standard frequency bands θ , α , δ , and β , which was explained in detail in Section 4.3. Because each EEG sample has a length of 2.5 s, we used a periodogram with a non-overlapping rectangular window size of 2.5 s to estimate the average spectral power in our experiment. The average spectral power was then integrated over each EEG frequency band to generate the PSD data [93,94]. From eight EEG channels and four frequency bands, each PSD sample corresponded to a 32-dimensional vector.

4.5.2. Extracted Feature for Pupillometry

Mean and variance have been considered as the most commonly extracted features for pupillometry signals in the literature [96,97,98]. Therefore, in this paper, we decided to compute the mean and variance of PCPS, which was taken from the pupillometry signal explained in Section 4.2, to generate its two-dimensional vector extracted features.

4.5.3. Combinations of Extracted Features

To explore the effectiveness of the combination of multiple physiological signals in predicting workload levels, the extracted features of pupillometry and EEG were concatenated as the inputs for our learning models. Particularly, by concentrating the extracted features from pupillometry and EEG, each combined sample corresponds to a 34-dimensional vector. These combined signals will be used later in Section 6.2.2.

5. Machine Learning Methodologies

In this section, we employed different machine learning models to predict the cognitive workload under two distinct settings: (a) single-modality, i.e., using a single physiological signal as the input of learning models, and (b) multiple-modality, i.e., combining multiple physiological signals as the input of learning models. Since the models matter in assessing the classification performance on different types of data, we employed five machine learning models named k-nearest neighbor (k-NN), naive Bayes (NB), random forest (RF), support-vector machines (SVM), and a neural network-based model (NNM) to provide a better comparison. It is worth noting that these selected learning models have been widely used in literature for cognitive workload classification tasks. For instance, k-NN was used in [39,48,49,50], NB was used in [51,52], RF was used in [50,53], SVM was used in [4,54,55,56,57], and NNM was used in [49,50,59,60].
For model training and testing, we divided each dataset into two subsets, the training set and the test set, with the ratio of 80% and 20%, respectively. For hyper-parameter tuning, we utilized five-fold cross-validation [99] for all non-neural network-based models. Particularly, the training set was first split into five equally sized groups. Within these five groups, one group was alternatively selected as the validation set, while the rest of the four groups were considered as four sub-training sets. The model was trained on four sub-training sets and tested on the validation set five times until all groups were selected once as the validation set. The model that achieved the highest accuracy averaged over five validation sets was selected as the best model. After the determination of hyper-parameters from the best model, we trained from scratch using all training data and tested on the testing set to obtain the final classification accuracy. For the neural network-based model (NNM), we kept the setting the same except that we did not apply five-fold cross-validation. We performed model selection based on the validation accuracy produced by a fixed validation set and repeated the above procedure five times for all models with different random seeds. The details of our learning models are described below.

5.1. k-Nearest Neighbor

k-nearest neighbor is a supervised learning method that outputs the label of the tested sample based on the labels of its k-nearest neighbors, i.e., the k closest samples (from the training set) to the tested sample. In this paper, we consider the number of nearest neighbors k as a tunable hyper-parameter which will be scanned in the range of [ 1 , 30 ] .

5.2. Naive Bayes

Naive Bayes is a supervised learning technique that relies on the well-known Bayes’ theorem together with a simple (naive) assumption that each input (sample) conditioned on its label is independent of each other. In the naive Bayes algorithm, the conditional distribution of the data given label can be modeled using five common probabilistic models named Gaussian naive Bayes, multinomial naive Bayes, complement naive Bayes, Bernoulli naive Bayes, and categorical naive Bayes. For example, in the Gaussian naive Bayes model, it is assumed that the conditional distribution of the data given label follows the Gaussian distribution. Because the conditional distribution of data and its label varies in practice, we consider the probabilistic models as tunable hyper-parameters.

5.3. Random Forest

Random forest is a supervised learning method that combines multiple decision trees trained from different samples and subsequently takes their majority vote to output to determine final estimation. The number of decision trees (or the number of predictors) used in the random forest method is considered a tunable hyper-parameter. Here, we select the number of decision trees in the range of [ 1 , 50 ] .

5.4. Support-Vector Machine

A support-vector machine is a supervised learning method that maps training samples from the input space into the latent space to maximize the width of the gap between different categories (groups of samples having the same label). To deal with non-linear datasets, a kernel trick is usually applied to SVM to obtain an optimal projection such that the data is linearly separable in the projected space. Here, we considered the types of kernel function as tunable hyper-parameters and selected the best kernel among four common kernel functions named linear, polynomial, radial basis, and sigmoid.

5.5. Neural Network-Based Models (NNM)

Deep neural networks have been suggested to be efficient in extracting task-related characteristics and thus have been widely used in literature for cognitive workload classification [49,50,59,60,100,101,102]. Here, we mainly used multi-layer perceptron (MLP) neural networks to classify the cognitive workload for all types of signals. In addition, we utilize the recent developed time-series neural network models such as EEGNet [102] for EEG, multivariate long short-term memory fully convolutional network (MLSTM-FCN) [103] and InceptionTime [104] for both EEG and pupillometry. Details of MLP, EEGNet, MLSTM-FCN, and InceptionTime models are described below.

5.5.1. Multi-Layer Perceptron

In this study, we used a multi-layer perceptron neural network composed of three fully connected layers followed by a linear layer with the output dimension of T 1 T 2 T 3 T 4 n , where T 1 , T 2 , T 3 , T 4 denote the size of the first, the second, the third fully connected layer, and the last linear layer, respectively, while n denotes the number of label classes, i.e., the number of workload levels that need to be classified. We used rectified linear unit (ReLU) [105] as the activation function. To stabilize the training process and avoid overfitting, we applied batch normalization and dropout in each fully connected layer. For EEG and pupillometry, we set ( T 1 , T 2 , T 3 , T 4 ) = ( 2048 , 1024 , 512 , 64 ) and ( T 1 , T 2 , T 3 , T 4 ) = ( 512 , 256 , 128 , 128 ) , respectively. For other physiological parameters, such as fixation number, fixation duration, heart rate variability, and blood pressure variability, we used ( T 1 , T 2 , T 3 , T 4 ) = ( 32 , 32 , 16 , 16 ) . The architecture of our MLP neural network can be found in Table 1.

5.5.2. Additional Time-Series Learning Models

Motivated by the recently developed time-series neural network models in [102,103,104], we want to apply EEGNet [102] for EEG and use MLSTM-FCN [103] and InceptionTime [104] for both EEG and pupillometry.
EEGNet. In this paper, we adapted the architecture of EEGNet in [102] composed of three convolutional layers followed by a classifier to our EEG signal. Since EEGNet was specially designed for EEG only, we did not employ EEGNet for other physiological signals. As recommended in [102], the length of the temporal convolution of the first layer in EEGNet was set to 250, which is exactly half of the sampling rate of EEG (500 samples per second).
MLSTM-FCN. In this paper, we applied a multivariate long short-term memory fully convolutional network (MLSTM-FCN) to both EEG and pupillometry signals. The MLSTM-FCN model was built from two blocks: a long short-term memory (LSTM) block and a fully convolutional network (FCN) block [103]. The outputs from the FCN block and LSTM block were combined and fed into a linear classifier to produce the final classification output.
In our implementation, we kept all model parameters the same as the settings in [103]. However, to adapt to different types of signals, we changed the number of LSTM cells to eight for EEG and to one for pupillometry without applying a grid search.
InceptionTime. In this paper, we applied the InceptionTime learning model proposed in [104] to both EEG and pupillometry signals. The InceptionTime model was composed of five Inception networks. To stabilize the training model, each Inception network was initialized randomly [104]. In addition, the Inception network has integrated the ResNet modules [106] with the inception modules using different lengths of filters for diverse feature extraction [107]. The bottleneck layers [106] were further applied to reduce the model complexity and sped up the training process. In our implementation, we kept all the parameters of the model the same as the setting in [104].
It is worth noting that to stabilize the training process and avoid overfitting, batch normalization and dropout were employed for all neural network-based models including MLP, EEGNet, MLSTM-FCN, and InceptionTime. Finally, for hyper-parameter tuning, the Adam optimizer [108] was applied with the learning rate selected from { 1 × 10 5 , 5 × 10 5 , 1 × 10 4 , 5 × 10 4 } . The batch size was picked from { 32 , 64 , 128 , 256 } . For all selected pairs of batch size and learning rate, the number of epochs was always set to 300.

6. Results and Discussion

To evaluate the effectiveness of different physiological signals in assessing the mental workload levels, we first performed a statistical analysis of the performance of various parameters obtained from different types of signal modalities, such as human gaze parameters (APCPS, fixation number, and fixation duration), EEG, HRV, and BPV. Next, the classification accuracies of several machine learning methodologies in three workload classification tasks (levels 0-1, 0-2, and 1-2) were reported to verify these statistical analyses.

6.1. Statistical Analysis of Physiological Signals

6.1.1. Pupillometry

We examined the variations in three human gaze parameters obtained from pupillometry signals, APCPS, mean fixation number, and mean fixation duration, for three workload levels. Figure 3 shows the change in the APCPS for eight samples obtained from different participants and the change in the mean APCPS over all samples within a 2.5 s time interval for various workload levels. The starting point represents the onset of baseline, dialogue, and SOA events for Level 0, Level 1, and Level 2, respectively. Our results indicate that a stable pattern is observed during Level 0, while the APCPS signal has distinctive patterns during Level 1 and Level 2 for multiple samples. Moreover, the mean APCPS over all samples remains around 0% for the baseline event, while it increases up around 10% and 15% for the dialogue and the SOA events, respectively. These results demonstrate the capability of the APCPS signal in differentiating different workload levels.
Next, we calculated the mean APCPS, mean fixation number, and fixation duration over a 2.5 s time interval for each sample and investigated the outcomes to assess the change in these three human gaze parameters for various workload levels. To validate the direct correlation between mental effort and human gaze parameters, we performed a one-way ANOVA test for each human gaze parameter [109]. Table 2 illustrates the mean, median, standard deviation, and standard error of APCPS, average fixation number, and average fixation duration over 116 samples for three workload levels. Figure 4 also depicts the boxplot of the mean APCPS over 116 samples for three cognitive workload levels. The results demonstrate that the mean of the APCPS significantly increases with increased mental effort. The mean of the average fixation number is almost the same for Level 0 and Level 1, while it rises noticeably from Level 1 to Level 2. The mean of the average fixation duration decreases significantly from Level 0 to Level 1. However, there is a negligible difference between Level 1 and Level 2.
We further performed Tukey’s “honestly significant difference” (HSD) multiple pairwise comparison test [110] and the Benjamini–Hochberg test [111] to assess the differentiation between the three workload levels by leveraging three human gaze parameters. Table 3 illustrates p-values obtained from Tukey’s HSD and Benjamini–Hochberg tests for all pairs of workload levels at a significance level of 0.95 . The results from both of the tests demonstrate that APCPS is the most efficient parameter within the three human gaze parameters to classify different pairs of workload levels. There is a statistically marked difference between all pairs of workload levels (≪0.05) in the case of using APCPS. On the other hand, the fixation number is capable of differentiating between Levels 0-2 and Levels 1-2; however, it looks like using the fixation number to distinguish between Levels 0-1 is practically impossible due to the small p-values of 0.33 and 0.14 obtained from both Tukey’s HSD and Benjamini–Hochberg tests, respectively. Moreover, although the fixation duration is efficient in classifying Levels 0-1 and 0-2 (p-values ≪0.05), it does not perform well in differentiating Levels 1-2. Based on the results, APCPS is the most efficient method of assessing the cognitive workload among these three morphological characteristics of pupillometry signals.

6.1.2. Electroencephalography

To investigate the performance of EEG signal on workload prediction, we performed ANOVA, Tukey’s HSD multiple pairwise test, and the Benjamini–Hochberg test on four different frequency bands generated from PSD of EEG, δ , θ , α , and β . Table 4 illustrates the p-values of each pair of workloads for different frequency bands. The results demonstrate that within four EEG bands, α and β waves are more efficient in workload prediction and are capable of classifying only two pairs of workload levels, 0-1 and 0-2 (p-values < 0.05). While the θ wave is inefficient in discriminating any pairs of workload levels, the δ wave is able to differentiate the workload levels 1 and 2 with a p-value of 9 × 10 3 based on the Benjamini–Hochberg test. However, it is worth noting that none of the frequency bands is competent for the classification of all pairs of workload levels.

6.1.3. Heart Rate Variability

We performed statistical analysis on five time domain and six frequency domain HRV features to explore the efficiency of HRV features on workload classification. Figure 5 shows the fluctuations of HRV features in the presence of different workload levels. The HRV results indicate that only one time domain feature (SD HR) and two frequency domain features (LFnu and HFnu) have remarkable changes from Level 0 to Level 1. On the other hand, none of the HRV characteristics are capable of discriminating between Level 1 and Level 2. In addition, we run ANOVA Tukey’s HSD multiple pairwise test (with a confidence level of 0.95 ) on each HRV feature for every pair of workload levels to explore the efficiency of HRV features in classifying different pairs of workload levels. The p-values obtained from Tukey’s HSD multiple pairwise test for levels 0-1, 0-2, and 1-2 can be viewed in Table 5. The results demonstrate that only three HRV parameters (SD HR, LFnu, and HFnu) have p-values lower than 0.05 and can differentiate between Level 0 and Level 1, which is compatible with the results presented in Figure 5.

6.1.4. Blood Pressure Variability

We also ran a statistical analysis on blood pressure variability (BPV) parameters to investigate the performance of time domain and frequency domain features of the BPV signal in assessing the cognitive workload levels. Figure 6 illustrates the variations of time domain and frequency domain BPV features with different workload levels. The variations of BPV in time domain indicate that only SD and CV for diastolic pressure can separate the three workload levels, while the rest of the time domain BPV features cannot distinguish between Level 0, Level 1, and Level 2. The variations of BPV in the frequency domain show that LF and LFnu for diastolic pressure increase with increased workload levels, and HFnu for diastolic pressure decreases with increased workload levels. The rest of the frequency domain BPV features do not have compatible patterns with workload degree alteration. Moreover, we also ran ANOVA Tukey’s HSD multiple pairwise tests (with a confidence level of 0.95 ) on each BPV feature for every pair of workload levels. Table 6 indicates the p-values for workload level pairs of 0-1, 0-2, and 1-2 of 20 different BPV features. The results show that among 20 BPV features, only four features (SD DP, CV DP, and LF DP) can differentiate Levels 0-1 and 0-2 (p-value < 0.05 ). Furthermore, we observed that the features obtained from diastolic powers from arterial blood pressure signals are more efficient in classifying between different workload level pairs. There are only three BPV features acquired from systolic power that can differentiate Levels 0-1, which are SD SP (p-value = 0.02 ), SV SP (p-value = 0.04 ), and CV SP (p-value = 0.03 ). Furthermore, any of the BPV features are able to distinguish between Level 1 and Level 2.

6.2. Classification Performance

As previously discussed in Section 5, in this section, we employed five different machine learning models to evaluate the performance of mental workload classification under (a) single-modality settings and (b) multi-modality settings. To shorten the notations, we used PCPS to denote the percentage change in pupil size of pupil diameter signal, BPV to denote blood pressure variability parameters, HRV to denote heart rate variability parameters, and FIX to denote fixation parameters including fixation number and fixation duration.

6.2.1. Single-Modality Learning

Under the single-modality learning framework, we performed three classification tasks: (1) the 0-1 task, i.e., differentiating between workload level 0 and workload level 1, (2) the 0-2 task, i.e., distinguishing between workload level 0 and workload level 2, and (3) the 1-2 task, i.e., differentiating between workload level 1 and workload level 2.
Five learning methods, k-NN, NB, RF, SVM, and MLP, were applied for all types of signals, while MLSTM-FCN and InceptionTime were used for EEG and pupil diameter, and EEGNet was used for EEG only. All the experiments were repeated five times, and only the average accuracy together with its standard deviation will be reported.
The performance of the k-NN, NB, RF, SVM, and MLP models in task 0-1, task 0-2, and task 1-2 are shown in Table 7, Table 8, and Table 9, respectively. As seen, the highest accuracy in the 0-1 task is 74.47%, which was achieved by using PCPS with RF. The highest accuracy in the 0-2 task is 80.45%, which was achieved by using PCSP with SVM, while the highest accuracy in the 1-2 task is only 67.52%, which was achieved by using PCSP with MLP. The low accuracy in the 1-2 task can be explained by the way workloads level 1 and level 2 were constructed. Indeed, by the construction, level 2 was generated from level 1 by adding braking events that occurred frequently and did not cause a significant increase in cognitive workload in a short period of time (please see Table 1). Therefore, there was not much difference between the amount of workload contained in level 1 and level 2, leading to the low classification accuracy of task 1-2. Over three tasks, one can observe that PCPS outperformed other physiological signals regardless of learning models.
The performances of other learning models (MLSTM-FCN, InceptionTime, and EEGNet) when applied to EEG and pupil diameter signals are shown in Table 10, Table 11, and Table 12, respectively. Note that we used EEGNet for the EEG signal only but applied MLSTM-FCN and InceptionTime for both EEG and pupil diameter. Again, the PCPS can achieve a higher classification accuracy over all three workload classification tasks regardless of the learning models.
In conclusion, the observed results from Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 agreed with our previous analysis in Section 6.1, demonstrating that the eye gaze, particularly, the PCPS, is a sufficient modality for assessing human cognitive workload in interactive, multi-modal, multi-task settings.

6.2.2. Multi-Modality Learning

For multi-modality learning, we decided to utilize two common approaches named data-level fusion and feature-level fusion to jointly learn from multiple physiological signals. The data-level fusion approach aims to combine two time-series signals together as the input of the learning models. Specifically, we separately used two MLP networks to learn the useful features from two physiological signals; then the outputs of these two networks were fed into the third MLP network to output the final labels. On the other hand, in the feature-level fusion approach, we simply concentrated the handcrafted extracted features from two different signals as the input for our learning models. Particularly, we used PSD as the extracted features for EEG and used mean and variance as the extracted features for PCPS. The details of the feature extraction process can be found in Section 4. Due to EEG and pupil diameter (PCPS) being the two most common physiological indicators for assessing the cognitive workload which have demonstrated their efficiency in literature, we focused on integrating these two signals together.
The performances of multi-modality learning for three workload classification tasks, 0-1, 0-2, and 1-2, can be viewed from Table 13. As seen, combining EEG with PCPS provided an improvement over using EEG alone; however, this combination did not show any improvement compared to only using PCPS. Particularly, the highest accuracies for EEG over three classification tasks. 0-1, 0-2, and 1-2, were 62.92 8.64 , 63.75 5.63 , and 61.76 1.84 , which were increased to 65.50 0.83 , 77.08 3.22 , and 62.90 2.85 after being combined with pupil-diameter. In contrast, the highest accuracies of using pupil diameter alone over three classification tasks, 0-1, 0-2, and 1-2, were 74.47 3.01 , 80.45 3.15 , and 67.52 6.35 , which were substantially higher than the highest accuracy of using multi-modality learning, at 65.50 0.83 , 77.08 3.22 , and 62.90 2.85 , respectively.
The numerical results from Table 7, Table 8, Table 9, Table 10, Table 11, Table 12 and Table 13 supported our conclusion that combining multiple physiological signals did not significantly improve the quality of workload prediction over just using pupil diameter alone. It is worth noting that our conclusion agrees with the observations in [41,45,46].
For the convenience of the reader, our numerical results were summarized in Table 14, which indicates the highest accuracy of the tested signals (PCPS, EEG, FIX, BPV, HRV, and PCPS+EEG) and its standard deviation over three classification tasks, 0-1, 0-2, and 1-2, respectively.
Finally, the neural network-based models (multi-layer perceptron, EEGNet [102], multivariate long short-term memory fully convolutional network (MLSTM-FCN) [103] and InceptionTime [104]) required approximately 20 min to finish the whole training process, while the non-neural network-based models (k-nearest neighbor, naive Bayes, random forest, and support-vector machine) usually required around one minute. Note that our dataset was small with just 348 samples and we ran all models on a Macbook pro 2.3 GHz, Quad-Core Intel Core i7, 32 GB 3733 MHz.

7. Conclusions

The goal of this paper was to investigate the utility of multiple physiological signals such as EEG, pupillometry, and arterial blood pressure, alone or combined, for assessing human cognitive workload in multi-modal, multi-task human-robot interaction contexts. Our statistical analyses and modeling results showed that (1) pupil dilation is the most effective physiological signal for assessing different cognitive workload levels and that (2) no fusion model combining multiple features could improve the prediction accuracy over pupillometry alone in three workload classification tasks. Since eye gaze data is practically easy to obtain and process compared to other physiological signals such as EEG and arterial blood pressure, our findings suggested that eye gaze could be the modality of choice for interactive assistive robots to infer human workload and use it to improve human-robot interactions.

8. Limitations and Future Work

Our study has some limitations. First, the current design only provided an offline determination of cognitive workload. Future works will evaluate the utility of real-time assessments of human workload to adapt robot behavior in human-robot interaction contexts. Second, our study only focused on examining the performance of different sensor modalities in predicting cognitive workload. However, we believe that other cognitive states, such as a sense of urgency, mind wandering, and interference, could be detected in the dataset as well. Finally, our results were limited by the subjects used in the experiment who had an average age of 20 and were recruited mostly from undergraduate/graduate programs at Tufts University. Hence, while we believe that the results will generalize to other populations (e.g., as the cognitive load would only be exacerbated in older populations), additional experiments need to be performed to verify this conjecture.

Author Contributions

Conceptualization, A.A., T.N. and M.S.; methodology, A.A., T.N., S.A. and M.S.; software, A.A., T.N. and Z.H.; validation, T.N., M.S.; formal analysis, A.A. and T.N.; investigation, A.A., T.N. and M.S.; resources, Z.H.; data curation, Z.H.; writing—original draft preparation, T.N., A.A. and M.S.; writing—T.N. and M.S.; visualization, A.A. and T.N.; supervision, M.S. and S.A.; project administration, Z.H.; funding acquisition, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This project was in part supported by AFOSR grant #FA9550-18-1-0465.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of Tufts University (protocol number: 1802002 original approval 23 February 2018).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Howard, Z.L.; Innes, R.; Eidels, A.; Loft, S. Using Past and Present Indicators of Human Workload to Explain Variance in Human Performance. Psychon. Bull. Rev. 2021, 28, 1923–1932. [Google Scholar] [CrossRef] [PubMed]
  2. Heard, J.; Harriott, C.E.; Adams, J.A. A survey of workload assessment algorithms. IEEE Trans. Hum.-Mach. Syst. 2018, 48, 434–451. [Google Scholar] [CrossRef]
  3. Berka, C.; Levendowski, D.J.; Lumicao, M.N.; Yau, A.; Davis, G.; Zivkovic, V.T.; Olmstead, R.E.; Tremoulet, P.D.; Craven, P.L. EEG correlates of task engagement and mental workload in vigilance, learning, and memory tasks. Aviat. Space Environ. Med. 2007, 78, B231–B244. [Google Scholar] [PubMed]
  4. So, W.K.; Wong, S.W.; Mak, J.N.; Chan, R.H. An evaluation of mental workload with frontal EEG. PLoS ONE 2017, 12, e0174949. [Google Scholar] [CrossRef]
  5. May, J.G.; Kennedy, R.S.; Williams, M.C.; Dunlap, W.P.; Brannan, J.R. Eye movement indices of mental workload. Acta Psychol. 1990, 75, 75–89. [Google Scholar] [CrossRef]
  6. Greef, T.D.; Lafeber, H.; Oostendorp, H.V.; Lindenberg, J. Eye movement as indicators of mental workload to trigger adaptive automation. In Proceedings of the International Conference on Foundations of Augmented Cognition, San Diego, CA, USA, 19–24 July 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 219–228. [Google Scholar]
  7. Liu, Y.; Ayaz, H.; Shewokis, P.A. Multisubject “learning” for mental workload classification using concurrent EEG, fNIRS, and physiological measures. Front. Hum. Neurosci. 2017, 11, 389. [Google Scholar] [CrossRef]
  8. Scheutz, M.; Aeron, S.; Aygun, A.; de Ruiter, J.; Fantini, S.; Fernandez, C.; Haga, Z.; Nguyen, T.; Lyu, B.; Rife, J. Estimating Individual Cognitive States from a Mixture of Physiological and Brain Signals. Trends Cogn. Sci. 2022. under review. [Google Scholar]
  9. Aygun, A.; Lyu, B.; Nguyen, T.; Haga, Z.; Aeron, S.; Scheutz, M. Cognitive Workload Assessment via Eye Gaze and EEG in an Interactive Multi-Modal Driving Task. In Proceedings of the 24th ACM International Conference on Multi-Modal Interaction, Bengaluru, India, 7–11 November 2022. [Google Scholar]
  10. Grimes, D.; Tan, D.S.; Hudson, S.E.; Shenoy, P.; Rao, R.P. Feasibility and pragmatics of classifying working memory load with an electroencephalograph. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Florence, Italy, 5–10 April 2008; ACM: New York, NY, USA, 2008; pp. 835–844. [Google Scholar]
  11. Paas, F.; Tuovinen, J.E.; Tabbers, H.; Van Gerven, P.W. Cognitive load measurement as a means to advance cognitive load theory. Educ. Psychol. 2003, 38, 63–71. [Google Scholar] [CrossRef]
  12. Abd Rahman, N.I.; Dawal, S.Z.M.; Yusoff, N. Ageing drivers’ mental workload in real-time driving task based on subjective and objective measures. J. Eng. Res. 2021, 9, 272–284. [Google Scholar] [CrossRef]
  13. Hart, S.G.; Staveland, L.E. Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research. Adv. Psychol. 1988, 52, 139–183. [Google Scholar]
  14. Reid, G.B.; Nygren, T.E. The subjective workload assessment technique: A scaling procedure for measuring mental workload. In Advances in Psychology; Elsevier: Amsterdam, The Netherlands, 1988; Volume 52, pp. 185–218. [Google Scholar]
  15. Tsang, P.S.; Velazquez, V.L. Diagnosticity and multidimensional subjective workload ratings. Ergonomics 1996, 39, 358–381. [Google Scholar] [CrossRef]
  16. Tao, D.; Tan, H.; Wang, H.; Zhang, X.; Qu, X.; Zhang, T. A systematic review of physiological measures of mental workload. Int. J. Environ. Res. Public Health 2019, 16, 2716. [Google Scholar] [CrossRef] [Green Version]
  17. Lei, S.; Roetting, M. Influence of task combination on EEG spectrum modulation for driver workload estimation. Hum. Factors 2011, 53, 168–179. [Google Scholar] [CrossRef]
  18. Ryu, K.; Myung, R. Evaluation of mental workload with a combined measure based on physiological indices during a dual task of tracking and mental arithmetic. Int. J. Ind. Ergon. 2005, 35, 991–1009. [Google Scholar] [CrossRef]
  19. Qu, H.; Shan, Y.; Liu, Y.; Pang, L.; Fan, Z.; Zhang, J.; Wanyan, X. Mental workload classification method based on EEG independent component features. Appl. Sci. 2020, 10, 3036. [Google Scholar] [CrossRef]
  20. Reddy, A.G.; Narava, S. Artifact removal from EEG signals. Int. J. Comput. Appl. 2013, 77, 17–19. [Google Scholar]
  21. Jiang, X.; Bian, G.B.; Tian, Z. Removal of artifacts from EEG signals: A review. Sensors 2019, 19, 987. [Google Scholar] [CrossRef]
  22. Rogasch, N.C.; Biabani, M.; Mutanen, T.P. Designing and comparing cleaning pipelines for TMS-EEG data: A theoretical overview and practical example. J. Neurosci. Methods 2022, 371, 109494. [Google Scholar] [CrossRef]
  23. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef]
  24. Hoover, A.; Singh, A.; Fishel-Brown, S.; Muth, E. Real-time detection of workload changes using heart rate variability. Biomed. Signal Process. Control 2012, 7, 333–341. [Google Scholar] [CrossRef]
  25. Delliaux, S.; Delaforge, A.; Deharo, J.C.; Chaumet, G. Mental workload alters heart rate variability, lowering non-linear dynamics. Front. Physiol. 2019, 10, 565. [Google Scholar] [CrossRef]
  26. Shakouri, M.; Ikuma, L.H.; Aghazadeh, F.; Nahmens, I. Analysis of the sensitivity of heart rate variability and subjective workload measures in a driving simulator: The case of highway work zones. Int. J. Ind. Ergon. 2018, 66, 136–145. [Google Scholar] [CrossRef]
  27. Stuiver, A.; Brookhuis, K.A.; de Waard, D.; Mulder, B. Short-term cardiovascular measures for driver support: Increasing sensitivity for detecting changes in mental workload. Int. J. Psychophysiol. 2014, 92, 35–41. [Google Scholar] [CrossRef]
  28. Hjortskov, N.; Rissén, D.; Blangsted, A.K.; Fallentin, N.; Lundberg, U.; Søgaard, K. The effect of mental stress on heart rate variability and blood pressure during computer work. Eur. J. Appl. Physiol. 2004, 92, 84–89. [Google Scholar] [CrossRef]
  29. Ahlstrom, U.; Friedman-Berg, F.J. Using eye movement activity as a correlate of cognitive workload. Int. J. Ind. Ergon. 2006, 36, 623–636. [Google Scholar] [CrossRef]
  30. Palinko, O.; Kun, A.L.; Shyrokov, A.; Heeman, P. Estimating cognitive load using remote eye tracking in a driving simulator. In Proceedings of the 2010 Symposium on Eye-Tracking Research & Applications, Austin, TX, USA, 22–24 March 2010; ACM: New York, NY, USA, 2010; pp. 141–144. [Google Scholar]
  31. Palinko, O.; Kun, A.L. Exploring the effects of visual cognitive load and illumination on pupil diameter in driving simulators. In Proceedings of the Symposium on Eye Tracking Research and Applications, Santa Barbara, CA, USA, 28–30 March 2012; ACM: New York, NY, USA, 2012; pp. 413–416. [Google Scholar]
  32. Beatty, J. Task-evoked pupillary responses, processing load, and the structure of processing resources. Psychol. Bull. 1982, 91, 276. [Google Scholar] [CrossRef]
  33. Pfleging, B.; Fekety, D.K.; Schmidt, A.; Kun, A.L. A model relating pupil diameter to mental workload and lighting conditions. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; ACM: New York, NY, USA, 2016; pp. 5776–5788. [Google Scholar]
  34. Das, S.; Prudhvi, K.; Maiti, J. Assessing Mental Workload Using Eye Tracking Technology and Deep Learning Models. In Handbook of Intelligent Computing and Optimization for Sustainable Development; Wiley Online Library: New Jersey, NJ, USA, 2022; pp. 1–11. [Google Scholar]
  35. Bitkina, O.V.; Park, J.; Kim, H.K. The ability of eye-tracking metrics to classify and predict the perceived driving workload. Int. J. Ind. Ergon. 2021, 86, 103193. [Google Scholar] [CrossRef]
  36. Pang, L.; Fan, Y.; Deng, Y.; Wang, X.; Wang, T. Mental Workload Classification By Eye Movements In Visual Search Tasks. In Proceedings of the 2020 13th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Chengdu, China, 17–19 October 2020; pp. 29–33. [Google Scholar]
  37. Kosch, T.; Hassib, M.; Buschek, D.; Schmidt, A. Look into my eyes: Using pupil dilation to estimate mental workload for task complexity adaptation. In Proceedings of the Extended Abstracts of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–16 April 2018; ACM: New York, NY, USA, 2018; pp. 1–6. [Google Scholar]
  38. Appel, T.; Scharinger, C.; Gerjets, P.; Kasneci, E. Cross-subject workload classification using pupil-related measures. In Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications, Warsaw, Poland, 14–17 June 2018; ACM: New York, NY, USA, 2018; pp. 1–8. [Google Scholar]
  39. Khedher, A.B.; Jraidi, I.; Frasson, C. Predicting learners’ performance using EEG and eye tracking features. In Proceedings of the Thirty-Second International Flairs Conference, Sarasota, FL, USA, 19–22 May 2019. [Google Scholar]
  40. Rozado, D.; Dunser, A. Combining EEG with pupillometry to improve cognitive workload detection. Computer 2015, 48, 18–25. [Google Scholar] [CrossRef]
  41. Christensen, J.C.; Estepp, J.R.; Wilson, G.F.; Russell, C.A. The effects of day-to-day variability of physiological data on operator functional state classification. NeuroImage 2012, 59, 57–63. [Google Scholar] [CrossRef]
  42. Aghajani, H.; Garbey, M.; Omurtag, A. Measuring mental workload with EEG+ fNIRS. Front. Hum. Neurosci. 2017, 11, 359. [Google Scholar] [CrossRef]
  43. Liu, Y.; Ayaz, H.; Shewokis, P.A. Mental workload classification with concurrent electroencephalography and functional near-infrared spectroscopy. Brain-Comput. Interfaces 2017, 4, 175–185. [Google Scholar] [CrossRef]
  44. Herff, C.; Fortmann, O.; Tse, C.Y.; Cheng, X.; Putze, F.; Heger, D.; Schultz, T. Hybrid fNIRS-EEG based discrimination of 5 levels of memory load. In Proceedings of the 2015 7th International IEEE/EMBS Conference on Neural Engineering (NER), Montpellier, France, 22–24 April 2015; pp. 5–8. [Google Scholar]
  45. Borys, M.; Plechawska-Wójcik, M.; Wawrzyk, M.; Wesołowska, K. Classifying cognitive workload using eye activity and EEG features in arithmetic tasks. In Proceedings of the International Conference on Information and Software Technologies, Druskininkai, Lithuania, 12–14 October 2017; Springer: Cham, Switerland, 2017; pp. 90–105. [Google Scholar]
  46. Coffey, E.B.; Brouwer, A.M.; van Erp, J.B. Measuring workload using a combination of electroencephalography and near infrared spectroscopy. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2012, 56, 1822–1826. [Google Scholar]
  47. Debie, E.; Rojas, R.F.; Fidock, J.; Barlow, M.; Kasmarik, K.; Anavatti, S.; Garratt, M.; Abbass, H.A. Multimodal fusion for objective assessment of cognitive workload: A review. IEEE Trans. Cybern. 2019, 51, 1542–1555. [Google Scholar] [CrossRef] [PubMed]
  48. Blanco, J.A.; Johnson, M.K.; Jaquess, K.J.; Oh, H.; Lo, L.C.; Gentili, R.J.; Hatfield, B.D. Quantifying cognitive workload in simulated flight using passive, dry EEG measurements. IEEE Trans. Cogn. Dev. Syst. 2016, 10, 373–383. [Google Scholar] [CrossRef]
  49. Cheema, B.S.; Samima, S.; Sarma, M.; Samanta, D. Mental workload estimation from EEG signals using machine learning algorithms. In Proceedings of the International Conference on Engineering Psychology and Cognitive Ergonomics, Las Vegas, NV, USA, 15–20 July 2018; Springer: Cham, Switerland, 2018; pp. 265–284. [Google Scholar]
  50. Kaczorowska, M.; Plechawska-Wójcik, M.; Tokovarov, M. Interpretable machine learning models for three-way classification of cognitive workload levels for eye-tracking features. Brain Sci. 2021, 11, 210. [Google Scholar] [CrossRef]
  51. Hope, R.M.; Wang, Z.; Wang, Z.; Ji, Q.; Gray, W.D. Workload classification across subjects using EEG. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 2011, 55, 202–206. [Google Scholar]
  52. Duraisingam, A.; Palaniappan, R.; Andrews, S. Cognitive task difficulty analysis using EEG and data mining. In Proceedings of the 2017 Conference on Emerging Devices and Smart Systems (ICEDSS), Mallasamudram, India, 3–4 March 2017; pp. 52–57. [Google Scholar]
  53. Pandey, V.; Choudhary, D.K.; Verma, V.; Sharma, G.; Singh, R.; Chandra, S. Mental Workload Estimation Using EEG. In Proceedings of the 2020 Fifth International Conference on Research in Computational Intelligence and Communication Networks (ICRCICN), Bangalore, India, 26–27 November 2020; pp. 83–86. [Google Scholar]
  54. Almogbel, M.A.; Dang, A.H.; Kameyama, W. Cognitive workload detection from raw EEG-signals of vehicle driver using deep learning. In Proceedings of the 2019 21st International Conference on Advanced Communication Technology (ICACT), PyeongChang, Korea, 17–20 February 2019; pp. 1–6. [Google Scholar]
  55. Dimitrakopoulos, G.N.; Kakkos, I.; Dai, Z.; Lim, J.; de Souza, J.J.; Bezerianos, A.; Sun, Y. Task-independent mental workload classification based upon common multiband EEG cortical connectivity. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1940–1949. [Google Scholar] [CrossRef]
  56. Mazher, M.; Abd Aziz, A.; Malik, A.S.; Amin, H.U. An EEG-based cognitive load assessment in multimedia learning using feature extraction and partial directed coherence. IEEE Access 2017, 5, 14819–14829. [Google Scholar] [CrossRef]
  57. Yu, K.; Prasad, I.; Mir, H.; Thakor, N.; Al-Nashash, H. Cognitive workload modulation through degraded visual stimuli: A single-trial EEG study. J. Neural Eng. 2015, 12, 046020. [Google Scholar] [CrossRef]
  58. Singh, U.; Ahirwal, M.K. Mental Workload Classification for Multitasking Test using Electroencephalogram Signal. In Proceedings of the 2021 IEEE International Conference on Technology, Research, and Innovation for Betterment of Society (TRIBES), Raipur, India, 17–19 December 2021; pp. 1–6. [Google Scholar]
  59. Yin, Z.; Zhang, J. Cross-session classification of mental workload levels using EEG and an adaptive deep learning model. Biomed. Signal Process. Control 2017, 33, 30–47. [Google Scholar] [CrossRef]
  60. Zarjam, P.; Epps, J.; Lovell, N.H. Beyond Subjective Self-Rating: EEG Signal Classification of Cognitive Workload. IEEE Trans. Auton. Ment. Dev. 2015, 7, 301–310. [Google Scholar] [CrossRef]
  61. Hefron, R.G.; Borghetti, B.J.; Christensen, J.C.; Kabban, C.M.S. Deep long short-term memory structures model temporal dependencies improving cognitive workload estimation. Pattern Recognit. Lett. 2017, 94, 96–104. [Google Scholar] [CrossRef]
  62. Rahman, H.; Ahmed, M.U.; Barua, S.; Funk, P.; Begum, S. Vision-based driver’s cognitive load classification considering eye movement using machine learning and deep learning. Sensors 2021, 21, 8019. [Google Scholar] [CrossRef] [PubMed]
  63. Yang, S.; Yin, Z.; Wang, Y.; Zhang, W.; Wang, Y.; Zhang, J. Assessing cognitive mental workload via EEG signals and an ensemble deep learning classifier based on denoising autoencoders. Comput. Biol. Med. 2019, 109, 159–170. [Google Scholar] [CrossRef]
  64. Huang, J.; Liu, Y.; Peng, X. Recognition of driver’s mental workload based on physiological signals, a comparative study. Biomed. Signal Process. Control 2022, 71, 103094. [Google Scholar] [CrossRef]
  65. Islam, M.R.; Barua, S.; Ahmed, M.U.; Begum, S.; Flumeri, G.D. Deep learning for automatic EEG feature extraction: An application in drivers’ mental workload classification. In Proceedings of the International Symposium on Human Mental Workload: Models and Applications, Rome, Italy, 14–15 November 2019; Springer: Cham, Switerland, 2019; pp. 121–135. [Google Scholar]
  66. Charles, R.L.; Nixon, J. Measuring mental workload using physiological measures: A systematic review. Appl. Ergon. 2019, 74, 221–232. [Google Scholar] [CrossRef]
  67. Saeedpour-Parizi, M.R.; Hassan, S.E.; Shea, J.B. Pupil diameter as a biomarker of effort in goal-directed gait. Exp. Brain Res. 2020, 238, 2615–2623. [Google Scholar] [CrossRef]
  68. Wildemeersch, D.; Peeters, N.; Saldien, V.; Vercauteren, M.; Hans, G. Pain assessment by pupil dilation reflex in response to noxious stimulation in anaesthetized adults. Acta Anaesthesiol. Scand. 2018, 62, 1050–1056. [Google Scholar] [CrossRef]
  69. Smallwood, J.; Brown, K.S.; Tipper, C.; Giesbrecht, B.; Franklin, M.S.; Mrazek, M.D.; Carlson, J.M.; Schooler, J.W. Pupillometric evidence for the decoupling of attention from perceptual input during offline thought. PLoS ONE 2011, 6, e18298. [Google Scholar] [CrossRef]
  70. Prieur-Coloma, Y.; Reinoso-Leblanch, R.A.; Mayeta-Revilla, L.; Delisle-Rodríguez, D.; Bastos, T.; López-Delis, A.; Balart-Fernández, L.; Falk, T.H. Enhancing shoulder pre-movements recognition through EEG Riemannian covariance matrices for a BCI-based exoskeleton. In Proceedings of the 2020 IEEE International Conference on Human-Machine Systems (ICHMS), Rome, Italy, 7–9 September 2020; pp. 1–3. [Google Scholar]
  71. Yu, Y. A study on the classification of left-and righthanded eeg signals based on motor imagination. In Proceedings of the 2020 IEEE International Conference on Advances in Electrical Engineering and Computer Applications (AEECA), Dalian, China, 25–27 August 2020; pp. 28–31. [Google Scholar]
  72. Li, R.; Principe, J.C. Blinking artifact removal in cognitive EEG data using ICA. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 5273–5276. [Google Scholar]
  73. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  74. Cerliani, M. Tsmoothie. 2021. Available online: https://github.com/cerlymarco/tsmoothie (accessed on 15 June 2022).
  75. Chen, H.; Erol, Y.; Shen, E.; Russell, S. Probabilistic model-based approach for heart beat detection. Physiol. Meas. 2016, 37, 1404. [Google Scholar] [CrossRef]
  76. De Morais Borges, G.; Brusamarello, V. Bayesian fusion of multiple sensors for reliable heart rate detection. In Proceedings of the 2014 IEEE International Instrumentation and Measurement Technology Conference (I2MTC) Proceedings, Montevideo, Uruguay, 12–15 May 2014; pp. 1310–1313. [Google Scholar]
  77. Dias, D.; Silva, L.; Katayama, P.; Silva, C.; Salgado, H.; Fazan, R. Correlation between RR, inter-systolic and inter-diastolic intervals and their differences for the analysis of spontaneous heart rate variability. Physiol. Meas. 2016, 37, 1120. [Google Scholar] [CrossRef] [PubMed]
  78. Avram, R.; Tison, G.H.; Aschbacher, K.; Kuhar, P.; Vittinghoff, E.; Butzner, M.; Runge, R.; Wu, N.; Pletcher, M.J.; Marcus, G.M.; et al. Real-world heart rate norms in the Health eHeart study. NPJ Digit. Med. 2019, 2, 1–10. [Google Scholar] [CrossRef] [PubMed]
  79. Zhao, M.; Gao, H.; Wang, W.; Qu, J.; Chen, L. Study on the identification of irritability emotion based on the percentage change in pupil size. In Proceedings of the 2020 2nd International Conference on Image, Video and Signal Processing, Singapore, 20–22 March 2020; ACM: New York, NY, USA, 2020; pp. 20–24. [Google Scholar]
  80. Ameera, A.; Saidatul, A.; Ibrahim, Z. Analysis of EEG spectrum bands using power spectral density for pleasure and displeasure state. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK , 2019; Volume 557, p. 012030. [Google Scholar]
  81. Ng, W.B.; Saidatul, A.; Chong, Y.; Ibrahim, Z. PSD-based features extraction for EEG signal during typing task. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2019; Volume 557, p. 012032. [Google Scholar]
  82. Lim, W.L.; Sourina, O.; Liu, Y.; Wang, L. EEG-based mental workload recognition related to multitasking. In Proceedings of the 2015 10th International Conference on Information, Communications and Signal Processing (ICICS), Singapore, 2–4 December 2015; pp. 1–4. [Google Scholar]
  83. Matthews, G.; Reinerman-Jones, L.; Abich IV, J.; Kustubayeva, A. Metrics for individual differences in EEG response to cognitive workload: Optimizing performance prediction. Personal. Individ. Differ. 2017, 118, 22–28. [Google Scholar] [CrossRef]
  84. Chikhi, S.; Matton, N.; Blanchet, S. EEG power spectral measures of cognitive workload: A meta-analysis. Psychophysiology 2022, 59, e14009. [Google Scholar] [CrossRef]
  85. Foroozan, F.; Mohan, M.; Wu, J.S. Robust beat-to-beat detection algorithm for pulse rate variability analysis from wrist photoplethysmography signals. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2136–2140. [Google Scholar]
  86. Tarvainen, M.P.; Niskanen, J.P.; Lipponen, J.A.; Ranta-Aho, P.O.; Karjalainen, P.A. Kubios HRV–heart rate variability analysis software. Comput. Methods Programs Biomed. 2014, 113, 210–220. [Google Scholar] [CrossRef]
  87. Parati, G.; Stergiou, G.S.; Dolan, E.; Bilo, G. Blood pressure variability: Clinical relevance and application. J. Clin. Hypertens. 2018, 20, 1133–1137. [Google Scholar] [CrossRef]
  88. Xia, Y.; Wu, D.; Gao, Z.; Liu, X.; Chen, Q.; Ren, L.; Wu, W. Association between beat-to-beat blood pressure variability and vascular elasticity in normal young adults during the cold pressor test. Medicine 2017, 96, e6000. [Google Scholar] [CrossRef]
  89. Tian, G.; Xiong, L.; Leung, H.; Soo, Y.; Leung, T.; Wong, L.K.S. Beat-to-beat blood pressure variability and heart rate variability in relation to autonomic dysregulation in patients with acute mild-moderate ischemic stroke. J. Clin. Neurosci. 2019, 64, 187–193. [Google Scholar] [CrossRef]
  90. Zawadka-Kunikowska, M.; Rzepiński, Ł.; Newton, J.L.; Zalewski, P.; Słomko, J. Cardiac Autonomic Modulation Is Different in Terms of Clinical Variant of Multiple Sclerosis. J. Clin. Med. 2020, 9, 3176. [Google Scholar] [CrossRef]
  91. Qin, X.; Zheng, Y.; Chen, B. Extract EEG Features by Combining Power Spectral Density and Correntropy Spectral Density. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 2455–2459. [Google Scholar]
  92. Hossain, M.F.; Yaacob, H.; Nordin, A. Development of Unified Neuro-Affective Classification Tool (UNACT). In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2021; Volume 1077, p. 012031. [Google Scholar]
  93. Hamzah, N.; Norhazman, H.; Zaini, N.; Sani, M. Classification of EEG signals based on different motor movement using multi-layer Perceptron artificial neural network. J. Biol. Sci. 2016, 16, 265–271. [Google Scholar] [CrossRef]
  94. Al-Nafjan, A.; Hosny, M.; Al-Wabil, A.; Al-Ohali, Y. Classification of human emotions from electroencephalogram (EEG) signal using deep neural network. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 419–425. [Google Scholar] [CrossRef]
  95. Stoica, P.; Moses, R.L. Spectral Analysis of Signals; Pearson Prentice Hall: Upper Saddle River, NJ, USA, 2005; Volume 452. [Google Scholar]
  96. Rozado, D.; Duenser, A.; Howell, B. Improving the performance of an EEG-based motor imagery brain computer interface using task evoked changes in pupil diameter. PLoS ONE 2015, 10, e0121262. [Google Scholar] [CrossRef]
  97. Plechawska-Wójcik, M.; Borys, M. An analysis of EEG signal combined with pupillary response in the dynamics of human cognitive processing. In Proceedings of the 2016 9th International Conference on Human System Interactions (HSI), Portsmouth, UK, 6–8 July 2016; pp. 378–385. [Google Scholar]
  98. Lobo, J.L.; Ser, J.D.; De Simone, F.; Presta, R.; Collina, S.; Moravek, Z. Cognitive workload classification using eye-tracking and EEG data. In Proceedings of the International Conference on Human-Computer Interaction in Aerospace, Paris, France, 14–16 September 2016; ACM: New York, NY, USA, 2016; pp. 1–8. [Google Scholar]
  99. Stone, M. Cross-validation: A review. Stat. A J. Theor. Appl. Stat. 1978, 9, 127–139. [Google Scholar]
  100. Saadati, M.; Nelson, J.; Ayaz, H. Mental Workload Classification From Spatial Representation of FNIRS Recordings Using Convolutional Neural Networks. In Proceedings of the 2019 IEEE 29th International Workshop on Machine Learning for Signal Processing (MLSP), Pittsburgh, PA, USA, 13–16 October 2019; pp. 1–6. [Google Scholar]
  101. Mughal, N.E.; Khalil, K.; Khan, M.J. fNIRS Based Multi-Class Mental Workload Classification Using Recurrence Plots and CNN-LSTM. In Proceedings of the 2021 International Conference on Artificial Intelligence and Mechatronics Systems (AIMS), Bandung, Indonesia, 28–30 April 2021; pp. 1–6. [Google Scholar]
  102. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  103. Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate LSTM-FCNs for time series classification. Neural Netw. 2019, 116, 237–245. [Google Scholar] [CrossRef] [Green Version]
  104. Ismail Fawaz, H.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.A.; Petitjean, F. Inceptiontime: Finding alexnet for time series classification. Data Min. Knowl. Discov. 2020, 34, 1936–1962. [Google Scholar] [CrossRef]
  105. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  106. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  107. Ruiz, A.P.; Flynn, M.; Large, J.; Middlehurst, M.; Bagnall, A. The great multivariate time series classification bake off: A review and experimental evaluation of recent algorithmic advances. Data Min. Knowl. Discov. 2021, 35, 401–449. [Google Scholar] [CrossRef]
  108. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  109. St, L.; Wold, S. Analysis of variance (ANOVA). Chemom. Intell. Lab. Syst. 1989, 6, 259–272. [Google Scholar]
  110. Jaccard, J.; Becker, M.A.; Wood, G. Pairwise multiple comparison procedures: A review. Psychol. Bull. 1984, 96, 589. [Google Scholar] [CrossRef]
  111. Benjamini, Y.; Hochberg, Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. R. Stat. Soc. Ser. B 1995, 57, 289–300. [Google Scholar] [CrossRef]
Figure 1. Pupillometry pre-processing steps: (a) raw pupil diameter signal, (b) signal after applying amplitude thresholding, (c) signal after applying linear interpolation, and (d) signal after applying Butterworth low-pass filter.
Figure 1. Pupillometry pre-processing steps: (a) raw pupil diameter signal, (b) signal after applying amplitude thresholding, (c) signal after applying linear interpolation, and (d) signal after applying Butterworth low-pass filter.
Sensors 22 06834 g001
Figure 2. Arterial blood pressure (ABP) signal pre-processing and SP/DP point detection steps: (a) raw ABP signal, (b) signal after upsampling and applying law-pass filter, (c) the candidate SP/DP points, and (d) estimated SP/DP values.
Figure 2. Arterial blood pressure (ABP) signal pre-processing and SP/DP point detection steps: (a) raw ABP signal, (b) signal after upsampling and applying law-pass filter, (c) the candidate SP/DP points, and (d) estimated SP/DP values.
Sensors 22 06834 g002
Figure 3. The variations in APCPS obtained from eight samples and the variation in mean APCPS over 116 samples (upper left) for three different workload levels.
Figure 3. The variations in APCPS obtained from eight samples and the variation in mean APCPS over 116 samples (upper left) for three different workload levels.
Sensors 22 06834 g003
Figure 4. Boxplot of mean APCPS over all events for different cognitive workload levels.
Figure 4. Boxplot of mean APCPS over all events for different cognitive workload levels.
Sensors 22 06834 g004
Figure 5. Analysis of HRV time domain and frequency domain features for different workload levels.
Figure 5. Analysis of HRV time domain and frequency domain features for different workload levels.
Sensors 22 06834 g005
Figure 6. Analysis of BPV time domain and frequency domain features for different workload levels.
Figure 6. Analysis of BPV time domain and frequency domain features for different workload levels.
Sensors 22 06834 g006
Table 1. Architecture of the used multi-layer perceptron. N is the number of samples in one batch, T 1 , T 2 , T 3 , T 4 denote sizes of four layers, and n is the number of outputted classes.
Table 1. Architecture of the used multi-layer perceptron. N is the number of samples in one batch, T 1 , T 2 , T 3 , T 4 denote sizes of four layers, and n is the number of outputted classes.
LayerOperationOutput Size
Input-(N, T 1 )
The first FC layerLinear( T 1 , T 2 ) + BatchNorm + ReLU +
Dropout(0.25)
(N, T 2 )
The second FC layerLinear( T 2 , T 3 ) + BatchNorm + ReLU +
Dropout(0.25)
(N, T 3 )
The third FC layerLinear( T 3 , T 4 ) + BatchNorm + ReLU +
Dropout(0.25)
(N, T 4 )
The last linear layerLinear( T 4 , n)(N, n)
Table 2. Statistical results of mean, median, standard deviation, and standard error for APCPS, fixation number, and fixation duration for different workload levels over all events.
Table 2. Statistical results of mean, median, standard deviation, and standard error for APCPS, fixation number, and fixation duration for different workload levels over all events.
Workload LevelMeanMedianStdSE
APCPSLevel 00.240.216.080.56
Level 15.755.767.360.68
Level 28.858.596.710.62
Fixation NumberLevel 07.307.003.320.31
Level 17.998.003.700.34
Level 210.1010.004.050.38
Fixation DurationLevel 00.280.210.310.03
Level 10.190.130.160.01
Level 20.180.130.220.02
Table 3. Results of Tukey’s HSD multiple pairwise test and Benjamini–Hochberg (B & H) test for APCPS, fixation number, and fixation duration for different pairs of cognitive workload levels.
Table 3. Results of Tukey’s HSD multiple pairwise test and Benjamini–Hochberg (B & H) test for APCPS, fixation number, and fixation duration for different pairs of cognitive workload levels.
Workload Levelp-Value (Tukey’s HSD)p-Value (B–H)
APCPSLevel 0-Level 1<1 × 10 7 3.6 × 10 9
Level 0-Level 2<1 × 10 7 2.0 × 10 16
Level 1-Level 21 × 10 3 9.3 × 10 4
Fixation NumberLevel 0-Level 10.330.14
Level 0-Level 25 × 10 8 7.5 × 10 8
Level 1-Level 25 × 10 5 6.6 × 10 5
Fixation DurationLevel 0-Level 17 × 10 3 6 × 10 3
Level 0-Level 23 × 10 3 6 × 10 3
Level 1-Level 20.960.76
Table 4. p-values from Tukey’s HSD multiple pairwise test and Benjamini–Hochberg test performed on four frequency bands obtained from PSD of EEG for different pairs of cognitive workload levels.
Table 4. p-values from Tukey’s HSD multiple pairwise test and Benjamini–Hochberg test performed on four frequency bands obtained from PSD of EEG for different pairs of cognitive workload levels.
Workload Level δ θ α β
Tukey’s HSDLevel 0-Level 10.060.79 0.02 4 × 10 5
Level 0-Level 20.880.71 1 × 10 3 3 × 10 5
Level 1-Level 20.160.980.670.99
Benjamini–HochbergLevel 0-Level 10.050.84 0.03 3 × 10 4
Level 0-Level 20.690.84 9 × 10 3 3 × 10 4
Level 1-Level 2 9 × 10 3 0.860.160.93
Table 5. p-values from Tukey’s HSD multiple pairwise test for different pairs of workload levels obtained from HRV parameters.
Table 5. p-values from Tukey’s HSD multiple pairwise test for different pairs of workload levels obtained from HRV parameters.
HRV FeaturesLevel 0-Level 1Level 0-Level 2Level 1-Level 2
Mean IBI (ms)0.410.340.99
SD IBI (ms)0.330.660.85
Mean HR (bpm)0.410.380.99
SD HR (bpm)0.020.110.82
RMSSD (ms)0.970.990.94
VLF ( ms 2 )0.610.490.98
LF ( ms 2 )0.160.610.66
HF ( ms 2 )0.960.990.99
TP ( ms 2 )0.910.790.97
LFnu (%)0.040.220.73
HFnu (%)0.040.220.73
Table 6. p-values from Tukey’s HSD multiple pairwise test for different pairs of workload levels obtained from BPV parameters.
Table 6. p-values from Tukey’s HSD multiple pairwise test for different pairs of workload levels obtained from BPV parameters.
BPV FeaturesLevel 0-Level 1Level 0-Level 2Level 1-Level 2
SD SP (mmHg)0.020.140.75
SD DP (mmHg) 2.1 × 10 4 1 × 10 5 0.76
SV SP (mmHg)0.040.110.91
SV DP (mmHg)0.040.150.81
CV SP (%)0.030.130.85
CV DP (%) 8.8 × 10 3 1.8 × 10 4 0.53
ARV SP (mmHg)0.110.150.99
ARV DP (mmHg)0.060.150.93
VLF SP ( mmHg 2 )0.990.990.99
VLF DP ( mmHg 2 )0.840.990.78
LF SP ( mmHg 2 )0.170.390.88
LF DP ( mmHg 2 ) 1.3 × 10 3 2.5 × 10 4 0.90
HF SP ( mmHg 2 )0.260.910.48
HF DP ( mmHg 2 )0.070.530.50
TP SP ( mmHg 2 )0.840.960.96
TP DP ( mmHg 2 )0.380.850.71
LFnu SP (%)0.370.370.99
LFnu DP (%)0.260.030.62
HFnu SP (%)0.370.370.99
HFnu DP (%)0.260.030.62
Table 7. Classification accuracy of task 0-1 using k-NN, NB, RF, SVM, and MLP.
Table 7. Classification accuracy of task 0-1 using k-NN, NB, RF, SVM, and MLP.
Signalsk-NNNBRFSVMMLP
PCPS67.02 ∓ 4.8873.40 ∓ 5.5374.47 ∓ 3.0172.34 ∓ 2.6167.92 ∓ 6.84
BPV53.19 ∓ 3.9859.57 ∓ 5.4257.45 ∓ 7.8255.32 ∓ 6.8960.42 ∓ 6.42
HRV50.53 ∓ 2.3255.85 ∓ 4.3555.85 ∓ 5.5055.85 ∓ 2.5655.00 ∓ 6.35
Fix52.13 ∓ 1.0654.79 ∓ 2.3253.19 ∓ 2.6157.98 ∓ 2.3261.67 ∓ 4.80
EEG54.19 ∓ 4.2652.13 ∓ 3.1960.11 ∓ 3.8060.11 ∓ 9.0762.92 ∓ 8.64
Table 8. Classification accuracy of task 0-2 using k-NN, NB, RF, SVM, and MLP.
Table 8. Classification accuracy of task 0-2 using k-NN, NB, RF, SVM, and MLP.
Signalsk-NNNBRFSVMMLP
PCPS79.26 ∓ 6.4579.78 ∓ 7.4470.68 ∓ 8.4380.45 ∓ 3.1574.58 ∓ 3.42
BPV59.57 ∓ 3.8058.51 ∓ 4.3958.51 ∓ 4.3963.30 ∓ 6.4563.75 ∓ 3.16
HRV51.60 ∓ 3.8056.38 ∓ 8.0348.94 ∓ 3.9854.72 ∓ 1.7654.58 ∓ 3.73
Fix61.17 ∓ 3.8064.36 ∓ 9.3258.51 ∓ 2.3866.49 ∓ 5.0763.33 ∓ 6.18
EEG60.64 ∓ 1.8454.54 ∓ 6.4562.77 ∓ 7.7457.98 ∓ 6.2763.75 ∓ 5.63
Table 9. Classification accuracy of task 1-2 using k-NN, NB, RF, SVM, and MLP.
Table 9. Classification accuracy of task 1-2 using k-NN, NB, RF, SVM, and MLP.
Signalsk-NNNBRFSVMMLP
PCPS59.57 ∓ 4.6360.63 ∓ 3.1966.49 ∓ 10.0263.11 ∓ 2.3167.52 ∓ 6.35
BPV43.09 ∓ 2.7644.68 ∓ 8.3853.19 ∓ 3.3648.81 ∓ 5.4249.17 ∓ 1.14
HRV47.87 ∓ 6.2953.19 ∓ 7.3749.47 ∓ 1.7645.74 ∓ 6.2949.58 ∓ 3.42
Fix50.32 ∓ 3.3653.85 ∓ 3.1550.00 ∓ 6.8158.12 ∓ 3.7756.25 ∓ 6.07
EEG57.98 ∓ 2.3253.19 ∓ 5.8359.77 ∓ 3.1961.76 ∓ 1.8461.33 ∓ 6.77
Table 10. Classification accuracy of task 0-1 using MLSTM-FCN, InceptionTime, and EEGNet.
Table 10. Classification accuracy of task 0-1 using MLSTM-FCN, InceptionTime, and EEGNet.
SignalsMLSTM-FCNInceptionTimeEEGNet
EEG54.16 ∓ 7.8054.58 ∓ 2.7254.07 ∓ 4.03
PCPS64.58 ∓ 2.9567.50 ∓ 3.16-
Table 11. Classification accuracy of task 0-2 using MLSTM-FCN, InceptionTime, and EEGNet.
Table 11. Classification accuracy of task 0-2 using MLSTM-FCN, InceptionTime, and EEGNet.
SignalsMLSTM-FCNInceptionTimeEEGNet
EEG55.00 ∓ 5.2353.33 ∓ 2.8053.33 ∓ 6.69
PCPS73.28 ∓ 5.2457.98 ∓ 9.13-
Table 12. Classification accuracy of task 1-2 using MLSTM-FCN, InceptionTime, and EEGNet.
Table 12. Classification accuracy of task 1-2 using MLSTM-FCN, InceptionTime, and EEGNet.
SignalsMLSTM-FCNInceptionTimeEEGNet
EEG54.22 ∓ 8.7053.91 ∓ 4.1952.70 ∓ 3.27
PCPS63.38 ∓ 4.7556.25 ∓ 9.51-
Table 13. Classification accuracy of tasks 0-1, 0-2, and 1-2 using data-level fusion and feature-level fusion approaches.
Table 13. Classification accuracy of tasks 0-1, 0-2, and 1-2 using data-level fusion and feature-level fusion approaches.
SignalsTasksData-LevelFeature-Level
PCPS + EEG0-165.50 ∓ 0.8365.25 ∓ 6.52
PCPS + EEG0-277.08 ∓ 3.2266.51 ∓ 6.46
PCPS + EEG1-262.90 ∓ 2.8560.57 ∓ 8.52
Table 14. The highest accuracy and its standard deviation of tested signals (PCPS, EEG, FIX, BPV, HRV, and PCPS+EEG) over three classification tasks 0-1, 0-2, and 1-2, respectively.
Table 14. The highest accuracy and its standard deviation of tested signals (PCPS, EEG, FIX, BPV, HRV, and PCPS+EEG) over three classification tasks 0-1, 0-2, and 1-2, respectively.
Signals0-10-21-2
PCPS74.47 ∓ 3.0180.45 ∓ 3.1567.52 ∓ 6.35
EEG62.92 ∓ 8.6463.75 ∓ 5.6361.76 ∓ 1.84
FIX61.67 ∓ 4.8066.49 ∓ 5.0758.12 ∓ 3.77
BPV60.42 ∓ 6.4263.75 ∓ 3.1653.19 ∓ 3.36
HRV55.85 ∓ 4.3554.58 ∓ 3.7356.38 ∓ 8.03
PCPS+EEG65.55 ∓ 0.8377.08 ∓ 3.2262.90 ∓ 2.85
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Aygun, A.; Nguyen, T.; Haga, Z.; Aeron, S.; Scheutz, M. Investigating Methods for Cognitive Workload Estimation for Assistive Robots. Sensors 2022, 22, 6834. https://doi.org/10.3390/s22186834

AMA Style

Aygun A, Nguyen T, Haga Z, Aeron S, Scheutz M. Investigating Methods for Cognitive Workload Estimation for Assistive Robots. Sensors. 2022; 22(18):6834. https://doi.org/10.3390/s22186834

Chicago/Turabian Style

Aygun, Ayca, Thuan Nguyen, Zachary Haga, Shuchin Aeron, and Matthias Scheutz. 2022. "Investigating Methods for Cognitive Workload Estimation for Assistive Robots" Sensors 22, no. 18: 6834. https://doi.org/10.3390/s22186834

APA Style

Aygun, A., Nguyen, T., Haga, Z., Aeron, S., & Scheutz, M. (2022). Investigating Methods for Cognitive Workload Estimation for Assistive Robots. Sensors, 22(18), 6834. https://doi.org/10.3390/s22186834

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop