Next Article in Journal
1.1-µm Band Extended Wide-Bandwidth Wavelength-Swept Laser Based on Polygonal Scanning Wavelength Filter
Previous Article in Journal
Ganoderma boninense Disease Detection by Near-Infrared Spectroscopy Classification: A Review
Previous Article in Special Issue
Hyperparameter Optimization for COVID-19 Pneumonia Diagnosis Based on Chest CT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Negative Stress through Spectral Features of Electroencephalographic Recordings and a Convolutional Neural Network

by
Arturo Martínez-Rodrigo
1,2,*,
Beatriz García-Martínez
3,4,
Álvaro Huerta
2 and
Raúl Alcaraz
5
1
Research Group in Electronic, Biomedical and Telecommunication Engineering, Facultad de Comunicación, University of Castilla-La Mancha, 16071 Cuenca, Spain
2
Instituto de Tecnologías Audiovisuales de Castilla-La Mancha, University of Castilla-La Mancha, 16071 Cuenca, Spain
3
Departamento de Sistemas Informáticos, Escuela Técnica Superior de Ingenieros Industriales, University of Castilla-La Mancha, 02071 Albacete, Spain
4
Instituto de Investigación en Informática de Albacete, University of Castilla-La Mancha, 02071 Albacete, Spain
5
Research Group in Electronic, Biomedical and Telecommunication Engineering, Escuela Politécnica de Cuenca, University of Castilla-La Mancha, 16071 Cuenca, Spain
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(9), 3050; https://doi.org/10.3390/s21093050
Submission received: 1 March 2021 / Revised: 23 April 2021 / Accepted: 25 April 2021 / Published: 27 April 2021
(This article belongs to the Special Issue Biomedical Signal Processing for Disease Diagnosis)

Abstract

:
In recent years, electroencephalographic (EEG) signals have been intensively used in the area of emotion recognition, partcularly in distress identification due to its negative impact on physical and mental health. Traditionally, brain activity has been studied from a frequency perspective by computing the power spectral density of the EEG recordings and extracting features from different frequency sub-bands. However, these features are often individually extracted from single EEG channels, such that each brain region is separately evaluated, even when it has been corroborated that mental processes are based on the coordination of different brain areas working simultaneously. To take advantage of the brain’s behaviour as a synchronized network, in the present work, 2-D and 3-D spectral images constructed from common 32 channel EEG signals are evaluated for the first time to discern between emotional states of calm and distress using a well-known deep-learning algorithm, such as AlexNet. The obtained results revealed a significant improvement in the classification performance regarding previous works, reaching an accuracy about 84%. Moreover, no significant differences between the results provided by the diverse approaches considered to reconstruct 2-D and 3-D spectral maps from the original location of the EEG channels over the scalp were noticed, thus suggesting that these kinds of images preserve original spatial brain information.

1. Introduction

Nowadays, one of the major issues in advanced societies is negative stress, also known as distress, given its detrimental influence on health of people who suffer from it [1,2]. Inhabitants in developed countries are surrounded by an economic and social pressure and a frenetic rhythm of life that leads them to a continuous state of anxiety and nervousness [3]. Furthermore, the current coronavirus pandemic has contributed to generally increasing the level of uncertainty and distress in the global population, with multifaceted drastic repercussions for people’s lives [4]. Short-term negative stress appears as a fight or flight reaction for self-protection and integrity of the organism, and it may not be a risk factor for health [5]. However, a long-term exposure to distressful conditions has demonstrated the production of serious negative effects on physical and mental health, causing, or even aggravating, several disorders related to cerebral, immune and endocrine systems [6,7]. Consequently, the accurate detection and regulation of distress could be essential to maintaining a healthy functioning [8].
Negative stress produces a series of measurable alterations on different physiological systems, such as the brain [9]. Brain activity can be quantified by means of electroencephalographic (EEG) recording, which represents the electrical activity originated under the scalp due to neural connections. The brain is the physiological system firstly responding against any stimulus, and then this response spreading to the rest of the peripheral organs by means of the central nervous system [10]. Therefore, the assessment of EEG signals may reveal more relevant information than secondary effects of the brain’s activity in the rest of the body [10]. This could be why the study of EEG recordings is receiving increased interest for the recognition of emotional states in recent years [11].
Traditionally, EEG recordings have been studied from a frequency perspective, analyzing brain activity in different spectral bands [12]. Indeed, it is widely known that emotional processes may induce changes in the cognitive state of an individual, which are accompanied by alterations in the brain’s oscillatory activity [13]. The information of interest in emotional processes occurs between 4 and 45 Hz, corresponding to theta ( θ ), alpha ( α ), beta ( β ) and gamma ( γ ) frequency sub-bands [14]. To analyze this information, power spectral density (PSD) is usually computed using different algorithms, such as Fast Fourier transform or Welch’s periodogram [12]. Then, power features are extracted from the different frequency sub-bands and combined using machine learning methodologies, such as k-nearest neighbor (k-NN) [15], support vector machine (SVM) [16,17], and Bayes neural networks [18]. However, most previous works only focused on how to combine these single-frequency parameters, without exploring spatial information collected from locations of the electrodes over the scalp [19].
Recently, in addition to machine learning methods, deep learning algorithms have also been used to diagnose some mental disorders, such as epilepsy [20,21], dementia [22], depression [23] or Parkinson [24]. In this respect, convolutional neural networks (CNN) are the most-used techniques within the wider context of deep-learning [25], as well as in the emotion recognition field [26]. Unlike traditional machine learning methods, CNN automatically learns complex features using different convolutional filters and combining the weights to predict class membership. According to the literature, most of the studies related to emotion recognition with CNN used 2-D EEG spectrograms as input data [27,28,29]. However, these spectrograms represent the information from a single channel, with each brain region being evaluated separately. From this perspective, the global coordinated brain information is ignored even when it has been corroborated that the brain works as a network, and mental processes are based on the synchronized performance of different areas [30,31]. Hence, for a thorough assessment of the underlying brain dynamics under different emotions, simultaneous analysis of all brain regions has been suggested [32,33].
To take advantage of that global brain information, the present work evaluates, for the first time, 2-D and 3-D spectral images constructed from simultaneous 32 channels typically acquired by EEG recordings. More precisely, a well-known CNN-based model, such as AlexNet, is proposed to discern these images from emotional states of calm and distress. Moreover, transfer learning (TL) has proven to be an effective methodology in many applications to palliate some drawbacks originated when a CNN is trained from the beginning [34]. Thus, several CNN models conserving the hyper-parameters from the original pre-trained AlexNet and with randomized initial hyper-parameters have been analyzed in the present work. Additionally, since there is not a gold-standard method regarding the rearrangement of the common location of the EEG electrodes over the scalp into a 2-D map [35], diverse approaches found in the literature have been compared.
The structure of the paper is as follows. Section 2 describes the analyzed database, as well as how the EEG signals were preprocessed and their power spectral density was computed. The diverse options considered for the rearrangement of EEG electrode locations into 2-D maps, the trained CNN models and the experimental protocol are also included in this section. The results obtained are presented in Section 3 and discussed in Section 4. Finally, conclusions extracted from this study are summarized in Section 5.

2. Materials and Methods

2.1. Dataset

The EEG signals analyzed in this study were extracted from the Database for Emotion Analysis using Physiological Signals (DEAP) [36]. This publicly available database consists of a total of 1280 samples of various emotional states, obtained from 32 participants (50% males, mean age 26.9 years). Each subject visualized 40 one-minute length videoclips with emotional content, while EEG signals and other physiological variables were recorded. After each visualization, participants rated their emotional state by means of self-assessment manikins (SAM), which are a graphical representation of nine levels of intensity of different emotional parameters. More precisely, the ratings of two parameters, called valence (i.e., the degree of pleasantness or unpleasantness of a stimulus) and arousal (i.e., the level of activation or deactivation provoked by a stimulus), were considered in this study. Although the whole scales of valence and arousal were covered during the creation of the DEAP dataset, only samples corresponding to calm and distress emotions were selected for the present study. In accordance with previous works [37,38], samples from the distress group were selected as those with a valence lower than 3 and an arousal higher than 5. On the other hand, the calm group was formed from samples with a valence between 4 and 6, and an arousal level lower than 4. Hence, a total of 122 trials of distress and 137 of calm were finally analyzed.

2.2. Preprocessing of the EEG Recordings

The EEG recordings were obtained from 32 channels distributed over the scalp according to the international standard 10–20 system for electrodes location [39]. Before further analysis, the raw EEG signals were preprocessed in order to eliminate nuisance and interferences blurring neural information. Precisely, the Matlab toolbox EEGLAB, specifically created for processing EEG recordings, was applied for this purpose [40]. The signals were firstly downsampled from 512 to 128 Hz, and a new reference based on the average potential of all electrodes was established. Later, high-pass and low-pass forward/backward filters were applied at 3 and 45 Hz, respectively, with the aim of maintaining the frequency sub-bands of interest in the EEG spectrum [14]. These filtering approaches also eliminated baseline and power line interferences. Then, artifacts and other interferences not eliminated in previous steps were rejected by means of a blind source separation technique, called independent component analysis (ICA). It is based on the computation of independent components, such that those identified as artifactual were discarded, and only the information related to brain activity was maintained [41]. Briefly, ICA was applied for the removal of artifacts derived from physiological (eye blinks, facial movements, or heart bumps) and technical sources (electrode-pops, or bad contacts of the electrodes on the scalp). Finally, highly contaminated channels were eliminated and reconstructed by interpolating the adjacent electrodes [42]. Although these noisy signals were identified before ICA procedure, the interpolation was done after that step to avoid the nonlinearities derived from the influence of interpolation on ICA decomposition [43]. Hence, the rejection of artifacts was not affected by the existence of the noisy channels replaced by interpolation [43].

2.3. Power Spectral Density Computation

Although the EEG signals collected in the DEAP database had a duration of 60 s, only the last 30 ones were used for power computation from frequency sub-bands, as in previous studies [36,37]. This segment for each EEG recording was then divided into six nonoverlapped epochs of 5 s of length, such that a total number of 822 (137 trials × 6 epochs/trial) and 732 (122 trials × 6 epochs/trial) excerpts were analyzed for emotional states of calm and distress, respectively. PSD was computed from every segment using a Welch’s periodogram, with a Hamming window of 2 s-length, 50% of overlapping between adjacent windows and a resolution of 256 points. Then, the power for every EEG channel was individually obtained for each frequency sub-band ( θ , α , β and γ ) and for the whole band (4–45 Hz) as the area under the PSD curve within the corresponding frequency band. In the case of the frequency sub-bands θ , α , β and γ , the resulting power was normalized by the one obtained for the whole band, with the purpose of preserving the variations between subjects [17]. Equations (1)–(5) describe the power computation for all the band, and for each frequency sub-band, respectively, with P w ( f ) the PSD estimated for each EEG excerpt.
P t = f = 4 Hz f = 45 Hz | P w ( f ) |
P θ = 1 P t f = 4 Hz f = 8 Hz | P w ( f ) |
P α = 1 P t f = 8 Hz f = 13 Hz | P w ( f ) |
P β = 1 P t f = 13 Hz f = 30 Hz | P w ( f ) |
P γ = 1 P t f = 30 Hz f = 45 Hz | P w ( f ) |

2.4. Rearrangement of EEG Channels in 2-D and 3-D Maps

The power values P t , P θ , P α , P β and P γ obtained for the 32 EEG channels were then initially transformed into 2-D images to feed several CNN models. The original locations of the 32 EEG channels on the scalp presented in Figure 1a were mapped, following three different approaches. The first mapping scheme, called direct matrix distribution (DMD), consisted of the placement of the EEG channels in a 9 × 4 matrix, as shown in Figure 1b. Each element of the matrix directly represented one electrode; thus, the disposition of the channels in the matrix was most similar to their real locations over the scalp, without leaving blank spaces. Furthermore, the channels located between left and right brain hemispheres, i.e., Fz, Cz, Pz and Oz, were duplicated in the central columns of the matrix for a symmetrical representation of all electrodes.
The second mapping approach was called direct matrix distribution interpolated (DMDi). As can be observed in Figure 1c, the DMDi approach presented the same distribution of channels as DMD. The only difference was that DMDi interpolated the power values between channels, making use of a biharmonic spline interpolation scheme [44], and thus coloring the whole 2-D surface. On the contrary, DMD presented a grid in which each cell was filled with a color according to the level of power calculated for the corresponding channel.
The third mapping approach was based on the azimuthal equidistant projection (AEP) [45]. Considering the human head as a sphere, the spherical coordinates of each electrode can be converted into Cartesian coordinates. Therefore, locations in the space were projected over a 2-D surface, maintaining proportional distances between electrodes and directions from the central point of the head, which corresponds to channel Cz. This projection of electrodes in a 2-D map is shown in Figure 1d. A biharmonic spline interpolation was also applied to calculate the values between channels.
It should be noted that, in the three mapping approaches, the spectral power values were represented using a Jet colormap with 256 colors, ranging from dark blue (assigned to the minimum value) to dark red (assigned to the maximum value). Furthermore, given that the input layer of AlexNet requires a 227-by-227 input image [46] (as will be mentioned in the next subsection), the 2-D maps obtained by the three mapping approaches were appropriately rescaled. In the cases of DMDi and AEP, each pixel of the image had a different value due to interpolation. On the other hand, DMD images still maintained a 9 × 4 grid of cells, where each one covered a 25 pixel height (227/9) and 56 pixels width (227/4) with the same color. As an example, Figure 1e shows the same EEG segment represented with the three mapping approaches.
Finally, in order to explore complementary information between frequency sub-bands, 3-D images were also constructed by stacking the resulting 2-D maps from each of the three mapping approaches separately. More precisely, for every EEG excerpt, the 2-D images obtained from the parameters P t , P θ , P α , P β and P γ were piled to obtain 3-D cubes with a size of 227 × 227 × 5.

2.5. AlexNet-Based CNN Models

AlexNet is the first documented large-scale, pre-trained CNN architecture, which won the ImageNet Large-Scale Visual Recognition Challenge in 2012 [46]. Since then, AlexNet has been used in a variety of applications [25]. The original model was formed by eight layers (five convolutional and three fully connected ones) with weights and ability to learn. On the one hand, convolutional layers assist in the extraction of features, performing a convolution operation between the input data and a kernel with different settled weights. During the training process, a backpropagation algorithm regulates weights according to the target using a rectifier linear unit (ReLU) as an activation function. Note that data resulting from these layers are usually normalized and downsampled using a max-pooling operation to reduce the spatial dimension of the feature map, while retaining the relevant information and making the network less prone to overfitting [46]. On the other hand, the fully connected layers ensure that all the neurons in the previous layer are connected to all neurons in the current one, such that the number of fully connected neurons in the final layer defines the number of output classes.
The original architecture of 2-D AlexNet is presented in Figure 2. This network receives 2-D images with a size of 227 × 227 and three color channels as input data, which are convolved with 96 kernels of size 11 × 11 × 3. Next, the output is normalized and max-pooled before being transmitted to the second convolutional layer, which filters the resulting feature space with 256 kernels of size 5 × 5 × 48. Then, the feature space is also normalized and max-pooled before being filtered with 384 kernels of size 3 × 3 × 256. The resulting feature space is convolved, with two layers presenting similar kernel sizes of 3 × 3 × 192 and 384 and 256 kernels, respectively. No pooling and normalization operations are applied between the third and fifth convolutional layers. Furthermore, a ReLU function is used after every convolutional and fully connected layer. Finally, the resulting feature space is recombined using three fully connected layers with 4096 neurons each. It is worth noting that a dropout regularization operation is performed after the first two layers, randomly dropping out nodes during the training stage with the aim of decreasing overfitting and improving generalization errors [47].
Recently, this 2-D network has been extended to deal with 3-D cubes as input data [48,49]. This 3-D version of AlexNet has been achieved by modifying its original convolution and max-pooling layers. More precisely, the five convolutional layers maintained 96, 256, 384, 384 and 256 kernels, but with sizes of 11 × 11 × 11 × 3, 5 × 5 × 5 × 96, 3 × 3 × 3 × 256, 3 × 3 × 3 × 384, and 3 × 3 × 3 × 384, respectively. Similarly, the size of the three max-pooling layers was extended to 3 × 3 × 3.

2.6. Fine-Tuning and Learning Parameters of the AlexNet-Based Models

The original 2-D AlexNet was initially trained with a subset of the ImageNet database, composed by more than one million images, to discern among more than 1000 classes [46]. In the present work, the original layer structure of 2-D AlexNet was adopted, except for the last outcome layer, which was modified to deal with only two classes (i.e., calm and distress), as can be observed in Figure 2. Nonetheless, two different training schemes for this network were conducted. On the one hand, a pre-trained version of AlexNet was fine-tuned by taking advantage of TL. Thus, the original weights of the AlexNet were transferred to this study and fine-tuned during the training stage. On the other hand, a 2-D AlexNet network with initial random weights was trained from scratch. In both cases, training was conducted with 40 epochs, 103 learning iterations per epoch, and a mini batch size of 12 signals. For that purpose, a stochastic gradient descent algorithm with a momentum of 0.9 was used. Furthermore, the initial learning rate was set to a constant value of 0.001 during the 4120 total iterations and no learn rate drop factor was established during the training stage. Additionally, L2 weight decay for convolutional weights was set to 0.001 and both weights and bias were updated at each 103 iterations in the direction of the negative gradient of the loss [50]. To avoid overfitting, the training progress was continuously monitored, validating the network every 50 iterations by predicting the response of the test data and calculating the loss and global accuracy on training and test samples. Thus, it was seen that training loss outcomes were comparable with test loss in every checkpoint, while global training accuracy was not significantly higher than test accuracy.
In a similar way, two 3-D AlexNet models were also trained in this study. Firstly, the weights from the original pre-trained 2-D AlexNet network were stacked five times to deal with the 3-D cubes generated, as described in Section 2.4. Secondly, random weights were initially established in the 3-D network. For both cases, the training stage was conducted as aforementioned for the 2-D AlexNet models, but 250 epochs with 50 iterations per epoch were run.

2.7. Experimental Setup and Performance Analysis

Every set of 2-D images obtained from each frequency sub-band, i.e., θ , α , β and γ bands as well as the band covering 4–45 Hz, and from each mapping scheme proposed to distribute the EEG channels into 2-D maps, i.e., DMD, DMDi, and AEP, were used to train the two previously described 2-D AlexNet models. Similarly, the set of 3-D cubes obtained for the three mapping approaches were employed to train the 3-D AlexNet networks with known and random initial weights. To quantify the performance of the resulting CNN model in each case, several validation cycles following a 80/20 hold-out approach were conducted. More precisely, 10 iterations were run for each network, such that, in each one, 658 calm images (out of 822) and 586 distress images (out of 732) were randomly selected for training. Then, the remaining images from each class were used to test the model and values of sensitivity ( S e ), specificity ( S p ) and accuracy ( A c c ) were obtained. Finally, mean and standard deviation ( s t d ) of these performance metrics for the 10 iterations were computed. While S e was defined as the rate of correctly classified distress EEG segments, S p was defined as the percentage of properly identified calm EEG segments. Finally, A c c was computed as the total proportion of correctly detected EEG segments. These metrics were mathematically computed as
S e = T P T P + F N ,
S p = T N T N + F P , and
A c c = T N + T P T N + T P + F N + F P ,
where T P was the number of correctly identified distress EEG segments, T N was the amount of correctly classified calm EEG segments, F P the number of calm segments improperly classified as distress ones, and F N the amount of distress EEG intervals wrongly identified as calm segments.

3. Results

3.1. AlexNet-Based 2-D CNN Models

Table 1 summarizes the classification outcomes obtained by pre-trained 2-D AlexNet networks for each mapping methodology, described in Section 2.4, and each frequency sub-band analyzed. As can be observed, notable differences in classification were obtained among the different frequency bands. On the one hand, P θ , P α and P β achieved comparable global accuracy, ranging from 58.74% to 62.35% when discriminating between calm and distress images. In this respect, DMDi mapping showed a slightly higher performance for P θ and P β bands, almost similar to the score achieved by P α using DMD mapping. However, DMD exhibited lower dispersion among validation cycles than DMDi and AEP mappings. In this respect, values lower than 8% were reported for all the performance indices with DMD representation, whereas higher values of s t d were reported for DMDi and AEP schemes, especially for the S p index. On the contrary, s t d values lower than 3% were reported for A c c , regardless of the mapping scheme used. It is worth noting that average S e and S p metrics were unbalanced on these frequency sub-bands, showing the same trend for the three mapping approaches. Indeed, the ability to detect correctly distress images was between 10 and 15% higher than that used to detect calm images in α and β sub-bands, and around 30% higher for θ sub-band, when DMD and DMDi mapping schemes were used.
Conversely, the performance shown by the γ sub-band was notably higher than in the other frequencies. Firstly, average accuracies were comparable, ranging from 81.3% for DMDi mapping to 82.6% for AEP mapping. This improvement represents an increase of more than 20% with respect to the other frequency sub-bands. Secondly, s t d values lower than 3% were observed for S e and S p and than 1% for A c c . Finally, S e and S p metrics were well-balanced, reporting differences lower than 3%. The average performance outcomes obtained when the whole frequency band was considered ( P t ) were very similar to those obtained by P γ . However, the global accuracy increased by 2%, reaching a final value of 84.77% when DMD mapping was used.
Regarding the 2-D AlexNet-based CNN models with random initial weights, Table 2 summarizes the main classification outcomes in terms of mean and s t d for the 10 validation cycles. In general terms, lower values of A c c than the pre-trained CNN models were provided for all the images obtained from the diverse frequency sub-bands and mapping approaches. More precisely, decreases of about 1 and 2% in values of A c c were noticed for P θ , P α and P β when the DMD mapping was used. The decrease in A c c was still more notable for the AEP mapping, reaching values between 2 and 4% lower than with the pre-trained 2-D networks. Moreover, it is worth noting that average values of S e and S p were significantly unbalanced, especially when the AEP mapping was used. For this mapping approach, high s t d values among the 10 validation cycles were noticed in terms of S e and S p for most of the frequency sub-bands.
The same falling trend was also noticed in the classification performance exhibited by P γ . In this case, A c c values of 76.85%, 77.38% and 77.90% were obtained for DMD, DMDi and AEP mappings, which represented decreases of 6, 4 and 5%, respectively, in comparison with those obtained by the pre-trained 2-D AlexNet-based CNN networks. Nonetheless, in this frequency sub-band, values of S e and S p were balanced for the three mappings, and the dispersion among validation cycles was notably lower than in P θ , P α and P β . As in the case of the pre-trained CNN models, the outcomes obtained for the frequency sub-band covering 4–45 Hz were similar to those reported by the γ sub-band. Thus, a poorer performance between 2 and 7% was noticed in this case.

3.2. AlexNet-Based 3-D CNN Models

The classification results obtained by 3-D AlexNet-based CNN networks, both with initial weights transferred from the pre-trained 2-D AlexNet (i.e., making use of TL) and with random initial weights (i.e., without using TL), are presented in Table 3 for the three mapping approaches. No great differences were noticed among DMD, DMDi and AEP mapping schemes, but, in both cases, the best results were reported by the first mapping approach. It is also interesting to note that about 3% higher values of A c c were always obtained when 3-D CNN models were initialized with the weights transferred from the original pre-trained 2-D AlexNet. Moreover, lower values of s t d among the 10 validation cycles, as well as more balanced values of S e and S p (especially for DMDi and AEP mapping schemes), were also noticed in that case.

4. Discussion

The understanding of how emotions are generated and regulated has become the focus of interest in recent years, fostering the appearance of what is known as the science of emotion regulation [51]. In the literature, a wide range of works which face the challenge of emotion recognition from different perspectives can be found. Nonetheless, to the best of our knowledge, this is the first work focused on distress identification from spectral features of the EEG signal with 2-D and 3-D CNN-based models. The results obtained corroborate the hypothesis that evaluating all brain locations simultaneously with a CNN-based classifier represents an enhancement of the outcomes with respect to the analysis of single and isolated EEG channels, traditionally conducted with machine learning algorithms. To this respect, Table 4 shows information from recent studies dealing with stress recognition using EEG spectral features and machine learning classifiers. As can be observed, the classification results obtained by these works presented accuracy values between 57 and 80% [15,16,17,18,52], which are significantly lower than those obtained in the present study by all pre-trained 2-D and 3-D AlexNet-based networks.
It is important to remark that these studies must be compared with caution, because most of them used different methodologies and experimental protocols, where substantial changes in the number of participants, EEG electrodes, and emotion elicitation ways were found. Nonetheless, two studies included in Table 4, i.e., works [15,52], used the same public database as in the present work, and thus a direct and fair comparison could be established. In this case, the use of pre-trained 2-D and 3-D CNN-based classification models has reported an improvement of around 10–15% with respect to the accuracy obtained by common machine learning classifiers in [15,52]. This outcome is in line with previous studies where deep learning methods have outperformed the results reported by traditional machine learning classifiers, such as SVM and decision trees, for the recognition of different valence and arousal levels [53].
For the three mapping approaches, the best classification outcomes using 2-D CNN-based networks were reported by the relative power obtained from the frequency band covering 4–45 Hz, i.e., P t . The relevance of this power measure with respect to the sub-bands θ , α , β and γ has also been previously described in another study of emotion recognition from EEG signals with CNNs [54]. This could be explained by the fact that these frequency sub-bands are involved in emotional processes, although with a different degree of significance depending on the emotional state. Indeed, classification results obtained by P γ were notably higher than those obtained by the remaining sub-bands. Thus, an improvement of more than 20% of accuracy with respect to P θ , P α and P β was noticed for the three mapping schemes, when discerning between calm and distress. The brain activity in the sub-band γ has already been strongly related to emotional processes, with a special intensity for negative stimuli [55]. Indeed, the induction of negative emotions has reported a higher level of activity in this sub-band with respect to positive and neutral stimuli in healthy subjects [56,57]. Moreover, in previous works P γ has also reported a notable increment in frontal and parietal regions under stressful conditions during task-switching activities [58]. The waves γ reflected on the EEG have also allowed for the differentiation of worry from relaxation and baseline, with an increase in activity in the sub-band γ for the case of patients with anxiety disorder [59].
In the literature, other research has also applied 2-D CNN-based approaches for the classification of different emotions by means of EEG spectral features using similar mapping approaches. In this respect, Li et al. [19] used a mapping scheme similar to DMDi for the representation of spectral power values from four different emotional states. The maximum classification accuracy ranged between 50% and 75%, depending of the window time size selected to estimate spectral features from the EEG channels, and using a combination of CNN and long–short-term memory recurrent neural networks [19]. On the other hand, Li et al. [60] represented spectral features in AEP maps for the identification of depression using different CNN-based classifiers. The accuracy outcomes obtained for the spectral power of the whole frequency band ranged between 76 and 80%, depending on the hyper-parameters used in each CNN model [60]. Although comparisons should be carefully established, the present study outperformed the results from these works, achieving accuracy values of about 85%. This improvement could be a consequence of the different focuses of these works, since the emotional states under study are different in each case.
Another strength of the present work regarding the aforementioned studies is the use of pre-trained AlexNet-based networks, instead of CNN-based models constructed and trained from scratch [19,60]. According to the literature, the use of TL methodologies and fine-tuning parameters to adapt pre-trained networks to different problems could have some advantages over networks constructed from scratch [34]. On the one hand, the number of samples needed to train a CNN from scratch is large due to the high number of parameters that need to be settled. Instead, TL provides an effective way of training complex network architectures using scarce data without overfitting [61]. Furthermore, hyper-parameters have to be configured randomly and then readjusted when a CNN is constructed from scratch, thus requiring extensive computational and memory resources [62]. On the contrary, hyper-parameters only have to be refined when using pre-trained networks. Therefore, most weights are maintained and only a few are filtered to adapt the network to the current classification problem, thus achieving better generalization and outperforming other models constructed from scratch. Accordingly, the results obtained in the present work were better when TL was used to maintain initial weights in pre-trained 2-D and 3-D AlexNet-based networks than when random initial weights were established in similar classifiers, regardless of the mapping approach. Thus, in addition to obtaining values of A c c between 3 and 6%, better, more balanced values of S e and S p and lower dispersion among validation cycles were noticed in the first case.
To assess the criticality of representing neural information in 2-D images used as input data for CNN-based classifiers, three different mapping approaches have been evaluated. All DMD, DMDi, and AEP maps preserved the topology information in EEG channels, since they were distributed to resemble the actual location of the electrodes over the scalp. The difference is that DMD and DMDi were a direct reorganization of all EEG channels in a 9 × 4 matrix, without considering their relative locations, and the AEP model was a projection of the electrodes into a 2-D image. In this last case, the proportions of distance and direction of the EEG channels with respect to a reference point were regarded. In terms of classification performance, the results obtained by the three mapping approaches in each frequency sub-band were similar, with slight differences in values of A c c around 1–3%. Consequently, the spatial information provided by preservation of proportional locations of the EEG channels within AEP images seems not to imply a notable improvement regarding the matrix distribution of the electrodes in DMD and DMDi maps. Furthermore, interpolation between electrodes in the matrix distribution did not entail substantial changes between classification results reported by DMD and DMDi maps. As a result, selection of one of these three mapping approaches would not be crucial to discern between emotional states of calm and distress from EEG recordings. A similar finding has also been obtained in a previous work that compares different mapping approaches to detect high and low levels of arousal and valence with a CNN-based classifier [63]. More precisely, in that work, four matrix distributions with random positions of the EEG channels and two DMD-based schemes were analyzed [63]. No relevant differences were obtained among the six mapping options, thus suggesting that the distribution of EEG channels into 2-D maps does not play a key role in identifying different emotional processes [63]. On the contrary, the presence of information from all EEG channels, instead of only single locations, could be more essential than the distribution of the electrodes in the images.
Similarly, no great improvement in distress detection was noticed when 2-D images were stacked to construct 3-D cubes with simultaneous information from all analyzed frequency sub-bands (i.e., θ , α , β , γ , and the whole frequency band covering 4–45 Hz). Indeed, for the three mapping approaches, 3-D AlexNet-based CNN networks only reported values of A c c 1–2% greater than 2-D classifiers. This result was observed both when the networks maintained initial weights from previous pre-training and when they used random initial weights. The fact that the sub-band γ seems to contain the most relevant information to discern between emotional states of calm and distress could explain that outcome. Indeed, 2-D AlexNet-based classifiers have provided no relevant classification differences between using 2-D maps from the sub-band γ or from the frequency band covering 4–45Hz, thus suggesting that no complementary information exists among the analyzed spectral regions.
To bring any of our CNN-based systems to the real world, we will embed it into a programmable logic device. Although it is well-known that CNN-based algorithms are computationally intense and require vast computational resources and dynamic power for computation of convolutional operations, in recent years, some programmable devices have been specifically developed to run these kinds of algorithms in real-time [64]. In this respect, some researchers have already successfully tested a variety of hardware implementation methods for different CNN-based structures, mostly based on field-programmable gate arrays (FPGA) architectures [65,66]. In fact, at present there are available commercial FPGA-based systems designed for vision artificial intelligence, which can be configured for enhancing the acceleration of vision applications. Interestingly, they contain implemented CNN-based algorithms which can be adapted and tuned for diverse applications [67]. Hence, bearing in mind that technical requirements of our system would be considerably limited regarding those needed by common vision artificial intelligence applications, its implementation in such embedded devices will be feasible. Indeed, our system only analyzes a 227 × 227 × 3 image each five seconds, and its computational load can, therefore, be considered notably low in comparison with other applications requiring real-time image processing with CNN-based methodologies.
Finally, some limitations should be considered. Firstly, every CNN-based algorithm does not allow for understanding the rationale behind its classification results. Thus, it is not possible to give a clinical interpretation of the outcomes, since functional dependencies between input and output information are completely hidden [68]. On the other hand, the database from which the EEG signals were extracted was not specifically created for recognition of calm and distress, since it contains samples corresponding to emotions in the whole valence/arousal space. Moreover, although the DEAP database presents interesting advantages, such as it being freely available and able to obtain comparable results with other previous studies, it contains a limited number of subjects. Additionally, the number of calm and distress samples for each subject is not completely balanced, and some subjects do not present samples from both emotional states. These two aspects aimed for all samples to be considered together for training and testing the CNN-based models, regardless of the subject they came from, such as in some previous works [29]. Nonetheless, because this approach could lead to some overfitting, additional experiments considering separate subjects for training and testing every classifier were also conducted. In this respect, the pre-trained 2-D and 3-D AlexNet-based CNN networks were trained and tested through 10 validation cycles, where, in each one, the participants randomly selected for training had approximately the 80% of the samples, and the remaining ones were used for testing. In this case, for the three mapping approaches, only values of A c c about 3–4% and 1–2% lower than those presented in Section 3 were noticed for pre-trained 2-D and 3-D classifiers, respectively, thus suggesting that the proposed algorithms for distress recognition were not significantly overtrained.

5. Conclusions

The present work has introduced, for the first time, the use of 2-D and 3-D CNN-based classifiers to discern emotional states of calm and distress by taking advantage of the brain’s behavior as a synchronized network. Thus, receiving 2-D and 3-D spectral maps constructed from common 32 channel EEG recordings as input data, the algorithms have provided significantly better classification of both emotions than previous methodologies based on combining isolated information from different brain regions with machine learning algorithms. Moreover, the use of pre-trained CNN-based classification models has also improved the diagnostic accuracy reported by other similar deep learning algorithms trained from scratch, thus highlighting the usefulness of transfer learning in distress recognition. Finally, no significant differences between the results provided by the three approaches considered to reconstruct 2-D spectral maps were noticed, then suggesting an independency of the distribution of the EEG electrodes in the maps.

Author Contributions

A.M.-R., B.G.-M. and R.A. conceived and designed the study. A.M.-R. programmed the experiments and A.M.-R. and R.A. supervised the study. A.M.-R. and B.G.-M. wrote the manuscript, and A.M.-R., R.A. and Á.H. reviewed the manuscript and contributed to the final version. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by grants DPI2017–83952–C3 from MINECO/AEI/FEDER EU, 2018/11744 from European Regional Development Fund (FEDER/EU), and SBPLY/17/ 180501/000411 from Junta de Comunidades de Castilla-La Mancha. Beatriz García-Martínez holds FPU16/03740 scholarship from Spanish Ministerio de Educación y Formación Profesional.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Epel, E.S.; Crosswell, A.D.; Mayer, S.E.; Prather, A.A.; Slavich, G.M.; Puterman, E.; Mendes, W.B. More than a feeling: A unified view of stress measurement for population science. Front. Neuroendocrinol. 2018, 49, 146–169. [Google Scholar]
  2. O’Connor, D.B.; Thayer, J.F.; Vedhara, K. Stress and health: A review of psychobiological processes. Annu. Rev. Psychol. 2021, 72, 663–688. [Google Scholar] [CrossRef]
  3. Zhong, S.; Shalev, I.; Koh, D.; Ebstein, R.P.; Chew, S.H. Competitiveness and stress. Int. Econ. Rev. 2018, 59, 1263–1281. [Google Scholar] [CrossRef]
  4. Vinkers, C.H.; van Amelsvoort, T.; Bisson, J.I.; Branchi, I.; Cryan, J.F.; Domschke, K.; Manchia, M.; Pinto, L.; de Quervain, D.; Schmidt, M.V.; et al. Stress resilience during the coronavirus pandemic. Eur. Neuropsychopharmacol. 2020, 35, 12–16. [Google Scholar] [CrossRef]
  5. Dhabhar, F.S. The short-term stress response—Mother nature’s mechanism for enhancing protection and performance under conditions of threat, challenge, and opportunity. Front. Neuroendocrinol. 2018, 49, 175–192. [Google Scholar] [PubMed]
  6. Cruz-Pereira, J.D.; Rea, K.; Nolan, Y.M.; O’Leary, O.F.; Dinan, T.G.; Cryan, J.F. Depression’s unholy trinity: Dysregulated stress, immunity, and the microbiome. Annu. Rev. Psychol. 2019, 71, 49–78. [Google Scholar] [CrossRef]
  7. Robertson, A.R. What becomes of the frequent hospital attenders with centrally mediated gastrointestinal pain syndrome? Visc. Med. 2020, 36, 312–317. [Google Scholar] [CrossRef]
  8. Gross, J.J. Emotion regulation: Affective, cognitive, and social consequences. Psychophysiology 2002, 39, 281–291. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Wagh, K.P.; Vasanth, K. Electroencephalograph (EEG) based emotion recognition system: A review. In Innovations in Electronics and Communication Engineering; Springer: Berlin/Heidelberg, Germany, 2019; pp. 37–59. [Google Scholar]
  10. Egger, M.; Ley, M.; Hanke, S. Emotion recognition from physiological signal analysis: A review. Electron. Notes Theor. Comput. Sci. 2019, 343, 35–55. [Google Scholar] [CrossRef]
  11. Dzedzickis, A.; Kaklauskas, A.; Bucinskas, V. Human emotion recognition: Review of sensors and methods. Sensors 2020, 20, 592. [Google Scholar]
  12. Alarcao, S.M.; Fonseca, M.J. Emotions recognition using EEG signals: A survey. IEEE Trans. Affect. Comput. 2017, 10, 374–393. [Google Scholar] [CrossRef]
  13. Zeng, H.; Yang, C.; Dai, G.; Qin, F.; Zhang, J.; Kong, W. EEG classification of driver mental states by deep learning. Cogn. Neurodyn. 2018, 12, 597–606. [Google Scholar] [CrossRef]
  14. Ismail, W.W.; Hanif, M.; Mohamed, S.; Hamzah, N.; Rizman, Z.I. Human emotion detection via brain waves study by using electroencephalogram (EEG). Int. J. Adv. Sci. Eng. Inf. Technol. 2016, 6, 1005–1011. [Google Scholar] [CrossRef] [Green Version]
  15. Shon, D.; Im, K.; Park, J.H.; Lim, D.S.; Jang, B.; Kim, J.M. Emotional stress state detection using genetic algorithm-based feature selection on EEG signals. Int. J. Environ. Res. Public Health 2018, 15, 2461. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Jebelli, H.; Hwang, S.; Lee, S. EEG-based workers’ stress recognition at construction sites. Autom. Constr. 2018, 93, 315–324. [Google Scholar] [CrossRef]
  17. Ahn, J.W.; Ku, Y.; Kim, H.C. A novel wearable EEG and ECG recording system for stress assessment. Sensors 2019, 19, 1991. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Arsalan, A.; Majid, M.; Butt, A.R.; Anwar, S.M. Classification of perceived mental stress using a commercially available EEG headband. IEEE J. Biomed. Health Inform. 2019, 23, 2257–2264. [Google Scholar] [CrossRef]
  19. Li, Y.; Huang, J.; Zhou, H.; Zhong, N. Human emotion recognition with electroencephalographic multidimensional features by hybrid deep neural networks. Appl. Sci. 2017, 7, 1060. [Google Scholar] [CrossRef] [Green Version]
  20. Zhou, M.; Tian, C.; Cao, R.; Wang, B.; Niu, Y.; Hu, T.; Guo, H.; Xiang, J. Epileptic seizure detection based on EEG signals and CNN. Front. Neuroinform. 2018, 12, 95. [Google Scholar] [CrossRef] [Green Version]
  21. Raghu, S.; Sriraam, N.; Temel, Y.; Rao, S.V.; Kubben, P.L. EEG based multi-class seizure type classification using convolutional neural network and transfer learning. Neural Netw. 2020, 124, 202–212. [Google Scholar] [CrossRef]
  22. Ieracitano, C.; Mammone, N.; Bramanti, A.; Hussain, A.; Morabito, F.C. A convolutional neural network approach for classification of dementia stages based on 2D-spectral representation of EEG recordings. Neurocomputing 2019, 323, 96–107. [Google Scholar] [CrossRef]
  23. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adeli, H.; Subha, D.P. Automated EEG-based screening of depression using deep convolutional neural network. Comput. Methods Programs Biomed. 2018, 161, 103–113. [Google Scholar] [CrossRef] [PubMed]
  24. Oh, S.L.; Hagiwara, Y.; Raghavendra, U.; Yuvaraj, R.; Arunkumar, N.; Murugappan, M.; Acharya, U.R. A deep learning approach for Parkinson’s disease diagnosis from EEG signals. Neural Comput. Appl. 2018, 32, 10927–10933. [Google Scholar] [CrossRef]
  25. Aloysius, N.; Geetha, M. A review on deep convolutional neural networks. In Proceedings of the 2017 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 6–8 April 2017; pp. 0588–0592. [Google Scholar]
  26. Suhaimi, N.S.; Mountstephens, J.; Teo, J. EEG-based emotion recognition: A state-of-the-art review of current trends and opportunities. Comput. Intell. Neurosci. 2020, 2020, 8875426. [Google Scholar] [CrossRef] [PubMed]
  27. Kwon, Y.H.; Shin, S.B.; Kim, S.D. Electroencephalography based fusion two-dimensional (2D)-convolution neural networks (CNN) model for emotion recognition system. Sensors 2018, 18, 1383. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Li, X.; Song, D.; Zhang, P.; Hou, Y.; Hu, B. Deep fusion of multi-channel neurophysiological signal for emotion recognition and monitoring. Int. J. Data Min. Bioinform. 2017, 18, 1–27. [Google Scholar] [CrossRef]
  29. Mishra, A.; Ranjan, P.; Ujlayan, A. Empirical analysis of deep learning networks for affective video tagging. Multimed. Tools Appl. 2020, 79, 18611–18626. [Google Scholar] [CrossRef]
  30. Anzellotti, S.; Coutanche, M.N. Beyond functional connectivity: Investigating networks of multivariate representations. Trends Cogn. Sci. 2018, 22, 258–269. [Google Scholar] [CrossRef]
  31. Sala-Llonch, R.; Bartrés-Faz, D.; Junqué, C. Reorganization of brain networks in aging: A review of functional connectivity studies. Front. Psychol. 2015, 6, 663. [Google Scholar] [CrossRef] [Green Version]
  32. Deshpande, G.; Santhanam, P.; Hu, X. Instantaneous and causal connectivity in resting state brain networks derived from functional MRI data. Neuroimage 2011, 54, 1043–1052. [Google Scholar] [CrossRef] [Green Version]
  33. Rabinovich, M.I.; Muezzinoglu, M.K. Nonlinear dynamics of the brain: Emotion and cognition. Physics-Uspekhi 2010, 53, 357–372. [Google Scholar] [CrossRef]
  34. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 1–40. [Google Scholar] [CrossRef] [Green Version]
  35. Bashivan, P.; Rish, I.; Yeasin, M.; Codella, N. Learning Representations from EEG with Deep Recurrent-Convolutional Neural Networks. In Proceedings of the International Conference on Learning Representations 2016, San Juan, PR, USA, 2–4 May 2016. [Google Scholar]
  36. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A database for emotion analysis using physiological signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  37. García-Martínez, B.; Martínez-Rodrigo, A.; Zangróniz, R.; Pastor, J.M.; Alcaraz, R. Application of entropy-based metrics to identify emotional distress from electroencephalographic recordings. Entropy 2016, 18, 221. [Google Scholar] [CrossRef]
  38. García-Martínez, B.; Martínez-Rodrigo, A.; Zangróniz, R.; Pastor, J.M.; Alcaraz, R. Symbolic analysis of brain dynamics detects negative stress. Entropy 2017, 19, 196. [Google Scholar] [CrossRef] [Green Version]
  39. Klem, G.H.; Lüders, H.O.; Jasper, H.; Elger, C. The ten-twenty electrode system of the International Federation. The International Federation of Clinical Neurophysiology. Electroencephalogr. Clin. Neurophysiol. 1999, 52, 3–6. [Google Scholar]
  40. Delorme, A.; Makeig, S. EEGLAB: An open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J. Neurosci. Methods 2004, 134, 9–21. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Park, S.; Kwak, N. Independent component analysis by lp-norm optimization. Pattern Recognit. 2018, 76, 752–760. [Google Scholar] [CrossRef]
  42. Pedroni, A.; Bahreini, A.; Langer, N. Automagic: Standardized preprocessing of big EEG data. Neuroimage 2019, 200, 460–473. [Google Scholar] [CrossRef] [Green Version]
  43. Nolan, H.; Whelan, R.; Reilly, R.B. FASTER: Fully automated statistical thresholding for EEG artifact rejection. J. Neurosci. Methods 2010, 192, 152–162. [Google Scholar] [CrossRef]
  44. Sandwell, D.T. Biharmonic spline interpolation of GEOS-3 and SEASAT altimeter data. Geophys. Res. Lett. 1987, 14, 139–142. [Google Scholar] [CrossRef] [Green Version]
  45. Snyder, J.P. Map Projections–A Working Manual; US Government Printing Office: Washington, DC, USA, 1987; Volume 1395.
  46. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  47. Guo, Y.; Liu, Y.; Oerlemans, A.; Lao, S.; Wu, S.; Lew, M.S. Deep learning for visual understanding: A review. Neurocomputing 2016, 187, 27–48. [Google Scholar] [CrossRef]
  48. Polat, H.; Danaei Mehr, H. Classification of pulmonary CT images by using hybrid 3D-deep convolutional neural network architecture. Appl. Sci. 2019, 9, 940. [Google Scholar] [CrossRef] [Green Version]
  49. Zhang, J.; Li, X.; Li, Y.; Wang, M.; Huang, B.; Yao, S.; Shen, L. Three dimensional convolutional neural network-based classification of conduct disorder with structural MRI. Brain Imaging Behav 2020, 14, 2333–2340. [Google Scholar] [CrossRef]
  50. Pascanu, R.; Mikolov, T.; Bengio, Y. On the difficulty of training recurrent neural networks. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; Volume 28, pp. 1310–1318. [Google Scholar]
  51. Gross, J.J. Handbook of Emotion Regulation; Guilford Publications: New York, NY, USA, 2013. [Google Scholar]
  52. Hasan, M.J.; Kim, J.M. A hybrid feature pool-based emotional stress state detection algorithm using EEG signals. Brain Sci. 2019, 9, 376. [Google Scholar] [CrossRef] [Green Version]
  53. Tao, W.; Li, C.; Song, R.; Cheng, J.; Liu, Y.; Wan, F.; Chen, X. EEG-based emotion recognition via channel-wise attention and self attention. IEEE Trans. Affect. Comput. 2020. [Google Scholar] [CrossRef]
  54. Chen, J.X.; Zhang, P.W.; Mao, Z.J.; Huang, Y.F.; Jiang, D.M.; Zhang, Y.N. Accurate EEG-based emotion recognition on combined features using deep convolutional neural networks. IEEE Access 2019, 7, 44317–44328. [Google Scholar] [CrossRef]
  55. Güntekin, B.; Başar, E. A review of brain oscillations in perception of faces and emotional pictures. Neuropsychologia 2014, 58, 33–51. [Google Scholar] [CrossRef] [PubMed]
  56. Martini, N.; Menicucci, D.; Sebastiani, L.; Bedini, R.; Pingitore, A.; Vanello, N.; Milanesi, M.; Landini, L.; Gemignani, A. The dynamics of EEG gamma responses to unpleasant visual stimuli: From local activity to functional connectivity. Neuroimage 2012, 60, 922–932. [Google Scholar] [CrossRef] [PubMed]
  57. Li, Y.; Cao, D.; Wei, L.; Tang, Y.; Wang, J. Abnormal functional connectivity of EEG gamma band in patients with depression during emotional face processing. Clin. Neurophysiol. 2015, 126, 2078–2089. [Google Scholar] [CrossRef]
  58. Lin, C.T.; King, J.T.; Fan, J.W.; Appaji, A.; Prasad, M. The influence of acute stress on brain dynamics during task switching activities. IEEE Access 2018, 6, 3249–3255. [Google Scholar] [CrossRef]
  59. Oathes, D.J.; Ray, W.J.; Yamasaki, A.S.; Borkovec, T.D.; Castonguay, L.G.; Newman, M.G.; Nitschke, J. Worry, generalized anxiety disorder, and emotion: Evidence from the EEG gamma band. Biol. Psychol. 2008, 79, 165–170. [Google Scholar] [CrossRef] [Green Version]
  60. Li, X.; Zhang, X.; Zhu, J.; Mao, W.; Sun, S.; Wang, Z.; Xia, C.; Hu, B. Depression recognition using machine learning methods with different feature generation strategies. Artif. Intell. Med. 2019, 99, 101696. [Google Scholar] [CrossRef] [PubMed]
  61. Huang, Z.; Pan, Z.; Lei, B. Transfer learning with deep convolutional neural network for SAR target classification with limited labeled data. Remote Sens. 2017, 9, 907. [Google Scholar] [CrossRef] [Green Version]
  62. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Li, Y.; Wong, C.M.; Zheng, Y.; Wan, F.; Mak, P.U.; Pun, S.H.; Vai, M.I. EEG-based emotion recognition under convolutional neural network with differential entropy feature maps. In Proceedings of the 2019 IEEE International Conference on Computational Intelligence and Virtual Environments for Measurement Systems and Applications, CIVEMSA 2019—Proceedings, Tianjin, China, 14–16 June 2019; pp. 1–5. [Google Scholar]
  64. Pantho, M.J.H.; Bhowmik, P.; Bobda, C. Towards an Efficient CNN Inference Architecture Enabling In-Sensor Processing. Sensors 2021, 21, 1955. [Google Scholar] [CrossRef]
  65. Zhang, C.; Li, P.; Sun, G.; Guan, Y.; Xiao, B.; Cong, J. Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks. In Proceedings of the 2015 ACM/SIGDA International Symposium on Field-Programmable Gate Arrays, Monterey, CA, USA, 22–24 February 2015. [Google Scholar] [CrossRef]
  66. Wai, Y.J.; Yussof, Z.M.; Salim, S.I.M. A scalable FPGA based accelerator for Tiny-YOLO-v2 using OpenCL. Int. J. Embed. Syst. 2019, 8, 206. [Google Scholar] [CrossRef]
  67. Xilinx. Kria K26 System-on-Module Home Page. Available online: https://www.xilinx.com/products/som/kria/k26c-commercial.html (accessed on 26 April 2021).
  68. Mincholé, A.; Rodriguez, B. Artificial intelligence for the electrocardiogram. Nat. Med. 2019, 25, 22–23. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) Representation of 32 EEG channels on the scalp according to the 10–20 system. (b) Distribution of 32 EEG channels in the DMD mapping scheme. (c) Distribution of 32 EEG channels in the DMDi mapping scheme. (d) Distribution of 32 EEG channels in the AEP mapping scheme. (e) Example of P t values from the same segment represented with the three mapping approaches.
Figure 1. (a) Representation of 32 EEG channels on the scalp according to the 10–20 system. (b) Distribution of 32 EEG channels in the DMD mapping scheme. (c) Distribution of 32 EEG channels in the DMDi mapping scheme. (d) Distribution of 32 EEG channels in the AEP mapping scheme. (e) Example of P t values from the same segment represented with the three mapping approaches.
Sensors 21 03050 g001
Figure 2. Illustration of the sequential layer-based architecture of AlexNet used in this study.
Figure 2. Illustration of the sequential layer-based architecture of AlexNet used in this study.
Sensors 21 03050 g002
Table 1. Classification results reported by pre-trained AlexNet networks for 2-D images obtained from P θ , P α , P β , P γ and P t using DMD, DMDi and AEP mapping schemes.
Table 1. Classification results reported by pre-trained AlexNet networks for 2-D images obtained from P θ , P α , P β , P γ and P t using DMD, DMDi and AEP mapping schemes.
DMDDMDiAEP
Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)
P θ Mean77.2643.3661.2974.7648.4262.3566.9553.4960.61
Std2.423.172.433.3511.481.7513.6816.832.07
P α Mean65.3056.1661.0068.2352.6760.9065.1251.5858.74
Std3.587.150.4212.7919.201.544.494.320.09
P β Mean68.2950.4859.9063.2358.9761.2361.9559.7960.94
Std6.457.521.053.936.740.4810.1612.761.02
P γ Mean81.883.282.479.982.981.382.582.782.6
Std1.982.770.691.491.370.452.182.230.53
P t Mean85.2484.2584.7780.6782.6081.5883.6684.5284.06
Std0.623.020.441.522.860.660.390.950.35
Table 2. Classification results reported by AlexNet networks with random initial weights for 2-D images obtained from P θ , P α , P β , P γ and P t using DMD, DMDi and AEP mapping schemes.
Table 2. Classification results reported by AlexNet networks with random initial weights for 2-D images obtained from P θ , P α , P β , P γ and P t using DMD, DMDi and AEP mapping schemes.
DMDDMDiAEP
Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)
P θ Mean69.1347.7759.0780.9532.1157.9481.4829.2056.85
Std10.7111.181.2511.3729.331.4420.2228.481.39
P α Mean74.7044.8660.6574.7044.8660.6087.0421.3256.09
Std13.2039.652.2713.2039.652.2714.4839.411.65
P β Mean55.4964.7359.8471.4939.7356.5361.2053.0057.34
Std8.319.500.7221.3238.920.665.4310.860.48
P γ Mean76.8376.8876.8580.5673.8077.3875.8480.2277.90
Std3.646.180.221.141.311.540.201.580.29
P t Mean76.7580.9178.7180.0379.7979.9276.8376.8876.85
Std2.291.430.181.114.752.293.646.180.23
Table 3. Classification results reported by AlexNet-based CNN networks with initial weights transferred from the pre-trained 2-D AlexNet and with random initial weights for 3-D cubes obtained with DMD, DMDi and AEP mapping schemes.
Table 3. Classification results reported by AlexNet-based CNN networks with initial weights transferred from the pre-trained 2-D AlexNet and with random initial weights for 3-D cubes obtained with DMD, DMDi and AEP mapping schemes.
Initials
Weights
DMDDMDiAEP
Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)Se (%)Sp (%)Acc (%)
KnownMean87.1184.7786.1285.2584.3284.8783.4885.2584.18
(with TL)Std0.711.181.251.371.771.440.881.481.39
RandomMean82.4283.2782.8777.3284.7181.2375.3285.1580.52
(without TL)Std5.396.904.053.422.863.63.424.312.14
Table 4. Comparison of the present study with other works detecting stress from power spectral features of EEG signals and machine learning classification algorithms.
Table 4. Comparison of the present study with other works detecting stress from power spectral features of EEG signals and machine learning classification algorithms.
WorkExperimentClassifierResults
Jebelli et al. [16]7 subjects
14 EEG channels
Construction work
SVMMax A c c : 80.32%
Shon et al. [15]32 subjects
32 EEG channels
Videoclips
k-NNMax A c c : 71.76%
Ahn et al. [17]7 subjects
2 EEG channels
Eyes open and close
SVMMax A c c : 77.90%
Arsalan et al. [18]28 subjects
4 EEG channels
Resting pre and post-activity
SVM, NB 1 and MLP 2 57–71% with all frequency bands
Hasan & Kim [52]32 subjects
32 EEG channels
Videoclips
k-NNMax A c c : 73.38%
This work32 subjects
32 EEG channels
Videoclips
2-D AlexNet
3-D AlexNet
Max A c c : 84.77%
Max A c c : 86.12%
1 NB: Naive Bayes; 1 MLP: Multilayer perceptron.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Martínez-Rodrigo, A.; García-Martínez, B.; Huerta, Á.; Alcaraz, R. Detection of Negative Stress through Spectral Features of Electroencephalographic Recordings and a Convolutional Neural Network. Sensors 2021, 21, 3050. https://doi.org/10.3390/s21093050

AMA Style

Martínez-Rodrigo A, García-Martínez B, Huerta Á, Alcaraz R. Detection of Negative Stress through Spectral Features of Electroencephalographic Recordings and a Convolutional Neural Network. Sensors. 2021; 21(9):3050. https://doi.org/10.3390/s21093050

Chicago/Turabian Style

Martínez-Rodrigo, Arturo, Beatriz García-Martínez, Álvaro Huerta, and Raúl Alcaraz. 2021. "Detection of Negative Stress through Spectral Features of Electroencephalographic Recordings and a Convolutional Neural Network" Sensors 21, no. 9: 3050. https://doi.org/10.3390/s21093050

APA Style

Martínez-Rodrigo, A., García-Martínez, B., Huerta, Á., & Alcaraz, R. (2021). Detection of Negative Stress through Spectral Features of Electroencephalographic Recordings and a Convolutional Neural Network. Sensors, 21(9), 3050. https://doi.org/10.3390/s21093050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop