Next Article in Journal
WiFi Signal-Based Gesture Recognition Using Federated Parameter-Matched Aggregation
Next Article in Special Issue
A Hidden Markov Ensemble Algorithm Design for Time Series Analysis
Previous Article in Journal
Sustainable Solutions for Advanced Energy Management System of Campus Microgrids: Model Opportunities and Future Challenges
Previous Article in Special Issue
Temporal Monitoring and Predicting of the Abundance of Malaria Vectors Using Time Series Analysis of Remote Sensing Data through Google Earth Engine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Feature Extraction on AsMap for Emotion Classification Using EEG

by
Md. Zaved Iqubal Ahmed
1,*,
Nidul Sinha
2,
Souvik Phadikar
2 and
Ebrahim Ghaderpour
3,*
1
Department of Computer Science & Engineering, National Institute of Technology, Silchar 788010, India
2
Department of Electrical Engineering, National Institute of Technology, Silchar 788010, India
3
Department of Geomatics Engineering, University of Calgary, Calgary, AB T2N 1N4, Canada
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(6), 2346; https://doi.org/10.3390/s22062346
Submission received: 24 February 2022 / Revised: 14 March 2022 / Accepted: 14 March 2022 / Published: 18 March 2022
(This article belongs to the Special Issue Advances in Time Series Analysis)

Abstract

:
Emotion recognition using EEG has been widely studied to address the challenges associated with affective computing. Using manual feature extraction methods on EEG signals results in sub-optimal performance by the learning models. With the advancements in deep learning as a tool for automated feature engineering, in this work, a hybrid of manual and automatic feature extraction methods has been proposed. The asymmetry in different brain regions is captured in a 2D vector, termed the AsMap, from the differential entropy features of EEG signals. These AsMaps are then used to extract features automatically using a convolutional neural network model. The proposed feature extraction method has been compared with differential entropy and other feature extraction methods such as relative asymmetry, differential asymmetry and differential caudality. Experiments are conducted using the SJTU emotion EEG dataset and the DEAP dataset on different classification problems based on the number of classes. Results obtained indicate that the proposed method of feature extraction results in higher classification accuracy, outperforming the other feature extraction methods. The highest classification accuracy of 97.10% is achieved on a three-class classification problem using the SJTU emotion EEG dataset. Further, this work has also assessed the impact of window size on classification accuracy.

1. Introduction

Human emotions play a central role in decision making, social interaction, diagnosis of mental conditions such as depression, etc. [1,2]. Traditionally, humans identify emotions using facial expressions, audio signals, body pose, gesture, etc. [3]. In contrast, machines cannot understand the feelings of an individual. In this context, affective computing aims to improve communication among individuals and machines by recognizing human emotions, thus making this interaction more accessible, usable, and effective [4].
Emotional experience is associated with physiological changes in the body. Therefore, the knowledge of the physiological reaction of every emotion is essential to emotion analysis [5]. Thus, research works have been conducted to recognize emotions using physiological signals. The physiological signals [6,7] are internal signals, such as electroencephalogram (EEG), electrocardiogram, heart rate, electromyogram (EMG), and galvanic skin response (GSR). According to Cannon’s theory [8], the emotion changes are associated with quick responses in physiological signals coordinated by the autonomic nervous systems. This makes the physiological signals not easily controlled and overcomes the shortcomings of bodily expressions [7].
The advancement of brain–computer interface (BCI) devices and their ease of operation has motivated research on emotion recognition using EEG signals. Some of the non-invasive EEG devices are Emotiv Epoc, Emotiv Insight, Neurosky MindWave, InteraXon Muse, and OpenBCI. These devices are low-cost and portable, thus making EEG signals highly accessible. These devices are accompanied by tools for various BCI applications as well. The EEG signals are captured from individuals (or subjects) using the BCI devices and analyzed using computers to identify the emotion class. At the heart of emotion recognition lies the task of emotion classification. Emotion classification is the process of distinguishing one emotion from another. Emotions are categorized based on two types of models: categorical models and dimensional models. The categorical model categorizes emotions into discrete classes, commonly anger, disgust, fear, joy, sadness, and surprise [9]. Based on facial expression, Ekman listed six basic emotions: happiness, anger, fear, sadness, surprise, and disgust [10]. On the other hand, the dimensional emotion model suggests that emotions can be placed in one or more dimensions rather than in categories. One of the popular dimensional models is the Circumplex model, where emotions are placed into two dimensions: valence (a continuum that varies from negative to positive) and arousal (a continuum that varies from low to high) [11].
Considering the emotion models, various research works have been conducted to trigger emotional events using images, music, audio-visual cues, etc., and subsequently record the EEG signals from individuals. Some of the popular publicly available EEG datasets prepared by applying audio-visual stimuli are DEAP [12] and SEED [13]. The EEG signals from the datasets are used by machine learning models in order to learn how to classify different emotions. Traditional machine learning approaches such as support vector machine [14,15,16,17,18], linear discriminant analysis [19,20], quadratic discriminant analysis [21], k-nearest neighbors [16,21,22,23], Naïve Bayes [20], feed-forward neural network [24], deep belief network [25], multi-layer perceptron neural network [22], etc., are commonly used in EEG-based emotion classification.
In this context, raw time-domain EEG signals are very complex to be handled by the machine learning models as the signals are non-stationary and contaminated by artifacts. Some of the significant physiological artifacts in EEG signals are eye movement, muscle activity, and eye blinks. Various research works have been conducted to remove artifacts from EEG signals [26]. Recently, automatic artifact removal techniques have gained much popularity [27,28]. After removal of artifacts, the most important task is feature extraction. Feature extraction methods are applied to reduce the complexity as well as the dimensionality of input data to the learning models. Features are commonly extracted from the delta, theta, alpha, beta, and gamma frequency bands. Some of the feature extraction methods available in the literature are the asymmetry measure [16], power spectral density (PSD) [14], differential entropy (DE) [16], wavelet transform [22,29,30], higher-order crossings [21], common spatial patterns [15], asymmetry index [31], differential asymmetry (DASM), relative asymmetry (RASM), and differential caudality (DCAU) [25]. Most feature extraction methods are manual and the selection of an appropriate method for emotion classification is still a challenging task [32].
In recent years, research works on automatic feature extraction using deep learning models have been explored in various problems such as speech recognition, vision system, pattern recognition, etc. [33]. Convolutional neural networks (CNNs) have shown tremendous capability in extracting spatial features from input data such as images, etc. Various research works [34,35,36,37,38,39] claim that deep learning models have shown their ability in emotion classification using EEG over traditional approaches. The authors in [34] proposed a feature extraction method that combines CNN and RNN. The CNN is used to extract spatial features and RNN is employed to extract temporal features. Both the feature vectors obtained from CNN and RNN are concatenated and given as input to the learning model. Classification accuracy of 90.80% and 91.03% was achieved for valence and arousal classification, respectively, on the DEAP dataset. In [35], raw EEG data are given as input to a CNN architecture having 3D convolution kernels. The automated features extracted using 3D-CNN result in arousal and valence classification accuracy of 73.1% and 72.1%, respectively, on the DEAP dataset. Moon et al. in [39] proposed a CNN-based approach for automated feature extraction. Three connectivity features, namely the Pearson correlation coefficient, phase-locking value, and phase lag index, are used to measure the cross-electrode relationship. Each connectivity feature is transformed into a 2D vector and given as input to different CNN models, such as CNN-2, CNN-5, and CNN-10, for automated feature extraction. The authors claimed accuracy of 99.72% for valence classification on the DEAP dataset using CNN-5 with phase-locking value matrices. The authors in [38] proposed an automated emotion classification method using the CNN model on time-domain and frequency-domain features.
In this work, a novel feature extraction method for emotion classification has been proposed. The EEG signals are first segmented into segments of fixed window size, and on each segment, DE features are calculated on five frequency bands. The method then generates a 2D feature map, termed the asymmetric map (AsMap), from the DE features obtained from an EEG segment. The AsMap features are then fed into a CNN for automated feature learning. The DE features give a measure of the randomness in the EEG signal. The DE of an EEG segment is considered to be equivalent to the logarithm energy spectrum of a specific frequency band [16]. The mathematical aspects of DE have been further discussed in Section 2.2.1. Other feature extraction methods such as DASM, RASM, and DCAU are derived from DE features. DASM is the difference in DE features on channels between two brain hemispheres. On the other hand, RASM is the ratio in DE features on channels between two brain hemispheres. In DCAU, the difference between the DE features on frontal and posterior brain regions is calculated. However, the AsMap represents the difference between DE features between every channel pair in a 2D vector. Thus, capturing all the possible inter-channel asymmetry in the spatial domain results in more discriminating features compared to other methods such as DASM, RASM, etc. Further, the windowing/segmentation process also provides time-domain resolution for each AsMap. Thus, the AsMap captures both temporal as well as spatial features from all brain regions. The proposed method has been tested on the SEED as well as on the DEAP dataset and compared with other features such as DE, DASM, RASM, and DCAU. Different classification scenarios have been tested on the proposed method.
The rest of the paper is organized as follows. In Section 2, the materials and methods used in automated feature extraction for emotion classification using the AsMap are discussed. Later, in Section 3, the results obtained during the experiment are presented. Section 4 provides a discussion of the contributions and the limitations of the proposed method. Lastly, Section 5 gives the conclusions and future work.

2. Materials and Methods

2.1. Public Datasets

2.1.1. SJTU Emotion EEG Dataset (SEED)

Zheng et al. [25] prepared an EEG emotion dataset in the Center for Brain-Like Computing and Machine Intelligence Laboratory by recording EEG signals. At the same time, participants were subjected to audio-visual stimuli. A total of 15 participants, comprising 7 males and 8 females, were part of the experiment. The SEED dataset considers three basic human emotions named positive, negative, and neutral. Positive emotion describes a pleasant or desirable state of mind, ranging from interest to contentment. On the other hand, a negative emotion depicts an unpleasant or unhappy state. Finally, the neutral emotion is associated with the feeling of indifference, nothing in particular, and a lack of preference. These emotions were elicited using 15 Chinese movie clips of length of around 4 min. Each trial of the experiment had 5 s indicating the start, followed by the presentation of the movie clip. After completion of the movie, each participant was allotted 45 s for their self-assessment, and lastly, a 5 s resting time was provided. The self-assessment involved the following questions: (1) what did they feel after watching the movie clip? (2) is he/she familiar with the movie clip? (3) have they understood the movie clip?
The EEG signals were captured using 62 electrodes placed according to the 10–20 system. The SEED dataset contains two parts: the first part contains the processed EEG recordings and the second part contains some extracted features. In the first part, the EEG recordings are down-sampled to 200 Hz, and EEG recordings containing artifacts such as EOG and EMG were visually checked. The recordings seriously contaminated by EMG and EOG were removed manually. In order to filter the noise and remove the artifacts, a bandpass frequency filter from 0.3 to 50.0 Hz was applied. The dataset includes only the EEG captured while watching the movie clip, with the rest eliminated. For the second part, each channel of the EEG data was divided into same-length epochs of 1 s without overlapping. There were around 3300 clean epochs for one experiment. Features such as PSD, DE, DASM, RASM, and DCAU were computed on each epoch of the EEG data. The dimensions of PSD, DE, DASM, RASM, and DCAU features obtained were 310, 310, 135, 135, and 115, respectively. In order to further filter out irrelevant components, each feature vector was further smoothed using conventional moving averages and linear dynamic systems, which are then provided as separate feature vectors.
One of the limitations of the SEED is that it was prepared on very few participants. Moreover, the annotation of the video clips with emotion classes was not done by the participants. Thus, the participants’ assessments after watching the videos were not considered for annotation in this dataset.

2.1.2. Database for Emotion Analysis Using Physiological Signals (DEAP)

Sander Koelstra et al. [12] prepared a multimodal dataset called DEAP containing EEG and physiological signals. The dataset was prepared from the recordings of 32 participants aged between 19 and 37 and had a balanced male–female ratio. Each participant was presented with 40 videos having emotional content. The 40 videos were selected out of 120 music videos, which were collected from the website last.fm, having affective tags and a manual procedure. The selection procedure for the videos involved a web-based subjective emotion assessment interface. All the videos were of 1-min length and contained music videos. EEG was recorded at a sampling rate of 512 Hz using 32 active AgCl electrodes (placed according to the international 10–20 system). Thirteen peripheral physiological signals, such as GSR, respiration amplitude, skin temperature, electrocardiogram, blood volume by plethysmograph, electromyograms of Zygomaticus and Trapezius muscles, and electrooculogram (EOG), etc., were also recorded.
The synchronization of the EEG with emotion data was done by first displaying a fixation cross on the screen and asking the participant to relax for 2 min. After that, 40 videos of 1-min length were presented in trials to each participant, and before each trial, a 2-s screen displayed the progress, and then a 5-s fixation cross was displayed to relax the participant. It is very difficult to find markers in EEG signals for transition status in emotions, as the transition status is highly subjective in nature. Therefore, the participant ratings were used to mark the induced emotion.
The DEAP dataset contains the processed EEG recordings, which were further downsampled to 128 Hz, and the eye blink artifact was removed using blind source separation. A bandpass frequency filter from 4.0 to 45.0 Hz was also applied. The data were averaged to the common reference and they were segmented into 60-s trials and a 3-s pre-trial baseline (out of the 5-s baseline recording). Moreover, the participant ratings were supplied separately for valence, arousal, and dominance.
DEAP and SEED are the two most popular publicly available EEG emotion datasets. Both the datasets used audio-visual stimuli for emotion elicitation. The DEAP dataset has a greater number of EEG recordings compared to the SEED dataset as the numbers of participants and videos are higher than in the SEED dataset. Unlike the SEED dataset, the DEAP dataset recorded physiological signals apart from the EEG. However, the EEG recordings of the SEED dataset have higher spatial resolution compared to the DEAP dataset, as a higher number of electrodes were used in the SEED dataset to capture EEG signals. The DEAP dataset used 40 different 1-min video clips to induce emotion in the participants but SEED used 15 different movie clips of a maximum duration of 4 min. Lastly, the SEED dataset used a categorical emotion model, whereas the DEAP dataset used a dimensional emotion model. The proposed feature extraction method was experimented on both the datasets.

2.2. Proposed Methodology

This section discusses the methodology behind applying the deep learning technique for automated feature learning from EEG data for emotion classification. The method involves three steps as given below:
  • Manual Feature Extraction;
  • Generation of Asymmetric Map;
  • Automated Feature Extraction.

2.2.1. Manual Feature Extraction

As EEG signals are complex and non-stationary, introducing EEG signals directly for automated feature learning can lead to sub-optimal performance. Therefore, in this work, DE features are extracted from the EEG signals. Considering an EEG signal from a channel as a continuous random variable, DE gives the measure of the randomness in the EEG signal. The DE of an EEG segment is considered to be equivalent to the logarithm energy spectrum of a specific frequency band [40]. The DE equation on a random variable is given as
h ( X ) = X f ( x ) log ( f ( x ) ) d x
To extract the DE features, the frequency spectrum of an EEG signal in a channel is first obtained using a 256-point short-time Fourier transform (STFT) with a non-overlapping Hanning window of 1 s. As different frequency ranges in EEG signals resemble different brain states, various research works pre-dominantly subdivide the waveforms into frequency bands such as delta, theta, alpha, beta, and gamma. Frequencies ranging from 1 Hz to 3 Hz are named the delta band, which indicates a sleep state. The theta band comprises frequencies ranging from 4 to 7 Hz and resembles a deeply relaxed state. The frequency band 8 to 13 Hz is named the alpha band and indicates a very relaxed and passive attention state. The beta band, comprising frequencies ranging from 14 to 30 Hz, resembles anxiety, external attention, and an active state. Frequencies ranging from 31 to 50 Hz, named the gamma band, represent a state of concentration and focus. The difference in the frequency ranges at low and high frequency is attributed to the rhythmic patterns associated with the brain states. The DE features are extracted for each frequency band in every epoch, thus retaining the temporal characteristics. The DE features are further smoothed using moving average in order to eliminate any unintended component introduced in the features. Figure 1a gives a pictorial representation of the manual feature extraction process.

2.2.2. Generation of Asymmetric Map

After manual feature extraction, the next important step is to generate the AsMap. Previous works have shown that the asymmetrical brain activity seems to be effective in discriminating EEG signals induced by different emotions [41,42]. Here, the DE features of each frequency band in n consecutive epochs in an EEG segment are grouped in fixed-sized, non-overlapping windows, and we average the DE features under a window to form a vector of size m. As there are 62 channels, we obtain a 62 × m vector for each frequency band. Each column in the 2D vector further undergoes transformation to generate an AsMap on the kth frequency band using Equation (2).
A s M a p ( i , j , k ) = D E ( i , k ) D E ( j , k )
Here, D E ( i , k ) represents DE features on the kth frequency band of the ith channel and D E ( j , k ) represents DE features on the kth frequency band of the jth channel.
Normalization is also performed on the AsMap to transform the data in such a way that each AsMap has distributions in a common scale from 0 to 1. The AsMap captures the difference in DE between all possible pairs of channels, as shown in Figure 1b. In the AsMap, the difference in DE features among all channel pairs gives a quantitative measure of the low-level asymmetry in different brain regions irrespective of their spatial location. For illustration, the AsMap of the gamma band for a slot in an EEG segment corresponding to positive, negative, and neutral emotion in the SEED dataset is presented as grayscale images in Figure 2.

2.2.3. Automated Feature Extraction

After obtaining the AsMap, we perform automated feature extraction on AsMaps of a subset of frequency bands to obtain patterns in the asymmetry of different brain regions across frequency bands. For this purpose, we use CNN on a subset of AsMaps to obtain a 1D feature vector. The CNN model has two 2D convolutional layers with a kernel size of 3 × 3 for spatial feature extraction. Further, each convolution layer uses the rectified linear unit (ReLU) activation function. The use of the 3 × 3 kernel and ReLU activation in this work is inspired by various models in the computer vision field. Initially, the first convolutional layer has 32 feature maps, but in the subsequent convolutional layer, the feature maps are halved to 16 feature maps. After each convolutional layer, we have a max pooling layer that strides a two-dimensional filter of size (2 × 2) over each channel of the feature maps and calculates the maximum or largest of the features lying within the region covered by the filter. It reduces the dimensions of the feature maps generated in the convolutional layer. The max pooling layer is followed by a dropout layer, where we randomly shut down 25% of a layer’s neurons at each training step by zeroing out the neuron values. Finally, the feature maps from the last max pooling layer are flattened to obtain a 1D feature vector. Different layers of the CNN model used in this work are presented in Figure 3.

3. Results

3.1. Experimental Setup

During the experiment, an Acer Desktop with Intel Core i3 7th gen processor and 4GB RAM was used. Anaconda 3, which is a free and open-source distribution of the Python and R programming language, was used to perform the scientific computing. Python libraries such as Numpy, Pandas, and Scikit-Learn are some of the most important libraries used for data handling during the experimentation. The proposed method for feature extraction was tested on both the SEED and DEAP datasets. The experiment conducted on the SEED dataset used the pre-extracted DE features. The DE features were used to generate the AsMap. As EEG recording in the SEED dataset contains signals from 62 channels, the dimension of the AsMap is 62 × 62 × k for all frequency bands together. Here, k is the number of frequency bands. As the SEED dataset presents three classes of emotion (positive, negative, and neutral), a three-class classification problem on the SEED dataset was formulated. The classification problem was formulated to classify between positive, negative, and neutral emotions. Further, experiments were conducted on the DEAP dataset, and AsMap features were extracted from the 32-channel EEG recordings. The dimension of the AsMap features extracted from the DEAP dataset was 32 × 32 × k for all frequency bands together. Based on the valence and arousal ratings provided in the DEAP dataset, two different classification problems were formulated: two-class classification (valence classification and arousal classification) and four-class classification. The two-class classification on valence was to classify between high valence and low valence. Meanwhile, the two-class classification on arousal was to classify between high arousal and low arousal. During the preparation of the DEAP dataset, participants provided a rating from 1 to 9 for valence and arousal after watching each video. Based on the distribution of the subjective ratings [12], these ratings were considered as an estimate for valence and arousal. The classes were obtained in the following manner: the participants’ ratings from 5.5 to 9 were categorized as the high-valence (HV) class and ratings from 1 to 5.5 were categorized as the low-valence (LV) class. Similarly, the participants’ ratings from 5.5 to 9 were categorized as the high-arousal (HA) class and ratings from 1 to 5.5 were categorized as the low-arousal (LA) class. In the four-class classification problem, both valence and arousal classes were combined together to classify four different classes of emotion. The class labels for the four-class classification problem were high valence–high arousal (HVHA), high valence–low arousal (HVLA), low valence–high arousal (LVHA), and low valence–low arousal (LVLA).
The 1D feature vector obtained in the automated feature learning process was used to train a fully connected neural network having two hidden layers with 512 neurons. Each hidden layer used the ReLU activation function. The output layer had a number of neurons equal to the number of classes, and the softmax activation function was used to classify the different classes of emotion. For comparison, other feature extraction methods such as DE, DASM, RASM, and DCAU were also used to train the classifier separately.
In order to analyze the proposed method on both the SEED and DEAP datasets, the classification accuracy using AsMap+CNN features was compared with DE and other DE-based features such as DASM, RASM, and DCAU. The features were obtained on different frequency bands such as delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). Experiments were also conducted on varying window sizes, where the window size was set to 3 s, 6 s, 12 s, 30 s, respectively.

3.2. Three-Class Classification on SEED

Table 1 presents the three-class emotion classification accuracy using different feature extraction methods such as DE, DASM, RASM, DCAU, and AsMap+CNN on delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). The proposed method outperformed all the DE-based feature extraction methods on delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). The highest classification accuracy of 97.10% was obtained using AsMap+CNN on the γ band with the use of a 3-s window size. It was also observed that the classification accuracy obtained using all the other feature extraction methods remained between 93% and 96% on the γ band. Further, the features on β and ALL BAND from all the feature extraction methods resulted in classification accuracy above 91%, except for DE and RASM. It was also observed that the classification accuracy using different feature extraction methods on delta ( δ ), theta ( θ ), and alpha( α ) remained below 70%.
The classification accuracy using AsMap+CNN on different frequency bands and window sizes is presented in Figure 4. It can be observed that an increase in window size has a negative impact on the classification accuracy. Using AsMap+CNN features on β , γ , and ALL BAND, the classification accuracy remained above 85% for window sizes smaller than or equal to 12 s. The classification accuracy obtained from features calculated on γ , β , and ALL BAND showed linear degradation, and the accuracy remained above 75% until a 30 s window size. However, features obtained on delta ( δ ), theta ( θ ), and alpha ( α ) did not show a linear degradation in accuracy. The figure also clearly illustrates that features on γ , β , and ALL BAND had greater discriminating ability than those of other bands.

3.3. Two-Class Classification on DEAP

On the DEAP dataset, valence and arousal classification accuracy were analyzed on different feature extraction methods. Table 2 presents the valence classification accuracy obtained using different feature extraction methods on delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). In this experiment also, the window size was set to 3 s. The highest valence classification accuracy was achieved on ALL BAND using AsMap+CNN features, which was 95.45%. However, the classification accuracy achieved by using DASM features on ALL BAND was very close to the accuracy using AsMap+CNN features. Further, the classification accuracy obtained by using DE, DASM, DCAU, and AsMap+CNN on ALL BAND was higher than that obtained with features on other frequency bands. In the β and γ bands, AsMap+CNN features generated the highest classification accuracy compared with other feature extraction methods. However, in the δ , θ , and α bands, the DE features yielded higher classification accuracy compared to other features. Table 3 presents the arousal classification accuracy obtained using different feature extraction methods on delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). The highest arousal classification accuracy was achieved on ALL BAND using AsMap+CNN features, which was 95.21%. However, the classification accuracy achieved using DCAU and DASM features on ALL BAND remained above 94%. In comparison to valence classification, similar observations were made wherein the arousal classification accuracy obtained by using DE, DASM, DCAU, and AsMap+CNN on ALL BAND was higher than that obtained with features on other frequency bands. In the θ , β , and γ bands, AsMap+CNN features generated the highest classification accuracy compared with other feature extraction methods. However, in the δ and α bands, the DE features obtained higher classification accuracy compared to other features.
The valence and arousal classification accuracy using AsMap+CNN on different frequency bands and window sizes are presented in Figure 5 and Figure 6, respectively. Both the figures show a similar trend, where, with the increase in window size, the classification accuracy decreases. Using AsMap+CNN features on ALL BAND, the valence and arousal classification accuracy remained above 90% for window sizes smaller than or equal to 12 s. The valence and arousal classification accuracy obtained showed linear degradation, and the accuracy remained above 68% until a 30 s window size. Both Figure 5 and Figure 6 clearly show that AsMap+CNN features on ALL BAND together have greater discriminating ability compared to other bands for valence and arousal classification.

3.4. Four-Class Classification on DEAP

In order to further test the capability of the AsMap+CNN feature extraction method, a four-class classification problem was formulated using the valence and arousal classes on the DEAP dataset. The four-class classification accuracy was also analyzed on other feature extraction methods. Table 4 presents the four-class classification accuracy obtained using different feature extraction methods on delta ( δ ), theta ( θ ), alpha ( α ), beta ( β ), gamma ( γ ), and all frequency bands together (ALL BAND). In this experiment also, the window size was set to 3 s. The highest classification accuracy of 93.41% was achieved on ALL BAND using AsMap+CNN features. However, the classification accuracy achieved by using DASM features on ALL BAND was 92.23%, which is close to the accuracy achieved using AsMap+CNN features. Similar to two-class classification, the four-class classification accuracy obtained using DE, DASM, DCAU, and AsMap+CNN on ALL BAND was higher than that obtained with features on other frequency bands. In the β and γ bands, AsMap+CNN features generated the highest classification accuracy compared with other feature extraction methods. However, in the δ , θ , and α bands, the DE features obtained higher classification accuracy compared to other features.
The four-class classification accuracy using AsMap+CNN on different frequency bands and window sizes are presented in Figure 7. Similar to the observations in two-class and three-class classification, it was observed that the window size has a negative impact on classification accuracy. Using AsMap+CNN features on ALL BAND, the classification accuracy remained above 85% for window sizes smaller than or equal to 12 s. However, the classification accuracy obtained on all frequency bands showed linear degradation, and the accuracy remained above 55% until a 30 s window size. Figure 7 clearly shows that AsMap+CNN features on ALL BAND together have greater discriminating ability compared to other bands for complex classification problems having four classes.

4. Discussion

In this experiment, the proposed hybrid feature extraction method (AsMap+CNN) outperformed other DE-based feature extraction methods in terms of classification accuracy. The proposed method was compared in competing scenarios where the window size was varied from 3 to 30 s. The accuracy of classification using the features was tested on different datasets and on a varying number of classes. On the DEAP dataset, AsMap+CNN features from all frequency bands achieved the highest valence and arousal classification accuracy of 95.45% and 95.21%, respectively. Further, experiments were conducted to increase the difficulty level by formulating a four-class classification problem on the DEAP dataset, and the highest classification accuracy of 93.41% was achieved on ALL BAND using AsMap+CNN features. The highest classification accuracy of 97.10% was achieved on the SEED dataset using AsMap+CNN features from the gamma band. One of the critical findings of this work is that AsMap+CNN on the gamma band generated more discriminative features than features from all bands together in classifying positive, negative, and neutral emotions on the SEED dataset. This indicates that emotional experience has a higher correlation with asymmetry in different brain regions on higher frequency bands. However, on the DEAP dataset, it was observed that features on all bands together provided higher classification accuracy than features on individual frequency bands. The DEAP dataset was prepared on 32 EEG channels, compared to 62 EEG channels for the SEED dataset. The features generated have a lower spatial resolution, and features on individual bands do not provide classification accuracy above 90%. Thus, with the power of CNN in learning hidden features, the classification accuracy increases by extracting hidden features from the AsMap on all bands.
In contrast to other feature extraction methods, the AsMap captures the asymmetry among all the brain regions in a 2D vector. This work is the first attempt to generate AsMaps using DE features and feed them into a CNN for feature engineering, to the best of our knowledge. One of the limitations of this method is that the size of the AsMap increases with the increase in the number of EEG channels, which introduces a higher computational overhead on the CNN model. It was also observed that the classification accuracy shows linear degradation with the increase in window size. This is due to the fact that an increase in window size compromises the frequency resolution in STFT. Moreover, the window size is fixed while passing through the entire frequency spectrum. A viable solution to this is to use least-squares wavelet analysis (LSWA) or continuous wavelet transform (CWT) instead of STFT for more accurate estimation of frequencies and amplitudes [29,30]. In LSWA or CWT, the window size decreases as the frequency increases, allowing one to capture the high-frequency components with short duration or with varying amplitude over time or frequency. The investigation of a frequency-dependent window length is subject to future work. The degradation in classification accuracy for large window sizes can also be attributed to the combination of more than one emotion feature in large windows. Investigation of the temporal features in the EEG data for a particular window can be a viable solution to the degradation in classification accuracy with an increase in window size.
This work highlights the importance of hybrid feature extraction in emotion classification, as the accuracy of the classifier is directly dependent on the quality of features. The results demonstrate that the hybrid method of manual and automated feature extraction provides an advantage over the existing state-of-the-art feature extraction methods in emotion recognition systems using EEG. The proposed method’s ability to classify discrete emotions in a valence–arousal coordinate space provides scope for advancement in EEG-based emotion recognition.

5. Conclusions

This work presented a deep learning approach for automated feature extraction for EEG-based emotion classification. As CNNs have shown potential in image classification, the DE features are transformed into a 2D feature vector called an AsMap. The automated features obtained using the AsMap on the CNN model provide the highest classification accuracy of 97.10%, using a 3 s window size. The AsMap+CNN for feature extraction outperformed other feature extraction methods such as DE, DASM, RASM, and DCAU in terms of classification accuracy. The AsMap+CNN features capture the spatial correlation among different brain regions, thus resulting in higher classification accuracy. Results also indicated that the gamma band features give higher classification accuracy than other frequency bands on the SEED dataset. Further, experiments revealed that an increase in window size results in lower classification accuracy.

Author Contributions

Conceptualization: M.Z.I.A. and N.S.; methodology: M.Z.I.A.; Original draft preparation: M.Z.I.A.; Review and editing: E.G., N.S. and S.P.; supervision: N.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval are not applicable for this study due to the use of public datasets.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the editors and the reviewers for their time and constructive comments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this paper:
AsMapasymmetric map
BCIbrain–computer interface
CNNconvolutional neural network
CWTcontinuous wavelet transform
DASMdifferential asymmetry
DCAUdifferential caudality
DEdifferential entropy
EEGelectroencephalogram
EMGelectromyogram
EOGelectrooculogram
GSRgalvanic skin response
HAhigh arousal
HVhigh valence
HVHAhigh valence–high arousal
HVLAhigh valence–low arousal
LSWAleast-squares wavelet analysis
LAlow arousal
LVlow valence
LVHAlow valence–high arousal
LVLAlow valence–low arousal
PSDpower spectral density
RASMrelative asymmetry
ReLUrectified linear unit
RNNrecurrent neural network
SEEDSJTU Emotion EEG Dataset
STFTshort-time Fourier transform

References

  1. Pan, C.; Shi, C.; Mu, H.; Li, J.; Gao, X. EEG-based emotion recognition using logistic regression with Gaussian kernel and Laplacian prior and investigation of critical frequency bands. Appl. Sci. 2020, 10, 1619. [Google Scholar] [CrossRef] [Green Version]
  2. Deshpande, M.; Rao, V. Depression detection using emotion artificial intelligence. In Proceedings of the 2017 International Conference on Intelligent Sustainable Systems (ICISS), Palladam, India, 7–8 December 2017; pp. 858–862. [Google Scholar] [CrossRef]
  3. Wioleta, S. Using physiological signals for emotion recognition. In Proceedings of the 2013 6th International Conference on Human System Interactions (HSI), Sopot, Poland, 6–8 June 2013; pp. 556–561. [Google Scholar] [CrossRef]
  4. Rached, T.S.; Perkusich, A. Emotion recognition based on brain-computer interface systems. In Brain-Computer Interface Systems-Recent Progress and Future Prospects; Intechopen: Rijeka, Croatia, 2013; pp. 253–270. Available online: https://www.intechopen.com/chapters/44926 (accessed on 10 February 2022).
  5. Sheykhivand, S.; Mousavi, Z.; Rezaii, T.Y.; Farzamnia, A. Recognizing emotions evoked by music using CNN-LSTM networks on EEG signals. IEEE Access 2020, 8, 139332–139345. [Google Scholar] [CrossRef]
  6. Bota, P.J.; Wang, C.; Fred, A.L.; Da Silva, H.P. A review, current challenges, and future possibilities on emotion recognition using machine learning and physiological signals. IEEE Access 2019, 7, 140990–141020. [Google Scholar] [CrossRef]
  7. Shu, L.; Xie, J.; Yang, M.; Li, Z.; Li, Z.; Liao, D.; Xu, X.; Yang, X. A review of emotion recognition using physiological signals. Sensors 2018, 18, 2074. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Cannon, W.B. The James-Lange theory of emotions: A critical examination and an alternative theory. Am. J. Psychol. 1927, 39, 106–124. [Google Scholar] [CrossRef]
  9. PS, S.; Mahalakshmi, G. Emotion models: A review. Int. J. Control Theory Appl. 2017, 10, 651–657. [Google Scholar]
  10. Ekman, P. Are There Basic Emotions? Psychol. Rev. 1992, 99, 550–553. [Google Scholar] [CrossRef]
  11. Russell, J.A. A circumplex model of affect. J. Personal. Soc. Psychol. 1980, 39, 1161. [Google Scholar] [CrossRef]
  12. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A database for Emotion Analysis; using Physiological Signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  13. SEED Dataset. Available online: http://bcmi.sjtu.edu.cn/~seed/seed.html (accessed on 26 April 2020).
  14. Liu, Y.J.; Yu, M.; Zhao, G.; Song, J.; Ge, Y.; Shi, Y. Real-time Movie-induced Discrete Emotion Recognition from EEG signals. IEEE Trans. Affect. Comput. 2018, 9, 550–562. [Google Scholar] [CrossRef]
  15. Li, M.; Lu, B.L. Emotion Classification based on Gamma-band EEG. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Berlin, Germany, 23–27 July 2009; pp. 1223–1226. [Google Scholar]
  16. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential Entropy Feature for EEG-based Emotion Classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar]
  17. Lin, Y.P.; Wang, C.H.; Wu, T.L.; Jeng, S.K.; Chen, J.H. EEG-based Emotion Recognition in Music Listening: A Comparison of Schemes for Multiclass Support Vector Machine. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 489–492. [Google Scholar]
  18. Ackermann, P.; Kohlschein, C.; Bitsch, J.A.; Wehrle, K.; Jeschke, S. EEG-based Automatic Emotion Recognition: Feature Extraction, Selection and Classification methods. In Proceedings of the 2016 IEEE 18th International Conference on e-Health Networking, Applications and Services (Healthcom), Munich, Germany, 14–16 September 2016. [Google Scholar]
  19. Ramirez, R.; Vamvakousis, Z. Detecting Emotion from EEG signals using the Emotive EPOC device. In International Conference on Brain Informatics; Springer: Berlin/Heidelberg, Germany, 2012; pp. 175–184. [Google Scholar]
  20. Mehmood, R.M.; Du, R.; Lee, H.J. Optimal Feature Selection and Deep learning Ensembles method for Emotion Recognition from Human Brain EEG sensors. Cities 2017, 4, 5. [Google Scholar] [CrossRef]
  21. Petrantonakis, P.C.; Hadjileontiadis, L.J. Emotion Recognition from Brain Signals using Hybrid Adaptive Filtering and Higher Order Crossings analysis. IEEE Trans. Affect. Comput. 2010, 1, 81–97. [Google Scholar] [CrossRef]
  22. Özerdem, M.S.; Polat, H. Emotion Recognition based on EEG features in Movie Clips with Channel Selection. Brain Inform. 2017, 4, 241. [Google Scholar] [CrossRef] [PubMed]
  23. Pham, T.D.; Tran, D. Emotion Recognition using the Emotiv EPOC device. In International Conference on Neural Information Processing; Springer: Berlin/Heidelberg, Germany, 2012; pp. 394–399. [Google Scholar]
  24. Khosrowabadi, R.; Wahab, A.; Ang, K.K.; Baniasad, M.H. Affective Computation on EEG Correlates of Emotion from Musical and Vocal Stimuli. In Proceedings of the 2009 International Joint Conference on Neural Networks, Atlanta, GA, USA, 14–19 June 2009; pp. 1590–1594. [Google Scholar]
  25. Zheng, W.L.; Lu, B.L. Investigating Critical Frequency bands and channels for EEG-based Emotion Recognition with Deep Neural Networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  26. Jiang, X.; Bian, G.B.; Tian, Z. Removal of artifacts from EEG signals: A review. Sensors 2019, 19, 987. [Google Scholar] [CrossRef] [Green Version]
  27. Phadikar, S.; Sinha, N.; Ghosh, R. Automatic EEG eyeblink artefact identification and removal technique using independent component analysis in combination with support vector machines and denoising autoencoder. IET Signal Process. 2020, 14, 396–405. [Google Scholar] [CrossRef]
  28. Phadikar, S.; Sinha, N.; Ghosh, R. Automatic eyeblink artifact removal from EEG signal using wavelet transform with heuristically optimized threshold. IEEE J. Biomed. Health Inform. 2020, 25, 475–484. [Google Scholar] [CrossRef]
  29. Ghaderpour, E. JUST: MATLAB and python software for change detection and time series analysis. GPS Solut. 2021, 25, 1–7. [Google Scholar] [CrossRef]
  30. Ghaderpour, E.; Pagiatakis, S.D.; Hassan, Q.K. A survey on change detection and time series analysis with applications. Appl. Sci. 2021, 11, 6141. [Google Scholar] [CrossRef]
  31. Petrantonakis., P.C.; Hadjileontiadis, L.J. A novel Emotion Elicitation Index using Frontal Brain Asymmetry for Enhanced EEG-based Emotion Recognition. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 737–746. [Google Scholar] [CrossRef]
  32. Phadikar, S.; Sinha, N.; Ghosh, R. A survey on feature extraction methods for EEG based emotion recognition. In International Conference on Innovation in Modern Science and Technology; Springer: Berlin/Heidelberg, Germany, 2019; pp. 31–45. [Google Scholar]
  33. Liu, W.; Wang, Z.; Liu, X.; Zeng, N.; Liu, Y.; Alsaadi, F.E. A survey of deep neural network architectures and their applications. Neurocomputing 2017, 234, 11–26. [Google Scholar] [CrossRef]
  34. Yang, Y.; Wu, Q.; Qiu, M.; Wang, Y.; Chen, X. Emotion Recognition from Multi-Channel EEG through Parallel Convolutional Recurrent Neural Network. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018. [Google Scholar]
  35. Wang, Y.; Huang, Z.; McCane, B.; Neo, P. EmotioNet: A 3-D convolutional neural network for EEG-based emotion recognition. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018. [Google Scholar]
  36. Donmez, H.; Ozkurt, N. Emotion classification from EEG signals in convolutional neural networks. In Proceedings of the 2019 Innovations in Intelligent Systems and Applications Conference (ASYU), Izmir, Turkey, 31 October–2 November 2019; pp. 1–6. [Google Scholar]
  37. Keelawat, P.; Thammasan, N.; Numao, M.; Kijsirikul, B. Spatiotemporal emotion recognition using deep CNN based on EEG during music listening. arXiv 2019, arXiv:1910.09719. [Google Scholar]
  38. Chen, J.; Zhang, P.; Mao, Z.; Huang, Y.; Jiang, D.; Zhang, Y. Accurate EEG-based emotion recognition on combined features using deep convolutional neural networks. IEEE Access 2019, 7, 44317–44328. [Google Scholar] [CrossRef]
  39. Moon, S.E.; Jang, S.; Lee, J.S. Convolutional neural network approach for EEG-based emotion recognition using brain connectivity and its spatial information. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2556–2560. [Google Scholar]
  40. Shi, L.; Jiao, Y.; Lu, B. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 6627–6630. [Google Scholar] [CrossRef]
  41. Davidson, R.J.; Fox, N.A. Asymmetrical brain activity discriminates between positive and negative affective stimuli in human infants. Science 1982, 218, 1235–1237. [Google Scholar] [CrossRef] [PubMed]
  42. Wu, D.; Courtney, C.G.; Lance, B.J.; Narayanan, S.S.; Dawson, M.E.; Oie, K.S.; Parsons, T.D. Optimal Arousal Identification and Classification for Affective Computing Using Physiological Signals: Virtual Reality Stroop Task. IEEE Trans. Affect. Comput. 2010, 1, 109–118. [Google Scholar] [CrossRef]
Figure 1. Pictorial representation of the steps involved in (a) manual feature extraction and (b) generation of AsMap.
Figure 1. Pictorial representation of the steps involved in (a) manual feature extraction and (b) generation of AsMap.
Sensors 22 02346 g001
Figure 2. AsMap of gamma band on a slot corresponding to positive, negative, and neutral emotion respectively.
Figure 2. AsMap of gamma band on a slot corresponding to positive, negative, and neutral emotion respectively.
Sensors 22 02346 g002
Figure 3. Different layers in the CNN model.
Figure 3. Different layers in the CNN model.
Sensors 22 02346 g003
Figure 4. 3-class classification accuracy on varying window size using AsMap+CNN features.
Figure 4. 3-class classification accuracy on varying window size using AsMap+CNN features.
Sensors 22 02346 g004
Figure 5. Valence classification accuracy on varying window size using AsMap+CNN features.
Figure 5. Valence classification accuracy on varying window size using AsMap+CNN features.
Sensors 22 02346 g005
Figure 6. Arousal classification accuracy on varying window size using AsMap+CNN features.
Figure 6. Arousal classification accuracy on varying window size using AsMap+CNN features.
Sensors 22 02346 g006
Figure 7. 4-class classification accuracy on varying window size using AsMap+CNN features.
Figure 7. 4-class classification accuracy on varying window size using AsMap+CNN features.
Sensors 22 02346 g007
Table 1. Three-class classification accuracy obtained using different feature extraction techniques on frequency bands.
Table 1. Three-class classification accuracy obtained using different feature extraction techniques on frequency bands.
Method δ θ α β γ ALLBAND
DE60.80%47.41%57.07%88.09%95.09%88.28%
RASM53.07%49.56%60.49%88.53%93.12%90.62%
DCAU59.79%55.15%64.02%91.31%95.12%94.70%
DASM57.44%52.54%63.58%91.41%95.87%94.34%
AsMap+CNN62.18%56.20%69.56%93.99%97.10%96.25%
The window size was set to 3 s.
Table 2. Valence classification accuracy obtained using different feature extraction techniques on frequency bands.
Table 2. Valence classification accuracy obtained using different feature extraction techniques on frequency bands.
Method δ θ α β γ ALLBAND
DE80.44%86.57%86.46%74.52%80.20%86.87%
RASM56.71%56.48%57.60%74.19%70.69%56.24%
DCAU70.68%74.84%72.35%74.07%74.78%93.20%
DASM72.59%78.61%78.43%78.48%80.74%95.08%
AsMap+CNN79.61%85.64%86.15%86.83%86.57%95.45%
The window size was set to 3 s.
Table 3. Arousal classification accuracy obtained using different feature extraction techniques on frequency bands.
Table 3. Arousal classification accuracy obtained using different feature extraction techniques on frequency bands.
Method δ θ α β γ ALLBAND
DE82.01%88.10%87.78%77.96%80.65%88.47%
RASM57.55%58.06%64.08%76.34%74.49%59.42%
DCAU71.96%75.90%75.35%75.27%74.52%94.60%
DASM75.13%81.03%79.64%79.31%81.06%94.17%
AsMap+CNN81.38%88.27%87.24%88.94%89.00%95.21%
The window size was set to 3 s.
Table 4. Four-class classification accuracy obtained using different feature extraction techniques on frequency bands.
Table 4. Four-class classification accuracy obtained using different feature extraction techniques on frequency bands.
Method δ θ α β γ ALLBAND
DE70.23%80.33%80.89%76.76%79.31%86.30%
RASM30.97%30.23%47.15%62.11%59.11%38.61%
DCAU53.20%62.71%59.47%58.87%61.89%90.48%
DASM60.38%69.65%67.08%67.57%70.51%92.23%
AsMap+CNN67.86%79.43%79.15%81.66%82.16%93.41%
The window size was set to 3 s.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ahmed, M.Z.I.; Sinha, N.; Phadikar, S.; Ghaderpour, E. Automated Feature Extraction on AsMap for Emotion Classification Using EEG. Sensors 2022, 22, 2346. https://doi.org/10.3390/s22062346

AMA Style

Ahmed MZI, Sinha N, Phadikar S, Ghaderpour E. Automated Feature Extraction on AsMap for Emotion Classification Using EEG. Sensors. 2022; 22(6):2346. https://doi.org/10.3390/s22062346

Chicago/Turabian Style

Ahmed, Md. Zaved Iqubal, Nidul Sinha, Souvik Phadikar, and Ebrahim Ghaderpour. 2022. "Automated Feature Extraction on AsMap for Emotion Classification Using EEG" Sensors 22, no. 6: 2346. https://doi.org/10.3390/s22062346

APA Style

Ahmed, M. Z. I., Sinha, N., Phadikar, S., & Ghaderpour, E. (2022). Automated Feature Extraction on AsMap for Emotion Classification Using EEG. Sensors, 22(6), 2346. https://doi.org/10.3390/s22062346

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop