Next Article in Journal
Existence of Classical Solutions for Nonlinear Elliptic Equations with Gradient Terms
Next Article in Special Issue
Information Geometry Theoretic Measures for Characterizing Neural Information Processing from Simulated EEG Signals
Previous Article in Journal
A New Deep Learning Method with Self-Supervised Learning for Delineation of the Electrocardiogram
Previous Article in Special Issue
Musical Emotions Recognition Using Entropy Features and Channel Optimization Based on EEG
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals

1
Department of Electronics and Information Engineering, Changchun University of Science and Technology, Changchun 130000, China
2
Economics School, Jilin University, Changchun 130000, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1830; https://doi.org/10.3390/e24121830
Submission received: 27 October 2022 / Revised: 7 December 2022 / Accepted: 8 December 2022 / Published: 15 December 2022
(This article belongs to the Special Issue Entropy Applications in Electroencephalography)

Abstract

:
As a major daily task for the popularization of artificial intelligence technology, more and more attention has been paid to the scientific research of mental state electroencephalogram (EEG) in recent years. To retain the spatial information of EEG signals and fully mine the EEG timing-related information, this paper proposes a novel EEG emotion recognition method. First, to obtain the frequency, spatial, and temporal information of multichannel EEG signals more comprehensively, we choose the multidimensional feature structure as the input of the artificial neural network. Then, a neural network model based on depthwise separable convolution is proposed, extracting the input structure’s frequency and spatial features. The network can effectively reduce the computational parameters. Finally, we modeled using the ordered neuronal long short-term memory (ON-LSTM) network, which can automatically learn hierarchical information to extract deep emotional features hidden in EEG time series. The experimental results show that the proposed model can reasonably learn the correlation and temporal dimension information content between EEG multi-channel and improve emotion classification performance. We performed the experimental validation of this paper in two publicly available EEG emotional datasets. In the experiments on the DEAP dataset (a dataset for emotion analysis using EEG, physiological, and video signals), the mean accuracy of emotion recognition for arousal and valence is 95.02% and 94.61%, respectively. In the experiments on the SEED dataset (a dataset collection for various purposes using EEG signals), the average accuracy of emotion recognition is 95.49%.

1. Introduction

Emotion recognition is a multi-curricular field of study that combines cognitive science, social psychology, and electronic information science [1,2,3]. It is a complex problem and a hot spot in the cognitive science industry. Emotion identification methods are divided into two categories. One is emotion identification based on non-physiological data signals, such as words, language expressions, facial emotions, gestures, etc. [4]. However, this method is relatively subjective and cannot ensure the authenticity and effectiveness of emotions. Another emotion recognition is based on physiological data signals [5]. With the development of wearable and non-invasive physiological signal acquisition devices, the real-time performance and accuracy of the physiological signal collection have been greatly improved, promoting the development of physiological signals in emotion recognition. Nervous system physiology and social psychology studies have shown that brain electrical signals can reflect various brain wave thematic activities and effects of the brain. Since the occurrence of emotion is related to the activities of the cerebral cortex, the brain’s electrical signals can also reflect the efficient information content of people’s emotional state [6]. Because of the advantages of strong universality and high classification and identification accuracy, brain electronic signals have been gradually introduced into the field of emotion identification research.
Traditional EEG emotion identification methods usually capture handcrafted emotion characteristics, followed by shallow machine learning for emotion classification. The main methods include the K-nearest neighbor algorithm, mixed Gaussian algorithm, artificial neural network method, Bayesian algorithm, support vector machine, and improvements in these methods [7]. At this stage, the key characteristics of emotional discrimination of EEG signals are time-domain characteristics, frequency domain characteristics, and time frequency domain characteristics [8]. Time domain features are mainly used to directly extract the waveform features of EEG signals, including event related potential (ERP), signal statistics, energy, power, high order zero crossing analysis, non-stationary index (NSI), and fractal dimension (FD) [9]. Since the time-domain feature cannot display the frequency information of the signal, the researchers added frequency domain analysis. First, the original time domain signal is converted into the frequency domain to obtain the frequency spectrum. Then, the frequency band is decomposed into five sub-frequency bands (δ, θ, α, β, γ) that are closely related to human psychological activities [10]. Extract features from these five frequency bands. These features usually include power spectral density (PSD) [11], differential entropy (DE) [12,13], differential asymmetry (DASM) [14], and rational asymmetry (RASM) [15,16]. EEG signal is a non-stationary and non-linear random signal [17]. Since the scope of the Fourier transform is the entire time domain, it lacks local generalization ability. It is impossible to confirm the time corresponding to each frequency domain component of the non-stationary signal, so the time frequency domain is introduced to analyze the EEG signal [18]. Usually, the time frequency domain signal transformation mainly adopts methods, such as short time Fourier transform (STFT), wavelet transform (WT), and wavelet packet transform (WPT) or Hilbert–Huang [19]. However, these features need to be designed and selected according to specific purposes, which also limits their performance in EEG emotion classification.
The rapid development of deep learning technology in computer vision [20], speech [21], and natural language processing [22], methods based on deep neural networks, have begun to get the attention of researchers in this field. EEG usually contains much helpful information and has a wide range of applications. The related work is summarized in Table 1. Komolovaitė et al. [23] contributed to the application of Alzheimer’s disease. They used convolutional neural networks to learn discriminant features in EEG to classify familiarity and emotion. Thammasan used deep belief networks combined with fractal dimension, power spectrum, and discrete wavelet transform features to classify the emotional categories of EEG under music stimulation [24]. Tripathi gives the effect of the convolutional neural network in the two classifications (high/low) and three classifications (high/normal/low) of EEG emotions [25]. Elham scientifically investigated a three-dimensional convolutional neural network (3D-CNN) for emotion discrimination based on multichannel EEG data information, and the discrimination accuracy of valence and arousal were 87.44% and 88.49%, respectively [26]. Yang obtained the differential entropy characteristics of different frequency bands and applied the three-dimensional convolutional neural network to classify emotions on DEAP data at the level of valence and arousal. The mean accuracy is 89.45% and 90.24%, respectively [27].
Although the deep learning emotion recognition model has achieved higher accuracy than the shallow model, how to integrate more helpful EEG information to perform emotion recognition better and improve the validity and reliability of predictions made with machine learning remain challenging [28]. Firstly, regarding the relationship between EEG signal frequency bands and emotion types, Zheng et al. [29] applied a deep belief network model to select meaningful frequency bands of key channels by training the weight distribution of DBNs. The results showed that the Gamma and Beta bands exhibited better classification performance in emotion recognition. Secondly, EEG signal acquisition equipment generally includes 32 channels, 62 channels, and 128 channels. It may contain some EEG signals that have nothing to do with emotions, which can bring noise and interference to emotion recognition. In response to this problem, Meng-meng extracted the frequency domain features on each electrode and then combined the common spatial patterns (CSP) algorithm with the spatial domain to calculate the spatial frequency domain features. The average accuracy rate in the three-category emotion recognition has reached 87.54% [30]. Finally, combining the temporal and spatial characteristics of EEG signals, Liu designed the RCNN LSTM hybrid deep neural network entity model, using the circulatory system neuron network to obtain the characteristics of brain electrical signals, then integrating the information content of the LSTM time concept characteristics to develop a fusion model [31].
In summary, the frequency, time, and space characteristics of EEG signals are critical for emotion identification. However, most of the literature only considers one or two of these three functions. Therefore, it is possible to adaptively capture important frequency, spatial and temporal characteristics in a unified network by integrating information from different fields, which is of crucial significance for improving the accuracy of EEG signal emotion recognition. The method proposed in this paper makes full use of frequency, temporal, and spatial information features while also making an outstanding contribution to the computational performance of the network, filling a gap in integrating information while improving computational performance in EEG signal emotion recognition.
A single feature extracted by traditional emotion research methods is not enough to describe the rich information contained in EEG, and a simple feature combination may cause redundant features in the feature space, thereby affecting model accuracy and increasing model complexity. This paper points out a new method of EEG emotion discrimination based on multi-domain feature fusion. The general flowchart of the proposed method is shown in Figure 1. Firstly, the EEG signal is converted into a 4D feature structure composed of several time slices, which contain frequency, spatial and temporal information. Secondly, because the EEG data training samples are limited and the feature dimension is large, we use the depthwise separable convolution for dimensionality reduction, which balances the width and depth of the network and can reduce the amount of computation. The frequency and spatial features of EEG signals are fully extracted by using the feature of depthwise separable convolution. Finally, since the traditional LSTM does not consider the order information of updating neurons during feature learning, it means that each neuron is independent. Therefore, the knowledge learned from the perspective of spatial structure is discrete, which makes it difficult to distinguish the information level between neurons, which causes the loss of information at the spatial level. Therefore, this paper uses ON-LSTM to increase the induced bias towards structural information by sorting neurons, which enables the model to integrate the emotional hierarchy of EEG signals into LSTM through a specific neuron order without destroying its sequential form. Thereby, the feature information of the emotional hierarchy of the EEG signal is learned to express the hidden emotional information in the time series.
The key dedications in the text are as follows:
  • We adopt a four-dimensional (4D) feature structure, including the frequency, space, and time information of EEG signals, as the feature input of EEG signal emotion recognition.
  • We use the FSTception model based on a depthwise separable convolutional neural network to solve the problems of a few training samples of EEG data, large feature dimensions, and feature extraction.
  • In particular, we adopt the ON-LSTM structure to deal with the deep emotional feature extraction hidden on time series in input features with a 4D structure.
Other parts of the organization are listed below. We will introduce the method proposed in this article in Section 2. Section 3 introduces the data set, experimental settings, results, and discussion. Finally, in Section 4, we conclude that we have achieved advanced performance on both the SEED and DEAP data sets.

2. Materials and Methods

This section mainly describes the 4-dimensional feature organization, FSTception model, and classification. The small details of each part will be introduced in detail.

2.1. 4D Frequency Spatial Temporal Representation

The original EEG signal can be divided into several different frequency patterns according to the in-band correlation of different behavioral states. The frequency modes and corresponding characteristics of EEG are listed in Table 2. The awareness increases with the increase of the frequency band. We believe that emotions are generated when humans are highly awake. This statement also conforms to our intuitive understanding. If a person is unconscious, he/she is unlikely to have specific emotions. To extract the feature entirely, we constructed a 4D feature structure to integrate the three information features of EEG signals, as shown in Figure 2. In data processing, just like the previous work [26,32], because the Butterworth filter has the smoothest response curve in the passband and is simple in design and easy to implement, we use the Butterworth filter to divide the EEG signal into five frequency bands (δ[1~4 Hz], θ[4–8 Hz], α[8~13 Hz], β[13~30 Hz], γ[30~50 Hz]). The frequency response of the filter is shown in Figure 3. After Butterworth filtering decomposition, the EEG data of a participant varied from 40 × 8064 × 32 (video × sample × channel) to 40 × 8064 × 5 × 32 (video × sample × band × channel). The EEG raw signal and its five frequency bands in the DEAP dataset are shown in Figure 4.
Since our focus is on identifying emotional states at the segmented level, the EEG signal is cut into l segments of degree n . Our goal is to categorize these fragments into the correct tags. The segmented EEG data is converted into 40 × l × n × 5 × 32 (video × segment × length × band × channel). When 0.5s is taken as the cutting window, l = 120 is the number of segments to be cut, and n = 64 is the sample points contained in a segment. Differential entropy can accurately reflect the response degree of the stimulus because of its ability to distinguish the balance between the low frequency band and the high frequency band of the EEG mode [33]. Therefore, the differential entropy algorithm is used in this experiment to represent the complexity of a fixed length EEG signal. Differential entropy is the generalized form of Shannon’s information entropy on continuous variables, where X is a random variable, and its original expression is
H ( X ) = f ( x ) log [ f ( x ) ] d x
where f(x) is the probability density function of the X . Shi et al. [34] proved that the differential entropy of a random variable approximately obeying the Gaussian distribution N(θ, σ2) is equal to the logarithm of its energy spectrum in a specific frequency band. The specific expression is as follows
H ( X ) = + 1 2 π σ 2 e ( x μ ) 2 2 σ 2 log [ 1 2 π σ 2 e ( x μ ) 2 2 σ 2 ] d x = 1 2 log 2 π e σ 2
The original EEG signal is complex and does not obey the normal distribution, but the bandpass filtering of a specific range approximates the Gaussian distribution. In order to test whether the EEG after band-pass filtering follows the Gaussian distribution, 4800 EEG segments of 0.5 s length in each sub-band signal were randomly selected from 32 subjects. The Kolmogorov–Smirnov test was then applied to all 4800 EEG data segments to test whether each sub-band was subject to Gaussian distribution, and the significance level α was set at 0.05. It can be proven that the probability of a sub-band signal satisfying the Gaussian distribution hypothesis is greater than 90%.
Therefore, in the fixed frequency band i , the DE feature of each EEG segment in each frequency band is respectively calculated as
H i ( X ) = 1 2 log ( 2 π e σ i 2 )
where H i and σ i 2 represent the differential entropy and signal variance of corresponding EEG signals in the frequency band i , respectively.
From Equation (3), where π , e are constants, we only need to know σ i 2 to get the differential entropy of X . For the frequency band of the EEG signal, as a result of the zero mean, the variance can be estimated as
σ ^ 2 = 1 N i = 1 N x i 2
Equation (4) shows that the variance estimate of the signal sequence X is its average energy. Moreover, due to Parseval’s theorem, the average energy is related to the energy spectrum. It can be written as Equation (5) For the Discrete Fourier transform
n = 0 N 1 | x i 2 | = 1 N k = 0 N 1 | X k 2 | = P i
where X k is the discrete Fourier transform coefficient of signal sequence X i . N is the length of the fixed time windows. P i can be thought of as the energy spectrum, equal to the value of the signal variance σ i 2 multiplied by a constant coefficient N .
Therefore, after the bandpass filter to a specific band i, its variance can be estimated by P i N . The relationship between the logarithmic energy spectrum and differential entropy can be expressed as
H i ( x ) = 1 2 log ( 2 π e σ i 2 ) = 1 2 log ( N σ i 2 ) + 1 2 log ( 2 π e N ) = 1 2 log ( P i ) + 1 2 log ( 2 π e N )
It can be seen that for EEG series of fixed length, the estimate of differential entropy is equivalent to the logarithmic energy spectrum of a specific frequency band.
EEG signals have characteristic frequency ranges and spatial distribution, which are often associated with different functional states of the brain. In the electroencephalogram, each electrode is physically adjacent to the other and records the electrical signals in one area of the brain. In order to maintain spatial information between multiple adjacent channels, according to the mapping method shown in Figure 5, the one-dimensional EEG signal is constructed into a two-dimensional matrix with a size of h × w . The values of h and w are the maximum values of the electrodes in the vertical and horizontal directions, respectively. Among them, a value of zero indicates that the signal of the channel is not used. Then, the two-dimensional feature matrices of different frequency bands are superimposed at the same time to obtain a three-dimensional matrix h × w × a , where a indicates the total number of frequency bands. The 4D structure can be expressed as X n R h × w × a × 2 T , n = 1 , 2 , , S , where S is the total number of samples. In this work, the values of h and w were determined according to the two-dimensional matrix mapping of the distribution positions of EEG acquisition electrodes. Two methods were used to select a compact graph (8 × 9) and a sparse graph (19 × 19).

2.2. The Structure of FSTception

The structure of the FSTception model proposed in this paper consists of two parts, the first is that we use a deep separable convolutional network to further extract the frequency and spatial information of the 4D input feature structure. Secondly, the LSTM with ordered neuron structure is used to extract the contextual correlation of the time-series EEG signals, thus mining the deep emotional information of the EEG signals and improving the accuracy of emotion recognition.

2.2.1. Frequency Spatial Characteristic Learning

Both spatial and temporal information is vital for time-series EEG signals, which represent activity relationships at different locations in the brain and the dynamics of brain patterns over time. In order to better represent the spatial features of EEG information in limited training samples, we propose the FSTception model, a CNN structure based on depthwise separable convolutions. For the 4D frequency space time structure expressed as X n , we apply the FSTception model to obtain frequency and spatial information from each time slice. The structure is shown in Figure 6. First, since the size of the 2D feature maps in the input samples is 8 × 9, the two-dimensional map size is relatively small. In order to better preserve all the information, we set up two convolutional layers, the first convolutional layer has 64 feature maps, and the filter size is 1 × 1, the second convolutional layer has 128 feature maps, the filter size is 3 × 3, and the deep features are extracted using convolution kernels of different scales. Then, we use deep convolution to learn spatial features. The main benefit of deep convolution is to reduce the total number of parameters in the network training process because the convolution is not fully connected to the previous feature map but are connected to each feature map separately to learn spatial features of specific frequencies. Importantly, when used in specific EEG applications, this operation provides a direct method to learn the spatial characteristics of each time slice so that the spatial characteristics of specific frequencies can be effectively extracted. Finally, we used separable convolution, which is a combination of deep convolution and pointwise convolution. The main advantages are: (1) reduce the number of fitting parameters; (2) summarize its feature maps separately by learning each core, and then merge the optimal results and output, thereby expressing the relationship between the internal and cross-feature maps of the feature map. When performing emotion recognition on EEG, this operation is used to summarize a single feature map (depthwise convolution) and optimize the combined feature map (pointwise convolution), because different feature maps can represent information at different time scales. For the convolutional layers of this network, zero-filling and rectified linear unit (ReLU) activation functions and normalization operations are used. To better reduce overfitting and improve the robustness of the network, we employ a max pooling layer with a size of 2 × 2 and a step size of 2 after the operation of the convolution neural network. Finally, the output of the max-pooling layer is unrolled and sent to a fully connected layer with 512 units. The frequency and spatial features of the final output EEG are expressed as Q n = R 512 × 2 T .

2.2.2. Temporal Characteristic Learning

Since the EEG signal contains dynamic content, the changes between time slices in the 4-dimensional structure may hide other information, which may help perform more accurate sentiment classification. Considering that the long short-term memory (LSTM) network has better modeling capabilities for time series information, this paper uses ON-LSTM to integrate neurons into the LSTM through specific sorting, thereby allowing ON-LSTM to learn the hierarchical structure information automatically [35]. The detailed structure of ON-LSTM is shown in Figure 7. ON-LSTM is actually a change in the overall design of the LSTM cell state update. The cell state is done in a hierarchical operation, and the traditional LSTM is to update all the cell states. The cell state is out of order, while ON-LSTM to cell state is updated by two integers for the hierarchy and then hierarchical for computation. This paper is the first innovative use of ON-LSTM in the temporal analysis of EEG emotional features. ON-LSTM is an improved network structure based on LSTM, given a CNN output encoding sequence Qn = (q1, q2, q2T), where qt = R512 and t = 1, 2, 2T. We apply an ON-LSTM layer with 128 data stores to discover the temporal dependence of internal structural segments. The gate structure and output structure of ON-LSTM are the same as those of general LSTM:
f t = σ ( W f q t + U f h t 1 + b f )
i t = σ ( W i q t + U i h t 1 + b i )
o t = σ ( W o q t + U o h t 1 + b o )
c ^ t = tan h ( W c q t + U c h t 1 + b c )
h t = o t ° tan h ( c t )
where historical information h t 1 and current information are input, the input gate i t determines the amount of change of the input vector q t to the information in the storage unit at the current moment. Output gates o t manipulate the output of information contained in the current storage unit. The forget gate f t determines the degree of influence of the historical information h t 1 at the last moment on the information in the current storage. h t is the final output of the ON-LSTM unit at a time t . σ ( ) is the logistic sigmoid function, tan h ( ) is the hyperbolic tangent function, W and U are the weight matrix corresponding to each gate. b the bias term corresponding to each gate. ° is the corresponding element multiplication.
The improvement of ON-LSTM relative to LSTM is different in the update mechanism from c ^ t to c t , thus introducing a hierarchical structure of based on LSTM.
f ˜ t = c s ( s o f t m a x ( W j q t + U j h t 1 + b f ) )
i ¯ t = c s ( s o f t m a x ( W i q t + U i h t 1 + b i ) )
w t = f ˜ t ° i ˜ t
c t = w t ° ( f t ° c t 1 + i ˜ t ° c ˜ t ) + ( f ˜ t w t ) ° c t 1 + ( i ˜ t w t ) ° c ˜ t
where c s is cumsum from left to right, and c s is cumsum from right to left. The main forget gate f ˜ t determines the level of influence to the hierarchical data in the current data storage from the hierarchical information content included in the historical in-formation ht−1 at the last moment. The main input gate i ˜ t determines the amount of change of the level information contained in the input vector qt at the current moment to the level information in the storage unit. After ON-LSTM sorts the neurons, the information level is represented by the front and back of the position. Then, when updating the neuron, first predict the historical level ht−1 and the input level qt, respectively, and update the neurons between partitions through these two levels. ON-LSTM partition update is shown in Figure 8. In this way, the high-level information may retain a considerable distance because the high-level will directly copy the historical information so that the historical information is continuously copied without being changed, and the low-level information may be updated at each step of input. Because the lower layer directly copies the input, and the input is constantly changing, a hierarchical structure is formed through information classification. More generally speaking, it is a group update. Higher group information is transmitted farther (larger span), and lower group span is more minor. These different spans form the hierarchical structure of the input sequence. By learning the hierarchical structure information, it can better mine the hidden connections between each time slice and then achieve more accurate sentiment classification.
The final node output of ON-LSTM is used to represent the highest level of the EEG segment, YnR128, which integrates the frequency, time, and space features of the EEG segment.

2.3. Classification

According to the final characteristics representation Yn, we apply the fully connected layer and the softmax activation function to predict and analyze the sample label of the EEG segment Xn. The calculation is as follows
O u t = A Y n + b = [ o u t 1 , o u t 2 , , o u t c ]
where A represents the transform matrix, b represents the bias, and m represents the number of emotion categories. Then, according to the output result, it is sent to the softmax classifier to obtain the final emotion discrimination result., which can be described as
P ( c | X n ) = exp ( o u t j ) i = 1 i = c exp ( o u t i ) ( j = 1 , , c )
where represents the normalized prediction results of each category. In back propagation, the Adam gradient algorithm is used.

3. Results

3.1. Experimental Setup

The tasks in this paper are done on TensorFlow on a suitable GPU (Nvidia Quadro P5000) using Python 3.7 and Keras. For each trial, we completed a five-fold cross validation and measured the mean classification accuracy (ACC) and standard deviation (STD) for each experimenter. The final characteristics of everyone’s mode are shown as the mean ACC and STD of all subjects. Applying the Adam boost loss function, we set the learning rate to 0.003, the batch size to 120, and the epoch to 100. Codes are available at https://github.com/lixiaonian2022/code.git (accessed on 14 July 2022).
All EEG trials are divided into five groups randomly, then four folds of trials are combined as training data, and the left 1-fold is considered as test data. The experiments are repeated five times so that every fold has a chance to be the test data. The result is calculated as the average of five experiments.

3.2. Dataset

All subjects were in good physical and psychological health, and 32 conductor AgCl electrodes were used in the collection of data, with scalp electrodes positioned in accordance with the international standard 10–20 system.

3.2.1. SEED Dataset

The SEED data set [36] has a total of 15 physical and mental health subjects (seven men, eight women, mean age 23.27). Each subject performed three sets of experiments, and each experiment watched 15 movie clips (five positive segments, five negative segments, five neutral segments). The viewing process of each segment can be divided into four stages, including a 5-s start prompt, 4-min movie playback time, 45-s self-assessment, and 15-s rest time. During the film screening, two films with the same emotion will not be displayed consecutively. Each volunteer carried out three experiments, each time separated by one week, for a total of 45 experiments.

3.2.2. DEAP Dataset

The DEAP data set [37] contains 32 subjects, and each subject has 32 channels of EEG signals and 8 channels of peripheral physiological signals. The content of the DEAP dataset is shown in Table 3. The 32 channels EEG signal is used as the experimental data in this paper. The EEG data signal was first sampled at a sampling frequency of 512 Hz. Then, the sampling frequency was reduced to 128 Hz, and then filtered using a 4.0~45.0 Hz filter to remove EOG artifacts. Each subject watched 40 one-minute emotional music videos. After watching each short video, the subjects scored the level of arousal, valence, liking, and controlling influence. In this experiment, only two dimensions of arousal and potency were selected for testing.

3.3. EEG Data Preprocessing

In order to eliminate the influence of artificial interference, such as ECG, EMG, and power frequency interference, on emotion recognition, we retained more practical components of emotional EEG. The raw multichannel EEG data were first filtered using a Butterworth bandpass filter, using 1 Hz and 50 Hz limits to divide the frequency bands into five bands containing emotional information. These include the δ band (1 to 4 Hz), the θ band (4–8 Hz), the α band (8–13 Hz), the β band (13–30 Hz), and the γ band (30–50 Hz). Each trial contains an EEG signal of 63 s. Among them, the first 3 s is the baseline signal, and the 60 s is the emotional signal of the EEG.
The EEG of 63 s was divided into 126 segments with a length of 0.5 s on average, the first 3 s was the baseline signal, and the eigenvalue of each frequency band of each segment was calculated. In the experiment, in order to improve the accuracy of the experiment, the difference between the characteristic value of EEG recorded by the subject under the stimulus and the mean characteristic value of the baseline signal was used as the experimental data, and the difference at the moment was recorded as the average segmentation of the 63 s EEG into 126 segments with the length of 0.5 s, in which the first 3 s was the baseline signal. The eigenvalue of each frequency band of each segment is calculated. In the experiment, in order to improve the accuracy of the experiment, the difference between the characteristic value of EEG recorded by the subject under the stimulus and the mean characteristic value of the baseline signal was used as the experimental data, and the difference at time t was denoted as E E G _ r e m o v e d t . The calculation formula is as follows
E E G _ r e m o v e d t = f e a t u r e t b a s e _ f e a t u r e 1 + b a s e _ f e a t u r e 2 + + b a s e _ f e a t u r e 6 6
where f e a t u r e t is the EEG characteristic value extracted at time t and b a s e _ f e a t u r e i ( i = 1 , 2 , , 6 ) is the baseline signal characteristic value extracted at time i .

3.4. Results and Discussion

The research results of the 4D-FSTception method proposed in this article on the SEED data set is shown in Table 4, and as shown in bar Figure 9, numerical changes can be observed and compared more easily. The accuracy rate of all subjects exceeded 90%, and the average accuracy rate reached 95.49%. For different subjects, the classification accuracy rate is different. For subject 1, the average classification accuracy rate in the three time periods is 91.5% (the lowest), and for subject 4, the classification accuracy rate in the three time periods is 91.5%. The average classification accuracy rate is 99.59% (the highest), which reflects individual differences. However, for the same subject, the classification accuracy of different time periods fluctuates. For example, for subject 8, the classification accuracy rate in time period 2 is 94.81% (lowest), and the classification accuracy rate in time period 3 is 97.63% (highest). Among 15 subjects, eight items (#2, #4, #6, #8, #9, #10, #11, #13) are better than the mean classification accuracy.
The experimental results of the 4D-FSTception method on the DEAP dataset are shown in Table 5 and Table 6, respectively, and as shown in bar Figure 10. The average accuracy of arousal classification is 95.02%, and the standard deviation is 1.85%. Only two subjects (#2 and #22) had an accuracy rate less than 90%. In particular, the valence and arousal accuracy rates of tester #22 were 79.88% and 84%, respectively, which were lower than other testers. The reason for this may be that the subjects were inattentive during the test operation, or the level of subjective experience was not effectively reported after the test.
The accuracy, precision, recall, and F1-score metrics are used to evaluate the performance of the 4D-FSTception. The evaluation index is calculated based on all the predicted results in DEAP and SEED datasets, as shown in Table 7 and Table 8. The confusion matrix of different data sets is shown in Figure 11 and Figure 12. The superiority of the model in this paper in EEG signal emotion recognition was fully demonstrated.
In order to verify the computational performance of the methods in this paper, we will choose different methods for time cost analysis in both SEED and DEAP datasets. To be able to view and compare maximum and minimum values in a more straightforward way. Figure 13 and Figure 14, respectively, show the performance of accuracy and time cost by different methods in different datasets. As shown in Table 9, for the SEED dataset, a comparison between the HCNN method with a sparse input map of 19 × 19 and the 4D-CRNN method with the same input map of 8 × 9 as in this paper was used. After training, it can be seen that the accuracy of the compact graph is 95.49%, which is 6.89% higher than that of the sparse graph. Furthermore, in terms of their time cost, the compact map spends only 660 s per topic in the five-fold cross-validation, which is almost a quarter of the time cost of the sparse map. Moreover, it can be found that the accuracy of our compact map is close to that of the 4D-CRNN method, which is also a compact map. However, in terms of time cost, the method in this paper is 143 s faster than the 4D-CRNN method. On the DEAP dataset, our compact map outperforms the sparse map and outperforms the 4D-CRNN method with the same compact map in terms of accuracy and training time cost. For valence classification, the average accuracy of our method is 95.49%, which is 0.42% higher than that of the 4D-CRNN method with the same input map. In addition, the time cost of this paper’s method is even 15 s faster than that of the 4D-CRNN method. For arousal classification, we can obtain a similar conclusion. The reason for obtaining similar results may be that adding a sparse map of zeros between adjacent electrodes does not add any useful information. The lower time cost of the compact map may be due to its smaller size, which involves fewer convolution filters to compute. We obtained the model parameter quantities as well as the floating-point operations to be 12.7 M and 20.61 G, respectively, in the DEAP dataset. Compared with other methods, the algorithm in this paper has better performance. The results for both datasets lead us to believe that the method in this paper exhibits better computational performance when classifying emotions.
To verify the importance of the combined structure of our proposed CNN model and ON-LSTM, we performed two different combinations of CNN + LSTM (4D-FSTception (LSTM)) and CNN + ON-LSTM (4D-FSTception) in the SEED and DEAP datasets. The comparative experimental results are shown in Figure 15. Structure experiment. The average classification accuracy of 4D-FSTception in the SEED dataset and the DEAP dataset is 0.4% higher than the average classification accuracy of 4D-FSTception (LSTM). The results show that the ON-LSTM structure contributes to modeling time-series information in EEG signal emotion recognition. The hidden connection between each time slice can be better mined by learning the hierarchical structure information. In particular, the output results of the network proposed in this paper, after the EEG frequency and spatial information modeling are used as the input of the ON-LSTM structure, which can better capture the continuity of emotion changes in time for the emotion recognition of the EEG signal with the 4D structure, is of great significance.

3.5. Method Comparison

In order to show that the 4D-FSTception method is superior to other classification methods, this article chooses the following EEG emotion classification methods for comparison.
  • HCNN [38]: It uses a hierarchical CNN for EEG emotion classification and recognition and uses the differential entropy features of two-dimensional EEG as the input of the neural network model, which proves that the band and band are more suitable for emotion recognition. The method considers the spatial information and frequency information of the EEG signal.
  • RGNN [39]: It uses the adjacency matrix in the graph neural network to simulate the inter-channel relationship in the EEG signal and realizes the simultaneous capture of the relationship between the local channel and the global channel. The connection and sparseness of the adjacency matrix are determined by humans. Supported by the neurological theory of brain tissue, this method shows that the relationship between global channels and local channels in the left and right hemispheres plays an important role in emotion recognition.
  • BDGLS [40]: It uses differential entropy features as input data. By combining the advantages of dynamic graph convolutional neural networks and generalized learning systems, emotion recognition accuracy can be improved over the full frequency band of EEG features. This method considers the frequency information and spatial information of the EEG signal at the same time.
  • PCRNN [32]: It first uses the CNN module to obtain space characteristics from each 2D EEG topographic map, then uses LSTM to obtain time characteristics from the EEG vector sequence, and finally integrates space and time characteristics to carry out emotional classification.
  • 4D-CRNN [41]: It first extracts features from EEG signals to construct a four-dimensional feature structure, then uses convolutional recurrent neural networks to extract EEG signals to obtain spatial features and frequency features, and uses LSTM to extract time from EEG vector sequences features, and finally carry out EEG emotion classification.
  • 4D-aNN (DE) [42]: It uses 4D space-spectrum-time representations containing the space, frequency spectrum, and time information of the EEG signal as input. An attention mechanism is added to the CNN module and the bidirectional LSTM module. This method also considers the time, space, and frequency of EEG information.
  • ACRNN [43]: It adopts the convolutional recurrent neural network method based on the attention mechanism. It first uses the attention mechanism to distribute the weights between channels, then uses the CNN to extract the spatial information of the EEG, and finally uses the RNN to integrate and extract the temporal information features of the EEG. The method considers spatial information and temporal information for the emotion classification of EEG signals.
The above methods considered in the statistical validation experiments in this paper mainly include HCNN, RGNN, and BDGLS methods that used only frequency and spatial information for sentiment classification and recognition, as well as PCRNN and ACRNN methods that used spatial and temporal information features for sentiment classification. We also worked with 4D-CRNN and 4D-aNN methods that used frequency, spatial, and temporal features for sentiment recognition at the same time.
We reproduce these methods described above based on the structural parameters exploited in the original paper. We conduct five-fold cross-validation experiments for these approaches on the SEED and DEAP data. We list the mean accuracy and standard deviation for each method on the SEED data set and DEAP data set in Table 10. For the SEED data set, the ACC of 4D-FSTception is 95.49%, which is 6.89% higher than the HCNN method that only considers the spatial information of the EEG signal, and it is also 1.25% higher than the RGNN. For the frequency information of the EEG signal simultaneously and the BDGLS method of spatial information, this method is 1.83% higher than this method. Both the 4D-CRNN and 4D-aNN (DE) methods consider the frequency, space, and time features at the same time. We choose the same input feature structure, and the differential entropy feature is selected in the frequency feature extraction. The 4D-FSTception is 0.75% higher than 4D-CRNN and 0.6% higher than 4D-aNN. For the DEAP data set, the experimental method in this paper achieves good results in EEG emotion recognition; its valence and arousal are 94.61% and 95.02%, respectively, which are 0.39% and 0.44% higher than 4D-CRNN, respectively. It is 4.35% and 4.04% higher than PCRNN, which only considers frequency and temporal information, and 0.89% and 1.64% higher than ACRNN, which only considers spatial and temporal information. For the 4D-CRNN method that simultaneously considers frequency, spatial, and temporal features, our method is 0.39% and 0.44% higher than 4D-CRNN in valence and arousal, respectively.
To sum up, the method proposed in this paper shows the best emotion recognition and classification performance in any combination of frequency, time, and space information or in the combination of the three characteristics.

4. Conclusions

The paper points out a 4D-FSTception method for EEG emotion identification. The 4D input features include the frequency features of different frequency bands of the EEG signal, as well as the continuous-time features, while also maintaining the spatial information between channels. This method first extracts the differential entropy feature vectors of different frequency bands of the EEG signal at the same time. Secondly, according to the distribution position of the electrodes, the multichannel one-dimensional feature vector is converted into a two-dimensional feature matrix, which fully considers the spatial position information of the EEG channel. Then, the two-dimensional feature matrix is superimposed to form a three-dimensional matrix, and finally, the three-dimensional EEG sequence time segments divided into fixed lengths constitute 4D input features. Next, we introduce the FSTception model closely combined with CNN and ON-LSTM in detail. CNN is used to solve frequency and space data, and ON-LSTM is used to obtain time correlations from CNN output to reveal the underlying deep emotional characteristics. The proposed method achieves better performance on both SEED and DEAP datasets. By comparing with seven excellent studies, the 4D-FSTception model can effectively improve the accuracy of emotion discrimination in EEG signals. We hope to enhance our paper by building our own data set and verifying the validity of our approach in a real-world scenario. In future work, we can consider combining DE features with more features to further enhance the emotion recognition performance of EEG signals through feature complementarity.

Author Contributions

Material preparation, data collection, and analysis were performed by Q.L. (Qi Li), Q.L. (Quanyang Liu), Y.M. and X.Z. The first draft of the manuscript was written by Q.L. (Qi Li) and all authors commented on previous versions of the manuscript. Y.L., Q.Z. and F.Y. made critical revisions to work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Department Project of Jilin Province, grant number YDZJ202101ZYTS064.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Datasets are available in a publicly accessible repository that does not issue DOIs. These data can be found at the following address: http://www.eecs.qmul.ac.uk/mmv/datasets/deap/index.html (accessed on 20 July 2022), and https://bcmi.sjtu.edu.cn/home/seed/(accessed on 16 March 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bota, P.J.; Wang, C.; Fred, A.L.N.; Da Silva, H.P. A review, current challenges, and future possibilities on emotion recognition using machine learning and physiological signals. IEEE Access 2019, 7, 140990–141020. [Google Scholar] [CrossRef]
  2. Huang, Y.A.; Dupont, P.; Van de Vliet, L.; Jastorff, J.; Peeters, R.; Theys, T.; Van Loon, J.; Van Paesschen, V.; Van Den Stock, J.; Vandenbulcke, M. Network level characteristics in the emotion recognition network after unilateral temporal lobe surgery. Eur. J. Neurosci. 2020, 52, 3470–3484. [Google Scholar] [CrossRef]
  3. Egger, M.; Ley, M.; Hanke, S. Emotion recognition from physiological signal analysis: A review. Electron. Notes Theor. Comput. Sci. 2019, 343, 35–55. [Google Scholar] [CrossRef]
  4. Luo, J.; Tian, Y.; Yu, H.; Chen, Y.; Wu, M. Semi-Supervised Cross-Subject Emotion Recognition Based on Stacked Denoising Autoencoder Architecture Using a Fusion of Multi-Modal Physiological Signals. Entropy 2022, 24, 577. [Google Scholar] [CrossRef]
  5. Yao, L.; Wang, M.; Lu, Y.; Li, H.; Zhang, X. EEG-Based Emotion Recognition by Exploiting Fused Network Entropy Measures of Complex Networks across Subjects. Entropy 2021, 23, 984. [Google Scholar] [CrossRef]
  6. Keshmiri, S.; Shiomi, M.; Ishiguro, H. Entropy of the Multi-Channel EEG Recordings Identifies the Distributed Signatures of Negative, Neutral and Positive Affect in Whole-Brain Variability. Entropy 2019, 21, 1228. [Google Scholar] [CrossRef] [Green Version]
  7. Pan, L.; Wang, S.; Yin, Z.; Song, A. Recognition of Human Inner Emotion Based on Two-Stage FCA-ReliefF Feature Optimization. Inf. Technol. Control 2022, 51, 32–47. [Google Scholar] [CrossRef]
  8. Gao, Z.; Cui, X.; Wan, W.; Gu, Z. Recognition of Emotional States Using Multiscale Information Analysis of High Frequency EEG Oscillations. Entropy 2019, 21, 609. [Google Scholar] [CrossRef] [Green Version]
  9. Chao, H.; Dong, L.; Liu, Y.; Lu, B. Emotion recognition from multiband EEG signals using CapsNet. Sensors 2019, 19, 2212. [Google Scholar] [CrossRef] [Green Version]
  10. Catrambone, V.; Greco, A.; Scilingo, E.P.; Valenza, G. Functional Linear and Nonlinear Brain–Heart Interplay during Emotional Video Elicitation: A Maximum Information Coefficient Study. Entropy 2019, 21, 892. [Google Scholar] [CrossRef]
  11. Krishnan, P.T.; Joseph Raj, A.N.; Rajangam, V. Emotion classification from speech signal based on empirical mode decomposition and non-linear features. Complex Intell. Syst. 2021, 7, 1919–1934. [Google Scholar] [CrossRef]
  12. Danelljan, M.; Robinson, A.; Shahbaz Khan, F.; Felsberg, M. Beyond correlation filters: Learning continuous convolution operators for visual tracking. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 472–488. [Google Scholar] [CrossRef] [Green Version]
  13. Zhang, J.; Min, Y. Four-Classes Human Emotion Recognition Via Entropy Characteristic and Random Forest. Inf. Technol. Control 2020, 49, 285–298. [Google Scholar] [CrossRef]
  14. Cao, R.; Shi, H.; Wang, X.; Huo, S.; Hao, Y.; Wang, B.; Guo, H.; Xiang, J. Hemispheric Asymmetry of Functional Brain Networks under Different Emotions Using EEG Data. Entropy 2020, 22, 939. [Google Scholar] [CrossRef]
  15. Liu, Y.; Sourina, O. Real-time fractal-based valence level recognition from EEG. In Transactions on Computational Science XVIII; Springer: Berlin/Heidelberg, Germany, 2013; pp. 101–120. [Google Scholar] [CrossRef]
  16. Lin, Y.P.; Wang, C.H.; Jung, T.P.; Wu, T.L.; Jeng, S.K.; Duann, J.R.; Chen, J.H. EEG-based emotion recognition in music listening. IEEE Trans. Biomed. Eng. 2010, 57, 1798–1806. [Google Scholar] [CrossRef]
  17. Altan, G.; Yayık, A.; Kutlu, Y. Deep learning with ConvNet predicts imagery tasks through EEG. Neural Process. Lett. 2021, 53, 2917–2932. [Google Scholar] [CrossRef]
  18. Subha, D.P.; Joseph, P.K.; Acharya, U.R.; Lim, C.M. EEG signal analysis: A survey. J. Med. Syst. 2010, 34, 195–212. [Google Scholar] [CrossRef]
  19. Zou, X.; Feng, L.; Sun, H. Compressive Sensing of Multichannel EEG Signals Based on Graph Fourier Transform and Cosparsity. Neural Process. Lett. 2020, 51, 1227–1236. [Google Scholar] [CrossRef]
  20. Voulodimos, A.; Doulamis, N.; Doulamis, A.; Protopapadakis, E. Deep learning for computer vision: A brief review. Comput. Intell. Neurosci. 2018, 2018, 7068349. [Google Scholar] [CrossRef]
  21. Khalil, R.A.; Jones, E.; Babar, M.I.; Jan, T.; Zafar, M.H.; Alhussain, T. Speech emotion recognition using deep learning techniques: A review. IEEE Access 2019, 7, 117327–117345. [Google Scholar] [CrossRef]
  22. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent trends in deep learning based natural language processing. IEEE Comput. Intell. Mag. 2018, 13, 55–75. [Google Scholar] [CrossRef]
  23. Komolovaitė, D.; Maskeliūnas, R.; Damaševičius, R. Deep Convolutional Neural Network-Based Visual Stimuli Classification Using Electroencephalography Signals of Healthy and Alzheimer’s Disease Subjects. Life 2022, 12, 374. [Google Scholar] [CrossRef]
  24. Thammasan, N.; Fukui, K.I.; Numao, M. Application of deep belief networks in eeg-based dynamic music-emotion recognition. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 881–888. [Google Scholar] [CrossRef]
  25. Tripathi, S.; Acharya, S.; Sharma, R.D.; Mittal, S.; Bhattacharya, S. Using deep and convolutional neural networks for accurate emotion classification on deap dataset. In Proceedings of the Twenty-ninth IAAI Conference, San Francisco, CA, USA, 6–9 February 2017. [Google Scholar] [CrossRef]
  26. Salama, E.S.; El-Khoribi, R.A.; Shoman, M.E.; Shalaby, M.A.W. EEG-based emotion recognition using 3D convolutional neural networks. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 329–337. [Google Scholar] [CrossRef] [Green Version]
  27. Yang, Y.; Wu, Q.; Fu, Y.; Chen, X. Continuous convolutional neural network with 3D input for EEG-based emotion recognition. In International Conference on Neural Information Processing; Springer: Cham, Switzerland, 2018; pp. 433–443. [Google Scholar] [CrossRef]
  28. Riezler, S.; Hagmann, M. Validity, Reliability, and Significance; Springer: Cham, Switzerland, 2022. [Google Scholar] [CrossRef]
  29. Zheng, W.L.; Lu, B.L. Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Trans. Auton. Ment. Dev. 2015, 7, 162–175. [Google Scholar] [CrossRef]
  30. Meng-meng, Y.A.N.; Zhao, L.V.; Wen-hui, S.U.N. Extraction of spatial features of emotional EEG signals based on common spatial pattern. J. Graph. 2020, 41, 424. [Google Scholar]
  31. Liu, C.Y.; Li, W.Q.; Bi, X.J. Research on EEG emotion recognition based on RCNN-LSTM. Acta Autom. Sin. 2020, 45, 1–9. [Google Scholar] [CrossRef]
  32. Yang, Y.; Wu, Q.; Qiu, M.; Wang, Y.; Chen, X. Emotion recognition from multi-channel EEG through parallel convolutional recurrent neural network. In Proceedings of the 2018 International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  33. Duan, R.N.; Zhu, J.Y.; Lu, B.L. Differential entropy feature for EEG-based emotion classification. In Proceedings of the 2013 6th International IEEE/EMBS Conference on Neural Engineering (NER), San Diego, CA, USA, 6–8 November 2013; pp. 81–84. [Google Scholar] [CrossRef]
  34. Shi, L.C.; Jiao, Y.Y.; Lu, B.L. Differential entropy feature for EEG-based vigilance estimation. In Proceedings of the 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Osaka, Japan, 3–7 July 2013; pp. 6627–6630. [Google Scholar]
  35. Shen, Y.; Tan, S.; Sordoni, A.; Courville, A. Ordered neurons: Integrating tree structures into recurrent neural networks. arXiv 2018, arXiv:1810.09536. [Google Scholar]
  36. Zheng, W.L.; Liu, W.; Lu, Y.; Lu, B.L.; Cichocki, A. Emotionmeter: A multimodal framework for recognizing human emotions. IEEE Trans. Cybern. 2018, 49, 1110–1122. [Google Scholar] [CrossRef]
  37. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. Deap: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  38. Li, J.; Zhang, Z.; He, H. Hierarchical convolutional neural networks for EEG-based emotion recognition. Cogn. Comput. 2018, 10, 368–380. [Google Scholar] [CrossRef]
  39. Zhong, P.; Wang, D.; Miao, C. EEG-based emotion recognition using regularized graph neural networks. IEEE Trans. Affect. Comput 2020, 13, 1290–1301. [Google Scholar] [CrossRef]
  40. Wang, X.H.; Zhang, T.; Xu, X.M.; Chen, L.; Xing, X.F.; Chen, C.P. EEG emotion recognition using dynamical graph convolutional neural networks and broad learning system. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 1240–1244. [Google Scholar] [CrossRef]
  41. Shen, F.; Dai, G.; Lin, G.; Zhang, J.; Kong, W.; Zeng, H. EEG-based emotion recognition using 4D convolutional recurrent neural network. Cogn. Neurodynamics 2020, 14, 815–828. [Google Scholar] [CrossRef]
  42. Xiao, G.; Shi, M.; Ye, M.; Xu, B.; Chen, Z.; Ren, Q. 4D attention-based neural network for EEG emotion recognition. Cogn. Neurodynamics 2022, 16, 805–818. [Google Scholar] [CrossRef]
  43. Tao, W.; Li, C.; Song, R.; Cheng, J.; Liu, Y.; Wan, F.; Chen, X. EEG-based emotion recognition via channel-wise attention and self attention. IEEE Trans. Affect. Comput. 2020. [Google Scholar] [CrossRef]
Figure 1. The general flowchart of the proposed method.
Figure 1. The general flowchart of the proposed method.
Entropy 24 01830 g001
Figure 2. The generation of 4D inputs.
Figure 2. The generation of 4D inputs.
Entropy 24 01830 g002
Figure 3. The frequency response of the filter.
Figure 3. The frequency response of the filter.
Entropy 24 01830 g003
Figure 4. The raw EEG signal and its five frequency bands in the DEAP dataset.
Figure 4. The raw EEG signal and its five frequency bands in the DEAP dataset.
Entropy 24 01830 g004
Figure 5. The 2D map of 62 channels.
Figure 5. The 2D map of 62 channels.
Entropy 24 01830 g005
Figure 6. The structure of CNN module for frequency and spatial feature learning.
Figure 6. The structure of CNN module for frequency and spatial feature learning.
Entropy 24 01830 g006
Figure 7. The detailed structure of ON-LSTM.
Figure 7. The detailed structure of ON-LSTM.
Entropy 24 01830 g007
Figure 8. ON-LSTM partition update. The numbers are randomly generated, with historical information at the top, current input at the bottom, and currently integrated output in the middle. The yellow part at the top is the historical information level (the main forgetting gate), the green part at the bottom is the input information level (the main input gate), the yellow part in the middle is the directly copied historical information, the green is the directly copied input information, the purple is the intersection information fused according to the LSTM, and the white is the unrelated “blank area”.
Figure 8. ON-LSTM partition update. The numbers are randomly generated, with historical information at the top, current input at the bottom, and currently integrated output in the middle. The yellow part at the top is the historical information level (the main forgetting gate), the green part at the bottom is the input information level (the main input gate), the yellow part in the middle is the directly copied historical information, the green is the directly copied input information, the purple is the intersection information fused according to the LSTM, and the white is the unrelated “blank area”.
Entropy 24 01830 g008
Figure 9. Overall performance of the 4D-FSTception model on SEED dataset.
Figure 9. Overall performance of the 4D-FSTception model on SEED dataset.
Entropy 24 01830 g009
Figure 10. Overall performance of the 4D-FSTception model on DEAP dataset.
Figure 10. Overall performance of the 4D-FSTception model on DEAP dataset.
Entropy 24 01830 g010
Figure 11. Confusion matrices of our method on SEED dataset.
Figure 11. Confusion matrices of our method on SEED dataset.
Entropy 24 01830 g011
Figure 12. Confusion matrix: (left) arousal, (right) valence on DEAP dataset.
Figure 12. Confusion matrix: (left) arousal, (right) valence on DEAP dataset.
Entropy 24 01830 g012
Figure 13. Accuracy performance of different methods in different datasets.
Figure 13. Accuracy performance of different methods in different datasets.
Entropy 24 01830 g013
Figure 14. Time cost performance of different methods in different datasets.
Figure 14. Time cost performance of different methods in different datasets.
Entropy 24 01830 g014
Figure 15. CNN + LSTM and CNN + ON-LSTM compare the experimental results.
Figure 15. CNN + LSTM and CNN + ON-LSTM compare the experimental results.
Entropy 24 01830 g015
Table 1. The detailed comparison to other works.
Table 1. The detailed comparison to other works.
ArticleMethodApplicationDatasetAccuracy (%)
Komolovaitė et al. [23]EEGNet SSVEPAlzheimer researchFigshare website50.2
Thammasan et al. [24]Deep Belief Network
(DBN)
Emotion recognition in music listening15 recruited healthy students
(valence/arousal emotion types)
82.42/88.24
Tripathi et al. [25]Convolutional Neural Network (CNN)Automatic emotion recognitionDEAP
(valence/arousal emotion types)
73.36/81.41
Salama et al. [26]3-Dimensional Convolutional Neural Networks (3D-CNN)Emotion recognitionDEAP
(valence/arousal emotion types)
88.49/87.44
Yang et al. [27]Convolutional Neural Network (CNN)Automatic emotion recognitionDEAP
(valence/arousal emotion types)
90.24/89.45
Zheng et al. [29]Deep Belief Network
(DBN)
Emotion recognition15 recruited subjects86.65
Meng-meng et al. [30]Common Spatial Pattern (CSP)Emotion recognition6 recruited healthy students87.54
Liu et al. [31]Recurrent Convolutional Neural Network and Long Short Term
Memory (RCNN-LSTM)
Automatic emotion recognitionDEAP96.63
Table 2. The frequency modes and corresponding characteristics of EEG.
Table 2. The frequency modes and corresponding characteristics of EEG.
PatternsFrequencyBrain StateAwareness
( δ ) Delta1–4 HzDeep sleep patternLower
( θ ) Theta4–8 HzLight sleep patternLow
( α ) Alpha8–13 HzClosing the eyes, relax stateMedium
( β ) Beta13–30 HzActive thinking, focus, high alert, anxiousHigh
( γ ) Gamma30–50 HzMentally active and hypertensiveHigher
Table 3. DEAP dataset content.
Table 3. DEAP dataset content.
NameSizeContents
Data40 × 40 × 8064videos × channels × data
Labels40 × 4videos × labels
(valence, arousal, dominance, liking)
Table 4. Recognition Accuracy for each subject on SEED dataset.
Table 4. Recognition Accuracy for each subject on SEED dataset.
SubjectsAccuracySubjectsAccuracySubjectsAccuracySubjectsAccuracy
192.45%595.06%995.89%1396.12%
296.86%697.63%1096.83%1495.32%
394.17%793.99%1195.83%1591.50%
499.59%896.15%1295.03%
Table 5. Recognition Accuracy for each subject on “Arousal”.
Table 5. Recognition Accuracy for each subject on “Arousal”.
SubjectsAccuracySubjectsAccuracySubjectsAccuracySubjectsAccuracy
198.13%990.50%1791.50%2598.13%
290.00%1096.00%1895.38%2692.63%
397.25%1196.50%1993.50%2798.38%
490.50%1297.50%2096.88%2896.00%
595.00%1399.86%2197.13%2993.38%
698.75%1497.00%2279.88%3090.88%
798.00%1594.63%2398.38%3195.00%
896.25%1698.25%2496.00%3293.63%
Table 6. Recognition Accuracy for each subject on “Valence”.
Table 6. Recognition Accuracy for each subject on “Valence”.
SubjectsAccuracySubjectsAccuracySubjectsAccuracySubjectsAccuracy
197.88%993.75%1791.25%2596.75%
289.88%1096.13%1896.50%2692.13%
395.63%1192.00%1992.00%2798.75%
493.00%1296.63%2098.00%2895.13%
587.50%1398.00%2196.00%2994.88%
697.13%1497.25%2284.00%3090.13%
798.25%1595.50%2398.63%3196.25%
895.88%1698.00%2489.88%3294.75%
Table 7. The score of 4D-FSTception evaluation index is on SEED dataset.
Table 7. The score of 4D-FSTception evaluation index is on SEED dataset.
ClassesAccuracyPrecisionRecallF1-Score
Negative96.67%94.87%92.50%93.67%
Neutral93.42%94.29%93.94%96.12%
Positive96.38%98.96%95.00%96.94%
Table 8. The score of 4D-FSTception evaluation index is on DEAP dataset.
Table 8. The score of 4D-FSTception evaluation index is on DEAP dataset.
Valence/ArousalClassAccuracyPrecisionRecallF1-Score
ValenceHigh94.90%98.48%93.94%96.47%
Low93.42%98.41%93.18%96.12%
ArousalHigh96.32%98.02%91.67%94.74%
Low93.72%93.30%95.37%96.71%
Table 9. Comparison of computational performance with other methods.
Table 9. Comparison of computational performance with other methods.
MethodMap ShapeSEEDDEAP-ValenceDEAP-Arousal
Acc (%)Time Cost (s)Acc (%)Time Cost (s)Acc (%)Time Cost (s)FLOPS (G)Params (M)
HCNN [38]19 × 1988.603600------
4D-CRNN [42]8 × 994.7481194.2239594.5840018.6315.8
4D-FSTception (LSTM) (ours)8 × 995.4466894.2928494.6029019.9414.4
4D-FSTception (ours)8 × 995.4966094.6128095.0227920.6112.7
Table 10. The performances (average ACC ± STD (%)) of the compared methods on SEED and DEAP datasets.
Table 10. The performances (average ACC ± STD (%)) of the compared methods on SEED and DEAP datasets.
MethodInformationACC ± STD (%)
SEEDDEAP-ValenceDEAP-Arousal
HCNN [38]Frequency + spatial88.60 ± 2.60--
RGNN [39]Frequency + spatial94.24 ± 5.95--
BDGLS [40]Frequency + spatial93.66 ± 6.11--
PCRNN [32]Spatial + temporal-90.26 ± 2.8890.98 ± 3.09
ACRNN [43]Spatial + temporal93.723.2193.38
4D-CRNN [41]Frequency + spatial + temporal94.74 ± 2.3294.22 ± 2.6194.58 ± 3.69
4D-aNN (DE) [42]Frequency + spatial + temporal95.39 ± 3.05--
4D-FSTception (LSTM) (ours)Frequency + spatial + temporal95.44 ± 0.3294.29 ± 1.8994.60 ± 2.08
4D-FSTception (ours)Frequency + spatial + temporal95.49 ± 3.0194.61 ± 2.8395.02 ± 2.85
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Q.; Liu, Y.; Liu, Q.; Zhang, Q.; Yan, F.; Ma, Y.; Zhang, X. Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals. Entropy 2022, 24, 1830. https://doi.org/10.3390/e24121830

AMA Style

Li Q, Liu Y, Liu Q, Zhang Q, Yan F, Ma Y, Zhang X. Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals. Entropy. 2022; 24(12):1830. https://doi.org/10.3390/e24121830

Chicago/Turabian Style

Li, Qi, Yunqing Liu, Quanyang Liu, Qiong Zhang, Fei Yan, Yimin Ma, and Xinyu Zhang. 2022. "Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals" Entropy 24, no. 12: 1830. https://doi.org/10.3390/e24121830

APA Style

Li, Q., Liu, Y., Liu, Q., Zhang, Q., Yan, F., Ma, Y., & Zhang, X. (2022). Multidimensional Feature in Emotion Recognition Based on Multi-Channel EEG Signals. Entropy, 24(12), 1830. https://doi.org/10.3390/e24121830

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop