Next Article in Journal
Optimization of Deep Neural Networks Using SoCs with OpenCL
Next Article in Special Issue
Staged Incentive and Punishment Mechanism for Mobile Crowd Sensing
Previous Article in Journal
A CMOS Self-Contained Quadrature Signal Generator for SoC Impedance Spectroscopy
Previous Article in Special Issue
Weak Defect Identification for Centrifugal Compressor Blade Crack Based on Pressure Sensors and Genetic Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electroencephalography Based Fusion Two-Dimensional (2D)-Convolution Neural Networks (CNN) Model for Emotion Recognition System

Department of Computer Science, Yonsei University, 50 Yonsei-ro, Seodaemun-gu, Seoul 03722, Korea
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(5), 1383; https://doi.org/10.3390/s18051383
Submission received: 20 March 2018 / Revised: 25 April 2018 / Accepted: 26 April 2018 / Published: 30 April 2018

Abstract

:
The purpose of this study is to improve human emotional classification accuracy using a convolution neural networks (CNN) model and to suggest an overall method to classify emotion based on multimodal data. We improved classification performance by combining electroencephalogram (EEG) and galvanic skin response (GSR) signals. GSR signals are preprocessed using by the zero-crossing rate. Sufficient EEG feature extraction can be obtained through CNN. Therefore, we propose a suitable CNN model for feature extraction by tuning hyper parameters in convolution filters. The EEG signal is preprocessed prior to convolution by a wavelet transform while considering time and frequency simultaneously. We use a database for emotion analysis using the physiological signals open dataset to verify the proposed process, achieving 73.4% accuracy, showing significant performance improvement over the current best practice models.

1. Introduction

Multimodal human and computer interaction (HCI) has been actively researched over the last few years. One outstanding issue is affective computing, designing devices that communicate with humans by interpreting emotions. Emotion recognition has been attracting attention as a next-generation technology in many fields, from the development of humanistic robots to consumer analysis and safe driving. Most previous research has classified emotions using only facial expressions. However, facial expressions only represent part of the overall human emotional response, and emotion discriminators can sometimes make significant mistakes. For example, classifying an athlete's image as displaying happy emotion, when actually the smiling athlete is nervous prior to an important game [1]. On the other hand, biological signals from the central (CNS) and the peripheral (PNS) nervous systems are hard for humans to mentally control, and can accurately represent emotions. Previous studies have shown that changes in skin signals (i.e., galvanic skin response (GSR)) are closely related to changes in peripheral nerves with emotional changes [2], and electroencephalogram (EEG) signals from the frontal lobe are strongly related to emotional changes [3,4]. Therefore, the current study classified emotions using biological signals, including EEG and GSR.
Electroencephalogram signals are used in brain computer interface research, measuring brain electrical activity using an electrode that is attached to the scalp. However, reduced accuracy due EEG signal instability remains a major problem, and EEG signals are untrustworthy, even when they are employing expensive and reliable equipment. The solution is to use as many heterogeneous sensors as possible to provide reliable multiple data. Therefore, we designed a data adaptive CNN model to improve the emotion classification accuracy, reducing current model instabilities, using EEG and GSR data. We also implemented effective spectrogram feature extraction and designed a multimodal classifier that takes two features as input at the first layer of a fully connected network.
This paper is organized as follows. Section 2 discusses previous research methodologies and results. Section 3 discusses the current paper’s main contributions, including the details of label processing, EEG signal transformation, GSR data feature extraction, and introduces the proposed CNN model architecture and training strategy. Section 4 analyzes the results and compares them with the current best practice models. Finally, Section 5 summarizes and concludes the paper.

2. Related Work

The related research fields of emotion classification and EEG preprocessing have achieved remarkable results. In general, preprocessing EEG data consists of selecting data while considering the frequency and the location of the brain. Fast Fourier transform (FFT) is the most common frequency analysis method for raw EEG data [5,6,7,8,9], and it was adopted here to extract EEG features. However, FFT cannot reflect temporal information in the frequency data, requiring additional methods to recognize emotions over time. Therefore, short time Fourier transform (STFT), which can express frequency per hour [10,11,12,13], was also used to analyze EEG signals.
Classifying EEG features by frequency is the most common method to differentiate alpha, beta, theta, and gamma waves. Liu et al. [14] presented a table of emotions by frequency and electrode location within the brain region. Figure 1 shows the location of the electrodes that are attached to the scalp using the 10-20 system, which is the international standard. Electrodes F3 and F4 distinguish between negative and positive emotional states, and AF3 andAF4 distinguish positive emotions from the surrounding emotions. Wavelet analysis is one of the best ways to express frequency and time, and has also been employed in EEG classification [15,16,17,18].
Various previous studies considered emotional classification methods. Mollahosseini et al. [19] designed a CNN based face recognition module. Gerard Pons et al. [20] enhanced facial image classification performance by supervised hierarchical learning. Ding et al. [21] performed deep face recognition that was based on a two steps model. Poria et al. [22] implemented multimodal visual and audio data analysis beyond the focus on text-based emotional analysis. They also succeeded in feature fusion through deep learning based heterogeneous data dimension reduction.
The Database for Emotion Analysis using Physiological signals (DEAP) dataset has been widely employed for emotion classification models using biomedical signals. Koelstra et al. [23] used the DEAP data set to classify PNS and CNS sensor data, and measured the emotional classification performance. Liu and Sourina [24] studied EEG valence levels for real-time applications. Naser et al. [25] predicted emotions extracted from music videos. Chen et al. [26] applied ontology and datamining techniques for EEG based emotion analysis.
Bayesian networks, unsupervised deep running, and deep belief networks have also been applied [27,28,29].

3. Methods

3.1. Multiple Label Classification

A label was constructed using the self-assessment value that was provided in the DEAP dataset, including valence, arousal, dominance, liking, and familiarity. Emotional states are typically evaluated using arousal and valence, and are divided into four sections: high arousal, high valence (HAHV); high arousal, low valence (HALV); low arousal, low valence (LALV); and, low arousal, high valence (LAHV) [30], as shown in Figure 2. Thus, emotional states can be classified according to arousal and valence levels.
Labeling was based on a threshold value for the two-dimensional (2D) plane. We implemented k-means clustering on self-assessed arousal and valence levels to find the most appropriate threshold. Previous studies have employed one shot encoding for labeling as a 2D vector, i.e., [HV, LV] and [HA, LA] using k-means clustering with k = 2 [31]. Therefore, we performed independent valence and arousal classifications in order to compare with previous models.
However, independent classification fails to consider arousal and valence correlations, and since the data is arousal and valence levels, rather than emotion level, it cannot be implemented for end to end learning, since it must be mapped onto the two-dimensional (2D) plane (Figure 2) for emotion judgment.
Therefore, we propose k-means clustering with k = 4 to provide a four-dimensional (4D) label vector. Figure 3 compares clustering for k = 2 and k = 4. Point (5, 5) is the approximate center mean for both k = 2 and k = 4, hence we use (5, 5) as the threshold.
Thus, labeling included 2D and 4D vectors through one shot encoding for learning.

3.2. EEG Signal Transformation to Time to Frequency Axes

The data was preprocessed to reflect EEG temporal and frequency characteristics. Since the EEG data measuring human emotions are time series data, time information must be reflected in the frequency data. Although the STFT has been widely used to add time information to frequency data [10,11,12,13], it has disadvantages for time-frequency analysis, in that temporal resolution decreases as the window increases; and, frequency resolution decreases as window size decreases.
Therefore, we propose using a wavelet transform to represent the frequency axis, using the open toolbox EEG lab. The extracted spectrogram was 42 × 200 pixel, width × height, where width (200 pixel) represents time, and height (42 pixel) represents EEG sensor frequency (4.0–45 Hz), as shown in Figure 4. Total transformed data include 40960 spectrograms. At this time, the number of batch data for training is 32 spectrogram data that means 32 electrodes that were derived from one stimulus. Therefore, the total amount of data set used in this study is 1280, with data labels, as shown in Table 1 and Table 2.
Conventional EEG based emotion classification analyzes the degree of activity in a specific area of the brain (e.g. the frontal lobe), using electrodes that were attached to the head close to the frontal lobe and some other lobes (e.g., AF3, AF4, P7). Frequency bands for specific electrodes were typically subdivided into alpha, beta wave, gamma, etc. waves to allow for simple and shallow classification models, such as support vector machines (SVMs). However, sensor selection and subdivision ignores emotion related signal changes in other brain regions. Recent advanced deep learning techniques can improve emotional analysis accuracy by incorporating all sensor data for each experiment.

3.3. GSR Preprocessing Using Short Time Zero Crossing Rate

To extract the feature, we divide the GSR waveform into defined windows and calculate the short time zero crossing rate (STZCR), i.e., the number of times the signal crosses zero within a given window. That is, we intend to use the change in amplitude of the GSR as the input feature vector for deep running. STZCR indicates the rate of signal change,
S T Z C R = 1 N n m | s i g { s ( n ) } s i g { s ( n 1 ) } | 2 w ( m n )
where N is the sampled signal, and w represents the window. We highlighted features using the extracted zero crossing rate vector with threshold
T = G S R s t z c r N s t z c r
where G S R s t z c r is a vector column and N s t z c r is the number of vectors. If the data is greater than the threshold, it outputs max, otherwise it outputs zero. GSR amplitude is generally sensitive to arousal changes and less sensitive to valence changes, hence, it can positively affect EEG features to focus on arousal in the classifier model.

3.4. Fusion Convolution Neural Network Model for EEG Spectrograms and GSR Features

Many neural networks have been developed for classification in recent studies. The first thing to consider when designing a CNN is data characteristics. Therefore, we designed the CNN to use the spectrogram image from the wavelet transformation of all the channels.
Tabar and Halici [32] considered CNN classification problems using EEG spectrograms, and designed a single layer CNN using one-dimensional filtering to provide good classification performance based on motor imagery EEG signals. However, a single filtering through the single convolutional layer does not efficiently extract features for emotion classification, since it is not deep enough to extract emotion data.
Therefore, we propose a neural network based on the extracted data as described above, which allows for deep convolution layers, while also reflecting temporal effects, as shown in Figure 5. We first normalized the data, making the cost function a spherical contour, and helping to increase the learning rate. We then designed a deep convolution layer that reflects time, using a 3 × 2 filter rather than conventional square filters, such as 2 × 2 or 3 × 3. The spectrogram frequency per hour can be reflected by increasing the filter height. Since the filter is a feature identifier that extracts the information from the manifold state, the shape of the filter is related to the content of the feature to be extracted from the receptive field. Our proposed filter can identify data in a region that is relatively longer than a square filter. Thus, the data containing the vertical meaning is repeatedly transmitted to the input of the next layer. As a result, the frequency per hour of the spectrogram image can be learned in CNN. Setting stride = [2, 1] with no padding, the filter can be extracted based only on the image time base. We used a fully connected layer for the final classification. The classifier is trained on the spectrogram features of 32 electrodes extracted through CNN. In continuous training, the classifier learns similar patterns extracted from 32 individual electrodes, and can be classified as a label through the last softmax layer. The entire model consists of four convolutional layers and seven fully connected layers.
Batch normalization [33] was performed before each value was passed to the activation function, except for the last convolutional layer, in order to prevent the model gradient vanishing during training. It has the effect of preventing internal covariance shift by reducing activation function variation that is caused by the previous layer’s variation. Batch normalization was implemented, as follows.
(1)
Normalize the batch data using the batch mean, μ β , and variance, σ β 2 ,
x ^ i = x i μ β σ β 2 + ϵ
(2)
Use the r and d values for scale and shift operations,
y i = γ x ^ i + β
Updating γ and β by training allows for the CNN to better reflect the model characteristics model in normalized variables, rather than simple normalization, such as whitening. Testing uses average γ and β obtained.
Feature maps are generated as the image passes through each convolution layer. The layer activation function is a rectified linear unit (ReLU), which is a function that makes the value of the part where x < 0 in the linear function y = x is 0,
R e L U ( u ) = m a x ( u , 0 ) , m a x ( u , 0 ) = { u , i f   u > 0 0 , o t h e r w i s e
The ReLU function is computationally efficient because its activation is not restricted to [−1, 1], as for the hyperbolic tangent function, but is used as it is. Therefore, training speed for large spectrogram images is increased, and outputting 0 prevents overfitting due to training many weights, hence training regularization can be expected.
After passing through the final 2 × 2 pooling layer, the image is flattened and combined with GSR. To positively influence EEG data performance classification, GSR data uses the data average as the thresh hold to remove noise. It also reduces the computation burden for training a fully connected network by transmitting a zero value to each neuron’s perceptron. The final layer neuron returns the model’s probability distribution using softmax, and performs classification by changing the number of neurons according to the experimental environment, such as [HV, LV], [HA, LA], or [HVHA, LVHA, LVLA, HVAL].

3.5. Training Strategy

We use maximum likelihood estimation (MLE) in order to train the proposed CNN model. MLE maximizes P ( Y | X ; θ ) by optimizing θ in the probability model for a given data point X and label Y . Cross entropy is the most commonly used MLE loss function, and it calculates the difference between two probability distributions. Let p ( x ) be the actual and q ( x ) be the predicted probability distribution for the label. Then, cross entropy, L ( p , q ) , is
L ( p , q ) = p ( x ) · ln q ( x ) d x
CNN training proceeds by back propagation using the gradient decent. We update the weights using the partial derivative of cross entropy loss L for weight matrix W ,
W = W γ · L W
where z j = w i j o i + b is the sum of inner products, and we calculate the gradient as
L W = L p ( z j ) · p ( z j ) z j · z j w i j
where L p ( z j ) is the magnitude of the influence of function p on L and p ( z j ) is the softmax result.
Generally, to find the optimal training point, we find the bias variance trade off point using validation loss, as shown in Figure 6 for the 4 class case.
After 400 iterations, validation loss increases, whereas training continues to decrease. Thus, we can conclude the model becomes over-fitted beyond 400 iterations, providing the optimal training point. Test data should be applied with this level of iteration to measure model accuracy.

4. Results and Discussions

4.1. Experiment Environment

Table 3 shows the hardware and framework specifications for the experiment.

4.2. Dataset

The DEAP dataset [23] was used to provide bio-signal data, containing CNS and PNS data. PNS data comprised GSR, skin temperature, respiration, blood volume (by plethysmograph), and electrooculogram (EOG). GSR data was the skin resistance of the middle and forefinger, skin temperature, and breath change by emotion, including body tension and irritating fear. Plethysmograph measured blood flow changes in the finger. EOG signal was measured by eye blinking, which is related to anxiety. CNS data was the EEG signal.
Data were collected from 32 subjects for 1 m for each of 40 selected music videos. Data was recorded on 48 channels with 512 Hz sampling frequency. We used preprocessed data version of MATLAB and numpy formats that were provided by the DEAP dataset, down-sampled to 128 Hz with a 4.0–45 Hz band pass filter applied.

4.3. Performance Analysis

In this section, we analyzed the performance of the model in two ways. In first evaluation, we analyzed the classification performance for each label using hold-out validation. To construct a hold-out validation set, test, verification, and learning datasets were created 1:1:9 ratio for each label, with batch size = 32 to reflect data from one stimulus. Second, for the LOOCV, we constructed the dataset that was measured by one-video as test set and the other video data as training set. The DEAP dataset consists of a data set for 40 videos per participants. In other words, the second evaluation was performed with 39 video stimuli as training dataset, and the data that was extracted by the other one stimulus was used as a test dataset.
The desired ideal model would accurately distinguish data patterns and generalize them even when testing data are considered, i.e., we want to find a model between over and under fitting. The proposed model does not apply L2 regularization to prevent overfitting, because there is a batch normalization layer. In addition, cross entropy loss was measured for each iteration to find the optimal training point, as shown in Figure 6. Table 4 shows the predicted accuracy for methods of label based and video based classification using each validation method.

4.4. Comparison with Existing Models

We used two class labels that were commonly adopted in previously studies to compare performance, as measured by arousal and valence classification accuracy for the DEAP dataset. Table 5 shows the performance compared with the existing models measured using the same dataset. The performance of our model is shown by the result of LOOCV in Section 4.3, to validate the generalized performance of the model.
The considered methods used a variety of approaches: Koelstra et al. [23] used CNS and PNS sensors; Liu and Sourina [24] used a fractal algorithm to reflect signal complexity that was based on a threshold value; Naser and Saha [25] extracted features using a dual-tree complex wavelet transform and used SVM for classification; Chen et al. [26] used decision trees; Yoon and Chung [27] used Bayesian and perceptron convergence; and, Wang et al. [28] and Li et al. [29] used deep belief networks to automatically extract features and to classify them. Figure 7 show the proposed model has better performance than all compared models
Although EEG data is easier to classify into two classes [34], increasing the number of classes not only enables end-to-end learning, but it also includes correlations between arousal and valence. Therefore, we compared the proposed model performance against previous four class models. Generally, when data quantity is limited, the model accuracy decreases as the number of labels to be classified increases. The performance of our model is shown by the result of LOOCV in Section 4.3, in order to validate the generalized performance of the model. Table 6 shows that the proposed model has high performance when compared to current models.
A variety of approaches were employed in the comparison models: Zubair and Yoon [35] used a discrete wavelet transform, and also applied the mRMR algorithm to enhance the feature correlations; Jadhav et al. [36] extracted EEG features using the gray level co-occurrence matrix, and classified emotion using K-nearest neighbor; Hatamikia et al. [37] used using nonlinear extraction and self-organized classification; Martínez-Rodrigo et al. [38] extracted biological signal features using quadratic sample entropy, performed feature selection, and classified the extracted features by SVM; Zhang et al. [39] used wavelet feature extraction that was based on a smoothed pseudo Winger-Ville distribution and classification using SVM; Mei et al. [40] extracted features by constructing a connection matrix of the brain structure, with subsequent classification using CNN. Figure 8 show a bar graph that the proposed model has high performance when compared to the current models.

5. Conclusions

This study devised data labeling according to emotion criteria, and proposed a data preprocessing methodology to increase the emotional classification performance. Emotion classification was performed using single and multiple sensor based models. Particular focus was overall analysis and CNN filter design according to input data characteristics and noise removal for data processing.
Feature extraction performance was remarkably improved through the proposed filter design, providing significantly improved classification performance when compared with previous models.
This study paves the way for combining data and designing corresponding deep running models. Future research directions will investigate further changes to the emotion analysis framework, such combining multiple neural networks. One approach would be to improve concatenation of simple convolution layers. It may be possible to construct convolution layers for each data characteristic and improve the classification performance using multiple convolution layers.

Author Contributions

Yea-Hoon Kwon and Shin-Dug Kim conceived and designed the experiments; Yea-Hoon Kwon performed the experiments; Yea-Hoon Kwon and Sae-Byuk Shin analyzed the data; Yea-Hoon Kwon and Sae-Byuk Shin wrote the paper.

Funding

This work was supported by The Institute for Information & Communications Technology Promotion funded by the Korean Government (MSIP) (R0124-16-0002, Emotional Intelligence Technology to Infer Human Emotion and Carry on Dialogue Accordingly).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gunadi, I.G.A.; Harjoko, A.; Wardoyo, R.; Ramdhani, N. Fake smile detection using linear support vector machine. In Proceedings of the IEEE International Conference on Data and Software Engineering (ICoDSE), Yogyakarta, Indonesia, 25–26 November 2015; pp. 103–107. [Google Scholar]
  2. Wu, G.; Liu, G.; Hao, M. The analysis of emotion recognition from GSR based on PSO. In Proceedings of the IEEE International Symposium on Intelligence Information Processing and Trusted Computing (IPTC), Huanggang, China, 28–29 October 2010; pp. 360–363. [Google Scholar]
  3. Allen, J.J.B.; Coan, J.A.; Nazarian, M. Issues and assumptions on the road from raw signals to metrics of frontal EEG asymmetry in emotion. Biol. Psychol. 2004, 67, 183–218. [Google Scholar] [CrossRef] [PubMed]
  4. Schmidt, L.A.; Trainor, L.J. Frontal brain electrical activity (EEG) distinguishes valence and intensity of musical emotions. Cognit. Emot. 2001, 15, 487–500. [Google Scholar] [CrossRef]
  5. Khalili, Z.; Moradi, M.H. Emotion detection using brain and peripheral signals. In Proceedings of the Biomedical Engineering Conference, Cairo, Egypt, 18–20 December 2008. [Google Scholar]
  6. Mu, L.; Lu, B.-L. Emotion classification based on gamma-band EEG. In Proceedings of the Annual International Conference of the IEEE, Minneapolis, MN, USA, 3–6 September 2009. [Google Scholar]
  7. Liu, Y.; Sourina, O. EEG based dominance level recognition for emotion-enabled interaction. In Proceedings of the IEEE International Conference on Multimedia and Expo, Melbourne, Australia, 9–13 July 2012. [Google Scholar]
  8. Rozgic, V.; Vitaladevuni, S.N.; Prasad, R. Robust EEG emotion classification using segment level decision fusion. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
  9. Yin, Z.; Wang, Y.; Liu, L.; Zhang, W.; Zhang, J. Cross-subject EEG feature selection for emotion recognition using transfer recursive feature elimination. Front. Neurorobot. 2017, 11, 19. [Google Scholar] [CrossRef] [PubMed]
  10. Liu, Y.J.; Yu, M.; Zhao, G.; Song, J.; Ge, Y.; Shi, Y. Real-time movie-induced discrete emotion recognition from EEG signals. IEEE Trans. Affect. Comput. 2017. [Google Scholar] [CrossRef]
  11. Chanel, G.; Karim, A.-A.; Thierry, P. Valence-arousal evaluation using physiological signals in an emotion recall paradigm. In Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, Montreal, QC, Canada, 7–10 October 2007. [Google Scholar]
  12. Zheng, W.L.; Dong, B.N.; Lu, B.-L. Multimodal emotion recognition using EEG and eye tracking data. In Proceedings of the 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014. [Google Scholar]
  13. Nie, D.; Wang, X.-W.; Shi, L.-C.; Lu, B.-L. EEG based emotion recognition during watching movies. In Proceedings of the 5th International IEEE/EMBS Conference on Neural Engineering, Cancun, Mexico, 27 April–1 May 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 667–670. [Google Scholar]
  14. Inuso, G.; La Foresta, F.; Mammone, N.; Morabito, F.C. Brain activity investigation by EEG processing: Wavelet analysis, kurtosis and Renyi's entropy for artifact detection. In Proceedings of the International Conference on Information Acquisition, Seogwipo-si, Korea, 8–11 July 2007; pp. 195–200. [Google Scholar]
  15. Rossetti, F.; Rodrigues, M.C.A.; de Oliveira, J.A.C.; Garcia-Cairasco, N. EEG wavelet analyses of the striatum-substantia nigra pars reticulata-superior colliculus circuitry: Audiogenic seizures and anticonvulsant drug administration in Wistar audiogenic rats (WAR strain). Epilepsy Res. 2006, 72, 192–208. [Google Scholar] [CrossRef] [PubMed]
  16. Akin, M.; Arserim, M.A.; Kiymik, M.K.; Turkoglu, I. A new approach for diagnosing epilepsy by using wavelet transform and neural networks. In Proceedings of the IEEE 23rd Annual International Conference on Engineering in Medicine and Biology, Istanbul, Turkey, 25–28 October 2001; pp. 1596–1599. [Google Scholar]
  17. Adeli, H.; Zhou, Z.; Dadmehr, N. Analysis of EEG records in an epileptic patient using wavelet transform. J. Neurosci. Methods 2003, 123, 69–87. [Google Scholar] [CrossRef]
  18. Adolphs, R.; Tranel, D.; Hamannb, S.; Youngc, A.W.; Calder, A.J.; Phelps, E.A.; Andersone, A.; Leef, G.P.; Damasio, A.R. Recognition of facial emotion in nine individuals with bilateral amygdala damage. Neuropsychologia 1999, 37, 1111–1117. [Google Scholar] [CrossRef]
  19. Mollahosseini, A.; Chan, D.; Mahoor, M.H. Going deeper in facial expression recognition using deep neural networks. In Proceedings of the Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar]
  20. Pons, G.; Masip, D. Supervised Committee of Convolutional Neural networks in automated facial expression analysis. IEEE Trans. Affect. Comput. 2017. [Google Scholar] [CrossRef]
  21. Ding, H.; Zhou, S.K.; Chellappa, R. Facenet2expnet: Regularizing a deep face recognition net for expression recognition. arXiv, 2016; arXiv:1609.06591. [Google Scholar]
  22. Poria, S.; Chaturvedi, I.; Cambria, E.; Hussain, A. Convolutional MKL based multimodal emotion recognition and sentiment analysis. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining, Barcelona, Spain, 12–15 December 2016; pp. 439–448. [Google Scholar]
  23. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef]
  24. Liu, Y.; Sourina, O. EEG based valence level recognition for real-time applications. In Proceedings of the IEEE international Conference on Cyberworlds, Darmstadt, Germany, 25–27 September 2012; pp. 53–60. [Google Scholar]
  25. Naser, D.S.; Saha, G. Recognition of emotions induced by music videos using DT-CWPT. In Proceedings of the IEEE Indian Conference on Medical Informatics and Telemedicine (ICMIT), Kharagpur, India, 28–30 March 2013; pp. 53–57. [Google Scholar]
  26. Chen, J.; Hu, B.; Moore, P.; Zhang, X.; Ma, X. Electroencephalogram based emotion assessment system using ontology and data mining techniques. Appl. Soft Comput. 2015, 30, 663–674. [Google Scholar] [CrossRef]
  27. Yoon, H.J.; Chung, S.Y. EEG based emotion estimation using Bayesian weighted-log-posterior function and perceptron convergence algorithm. Comput. Biol. Med. 2013, 43, 2230–2237. [Google Scholar] [CrossRef] [PubMed]
  28. Wang, D.; Shang, Y. Modeling physiological data with deep belief networks. Int. J. Inf. Education Technol. 2013, 3, 505. [Google Scholar]
  29. Li, X.; Zhang, P.; Song, D.; Yu, G.; Hou, Y.; Hu, B. EEG based emotion identification using unsupervised deep feature learning. In Proceedings of the SIGIR2015 Workshop on Neuro-Physiological Methods in IR Research, Santiago, Chile, 13 August 2015. [Google Scholar]
  30. Yin, Z.; Zhao, M.; Wang, Y.; Yang, J.; Zhang, J. Recognition of emotions using multimodal physiological signals and an ensemble deep learning model. Comput. Methods Programs Biomed. 2017, 140, 93–110. [Google Scholar] [CrossRef] [PubMed]
  31. Christie, I.C.; Friedman, B.H. Autonomic specificity of discrete emotion and dimensions of affective space: A multivariate approach. Int. J. Psychophysiol. 2004, 51, 143–153. [Google Scholar] [CrossRef] [PubMed]
  32. Tabar, Y.R.; Halici, U. A novel deep learning approach for classification of EEG motor imagery signals. J. Neural Eng. 2016, 14, 016003. [Google Scholar] [CrossRef] [PubMed]
  33. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  34. Karyana, D.N.; Wisesty, U.N.; Nasri, J. Klasifikasi EEG menggunakan Deep Neural Network Dengan Stacked Denoising Autoencoder. Proc. Eng. 2016, 3, 5296–5303. [Google Scholar]
  35. Zubair, M.; Yoon, C. EEG Based Classification of Human Emotions Using Discrete Wavelet Transform. In Proceedings of the Conference on IT Convergence and Security 2017, Seoul, Korea, 25–28 September 2017; pp. 21–28. [Google Scholar]
  36. Jadhav, N.; Manthalkar, R.; Joshi, Y. Electroencephalography based emotion recognition using gray-level co-occurrence matrix features. In Proceedings of the International Conference on Computer Vision and Image Processing, Roorkee, India, 26–28 February 2016; pp. 335–343. [Google Scholar]
  37. Hatamikia, S.; Nasrabadi, A.M. Recognition of emotional states induced by music videos based on nonlinear feature extraction and som classification. In Proceedings of the IEEE 21st Iranian Conference on Biomedical Engineering (ICBME), Tehran, Iran, 26–28 November 2014; pp. 333–337. [Google Scholar]
  38. Martínez-Rodrigo, A.; García-Martínez, B.; Alcaraz, R.; Fernández-Caballero, A.; González, P. Study of Electroencephalographic Signal Regularity for Automatic Emotion Recognition. In Proceedings of the International Conference on Ubiquitous Computing and Ambient Intelligence, Philadelphia, PA, USA, 7–10 November 2017; pp. 766–777. [Google Scholar]
  39. Zhang, X.-Y.; Wand, W.-R.; Shen, C.Y.; Sun, Y.; Huang, L.-X. Extraction of EEG Components Based on Time-Frequency Blind Source Separation. In Proceedings of the International Conference on Intelligent Information Hiding and Multimedia Signal Processing, Matsue, Japan, 12–15 August 2017; pp. 3–10. [Google Scholar]
  40. Mei, H.; Xu, X. EEG based emotion classification using convolutional neural network. In Proceedings of the IEEE International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), Shenzhen, China, 15–17 December 2017; pp. 130–135. [Google Scholar]
Figure 1. The 10-20 system the international standard and location of the electrodes.
Figure 1. The 10-20 system the international standard and location of the electrodes.
Sensors 18 01383 g001
Figure 2. Arousal, valence two-dimension plane.
Figure 2. Arousal, valence two-dimension plane.
Sensors 18 01383 g002
Figure 3. K-means clustering results of Arousal-Valence self-assessment data (a) Clustered Arousal-Valence data result when k = 2; and, (b) Clustered Arousal-Valence data result when k = 4.
Figure 3. K-means clustering results of Arousal-Valence self-assessment data (a) Clustered Arousal-Valence data result when k = 2; and, (b) Clustered Arousal-Valence data result when k = 4.
Sensors 18 01383 g003
Figure 4. Wavelet transformed spectrogram for each electrode.
Figure 4. Wavelet transformed spectrogram for each electrode.
Sensors 18 01383 g004
Figure 5. Proposed convolution neural network combining electroencephalogram (EEG) and wavelet transformed galvanic skin response (GSR).
Figure 5. Proposed convolution neural network combining electroencephalogram (EEG) and wavelet transformed galvanic skin response (GSR).
Sensors 18 01383 g005
Figure 6. Four class loss to find the optimal training point.
Figure 6. Four class loss to find the optimal training point.
Sensors 18 01383 g006
Figure 7. Arousal and valence classification accuracy.
Figure 7. Arousal and valence classification accuracy.
Sensors 18 01383 g007
Figure 8. Four class classification accuracy.
Figure 8. Four class classification accuracy.
Sensors 18 01383 g008
Table 1. The number of extracted wavelet transformed data for two types of labels.
Table 1. The number of extracted wavelet transformed data for two types of labels.
Label 1Data quantity
HAHV458
HALV294
LAHV255
LALV273
Total1280
1 H: high, L: low, V: valence, A: arousal, e.g., HAHV: high arousal, high valence.
Table 2. The number of extracted wavelet transformed data for four types of labels.
Table 2. The number of extracted wavelet transformed data for four types of labels.
Label 1Data quantityLabel 2Data quantity
HA752HV713
LA528LV567
Total1280Total1280
1, 2 H: high, L: low, V: valence, A: arousal, e.g., HA: high arousal.
Table 3. Hardware specifications.
Table 3. Hardware specifications.
CPUIntel Core i5-6600
GPUNVIDIA GeForce GTX 1070 8GBytes
RAMDDR4 16GBytes
OSUbuntu 16.04.
FrameworksTensorflow1.3
MATLAB/ EEG toolbox
Table 4. Emotion classification accuracy.
Table 4. Emotion classification accuracy.
Results A 1Results B 2
Clssification MethodsTwo Class Classification Accuracy 3Four Class Classification Accuracy 4Two Class Classification Accuracy 3Four Class Classification Accuracy 4
ArousalValenceArousalValence
Proposed Fusion Model0.78120.81250.75000.76560.80460.7343
1 Results A: label based classification using hold-out validation.
2 Results B: video based classification using leave-out one cross validation.
3 Arousal: HA, LA; Valence: HV, LV.
4 HAHV, HALV, LALV, LAHV.
Table 5. Two class classification performance.
Table 5. Two class classification performance.
ModelAccuracy
ArousalValence
CNS feature based single modality[23]0.62000.5760
PNS feature based single modality[23]0.57000.6270
Liu and Sourina[24]0.76510.5080
Naser and Saha[25]0.66200.6430
Chen et al.[26]0.69090.6789
Yoon and Chung[27]0.70100.7090
Li et al.[29]0.64200.5840
Wang and Shang[28]0.51200.6090
Proposed fusion CNN model 0.76560.8046
Table 6. Four class classification performance.
Table 6. Four class classification performance.
Model Accuracy
M Zubair and C Yoon[35]0.4540
N Jadhav et al.[36]0.4625
Hatamikia et al.[37]0.5515
Martínez-Rodrigo et al.[38]0.7250
Zhang et al.[39]0.7162
Mei et al.[40]0.7310
Proposed fusion CNN model 0.7343

Share and Cite

MDPI and ACS Style

Kwon, Y.-H.; Shin, S.-B.; Kim, S.-D. Electroencephalography Based Fusion Two-Dimensional (2D)-Convolution Neural Networks (CNN) Model for Emotion Recognition System. Sensors 2018, 18, 1383. https://doi.org/10.3390/s18051383

AMA Style

Kwon Y-H, Shin S-B, Kim S-D. Electroencephalography Based Fusion Two-Dimensional (2D)-Convolution Neural Networks (CNN) Model for Emotion Recognition System. Sensors. 2018; 18(5):1383. https://doi.org/10.3390/s18051383

Chicago/Turabian Style

Kwon, Yea-Hoon, Sae-Byuk Shin, and Shin-Dug Kim. 2018. "Electroencephalography Based Fusion Two-Dimensional (2D)-Convolution Neural Networks (CNN) Model for Emotion Recognition System" Sensors 18, no. 5: 1383. https://doi.org/10.3390/s18051383

APA Style

Kwon, Y. -H., Shin, S. -B., & Kim, S. -D. (2018). Electroencephalography Based Fusion Two-Dimensional (2D)-Convolution Neural Networks (CNN) Model for Emotion Recognition System. Sensors, 18(5), 1383. https://doi.org/10.3390/s18051383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop