A Moth–Flame Optimized Echo State Network and Triplet Feature Extractor for Epilepsy Electro-Encephalography Signals
Abstract
:1. Introduction
- (1)
- A novel feature extraction called moth–flame optimized echo state network (MFO-ESN) is developed, which uses MFO to optimize the hyperparameters of ESN for fitting the specific tasks.
- (2)
- A new function based on the triplet is introduced to evaluate the distribution of features extracted by MFO-ESN without relying on specific classifiers.
- (3)
- MFO-ESN is verified on the real-world single-channel EEG signals classification task with an accuracy of 98.16%. The results also show that MFO-ESN with FDEF can promote the performances of many classifiers.
- (4)
- We also conduct experiments on the multi-channel EEG signals classification task with the highest specificity of both the patient specific and the cross-patient task. The cross-patient task simulates the real diagnosis situation with high specificity, proving the strong generalization ability of MFO-ESN.
2. Related Works
2.1. Echo State Network (ESN)
- (1)
- The number of neurons in the reservoir layer (P): P is the most important hyperparameter, which determines the complexity of ESN and directly affects its performance in various tasks. Generally, increasing P can increase the memory capacity of ESN [24] that is of significance to handling complex tasks. However, excessively increasing P will lead to overfitting.
- (2)
- Spectral radius (SR): SR is the maximum achievable value of the characteristic value of the reservoir matrix . To guarantee the echo state principle (ESP) of ESN, the value of SR should be less than 1 [25].
- (3)
- Connection density (CD): CD indicates the number of neurons in the reservoir layer that participate in random connections. The selection range of CD is usually between 0.01 and 0.2 [26]. In addition, a range of interesting initial structures is proposed to select proper CD, such as distance-based [13], small-world [27], or scale-free [28,29].
- (4)
- Input scaling coefficient (IS): It is important to choose an appropriate IS that can scale the input data to a proper range, so that the neurons of ESN can be well activated and fully take advantage of the non-linearity of the activation function [30].
2.2. Moth–Flame Optimization (MFO)
- (1)
- The moths can only fly within a limited search range.
- (2)
- The initialized positions of the moths are the starting points of the spiral flight.
- (3)
- The endpoints of the spiral flight are the positions of the flames.
- (4)
- The positions of the flames are adjusted during the process of searching.
3. Methodology
3.1. Feature Distribution Evaluation Function (FDEF)
- (1)
- FDEF is robust to the mean value of features. The mean values of the features extracted through different ESN feature extractors are different. Further, features with a bigger mean value always obtain a better result using Equation (11) without considering the distribution of features.
- (2)
- FDEF is not sensitive to the dimension of the feature. The number of neurons in the reservoir layer (P) is a key parameter that needs to be adjusted. The feature dimension is the same as P, which means the feature dimension changes during the training process of MFO-ESN. Equation (11) is sensitive to the varying feature dimensions that force the model to reduce P.
3.2. Moth–Flame Optimized ESN
Algorithm 1 MFO-ESN |
Input: the population number (N), the maximum times of iteration (T) |
Output: the best hyperparameters of ESN |
Steps: |
(a) Set the population number and maximum number of iterations. |
(b) Initialize the moth population. |
(c) Initialize the ESN using hyperparameters represented by moths. |
(d) Extract features using initialized ESN. |
(e) Calculate fitness values of moths according to Equation (12) and sort fitness values. |
(f) Update the moth position based on Equation (7). |
(g) Update the flame position to determine the current optimal solution. |
(h) Determine the number of moths and flames based on Equation (9). |
(i) Repeat step (c) to step (h) until the constraint is met. |
(j) The process ends. |
4. Experiments on the Bonn University EEG Data Set
4.1. A Brief Description of the Data Set
4.2. Feature Extraction of Epileptic Seizure EEG Signals
4.3. Epileptic EEG Signal Classification
5. Experiments on the CHB-MIT EEG Data Set
5.1. A Brief Description of the Data Set
5.2. Results and Discussion of the Multi-Channel Classification
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Megiddo, I.; Colson, A.; Chisholm, D.; Dua, T.; Nandi, A.; Laxminarayan, R. Health and economic benefits of public financing of epilepsy treatment in India: An agent-based simulation model. Epilepsia 2016, 57, 464–474. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Liu, Q.; Ganzetti, M.; Wenderoth, N.; Mantini, D. Detecting Large-Scale Brain Networks Using EEG: Impact of Electrode Density, Head Modeling and Source Localization. Front. Neuroinform. 2018, 12, 4. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Lehnertz, K. Epilepsy and Nonlinear Dynamics. J. Biol. Phys. 2008, 34, 253–266. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Chan, H.-L.; Kuo, P.-C.; Cheng, C.-Y.; Chen, Y.-S. Challenges and Future Perspectives on Electroencephalogram-Based Biometrics in Person Recognition. Front. Neuroinform. 2018, 12, 66. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Vijayan, A.E.; Sen, D.; Sudheer, A. EEG-Based Emotion Recognition Using Statistical Measures and Auto-Regressive Modeling. In Proceedings of the 2015 IEEE International Conference on Computational Intelligence & Communication Technology, IEEE, Ghaziabad, India, 13–14 February 2015; pp. 587–591. [Google Scholar] [CrossRef]
- Farihah, S.N.; Lee, K.Y.; Mansor, W.; Mohamad, N.B.; Mahmoodin, Z.; Saidi, S.A. EEG average FFT index for dyslexic children with writing disorder. In Proceedings of the 2015 IEEE Student Symposium in Biomedical Engineering & Sciences (ISSBES), IEEE, Shah Alam, Malaysia, 4 November 2015; pp. 118–121. [Google Scholar] [CrossRef]
- Yuyi, Z.; Surui, L.; Lijuan, S.; Zhenxin, L.; Bingchao, D. Motor imagery eeg discrimination using hilbert-huang entropy. Biomed. Res. 2017, 28, 727–733. [Google Scholar]
- Zhou, M.; Tian, C.; Cao, R.; Wang, B.; Niu, Y.; Hu, T.; Guo, H.; Xiang, J. Epileptic Seizure Detection Based on EEG Signals and CNN. Front. Neuroinform. 2018, 12, 95. [Google Scholar] [CrossRef] [Green Version]
- Mishra, S.; Birok, R. Sleep Classification using CNN and RNN on Raw EEG Single-Channel. In Proceedings of the 2020 International Conference on Computational Performance Evaluation (ComPE), IEEE, Shillong, India, 2–4 July 2020; pp. 232–237. [Google Scholar] [CrossRef]
- Hochreiter, S. The Vanishing Gradient Problem During Learning Recurrent Neural Nets and Problem Solutions. Int. J. Uncertain. Fuzziness Knowl.-Based Syst. 1998, 6, 107–116. [Google Scholar] [CrossRef] [Green Version]
- Sun, L.; Jin, B.; Yang, H.; Tong, J.; Liu, C.; Xiong, H. Unsupervised EEG feature extraction based on echo state network. Inf. Sci. 2019, 475, 1–17. [Google Scholar] [CrossRef]
- Jaeger, H. Echo state network. Scholarpedia 2007, 2, 2330. [Google Scholar] [CrossRef]
- Jarvis, S.; Rotter, S.; Egert, U. Extending stability through hierarchical clusters in Echo State Networks. Front. Neuroinform. 2010, 4, 11. [Google Scholar] [CrossRef] [Green Version]
- Wang, L.; Hu, H.; Ai, X.-Y.; Liu, H. Effective electricity energy consumption forecasting using echo state network improved by differential evolution algorithm. Energy 2018, 153, 801–815. [Google Scholar] [CrossRef]
- Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl. Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
- Mei, R.N.S.; Sulaiman, M.H.; Mustaffa, Z.; Daniyal, H. Optimal reactive power dispatch solution by loss minimization using moth-flame optimization technique. Appl. Soft Comput. 2017, 59, 210–222. [Google Scholar] [CrossRef] [Green Version]
- Shehab, M.; Abualigah, L.; Al Hamad, H.; Alabool, H.; Alshinwan, M.; Khasawneh, A.M. Moth–flame optimization algorithm: Variants and applications. Neural Comput. Appl. 2020, 32, 9859–9884. [Google Scholar] [CrossRef]
- Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
- Jaeger, H. Tutorial on Training Recurrent Neural Networks, Covering BPPT, RTRL, EKF and the “Echo State Network” Approach; GMD-Forschungszentrum Informationstechnik: Bonn, Germany, 2002. [Google Scholar]
- Bo, Y.-C.; Zhang, X. Online adaptive dynamic programming based on echo state networks for dissolved oxygen control. Appl. Soft Comput. 2018, 62, 830–839. [Google Scholar] [CrossRef]
- Wang, X.; Jin, Y.; Hao, K. Evolving Local Plasticity Rules for Synergistic Learning in Echo State Networks. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 1363–1374. [Google Scholar] [CrossRef]
- Wyffels, F.; Schrauwen, B.; Stroobandt, D. Stable Output Feedback in Reservoir Computing Using Ridge Regression. In International Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 2008; pp. 808–817. [Google Scholar] [CrossRef]
- Chouikhi, N.; Ammar, B.; Rokbani, N.; Alimi, A.M. PSO-based analysis of Echo State Network parameters for time series forecasting. Appl. Soft Comput. 2017, 55, 211–225. [Google Scholar] [CrossRef]
- Lukoševičius, M.; Jaeger, H. Reservoir computing approaches to recurrent neural network training. Comput. Sci. Rev. 2009, 3, 127–149. [Google Scholar] [CrossRef]
- Jaeger, H.; Maass, W.; Principe, J. Special issue on echo state networks and liquid state machines. Neural Netw. 2007, 20, 287–289. [Google Scholar] [CrossRef]
- Song, Q.; Feng, Z. Effects of connectivity structure of complex echo state network on its prediction performance for nonlinear time series. Neurocomputing 2010, 73, 2177–2185. [Google Scholar] [CrossRef]
- Watts, D.J.; Strogatz, S.H. Collective dynamics of ‘small-world’network. Nature 1998, 393, 440–442. [Google Scholar] [CrossRef] [PubMed]
- Deng, Z.; Zhang, Y. Collective Behavior of a Small-World Recurrent Neural System With Scale-Free Distribution. IEEE Trans. Neural Netw. 2007, 18, 1364–1375. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Barabási, A.-L.; Albert, R. Emergence of scaling in random networks. Science 1999, 286, 509–512. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Jaeger, H.; Haas, H. Harnessing Nonlinearity: Predicting Chaotic Systems and Saving Energy in Wireless Communication. Science 2004, 304, 78–80. [Google Scholar] [CrossRef] [Green Version]
- Strauss, T.; Wustlich, W.; Labahn, R. Design Strategies for Weight Matrices of Echo State Networks. Neural Comput. 2012, 24, 3246–3276. [Google Scholar] [CrossRef]
- Bianchi, F.M.; Livi, L.; Alippi, C. Investigating Echo-State Networks Dynamics by Means of Recurrence Analysis. IEEE Trans. Neural Netw. Learn. Syst. 2016, 29, 427–439. [Google Scholar] [CrossRef] [Green Version]
- Domingos, P. A Few Useful Things to Know about Machine Learning. Commun. ACM 2012, 55, 78–87. [Google Scholar] [CrossRef] [Green Version]
- Andrzejak, R.G.; Lehnertz, K.; Mormann, F.; Rieke, C.; David, P.; Elger, C.E. Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: Dependence on recording region and brain state. Phys. Rev. E 2001, 64, 061907. [Google Scholar] [CrossRef] [Green Version]
- Chauhan, V.K.; Dahiya, K.; Sharma, A. Problem formulations and solvers in linear SVM: A review. Artif. Intell. Rev. 2019, 52, 803–855. [Google Scholar] [CrossRef]
- Raj, S.; Ray, K.C. ECG Signal Analysis Using DCT-Based DOST and PSO Optimized SVM. IEEE Trans. Instrum. Meas. 2017, 66, 470–478. [Google Scholar] [CrossRef]
- Güler, N.F.; Übeyli, E.D.; Güler, I. Recurrent neural networks employing Lyapunov exponents for EEG signals classification. Expert Syst. Appl. 2005, 29, 506–514. [Google Scholar] [CrossRef]
- Guo, L.; Rivero, D.; Dorado, J.; Munteanu, C.R.; Pazos, A. Automatic feature extraction using genetic programming: An application to epileptic EEG classification. Expert Syst. Appl. 2011, 38, 10425–10436. [Google Scholar] [CrossRef]
- Shoeibi, A.; Ghassemi, N.; Alizadehsani, R.; Rouhani, M.; Hosseini-Nejad, H.; Khosravi, A.; Panahiazar, M.; Nahavandi, S. A comprehensive comparison of handcrafted features and convolutional autoencoders for epileptic seizures detection in EEG signals. Expert Syst. Appl. 2020, 163, 113788. [Google Scholar] [CrossRef]
- Tuncer, T.; Dogan, S.; Akbal, E. A novel local senary pattern based epilepsy diagnosis system using EEG signals. Australas. Phys. Eng. Sci. Med. 2019, 42, 939–948. [Google Scholar] [CrossRef]
- Raghu, S.; Sriraam, N.; Hegde, A.S.; Kubben, P.L. A novel approach for classification of epileptic seizures using matrix determinant. Expert Syst. Appl. 2019, 127, 323–341. [Google Scholar] [CrossRef]
- de la O Serna, J.A.; Paternina, M.R.A.; Zamora-Mendez, A.; Tripathy, R.K.; Pachori, R.B. EEG-Rhythm Specific Taylor–Fourier Filter Bank Implemented With O-Splines for the Detection of Epilepsy Using EEG Signals. IEEE Sens. J. 2020, 20, 6542–6551. [Google Scholar] [CrossRef]
- Goldberger, A.; Amaral, L.; Glass, L.; Hausdorff, J.; Ivanov, P.C.; Mark, R.; Mietus, J.; Moody, G.; Peng, C.; Stanley, H. Components of a new research resource for complex physiologic signals, PhysioBank, PhysioToolkit, and Physionet. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
- Shoeb, A.H. Application of Machine Learning to Epileptic Seizure Onset Detection and Treatment. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2009. Available online: https://dspace.mit.edu/handle/1721.1/7582 (accessed on 3 March 2023).
- Li, Y.; Liu, Y.; Cui, W.-G.; Guo, Y.-Z.; Huang, H.; Hu, Z.-Y. Epileptic Seizure Detection in EEG Signals Using a Unified Temporal-Spectral Squeeze-and-Excitation Network. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 782–794. [Google Scholar] [CrossRef]
- Chen, D.; Wan, S.; Xiang, J.; Bao, F.S. A high-performance seizure detection algorithm based on Discrete Wavelet Transform (DWT) and EEG. PLoS ONE 2017, 12, e0173138. [Google Scholar] [CrossRef] [Green Version]
- Wei, Z.; Zou, J.; Zhang, J.; Xu, J. Automatic epileptic EEG detection using convolutional neural network with improvements in time-domain. Biomed. Signal Process. Control. 2019, 53, 101551. [Google Scholar] [CrossRef]
- Hu, X.; Yuan, S.; Xu, F.; Leng, Y.; Yuan, K.; Yuan, Q. Scalp EEG classification using deep Bi-LSTM network for seizure detection. Comput. Biol. Med. 2020, 124, 103919. [Google Scholar] [CrossRef] [PubMed]
- Fergus, P.; Hussain, A.; Hignett, D.; Al-Jumeily, D.; Abdel-Aziz, K.; Hamdan, H. A machine learning system for automated whole-brain seizure detection. Appl. Comput. Inform. 2016, 12, 70–89. [Google Scholar] [CrossRef] [Green Version]
Methods | Feature Extractor | Classifier | Accuracy |
---|---|---|---|
Guler et al. (2005) [37] | Lyapunov Exponents + RNN | Logistic regression | 0.9697 |
Guo et al. (2011) [38] | Genetic Programming | KNN | 0.9350 |
Shoeibi et al. (2018) [39] | CNN | Logistic regression | 0.8870 |
Tuncer et al. (2019) [40] | LSP + NCA | SVM | 0.9650 |
Raghu et al. (2019) [41] | Matrix Determinant | MLP | 0.9652 |
Sema et al. (2020) [42] | 5 BPFs + TFEBE | SVM | 0.9488 |
this paper | ESN feature extractor | SVM | 0.9592 |
this paper | MFO-ESN without FDEF | SVM | 0.9319 |
this paper | MFO-ESN | SVM | 0.9816 |
Methods | Feature Extractors | Classifier | Accuracy |
---|---|---|---|
Guler et al. (2005) [37] | Lyapunov Exponents + RNN | Logistic regression | 0.9697 |
this paper | MFO-ESN | 0.9713 | |
Guo et al. (2011) [38] | Genetic Programming | KNN | 0.9350 |
this paper | MFO-ESN | 0.9760 | |
Raghu et al. (2019) [41] | Matrix Determinant | MLP | 0.9652 |
this paper | MFO-ESN | 0.9793 | |
Shoeibi et al. (2018) [39] | CNN | Logistic regression | 0.8870 |
this paper | MFO-ESN | 0.9426 | |
this paper | ESN feature extractor | SVM | 0.9392 |
this paper | MFO-ESN | 0.9816 |
Predict Class | |||
---|---|---|---|
Seizure Class | Normal Class | ||
Actual class | Seizure class | TP | FN |
Normal class | FP | TN |
Case | Patient-Specific | Cross-Patient | ||||||
---|---|---|---|---|---|---|---|---|
Sen. (%) | Spec. (%) | Acc. (%) | F1 (%) | Sen. (%) | Spec. (%) | Acc. (%) | F1 (%) | |
1 | 95.76 | 97.05 | 97.09 | 97.07 | 94.26 | 93.64 | 95.36 | 94.49 |
2 | 99.46 | 100.00 | 99.51 | 99.75 | 99.03 | 97.54 | 98.65 | 98.09 |
3 | 100.00 | 84.28 | 84.63 | 84.45 | 100.00 | 81.22 | 82.39 | 81.80 |
4 | 94.52 | 97.51 | 98.06 | 97.78 | 94.74 | 95.73 | 97.19 | 96.45 |
5 | 98.36 | 87.65 | 88.48 | 88.06 | 97.12 | 87.99 | 88.46 | 88.22 |
6 | 100.00 | 83.26 | 83.64 | 83.45 | 97.81 | 82.91 | 83.01 | 82.96 |
7 | 99.96 | 100.00 | 99.97 | 99.98 | 98.73 | 99.09 | 99.52 | 99.30 |
8 | 96.25 | 83.44 | 83.52 | 83.48 | 93.59 | 81.98 | 82.86 | 82.42 |
9 | 94.71 | 75.53 | 75.59 | 75.56 | 94.56 | 73.04 | 74.21 | 73.62 |
10 | 93.45 | 100.00 | 99.32 | 99.66 | 94.36 | 99.43 | 99.52 | 99.47 |
11 | 96.25 | 100.00 | 99.45 | 99.72 | 94.69 | 98.88 | 99.01 | 98.94 |
12 | 99.46 | 81.53 | 81.77 | 81.65 | 96.68 | 78.74 | 79.33 | 79.03 |
13 | 97.72 | 99.10 | 98.16 | 98.63 | 95.30 | 98.23 | 98.29 | 98.26 |
14 | 99.75 | 89.49 | 89.64 | 89.56 | 100.00 | 87.19 | 88.03 | 87.61 |
15 | 97.54 | 88.13 | 88.43 | 88.28 | 93.69 | 85.22 | 86.83 | 86.02 |
16 | 96.35 | 98.69 | 99.52 | 99.10 | 96.35 | 97.54 | 99.46 | 98.49 |
17 | 90.13 | 93.24 | 93.63 | 93.43 | 90.13 | 93.91 | 95.42 | 94.66 |
18 | 98.96 | 86.36 | 86.37 | 86.36 | 95.29 | 86.51 | 87.07 | 86.79 |
19 | 100.00 | 99.46 | 99.61 | 99.53 | 96.24 | 98.54 | 98.16 | 98.35 |
20 | 95.80 | 98.36 | 99.10 | 98.73 | 96.72 | 96.76 | 98.15 | 97.45 |
21 | 98.05 | 88.83 | 88.86 | 88.84 | 96.81 | 86.69 | 87.25 | 86.97 |
22 | 96.38 | 99.58 | 99.84 | 99.71 | 100.00 | 98.46 | 99.41 | 98.93 |
23 | 90.25 | 94.66 | 94.67 | 94.66 | 97.16 | 91.59 | 92.21 | 91.90 |
24 | 93.83 | 95.30 | 95.98 | 95.64 | 94.00 | 93.67 | 95.06 | 94.36 |
Ave. | 96.79 | 92.56 | 92.75 | 92.63 | 96.14 | 91.02 | 91.92 | 91.44 |
Methods | Detector Type | Sen. (%) | Spec. (%) | Acc. (%) | F1 (%) |
---|---|---|---|---|---|
MIDS + CNN [47] | Patient-specific | 74.08 | 92.46 | 83.27 | 87.62 |
Data argument + CNN [47] | Patient-specific | 72.11 | 95.89 | 84.00 | 89.55 |
LMD + Bi-LSTM [48] | Patient-specific | 93.61 | 91.85 | 92.66 | 92.25 |
CE-stSENet [45] | Patient-specific | 92.41 | 96.05 | 95.96 | 96.00 |
KNN [49] | Cross-patient | 88.00 | 88.00 | 88.00 | 88.00 |
Dyadic WT + SVM [46] | Cross-patient | 91.71 | 92.89 | 92.30 | 92.59 |
MFO-ESN + SVM | Patient-specific | 96.79 | 92.56 | 92.75 | 92.65 |
MFO-ESN + SVM | Cross-patient | 96.14 | 91.02 | 91.92 | 91.47 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Tang, X.-s.; Jiang, L.; Hao, K.; Wang, T.; Liu, X. A Moth–Flame Optimized Echo State Network and Triplet Feature Extractor for Epilepsy Electro-Encephalography Signals. Mathematics 2023, 11, 1438. https://doi.org/10.3390/math11061438
Tang X-s, Jiang L, Hao K, Wang T, Liu X. A Moth–Flame Optimized Echo State Network and Triplet Feature Extractor for Epilepsy Electro-Encephalography Signals. Mathematics. 2023; 11(6):1438. https://doi.org/10.3390/math11061438
Chicago/Turabian StyleTang, Xue-song, Luchao Jiang, Kuangrong Hao, Tong Wang, and Xiaoyan Liu. 2023. "A Moth–Flame Optimized Echo State Network and Triplet Feature Extractor for Epilepsy Electro-Encephalography Signals" Mathematics 11, no. 6: 1438. https://doi.org/10.3390/math11061438
APA StyleTang, X. -s., Jiang, L., Hao, K., Wang, T., & Liu, X. (2023). A Moth–Flame Optimized Echo State Network and Triplet Feature Extractor for Epilepsy Electro-Encephalography Signals. Mathematics, 11(6), 1438. https://doi.org/10.3390/math11061438