1. Introduction
The demand has been increasing for implementing fast automatic epileptic seizure detection, shortening healing time, raising the success rate of cure, protecting patients from accidents, and saving their lives. As such, automatic epileptic seizure detection has become a hot research topic [
1,
2,
3,
4,
5]. Recently, an incident report of epilepsy, one of the most common neurological disorders, stated that the prevalence is about 1 to 2% among the global population, and is rapidly increasing in the elderly population [
1,
4,
5,
6]. In the past several decades, although new anti-epileptic drugs had been invented, a third of epileptic patients are still suffering from epileptic seizures. Additionally, even though seizures are treated and controlled, self-reported life quality is significantly reduced by the unpredictability and consequences of epilepsy, as the disease causes many anomalies in terms of behavior, awareness, movement, and sensation [
3,
7,
8].
The significant differences between epileptic seizure and healthy states are generally highlighted as the frequency and patterns of neurons, meaning the spatial-temporal patterns of neurons gradually increase from the normal state to epileptic seizure-free state and then to the epileptic seizure state [
7,
9]. The World Health Organization (WHO) stipulates that epilepsy is caused by a group of brain cells with unexpected, uncontrolled electrical discharges, termed ’epileptic seizures’ [
6,
9,
10]. In 1929, Berger first measured the spontaneous electrical activity in the brain using electrodes, with electroencephalographic (EEG) signals being produced. His method is still considered a commonly used technology to monitor activity in the human brain. Afterwards, it was further used to detect epileptic seizure [
9,
10,
11,
12,
13,
14].
In the past and at present, well-trained and experienced neurophysiologists are employed universally to detect epileptic seizures through conventional methods by inspecting EEG signals over a long period [
6,
9,
10,
13,
14,
15,
16]. Practically, to ensure high accuracy, detailed evaluations are essential, such as brain imaging tests like magnetic resonance imaging (MRI), computed tomography (CT), long-term EEG tests, blood tests, and medical history examination [
15,
17,
18,
19,
20,
21,
22,
23,
24]. All of these evaluations are costly, which places a heavy financial burden on patients, especially those living in areas with frequent epileptic seizure breakout and poverty. Unfortunately, these methods have many limitations: for example, they are time-consuming, tedious, subjective, [
9,
10,
13,
14,
15,
16,
25], and face many challenges in terms of inevitable factors on reducing misjudgments [
13]. Computer-aided detection with EEG signals can therefore be considered a promising technology. It can be used to realize the goal of minimizing epileptic seizure activity to help control the disease [
16,
26].
Recently, the signal-processing-based method has been applied to extract features to analyze, recognize, and detect epileptic seizures [
1,
2,
3,
4,
5]. Between 2014 and 2017, many different methods were being used to extract and recognize features to achieve the automatic detection of epileptic seizures, such as transform techniques (e.g., discrete short-time Fourier transform, wavelet transform, generalized S-transform, shearlet and contourlet transforms), nonlinear dynamic analyses (e.g., various entropies and the Lyapunov exponent), and data-dependent time-frequency decomposition approaches (e.g., empirical mode decomposition and local mean decomposition) [
5,
7,
15,
17,
18,
19,
20,
21,
22,
23,
24,
26,
27,
28,
29,
30,
31,
32].
However, although numerous methods have been effectively used to achieve the automatic detection of epileptic seizures, there are still some problems affecting improved recognition accuracy. This fact illustrates the nonlinear and non-stationary problem in EEG signals, which causes many challenges in analyzing the signals with conventional methods [
30,
31,
33]. Due to these reasons, in this study, a novel hypothesis is proposed and applied to the publicly available EEG database of the University of Bonn, Germany (UoB). The results demonstrate the reasonableness and accuracy of our proposed method for detecting epileptic seizures.
The remainder of this paper is structured as follows:
Section 2 outlines the materials and problem statement, and also highlights the main contributions. In
Section 3, the methodology is described in detail, including the novel hypothesis, the Mel-Frequency Cepstrum Coefficients (MFCC) for well-solved nonlinear problems, local mean decomposition (LMD) for well-solved non-stationary problems, and a performance evaluation. The reasonableness and availability of our novel method is illustrated by figures in
Section 4, which also outlines the promising results and some comparisons between this method and other methods. Finally,
Section 5 draws a conclusion and lists probable future research directions.
3. Methodology
Generally, epileptic seizure detection can be formulated into a pattern recognition problem using the signal processing technique. Detection can be dismantled into three parts: feature extraction, selection, and feature recognition, as shown in
Figure 2 [
1,
2,
3,
4,
5,
44]. The idea for this method originated from a novel hypothesis which considers EEG signals as convolution signals and regards itself as the generation mechanism of EEG signals. According to the two problems mentioned before, the LMD algorithm is often used to solve non-stationary problems, while MFCC is employed to solve nonlinear problems in EEG signals.
The information provided in
Figure 2 illustrates the architecture of the method used in this study. Bioelectrical signals were considered the interactions among a great number of “atomic modes”, and the interactions were so complicated that they could not be described by existing mathematical language. Due to these extremely complicated spatial-temporal patterns of neurons, the bioelectrical signals often manifest with non-stationary and nonlinear characteristics, which were handled by LMD and MFCC in this article, respectively.
3.1. Local Mean Decomposition (LMD) for Accurately-Solved Non-Stationary Problems in EEG Signals
LMD is considered a data-dependent time-frequency decomposition approach with superior self-adaptive ability [
5,
26,
29,
35,
45]. Our team previously completed a preliminary study on feature extraction based on LMD [
5]. Following that study, LMD can be regarded as the most effective method based on time-frequency analysis from the highest frequency oscillation to the lowest frequency oscillation for non-stationary and nonlinear signals. Furthermore, the LMD method can be used to decompose EEG signals into a series of product functions (PFs). Then, the stationary statistical features of PFs can be obtained [
26,
45].
Figure 2 illustrates the first step of epileptic seizure detection, by decomposing the original EEG signals into several subcomponent signals to solve the non-stationary problem [
5,
26].
Generally, EEG signals can be demodulated into PFs, and each PF can be considered as the product of the multiplication of an envelope signal by a frequency modulated (FM) signal. The LMD procedure involves iteratively extracting subcomponents from the highest to the lowest frequency oscillations, with low computational complexity and high processing speed [
5,
26].
For an arbitrary signal
, the iterative decomposition can be briefly described as follows [
5,
26]:
Finally, the original signal
can be decomposed into numerous PFs and a residual component
. Furthermore, the original signal
can be reconstructed as follows:
3.2. Novel and Reasonableness Hypothesis for the Accurately-Solved Nonlinear Problem in EEG Signals
In order to accurately solve the nonlinear problem in EEG signals, it is important to understand the work and generation mechanism of EEG signals. Currently, there are no definitive theories or mathematical models to describe the process of EEG production. However, some basic facts have been discovered, which are simplified and described in
Figure 3. EEG signals are first produced by the nerve cells in the brain, as shown in
Figure 3a. Usually, the characteristics of electrical signals are decided by two elements: the structure of brain cells, and motivation signals, as shown in
Figure 3b. Neurons, connected by synapses, can be considered a type of cell that has a specific function of discharging electricity. The purpose of this electricity is to store and transmit information among different neurons. Neurons are often considered as the basic signal-processing unit of the human brain, whose main function is to receive information and transmit it to other cells, as shown in
Figure 3a.
Figure 3b demonstrates the hypothesis on the work and generation mechanism of EEG signals, which can be divided into three parts: motivation signals denoted by
, the working principle of neurons denoted by
, and the functional electrical signal output denoted by
. Clearly, the most important part is the working principle of neurons
, and, generally, each neuron is not isolated; they all connect with each other. Moreover, the connections are denoted by
, and the discharge function can be considered as two types of neurons: summation neuron ∑ and multiplication neuron ∏.
Usually, summation neuron ∑ refers to emphasizing signals according to linear functions, whereas multiplication neuron ∏ generates the new signals or enhances signal energy in accordance with some nonlinear pattern, which results in the EEG signals being nonlinear. Generally, the neuron design is a simple but extremely efficient architecture, which makes it extremely responsive to any minor electrical signals, and microscopically, it is non-stationary.
From the hypothesis on the construction of neurons, as shown in
Figure 3b, functional electrical signal output can be expressed as follows:
where
denote the motivation signals and function electrical signal output, respectively, and
is the working principle of neurons. It should be noted that the summation neuron ∑ was not considered because EEG signals are often non-stationary and nonlinear signals, which are generally caused by multiplication neurons ∏. Therefore, it was reasonable and suitable for the following research work to analyze EEG signals for automated epileptic seizure detection. From the hypothesis, the following were assumed to be true: EEG signals are quite different in different individuals due to differences in the construction of brain cells and the characteristics of EEG signals recorded in different brain activities, which can be discriminated by the features denoted by the coefficient
.
The frequency spectra of EEG signals are always aliasing, and generally cannot be obtained by a series of simple filters. Deconvolution may be an efficient method to separate EEG signals into various components. One frequency segment, extracted from the original EEG signals, can be considered as the input signal of the brain neuron system, and can be calculated as follows:
where
are the integration functions, which mean that the EEG signals from high frequency to low frequency can be regarded as the result of the convolution of two different functions, via the EEG signals being generated by the common effect of the neuronal structures and motivation signals. Therefore, the next step was to convert the convoluted signal into a time-domain linear signal.
We wish to illustrate is our hypothesis, which is the basis of this method. When solving different problems, LMD and MFCC were employed. All the methods used in this paper were built on this hypothesis instead of simply listing several methods, which is also the innovative contribution of this work. Accordingly, we refer to “this method” instead of “this proposed method” to avoid misunderstanding. To the best of our knowledge, this hypothesis can be regarded as a theoretical perspective that merits study.
MFCC was originally designed to study real speech registered by human ears. Since the frequency of speech signals is not linear, it requires the study of a new frequency unit that effectively represents the relationship between the size and the frequency. It must also satisfy the following three conditions: the general physical linear description, a low level of discrimination in the high frequency range, and a high level of discrimination in the low frequency range [
41,
43,
46,
47,
48,
49,
50,
51].
Step 1. The PFs are segmented into
N frames. The EEG signals can be set with a 23.6 s duration to guarantee short-time stationarity inside a frame. Generally, there is an overlap with a duration of one-half of a frame between two adjacent frames [
46,
47,
48,
50,
51]. The hamming window was applied to each frame to obtain windowed frames, and can be calculated as follows:
where
N is the number of points in a frame, and
denotes the parameters of the hamming window (
).
Step 2. The frequency spectrum
of the
i-th frame
can be calculated using the Fast Fourier Transform. The short-time power spectrum
can be calculated and filtered by a Mel-filter bank
, which consists of
M triangular band-pass filters and is located in the Mel-frequency domain in a uniformly spaced manner. The mapping relation from a linear frequency
f to Mel-frequency
can be calculated by the following equation:
where
denote the linear frequency and Mel-frequency, respectively, and
are the parameters (
).
Step 3. Then, the output of the short-time power spectrum
can be calculated via this Mel-filter bank:
Step 4. Finally, the output power
of
m-th can be calculated, and the Mel-Frequency Cepstrum
can be calculated by applying the updated Inverse Discrete Cosine Transform (IDCT) in the Mel-frequency coordinate spectrum, which can be described as follows, mathematically:
where
p is the dimension of MFCC,
denotes the
k-th MFCC, and
p is less than the number
M of Mel-filters.
3.3. Identification Mechanism-Categorical Classifications
Practical experience has demonstrated that the performance of epileptic seizure detection depends not only on feature extraction but also on the quality of features selected from EEG data. This procedure is often hidden in the categorical classification. The correlation among features is a good break point to achieve the reduction of feature dimension [
8,
15,
16,
32,
41,
43,
52,
53,
54]. Recognition, also called classification, can be considered an important research field, whose common and traditional methods mainly include support vector machine (SVM), K-nearest neighbor (KNN), neural network (NN) classifiers, forest classifiers, and decision trees (DTs) [
8,
9,
10,
14,
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40,
41,
42,
43,
44,
45,
46,
47,
48,
49,
50,
51,
52,
53,
54,
55,
56,
57,
58,
59,
60].
In the present work, a series of coefficient features, calculated using the MFCC algorithm, were gathered for classification using a wide range of classifiers. The recognition of EEG signals was perfectly achieved. In order to test and verify the effectiveness and correctness of our proposed hypothesis, multiple classification categories were employed in the current study, such as back-propagation neural network (BPNN), SVM, and KNN.
Back-Propagation Neural Network (BPNN) is a feed-forward neural network that can be considered a category of artificial neural networks (ANNs). Generally, the structure of BPNN contains three layers: an input layer, a hidden layer, and an output layer. The steepest descent algorithm is often employed in the learning rule, and the weights and thresholds can be gradually adjusted, so the minimum summation of squared errors can be reached [
5,
26,
32].
Support Vector Machine (SVM) is a potential methodology used for solving the classification problem. It is not only regarded as a linear classifier, but also as a nonlinear classifier. Usually, SVM can be employed to minimize the operational error and maximize the margin hyperplane to maximize the classification performance [
5,
7,
9,
10,
14,
25,
27,
35].
Linear Discriminant Analysis (LDA) has been successfully used in multidimensional reduction and pattern recognition [
26,
41]. LDA can be defined as classifying the dependence by separating the
m-dimensional descriptor space into two areas that are separated by a hyperplane. Concerning classification issues, LDA can be used to minimize the distances among vectors of the same class and maximize the distance of inconsistent class centers [
26,
41].
K-nearest neighbour (KNN) is often used to identify different EEG signals, and can be considered as a supervised and non-parametric machine learning technique, which is relatively simple, has low computational burden, and is easy to implement [
5,
11,
26,
40].
Decision Tree (DT) has been verified to have a high interpretability, and the rules, obtained by splitting the training dataset into subsets by testing attributes, are also readily compressible [
39].
In this paper, the recognition accuracy is calculated and simplified first as a summation and then as the average value, which can be described as follows:
where
represents the classification result of the
i-th classifier, and
N denotes the number of classifiers, and
denotes the recognition accuracy.
The goal of simultaneously using different types of classifiers involves two points: to attempt to mix several classifiers to obtain a stable result, and to verify the super generalization ability of EEG signal analysis. Once all recognition results were calculated for each testing dataset, and the average value for all was calculated, the result was more credible with universal significance.
3.4. Performance Evaluation-Confusion Matrix Metrics
Confusion matrix metrics are often used to quantitatively evaluate recognition performance. A confusion matrix generally involves the following parameters:
a is the number of healthy EEG records identified correctly,
b is the number of healthy EEG records mistakenly detected as epilepsy,
c is the number of epileptic EEG data mistakenly identified as healthy,
d is the number of epileptic EEG data detected correctly.
The following aggregate metrics can then be calculated based on the above parameters.
Accuracy (AC), the proportion of total number of correct predictions, is .
True positive rate (TP), the proportions of positive cases correctly identified, is .
False positive rate (FP), the proportion of negatives cases classified incorrectly, is .
True negative rate (TN), the proportion of negatives cases classified correctly, is .
False negative rate (FN), the proportion of positives cases classified incorrectly, is .
Precision (P), the proportion of the correctly predicted positive cases, is .
Sensitivity (SEN) is , specificity(SPE) is .
The criteria for performance evaluation are usually employed in biomedical studies, which include three parts: sensitivity (the proportion of the total number of labeled ictal EEGs that are correctly classified), specificity (the proportion of the total number of labeled inter-ictal EEGs that are correctly classified), and classification accuracy (the proportion of the total number of EEG signals that are correctly classified) [
2,
6,
7,
8,
9,
10,
13,
14,
15,
16,
25,
26,
27,
28,
29,
30,
31,
32,
34,
35,
38,
39,
40,
41,
49,
53,
59].
4. Results and Discussion
This section describes the results of the publicly available UoB EEG database to show the effectiveness and correctness of the method.
4.1. Experimental Setup
The majority of experiments used the train-to-test ratio (TTTR) of 7:3 in each EEG channel, meaning that 70% of the dataset was used for training, and the remaining 30% was used for testing. Many researchers have achieved the estimated performance using a 9:1 ratio, which makes the performance predictably high with such a high percentage of datasets for training. To the best of our knowledge, varying the TTTR merits investigation to test the robustness of the computational performance [
6]. Therefore, this experiment set the TTTRs as 9:1, 8:2, 7:3, 6:4, 5:5, 4:6, 3:7, 2:8, and 1:9.
This study implemented the algorithm using the MATLAB (R2017a) software (MathWorks, Natick, Massachusetts, USA) and its toolboxes, on a computer with a 3.20 GHz Intel-Core-i5-3470 CPU (Beijing, China), with 12 GB of RAM and running 64-bit Windows7.
In order to eliminate the influence of different time-space dimensions, the data had to be processed by normalization methods. After that, the features were of the same order of magnitude and the data were then suitable for comprehensive comparison and evaluation. This experiment was carried out on EEG signals using the following equation, which is usually called the Z-core normalization method:
where
are the mean and standard deviation of sample EEG segment
, respectively, and
is the normalized
.
4.2. Signal Subcomponents (Product Functions (PFs)) Solving the Non-Stationary Problem
The iterative decomposition algorithm LMD has shown great potential for handling various kinds of signals. However, the number of obtained PFs is not completely consistent when applied to EEG signals. This phenomenon may be caused by the endpoint effect during the decomposition procedure. Therefore, an approach known as the mirror extending technique was applied to address this problem. This technique involves the boundaries of signals at both ends being counted as two symmetrical axes, so that the original segment can be stretched into a triple-length sequence. Finally, the LMD can be subsequently executed on this extended sequence.
Sample EEG signals of these five EEG datasets (datasets A, B, C, D, and E) and their corresponding PFs (PF1 to PF5) are shown in
Figure 4. The subcomponents from top to down are denoted by PF1 to PF5, and PF1 generally contains more high-frequency elements than the others. Considerably more noise is always contained in the highest frequency component, and similarly the lowest component always contains large amounts of useless information. The decomposed PFs of each EEG segment were considered as the high frequency component (PF1) to the low frequency component (PF5). Therefore, the subcomponents from PF1 to PF5, meaning from high-frequency to low-frequency components, were employed to represent the functional information generated by one or several neurons.
4.3. Mel-Frequency Cepstrum Coefficients (MFCC) Solving the Nonlinear Problem
The EEG signals can be deconvoluted into a time-domain signal using the MFCC algorithm; its first series of typical coefficient results are provided in
Figure 5 and
Figure 6. The MFCCs in PF1 from top to down are clearly very different. It is foreseeable that the MFCCs can be fed into some classifiers, and then the classifiers, containing hundreds of dimensional features, will be built up and can be used to achieve a super recognition result.
Generally, as mentioned before, the first subcomponent PF1 contains more information than the others, and this opinion was verified by the results of five datasets (datasets A, B, C, D, and E), as shown in
Figure 5, and the differences among these five datasets (datasets A, B, C, D, and E) are obvious in MFCC (PF1). The MFCCs of the extracted features are not just these exhibitions, but were numerous (about 144 features from PF1 to PF4). Furthermore, when the datasets were fed into some classifiers, the recognition results (provided in the next subsection in details) were obtained quickly with impressive results. To explaining the causes of this phenomenon, we need to return to our hypothesis. The construction of brain cells is often unique and the motivation signals are large in number. This is consistent with the result obtained from the current biomedical research.
Although there are some cluttered features that cannot be distinguished, the results of these five different types of classifiers demonstrate the efficiency of these features. One possible explanation is that the dimensions of these 144 features can be employed to form ultra-dimensional spatial-temporal space, which is very hard to describe.
Figure 5 and
Figure 6 best describe what we know so far.
4.4. Recognition Result
In this study, a total of 6368 EEG segments were considered, and can be considered as the most difficult recognition problem. Systematic analysis seems highly necessary, and accordingly the experiment was conducted by considering all the classification problems, using the four cases, which completely contained all the possible situations.
In order to evaluate the performance of this method with the four cases,
Table 2 illustrates the classification results with the five classifiers. Based on the results, the highest in Case 1 was 100%, which means that the combination, created with the characteristics as the key features in all the experiments, can be employed to obtain a perfect recognition result.
The results in
Table 2 suggest that Case 1 (healthy versus epileptic seizure) yielded a high accuracy by this method. In order to assess the performance of this proposed method, other cases were also used and classified, as shown in
Table 3,
Table 4,
Table 5 and
Table 6. The results can be used to provide enough evidence to show the efficiency of this method on extracting the useful features.
The results in
Table 3 demonstrate the detection ability for healthy EEG groups versus epileptic seizure groups, whereas
Table 4 illustrates the results of the epileptic seizure-free versus epileptic seizure groups.
Table 5 describes the three classification problem for healthy, epileptic seizure-free, and epileptic seizure EEG segments.
Table 7 tries to explain all the differences between the five EEG datasets. It should be noted that all of the classification problems were solved perfectly.
From the results in
Table 3,
Table 4,
Table 5 and
Table 6, an interesting phenomenon revealed that datasets C and D always contained some healthy EEG components, which lowered the classification accuracy in
Table 3,
Table 5 and
Table 6, and increased the classification accuracy in
Table 4. One possible explanation is that the structures of brain cells have a few common features, which imparts the EEG segment with some commonalities with healthy EEG segments when the motivation signals through these neurons. Similarly, the epileptic EEG segments also have some common features to those of healthy EEG segments, as shown in
Table 3,
Table 5 and
Table 6.
In order to eliminate the influence of the number of training samples on the recognition result, the majority of the results in
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6 were obtained using 7:3 as the TTTR. In order to test the robustness of the computational performance with varying TTTRs,
Table 7 provides the recognition results obtained with the BPNN classifier. As can be seen in
Table 7, the results obtained by this proposed method do not significantly rely on a high TTTR.
4.5. Performance Analysis
As presented in
Table 2,
Table 3,
Table 4,
Table 5,
Table 6,
Table 7,
Table 8, regardless of the case, all the classifiers could be used to provide high classification accuracy. On one hand, the proposed method can be employed to realize a maximal classification accuracy at 100% in Cases 1 and 2. On the other hand, it still provides a high classification accuracy of 99.98% in Case 3, as well as a maximum accuracy of 98.35% in Case 4, as shown in
Table 2. Furthermore, combining all the features from PF1 to PF4 resulted in better performance than the others in all cases, as shown in
Table 7. One plausible explanation is that the information transmission must be distributed to prevent the effects of noise, transmission channel blocking, etc. Another possible explanation is that PF1 contains more significant features than the others, such as healthy versus seizure, seizure-free versus seizure, healthy versus seizure-free versus seizure, etc.
Table 8 presents the confusion matrix corresponding to the four cases.
Generally, the performance of this automated seizure detection method can be analyzed using three criteria:
Detection sensitivity (SEN). This method successfully detected 1562, 1549, 1583, and 1406 of 1592 healthy events, which amount to detection sensitivities of 98.12%, 97.30%, 99.43%, and 88.32%, respectively. The detection sensitivity can be as high as 100% and as low as 99.94%.
False positives (specificity; SPE). This method successfully detected 1589, 1589, 1588, and 1585 of 1592 seizure events, which amount to detection sensitivities of 99.81%, 99.81%, 99.75%, and 99.56%, respectively. A good seizure detector must achieve maximum seizure detection sensitivity while minimizing false detections. Even with high sensitivity, a large amount of false detections can be problematic for the patient and medical staff in real-life applications. In this study, the false positives were as low as 0.00% and as high as 9.36%.
Classification accuracy (CA) is considered as a statistical measure by which the performance of a classification algorithm can be evaluated. It is computed by dividing the number of correctly classified samples by the total number of samples. This method was successfully used to obtain accuracies as high as 99.91% and as low as 93.41%.
Therefore, strong universality has been proven and can be considered as the greatest aspect of this method. Furthermore, the results of these four cases illustrate the robustness of this method, which accordingly has the potential to be applied to clinical cases with a high classification accuracy.
4.6. Comparison with Related State-of-the-Art Work
A comparison between this automatic epileptic seizure detection method and the existing state-of-the-art works from 2014 to 2017 is provided in
Table 9. The previously reported methods [
8,
9,
11,
16,
26] were tested with the identified EEG segment, thus making the comparison more feasible and providing additional perspective. All the recognition results for these methods [
8,
9,
11,
16,
26] were copied from their reports. The recognition results of this method include the averages of the five classifiers by setting the TTTR to 7:3.
Taking dataset A versus dataset E and dataset B versus dataset E as two examples, the maximum recognition accuracy of the presented method was as high as 100%, whereas some others also produced results as high as 100%. From
Table 9, the recognition accuracy obtained in classifying healthy and seizure classes (dataset A versus dataset E and dataset B versus dataset E) was as high as 100%. The seizure-free and seizure classes (dataset C versus dataset E and dataset D versus dataset E) obtained by this proposed method was at least as high as 99.83%, but the existing methods could only reach a maximum accuracy of 99.82%.
When considering all five of the EEG datasets, the classification task problems were datasets A, B, C, and D, versus dataset E; datasets A and B versus dataset C and dataset D versus dataset E; and dataset A versus dataset B versus dataset C versus dataset D versus dataset E. Fortunately, this proposed method can be utilized to achieve high accuracies, of 99.46%, 97.85%, and 97.43%, respectively, and the other methods were not as accurate.
5. Conclusions and Future Works
The goal of this paper was to attempt to solve the nonlinear and non-stationary problems of analyzing EEG signals to achieve recognition of healthy, epileptic seizure-free, and epileptic seizure patients. For this reason, an automated epileptic seizure detection method was constructed. The idea was sourced from a novel and reasonable hypothesis that mimics the generation mechanism of EEG signals to some sense.
From the novel hypothesis, the characteristics of EEG signals depended on the construction of the brain cells and these EEG signals involved the macroscopic performance of a large number of neurons. This means that the EEG signals should be first decomposed into several subcomponents that describe the different states of brain, from low frequency to high frequency, which can be employed to demonstrate the active or non-active areas of neurons. A local mean decomposition algorithm was able to attain this goal, and can be considered as the solution to the non-stationary problem.
The problem then becomes how to separate these subcomponents from active or non-active areas of neurons into the feature vectors, which can be used to achieve EEG segment recognition. The Mel-frequency cepstrum coefficient (MFCC) was used to achieve the deconvolution calculation to separate these various subcomponents, and can be considered as the solution to the nonlinear problem.
In future works, the suitable length of the window overlap to build a new EEG segment should be studied by determining the impacts of the length of the signal acquisition changes on the amount of information. Secondly, we hope to evaluate our proposed method for automatically analyzing long recording EEG signals. Thirdly, we will apply this method to solving the problem of classifying multichannel sleep EEG signals and the recognition of different individual behaviors using EEG signals.
Furthermore, ultrasound and optoacoustics, applied in epileptic seizure detection, can be considered as the recent advances in this field, which have been reported in mainstream journals [
17,
18,
19,
20,
21,
22,
23,
24]. Therefore, these two methods will be the research areas that require most of our attention and practical action.