Next Article in Journal
The Contribution of Visual and Auditory Working Memory and Non-Verbal IQ to Motor Multisensory Processing in Elementary School Children
Previous Article in Journal
Investigating White Matter Abnormalities Associated with Schizophrenia Using Deep Learning Model and Voxel-Based Morphometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convolutional Neural Network with a Topographic Representation Module for EEG-Based Brain—Computer Interfaces

College of Intelligence Science and Technology, National University of Defense Technology, Changsha 410073, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(2), 268; https://doi.org/10.3390/brainsci13020268
Submission received: 9 January 2023 / Revised: 2 February 2023 / Accepted: 3 February 2023 / Published: 5 February 2023
(This article belongs to the Topic Human–Machine Interaction)

Abstract

:
Convolutional neural networks (CNNs) have shown great potential in the field of brain–computer interfaces (BCIs) due to their ability to directly process raw electroencephalogram (EEG) signals without artificial feature extraction. Some CNNs have achieved better classification accuracy than that of traditional methods. Raw EEG signals are usually represented as a two-dimensional (2-D) matrix composed of channels and time points, ignoring the spatial topological information of electrodes. Our goal is to make a CNN that takes raw EEG signals as inputs have the ability to learn spatial topological features and improve its classification performance while basically maintaining its original structure. We propose an EEG topographic representation module (TRM). This module consists of (1) a mapping block from raw EEG signals to a 3-D topographic map and (2) a convolution block from the topographic map to an output with the same size as the input. According to the size of the convolutional kernel used in the convolution block, we design two types of TRMs, namely TRM-(5,5) and TRM-(3,3). We embed the two TRM types into three widely used CNNs (ShallowConvNet, DeepConvNet and EEGNet) and test them on two publicly available datasets (the Emergency Braking During Simulated Driving Dataset (EBDSDD) and the High Gamma Dataset (HGD)). Results show that the classification accuracies of all three CNNs are improved on both datasets after using the TRMs. With TRM-(5,5), the average classification accuracies of DeepConvNet, EEGNet and ShallowConvNet are improved by 6.54%, 1.72% and 2.07% on the EBDSDD and by 6.05%, 3.02% and 5.14% on the HGD, respectively; with TRM-(3,3), they are improved by 7.76%, 1.71% and 2.17% on the EBDSDD and by 7.61%, 5.06% and 6.28% on the HGD, respectively. We improve the classification performance of three CNNs on both datasets through the use of TRMs, indicating that they have the capability to mine spatial topological EEG information. More importantly, since the output of a TRM has the same size as the input, CNNs with raw EEG signals as inputs can use this module without changing their original structures.

1. Introduction

Brain–computer interfaces (BCIs) enable direct communication between humans and machines via electroencephalogram (EEG) signals [1]. EEG signals contain instinctive biometric information from the human brain. Through precise EEG decoding, BCIs can recognize the inner thoughts of users. The EEG applications in ML/DL-based disease, mental workload and sleep stage prediction have also been widely studied [2,3,4,5,6]. They are also based on the classification of EEG signals and are similar to BCIs in implementation and processing methods. In general, EEG decoding consists of five main stages: data collection, signal preprocessing, feature extraction, classification and data analysis [7]. Although these stages are essentially the same in a BCI paradigm, signal preprocessing [8], feature extraction [9] and classification methods [10] typically require substantial expertise and a priori knowledge about the BCI paradigm. Moreover, due to manual processing, some useful information may be excluded from the extracted features, which poses a challenge in the subsequent classification and data analysis steps.
Deep learning has largely alleviated the need for manual feature extraction with the combination of automatic feature extraction and classification. Convolutional neural networks (CNNs), in particular, have achieved great success in many challenging image classification tasks, outperforming approaches that rely on handcrafted features [11,12]. Inspired by the success of deep learning in areas such as computer vision and natural language processing, researchers have introduced it to the EEG decoding field [13]. CNNs are some of the most versatile deep learning methods in BCI research. Among all deep learning-based EEG decoding methods, those related to CNNs account for 53% of the total (43% for CNNs and 10% for hybrid CNNs) [14]. CNNs are typically composed of three structural blocks: convolutional layers, pooling layers and fully connected layers. A convolutional layer is an essential part of a CNN that performs the feature extraction function. A pooling layer provides a downsampling operation that both ensures the learning of more robust features and reduces the number of required computations. Fully connected layers are typically located at the bottom of the network and implement the combination of local features and classifier functions. The architecture of a CNN generally consists of layers arranged in a specific order, with earlier layers learning lower-level features and deeper layers learning higher-level features. Several studies have used CNN models, including light [15,16,17,18,19,20] and deep [17,21,22,23] architectures, as well as other varieties [24,25,26,27,28,29,30], to decode EEG signals. Waytowich et al. introduced a compact CNN for directly performing feature extraction and classification based on raw steady-state visually evoked potential (SSVEP) signals, with an average cross-subject accuracy of approximately 80% on a 12-class dataset [19]. By introducing batch normalization in the input and convolutional layers to cope with the overfitting problem, Liu et al. applied a CNN to the task of P300 signal detection and achieved state-of-the-art recognition performance on both Dataset IIb of BCI competition II and Dataset II of BCI competition III [20]. Tang et al. proposed a CNN model based on spatial–temporal features to classify single-motor-imagery (MI) tasks, and the results showed that compared with traditional methods, CNNs could further improve the classification performance [23]. To address the overfitting problem faced by traditional machine learning methods in EEG-based emotion reading, Li et al. used a hierarchical CNN with differential entropy features from different channels as inputs and achieved superior classification results compared to those of traditional methods [30]. Li et al. proposed MI-Visual Geometry Group (VGG) by modifying the VGG network to enable the effective recognition of spectral images generated by MI-EEG and obtained competitive results on three publicly available datasets [22].
EEG-based CNNs use both raw signals and features generated from raw signals as inputs. In this paper, we focus only on CNNs with raw 2-D EEG signals as inputs. Raw EEG signals refer to EEG data in the time domain, i.e., the C (channels) × TP (time points) matrix. Since deep learning-based CNN models have the ability to learn complex features from data without using handcrafted features and can achieve end-to-end learning, raw EEG signals are the most commonly used input formulations [31]. Based on their classification errors, CNNs simultaneously learn and optimize the feature representations of raw EEG signals. Several competitive CNN models using raw EEG signals as inputs have been proposed [16,17,23,28,32,33,34,35,36,37]. Schirrmeister et al. more systematically investigated the process of performing end-to-end learning from raw signals in EEG decoding using CNNs and designed two widely used network architectures, DeepConvNet and ShallowConvNet [17]. Test results obtained on two different datasets showed that the proposed networks achieved classification performance that was at least as good as that of the best traditional methods. In addition, a visualization of the learned features also showed that the two networks performed effective spatial mapping. Lawhern et al. introduced a compact CNN, EEGNet, which uses depthwise and separable convolutions. Test results obtained on four different types of BCI paradigms showed that EEGNet had better generalization ability while obtaining comparable classification performance to that of other methods [16]. Amin et al. used raw EEG signals without preprocessing or artifact removal as inputs, and the classification performance was significantly improved by fusing multiple CNN models with different architectures [36]. CNNs with raw EEG signals as inputs ignore the spatial intertopology of the electrodes; therefore, most of these networks contain a spatial (depth) convolutional layer to learn the weights of the electrodes, which is equivalent to a compensatory operation for ignoring the spatial topological EEG information. However, this compensatory operation can only learn some spatial topology at the global level.
In this paper, we introduce a topographic representation module (TRM) to address the spatial topological information loss induced by CNNs with raw EEG signals as inputs. Our goal is to make a CNN that takes raw the EEG signal as input have the ability to learn the topological information in EEG and improve its classification performance while basically maintaining its original structure. The TRM consists of (1) a mapping block from the raw EEG signals to a topographic map and (2) a convolutional block for transforming the topographic map into an output with the same size as the input. That is, the size of the EEG signals remains unchanged after passing through our TRM. CNNs that take raw 2-D EEG signals as inputs can use this module without changing their original network structure. Such a design takes advantage of both the spatial topological information of EEG signals and various excellent existing CNN architectures, making the TRM very versatile. The main contributions of this study are as follows:
  • The TRM is designed, and the dimension and size of EEG signals remain unchanged after passing through the TRM.
  • The TRM can be embedded into a CNN using the raw EEG signals as input without any adjustment to the network structure, which enables TRM to use the existing excellent EEG classification network.
  • The classification results show that with TRM, the results of three commonly used CNNs on two public datasets are better than those of the original network, indicating that TRM has the potential to improve CNN classification performance by learning EEG topological information.
The rest of this paper is organized as follows: Section 2 presents the materials and methods, including the utilized datasets, classification algorithms, TRM, implementation details and evaluation metrics. Section 3 presents the experimental results, i.e., the algorithmic performance achieved in terms of different evaluation metrics. Section 4 presents a discussion of the algorithms and the results. Finally, we provide a conclusion.

2. Materials and Methods

2.1. Datasets

The EEG signals in BCIs are generally classified into two types according to the presence or absence of external stimuli: evoked potentials and spontaneous EEG signals [7]. Evoked-potential BCIs have clear external stimuli, and their EEG signals exhibit certain time-locked characteristics. Usually, this type of BCI has a high classification accuracy. Unlike evoked-potential BCIs, spontaneous EEG-based BCIs rely on the subject’s spontaneous brain activity, usually without an external stimulus, and are generally more difficult to train. In this paper, we choose an evoked-potential BCI dataset and a spontaneous EEG-based BCI dataset.
The EBDSDD: The EBDSDD has been described in detail in [38]. Here, we briefly describe it as follows: The experiment was carried out on a simulated platform. There were two vehicles in total. One was controlled by a computer as a leading vehicle, with a speed of 100 km/h, and the other was controlled by the subject, following the leading vehicle in front and remaining less than 20 m away from it. The leading vehicle would decelerate randomly in an emergency every 20–40 s, and remind the following vehicle through the brake light. In order to avoid a collision, the subject needed to brake immediately after seeing the brake light of the vehicle in front flashing. The time when the leading vehicle’s brake light came on was marked as a stimulus signal, and the time when the subject started to press the brake pedal was a response signal. The EBDSDD is an evoked-potential EEG dataset containing 2 types of tasks (emergency braking and normal driving) obtained from 18 subjects, each performing approximately 210 emergency braking (target) trials. EEG signals were recorded using 59 electrodes placed on the scalp sites (international 10–20 system, reference at nose), low-pass filtered at 45 Hz and downsampled to 200 Hz. We refer to [38] for the data processing method. Data from 1300 ms before the subjects’ emergency braking actions to 200 ms after braking are chosen as the target segments. The nontarget segments are intercepted from normal driving EEG signals away from any stimulation and braking behavior for at least 3000 ms. Through a sliding window with a length of 1500 ms and a step length of 500 ms, we obtain the normal driving (nontarget) segments. Baseline correction is performed in a segmentwise manner using the data for the first 100 ms. Each subject has about 6628 normal driving (nontarget) trials. For each subject, we choose the same number of nontarget segments and target segments. Therefore, the number of emergency braking (target) and normal driving (nontarget) trials for each subject is about 210. In addition, electrodes FP1, FP2, AF3 and AF4 are susceptible to interference from the oculomotor potential, so we exclude these and use the remaining 55 electrodes. Therefore, the size of each target and nontarget segment is 55 (channels) × 280 (time points).
The HGD: The HGD has been described in detail in [17]. Here, we give a brief description. The HGD is a spontaneous EEG dataset containing 4 classes of movements (left hand, right hand, feet and rest) obtained from 14 healthy subjects, each possessing approximately 1040 trials of executed movements with lengths of 4 s. The experiment has 13 runs in total, and each run contains 80 movement cues. The order of presentation is pseudo-random, and every 4 trials show all the 4 types of movement cues. Ideally, each type of movement contains 260 trials. The first approximately 880 trials are the training set, and the last approximately 160 trials are the test set. The HGD is a 128-electrode dataset with a sampling rate of 500 Hz. For the data processing method, we refer to [17]. Forty-four electrodes covering the motor cortex (all central electrodes except Cz, which is used as the recording reference electrode) are selected. The EEG signals are filtered using a 4–125 Hz bandpass filter and downsampled to 250 Hz. We adopt the standard trialwise training strategy, using the entire duration of each trial, so the data matrix size for each trial is 44 (channels) × 1000 (time points).

2.2. Classification Algorithms

In this paper, we use three representative and widely used EEG-based CNNs: ShallowConvNet, DeepConvNet and EEGNet. They are CNNs specifically designed for EEG decoding tasks, and they have achieved better performance than traditional methods in many BCI applications.
ShallowConvNet: The design of ShallowConvNet is inspired by the filter bank common spatial pattern (FBCSP). It is similar to the FBCSP in terms of its EEG feature extraction process. ShallowConvNet has a simple architecture, consisting only of a temporal convolutional layer, a spatial convolutional layer, an average pooling layer and a dense classification layer. It has achieved better results than the best traditional methods in many EEG decoding tasks [16,17,21,39,40,41,42]. For more details about ShallowConvNet, please refer to [17].
DeepConvNet: The structure of DeepConvNet is inspired by the successful architectures of deep CNNs in computer vision; it aims to extract a wide range of features without relying on specific feature types. DeepConvNet is designed as a general CNN architecture with the hope of achieving competitive accuracy by using only a small amount of expert knowledge. It consists of four convolutional pooling blocks and a dense classification layer. The first convolutional pooling block is divided into a temporal convolutional layer, a spatial convolutional layer and a max pooling layer. The remaining convolutional pooling blocks consist of only one convolutional layer and one max pooling layer. DeepConvNet has achieved competitive classification accuracy compared to that of traditional methods in many EEG decoding tasks [16,17,21,39,41,42,43]. For more details about DeepConvNet, please refer to [17].
EEGNet: EEGNet is designed to find a single CNN architecture that can be applied to different types of EEG-based BCIs and make the network as compact as possible. The structure of EEGNet consists of 2 convolutional pooling blocks and a classification layer. Block 1 performs a temporal convolution and a depthwise convolution sequentially. Block 2 uses a separable convolution consisting of a depthwise convolution and a pointwise convolution. In the classification layer, the softmax method is used. Due to the use of depthwise and pointwise convolutions and the omission of the dense layer, this design reduces the number of trainable parameters by at least one order of magnitude relative to other CNNs. Related studies have shown that EEGNet has a reasonable structure and excellent performance in different types of BCI paradigms [16,21,39,40,41,42,43,44,45,46,47,48]. For a detailed description of EEGNet, please refer to [16].

2.3. TRM

To make the CNNs with raw EEG signals as inputs more effective in utilizing the spatial topological electrode information and to take advantage of various excellent EEG-based CNNs, we propose the EEG TRM. This module consists of (1) a mapping block from the raw EEG signals to a 3-D topographic map and (2) a convolution block from the 3-D topographic map to the processed output with the same size as the input, as shown in Figure 1.
Mapping block: According to the correspondence between the channels and electrode locations on the scalp, the raw EEG signals with a total size of C × TP are mapped into a 3-D EEG topographic map with a size of H (height) × W (width) × TP (time points). For the correspondence between the electrodes and the 2-D matrix coordinates with a size of H × W, we refer to [49,50] and adjust the size of the matrix according to the electrode distribution. Figure 2 shows the correspondence between the electrode locations and 2-D matrix coordinates for the EBDSDD and the HGD, respectively. The values of the yellow points in each matrix are the potential values of the corresponding electrodes, and the values of the gray points are set to 0. For the reason why the matrix coordinate without electrode mapping is assigned to 0, we refer to the practice of [50], and in the process of designing the TRM, we also found that setting to 0 can achieve better classification results than using interpolation operation. According to the number and distribution of electrodes used, we can adjust the size of the mapped matrix accordingly. The EEG signal at any time can be mapped into a 2-D matrix according to the corresponding relationship between the electrode and the matrix; that is, each matrix is a 2-D EEG topographic map for that moment. After mapping, the potential values of the raw EEG signals remain unchanged in the 2-D matrix, while for those coordinates without corresponding electrodes in the matrix, their values are set to 0. In this way, we transform the raw EEG signals at a certain time into a 2-D topographic map represented by electrode locations (matrix coordinates) and their corresponding potentials. These 2-D topographic maps are combined in temporal order to form a 3-D topographic map.
Convolution block: We perform convolution on the 3-D EEG topographic map to take advantage of its powerful feature learning capability. In addition, to take advantage of the various excellent EEG-based CNN architectures, we make the output of the TRM the same size as the input so that it can be embedded into these networks. Previous studies have shown that the size of the convolution kernel can affect the classification performance of neural networks. According to the sizes of the convolutional kernels used in the convolutional blocks, we design two types of TRMs, namely TRM-(5,5) and TRM-(3,3). In TRM-(5,5), the first layer uses C convolution kernels of size 5 × 5 with a step size of 1. The sizes of the convolution kernels in the subsequent layers are chosen according to the size of the feature map obtained after convolution in the previous layer. If the feature map is larger than 5 × 5, C convolution kernels of size 5 × 5 are used; otherwise, C convolution kernels with the same size as the feature map are used. Similarly, in TRM-(3,3), the first layer uses C convolution kernels of size 3 × 3 with a step size of 1. The sizes of the convolution kernels in the subsequent layers are chosen according to the size of the feature map obtained after convolution in the previous layer. If the feature map is larger than 3 × 3, C convolution kernels of size 3 × 3 are used; otherwise, C convolution kernels with the same size as the feature map are used. Thus, the 3-D EEG topographic map is convolved to produce an output of size C × TP, which is the same size as that of the input. After all convolutional layers, batch normalization is used to adjust the distribution of the data and speed up the training process.
The input and output of the TRM have the same size, that is, the size of the EEG signal remains unchanged after passing through the TRM, so it can be directly embedded in the front end of the CNN without any structural adjustment to the original network. Meanwhile, the internal structure of TRM is composed of mapping, convolution, batch normalization and resizing operations, which are common operations of CNNs, making it possible to conduct joint training with the CNN. Table A1 and Table A2 show the structure of TRM on the EBDSDD and HGD, respectively. The size of the 2-D matrix in the mapping block can be adjusted according to the number and distribution of electrodes used. For example, on the EBDSDD, we use a 7 × 9 matrix, while on the HGD, we use a 7 × 7 matrix. For the commonly used 64-lead electrodes, the size of the mapping matrix is usually not greater than 9 × 9. Therefore, we use the commonly used 3 × 3 and 5 × 5 convolution kernels. We choose C (number of channels) convolution kernels to ensure that the EEG signal can remain the same size after TRM.

2.4. Implementation Details

Figure 2 shows the correspondence between the electrode locations and 2-D matrix coordinates. The upper and lower panels show the correspondence between the electrode locations and matrix coordinates for the EBDSDD and HGD, respectively. Through the mapping method in TRM, the raw EEG signal can be mapped into the corresponding 2-D matrix. Each matrix is a 2-D EEG topographic map for that moment. By arranging these 2-D matrices according to the temporal orders of the EEG signals, a 3-D EEG topographic map is formed, as shown in Figure 1. Both datasets are used to train and test DeepConvNet, ShallowConvNet and EEGNet in their original forms and with TRM-(5,5) or TRM-(3,3) for each subject. Table A1 and Table A2 show the TRM architecture on the EBDSDD and HGD, respectively.
For the division of training set, verification set and test set, we adopt different methods according to the difference between the two datasets. For the EBDSDD, we use a 4-fold cross-validation approach, with 50% of each subject’s data as the training set, 25% as the validation set and the remaining 25% as the test set. For the HGD, the original data for each subject have been divided into a training set and a test set. We keep the test set unchanged and randomly select 80% of the original training set as the new training set and the remaining 20% as the validation set, and each subject’s data are used for 4 training and testing rounds, and the average is taken as the final result. For a fair comparison, we set the same training and testing conditions for all three CNNs and their varieties after using the TRM. For each algorithm, the settings are the same except for the use or nonuse of the TRM. All algorithms are trained and tested on a computer with an NVIDIA 2080 Ti graphics card.
Several aspects of the algorithms are set up as follows:
  • The Adam optimizer is used with a weight decay of 0.001, and the remaining parameters are set to their default values.
  • The cross-entropy loss is used as a criterion.
  • The batch size is set to 32.
  • The number of training epochs is set to 300, and the minimum validation loss strategy is used. The network is trained on the training set and validated on the validation set. After each training, the loss on the validation set is compared with the previous loss, and the model with reduced validation loss is saved. The algorithm model with the lowest loss on the validation set is saved for testing.
  • Both ShallowConvNet and EEGNet are trained and tested on both datasets using the codes officially provided by BrainDecode [17]. On the HGD, we use the code of DeepConvNet provided by BrainDecode, while on the EBDSDD, we adjust DeepConvNet with the settings recommended by [39] since the size of the input data does not meet the minimum length requirement of the original DeepConvNet.

2.5. Evaluation Metrics

We comprehensively compare these algorithms using metrics such as classification accuracy, training loss, validation loss, the number of training epochs leading to the lowest validation loss, and a time consumption analysis. Classification accuracy refers to the ratio of the number of correct classifications to the total number of classifications. The training loss and validation loss are the cross-entropy losses of the algorithms on the training and validation sets, respectively. We also count the number of training epochs that leads to the lowest validation loss for each algorithm with the aim of determining which algorithm converges most easily. For the time consumption analysis, we use the time required for training and validation for 300 epochs with each algorithm.

3. Results

3.1. Classification Accuracy

Table 1 shows the classification accuracies obtained by three CNNs (ShallowConvNet, DeepConvNet and EEGNet) on the EBDSDD with and without the TRM. The results for each subject are the mean value ± standard deviation of 4-fold cross-validation, and the p values in the table are calculated by a two-tailed paired t test. The average classification accuracies of DeepConvNet, DeepConvNet-TRM-(5,5) and DeepConvNet-TRM-(3,3) are 84.21%, 90.75% and 91.97%, respectively; for EEGNet, EEGNet-TRM-(5,5) and EEGNet-TRM-(3,3), the results are 93.39%, 95.11% and 95.10%, respectively; for ShallowConvNet, ShallowConvNet-TRM-(5,5) and ShallowConvNet-TRM-(3,3), the results are 93.21%, 95.28% and 95.38%, respectively. For DeepConvNet, the classification accuracies are improved for all 18 subjects after using TRM-(5,5), with a maximum improvement of 22.60% (Subject VPbba), and the average classification accuracy is improved by 6.54% (p < 0.001). The classification accuracies of 17 subjects are improved after using TRM-(3,3), with a maximum increase of 27.40% (Subject VPbba); only one subject has a slight decrease (Subject VPsaj with −0.46%), and the average classification accuracy is improved by 7.76% (p < 0.001). For EEGNet, after using TRM-(5,5), the classification accuracies of 17 subjects are improved, with a maximum increase of 2.92% (Subject VPbad); only one subject shows a slight decrease (Subject VPgaa with −0.42%), and the average accuracy is improved by 1.72% (p < 0.001). With TRM-(3,3), the classification accuracies of 15 subjects are improved, with a maximum increase of 4.17% (Subject VPbax); three subjects exhibit a decline (Subject VPgaa with −0.85%, Subject VPgab with −0.23% and Subject VPgal with −0.99%), and the average accuracy is improved by 1.71% (p < 0.001). For ShallowConvNet, after using TRM-(5,5), the classification accuracies increase for all 18 subjects with a maximum increase of 5.82% (Subject VPbba), and the average result is improved by 2.07% (p < 0.001); after using TRM-(3,3); the classification accuracies increase for 15 subjects with a maximum increase of 6.17% (Subject VPbba), decrease for 2 subjects (Subject VPbax with −0.44% and Subject VPja with −0.72%), and remain unchanged for 1 subject (Subject VPgab), and the average result is improved by 2.17% (p < 0.001).
Table 2 shows the classification accuracies achieved by the three CNNs on the HGD with and without the TRM. The average classification accuracies of DeepConvNet, DeepConvNet-TRM-(5,5) and DeepConvNet-TRM-(3,3) are 67.09%, 73.14% and 74.70%, respectively; for EEGNet, EEGNet-TRM-(5,5) and EEGNet-TRM-(3,3), the results are 76.48%, 79.50% and 81.54%, respectively; for ShallowConvNet, ShallowConvNet-TRM-(5,5) and ShallowConvNet-TRM-(3,3), the results are 81.83%, 86.97% and 88.11%, respectively. For DeepConvNet, the classification accuracies increase for 11 subjects after using TRM-(5,5), with a maximum increase of 22.33% (Subject 9), and decrease for 3 subjects (Subject 2 with −0.94%, Subject 11 with −2.66% and Subject 12 with −2.03%), and the average accuracy is improved by 6.05% (p < 0.01). After using TRM-(3,3), the classification accuracies increase for 12 subjects, with a maximum increase of 22.49% (Subject 9), and decrease for 2 subjects (Subject 10 with −2.82% and Subject 12 with −1.25%); the average result increases by 7.61%. For EEGNet, 13 subjects exhibit accuracy increases after using TRM-(5,5), with a maximum increase of 10.63% (Subject 6); one subject shows a decrease (Subject 9 with −7.81%), and the average accuracy is improved by 3.02% (p < 0.05). All 14 subjects exhibit classification accuracy increases after using TRM-(3,3), with a maximum increase of 11.72% (Subject 5), and the average classification accuracy is improved by 5.06% (p < 0.001). For ShallowConvNet, after using TRM-(5,5), the classification accuracies increase for 13 subjects with a maximum of 18.44% (Subject 11) and decrease for 1 subject (Subject 9 with −1.87%); the average classification accuracy is improved by 5.14% (p < 0.01). After using TRM-(3,3), the classification accuracies increase for 13 subjects with a maximum of 17.03% (Subject 11) and decrease for 1 subject (Subject 3 with −0.16%), and the average classification accuracy is improved by 6.28% (p < 0.001).

3.2. Training Loss

Figure 3 shows the average training cross-entropy loss curves for each algorithm. Algorithms with “-TRM” use the TRM. The left and right panels show the average training loss curves for 18 subjects on the EBDSDD and 14 subjects on the HGD, respectively. With the TRM, the training loss curves of all three CNNs follow roughly the same trends as their original forms on both datasets. DeepConvNet and its varieties have the fastest declines in their training loss curves, followed by ShallowConvNet, while the curves of EEGNet and its varieties are relatively flat. On the EBDSDD, the losses of all algorithms are relatively small after a period of training, which is related to the fact that this dataset is a two-classification task with a high classification accuracy. We find that the training loss curves are not smooth, especially those of DeepConvNet, ShallowConvNet and their varieties on the HGD, which is related to the use of weight decay. By imposing certain restrictions on the learning weights, the overfitting problem can be mitigated to some extent. The training loss curves of EEGNet and its varieties show relatively smooth downward trends on both datasets, which we believe is related to the fact that it has fewer trainable parameters than the other two algorithms.

3.3. Validation Loss

Figure 4 shows the average validation loss curves obtained by each algorithm. Compared with the training loss curve, the validation curve provides a more accurate reflection of the classification and convergence performance of an algorithm. Similar to what we see in the training loss curves, with either TRM-(5,5) or TRM-(3,3), the validation loss curves of all three CNNs have similar trends to those of their original forms on both datasets. After the first few epochs of training, the validation loss of DeepConvNet is significantly higher than those of the other two algorithms on both datasets, and even with the TRM, its validation loss is still higher than those of the other two methods; this result is consistent with the classification accuracy. On the EBDSDD, EEGNet, EEGNet-TRMs, ShallowConvNet and ShallowConvNet-TRMs all have good validation loss curves, while on the HGD, ShallowConvNet and ShallowConvNet-TRMs are better than EEGNet and EEGNet-TRMs. With the TRM, the validation loss curves of all three algorithms show decreasing trends, indicating that the TRM has the ability to improve classification performance.

3.4. The Number of Training Epochs That Yields the Lowest Validation Loss

Figure 5 shows the average number of training epochs that yields the lowest validation loss for each algorithm. The average number of epochs required on the EBDSDD is overall higher than that needed for the HGD. With TRM-(5,5), the average number of training epochs yielding the lowest validation loss decreases for DeepConvNet (p = 0.25) and ShallowConvNet (p = 0.22) and increases for EEGNet (p = 0.09) on the EBDSDD; on the HGD, the average number decreases for DeepConvNet (p = 0.07) and EEGNet (p = 0.72) and increases for ShallowConvNet (p = 0.06). With TRM-(3,3), the average number of training epochs required for the lowest validation loss increases for DeepConvNet (p = 0.87), EEGNet (p = 0.13) and ShallowConvNet (p = 0.46) on the EBDSDD; on the HGD, the average number decreases for DeepConvNet (p = 0.29) and increases for EEGNet (p = 0.23) and ShallowConvNet (p = 0.02). The p values are calculated by two-tailed paired t tests.

3.5. Time Consumption Analysis

Considering the practicality of BCIs, we need to consider not only the classification accuracy but also the execution time of each algorithm. Due to the addition of the TRM, the execution time of each algorithm definitely increases. Figure 6 shows the average time consumption levels of the algorithms for 300 training and validation epochs on each subject. On both datasets, ShallowConvNet consumes less time than DeepConvNet, and DeepConvNet consumes less time than EEGNet. With TRM-(5,5), the times required by ShallowConvNet, DeepConvNet and EEGNet for conducting training and validation for 300 epochs increase by approximately 86%, 48% and 43% on the EBDSDD and approximately 49%, 32% and 21% on the HGD, respectively. With TRM-(3,3), the times required by ShallowConvNet, DeepConvNet and EEGNet increase by approximately 158%, 89% and 67% on the EBDSDD and approximately 111%, 74% and 41% on the HGD, respectively. The execution time of ShallowConvNet-TRM-(5,5) is comparable to that of DeepConvNet, while that of DeepConvNet-TRM (5, 5) is less than that of EEGNet. ShallowConvNet-TRM-(3,3) has shorter execution times than EEGNet on both datasets, while DeepConvNet-TRM-(3,3) has a greater execution time than EEGNet on the EBDSDD and a smaller time on the HGD. The increased time consumption is relatively acceptable.

4. Discussion

A raw EEG signal is generally represented as a 2-D matrix form with C (channels) × TP (time points). In a CNN with raw EEG signals as inputs, if there is no representation module concerning the spatial topology of electrodes in the network, the EEG signals are processed as tensors similar to 2-D pictures, and the spatial topological information of the electrodes is ignored. A topographic map is a representation of an EEG signal as a 2-D or 3-D image, depending on the spatial topology of the electrodes (their locations on the scalp) [31]. Topographic maps can be constructed using either raw EEG signals [24,51,52] or extracted features [18,49,50,53,54]. If a CNN uses the EEG topographic map constructed from the extracted features as input, its performance depends on the quality of the utilized features, which often requires substantial expertise and a priori knowledge. In studies involving CNNs using topographic maps constructed from raw EEG signals, to the best of our current knowledge, all the networks were designed according to the needs of the associated tasks. In References [49,50,53], the generation process from the international 10–20 system to a 9 × 9 2-D matrix is described. The generation process not only basically preserves the relative position of the electrode on the scalp, but also maps the irregular electrode distribution into a regular 2-D matrix form, which can be seen as a form similar to the image, making it very suitable for processing using CNNs. Given that there are already many excellent CNNs that use raw EEG signals as inputs, it is possible to achieve the purpose of using the spatial topological information of EEG signals by simply adding a module without changing the structure of the original network.
Therefore, we design the TRM. By mapping raw EEG signals into a 3-D topographic map, we make the input contain the spatial topological information of the electrodes. For the points without corresponding electrodes, we adopt the practice used in [50] and directly set them to 0 instead of using interpolation because, in our experiments, we find that interpolation does not lead to performance improvement but instead increases the time consumption. We perform a convolution operation on the 3-D EEG topographic map to transform it into an output with the same size and dimensions as the input. Depending on the size of the convolutional kernels used, two convolutional strategies, TRM-(5,5) and TRM-(3,3), are chosen to analyze the impact of the size of the convolutional kernels on the classification performance of the employed algorithm. Such a design takes advantage of the powerful feature learning capabilities of deep learning while using a variety of existing excellent CNNs in EEG-based BCIs.
We choose three widely used CNNs, namely ShallowConvNet with a shallow structure, DeepConvNet with a deep structure, and the compact EEGNet. We use two datasets derived from different types of BCI paradigms. It is hoped that these practices make our results representative. We are very pleased to find that each CNN utilizing the TRM exhibits a strong similarity to its original network, in terms of both the training loss and the validation loss curves, indicating that the properties of the original CNN are largely preserved with the use of the TRM. Under the same training and test conditions, the validation loss curves of all three CNNs exhibit downward trends after using TRM-(5,5) or TRM-(3,3) on both datasets, which to some extent indicates the ability of the TRM to improve the classification performance of EEG-based CNNs.
As the size of the 3-D topographic map obtained from the raw EEG signals and the size of the utilized convolutional kernel vary, the number of trainable parameters in the TRM changes accordingly. Figure 7 shows the numbers of trainable parameters in the three CNNs with or without the TRM. In this paper, the number of trainable parameters for TRM-(5,5) is 46,860 when used on the EBDSDD and 18,612 on the HGD, and the number of trainable parameters for TRM-(3,3) is 64,130 when used on the EBDSDD and 35,332 when used on the HGD.
ShallowConvNet has a simple structure, consisting of only a temporal convolutional layer, a spatial convolutional layer, a pooling layer and a dense layer, and its number of parameters is relatively moderate, so it performs the fastest among all three CNNs. Even with TRM-(5,5), the time it consumes is still comparable to that of DeepConvNet and significantly less than that of EEGNet. When using TRM-(3,3), it consumes more time than DeepConvNet but still less time than EEGNet. On the EBDSDD, ShallowConvNet and its varieties achieve the best results on 10 out of 18 subjects, with ShallowConvNet-TRM-(5,5) accounting for 6 and ShallowConvNet-TRM-(3,3) accounting for 4. The average classification accuracy of ShallowConvNet is lower than that of EEGNet, while the accuracies of ShallowConvNet-TRM-(5,5) and ShallowConvNet-TRM-(3,3) are higher than that of EEGNet. Considering the classification accuracy and the execution time of the algorithm, we recommend ShallowConvNet-TRM-(5,5). On the HGD, ShallowConvNet and its varieties achieve the highest classification accuracy on 12 out of 14 subjects, with ShallowConvNet-TRM-(5,5) and ShallowConvNet-TRM-(3,3) each accounting for 6 of them, and ShallowConvNet-TRM-(3,3) achieves the highest average classification accuracy. Inspired by FBCSP, ShallowConvNet often has a better classification performance in spontaneous EEG decoding [16,39]. From the validation loss curves yielded on the HGD, we can also find that ShallowConvNet, ShallowConvNet-TRM-(5,5) and ShallowConvNet-TRM-(3,3) have smaller validation losses than the other methods. Thus, ShallowConvNet-TRM-(3,3) is the recommended algorithm to be used on the HGD.
DeepConvNet has a relatively deep structure and a large number of trainable parameters, so it often requires a large number of samples and uses certain skills for training. With fewer samples, it is often prone to overfitting [16,21,39]. The classification results of DeepConvNet are poor on both datasets, and the validation loss curves also indicate that it has a larger validation loss than the other methods. Although the classification accuracy of DeepConvNet is improved after using the TRM, it is still lower than that of ShallowConvNet and EEGNet, which may be related to the small number of samples. The time consumption of DeepConvNet is higher than that of ShallowConvNet. With TRM-(5,5), its time consumption increases but is still lower than that of EEGNet. When TRM-(3,3) is used, its time consumption exceeds that of EEGNet on the EBDSDD and remains lower than that of EEGNet on the HGD. DeepConvNet and its varieties are not recommended when the trainable sample size is small.
By using depthwise and separable convolutions and omitting the dense layer, the number of trainable parameters in EEGNet is at least one order of magnitude smaller than that in the other two algorithms, which greatly alleviates the overfitting problem that often occurs in deep learning [16]. Compared to the other two methods, EEGNet has smoother training loss and validation loss curves, and the same is true when the TRM is used. Although EEGNet has the fewest trainable parameters, it has the highest time consumption among the three CNNs. With the TRM, its trainable parameter quantity is significantly improved. On the EBDSDD, EEGNet and its variants with the TRM achieve the best results in 8 out of 18 subjects, with EEGNet, EEGNet-TRM-(5,5) and EEGNet-TRM-(3,3) accounting for 1, 2 and 5, respectively. Compared with that of EEGNet, the average classification accuracy of EEGNet-TRM-(5,5) is improved by 1.72% (p value < 0.001), which is very valuable. On the HGD, EEGNet and its varieties achieve the best results for 2 out of 14 subjects, and both of them are achieved by EEGNet-TRM-(3,3). The average classification accuracy of EEGNet with TRM-(3,3) is improved by 5.06% (p < 0.001), but it is still lower than that of ShallowConvNet.
We also notice that TRM has some disadvantages. As another EEG topographic representation module is added at the input end of the CNN, the newly generated network will inevitably increase the number of trainable parameters and execution time, as shown in Figure 6 and Figure 7. At the same time, we also found that when TRM was used for cross-subject training and testing, its performance was not significantly improved. We believe that this may be related to the obvious individual differences in EEG signals, and the topological information obtained from training for a single individual is not applicable to other individuals. The successful transfer of the model trained in one subject to other subjects requires further research. By analyzing the TRM structure, we find that a large portion of the time consumption is spent on mapping the raw EEG signals into the 3-D topographic map. If the EEG signals are collected as a 3-D matrix (3-D topographic map), the TRM can omit the mapping process. In this case, the additional time consumption caused by the TRM will be very little.

5. Conclusions

In this paper, we introduce the TRM. This module consists of a mapping block and a convolution block, and its output has the same size as the input. CNNs that take raw EEG signals as inputs can use the TRM without changing their structure. We select three representative CNNs and test them on two different types of datasets. The results show that the classification accuracies of all three CNNs are improved with the TRM. Next, we intend to use the TRM on more CNNs and datasets for further validation.

Author Contributions

Data curation, Y.L. (Yaru Liu); formal analysis, X.L. and Y.Y.; funding acquisition, Y.Y. and Y.L. (Yadong Liu); investigation, X.L.; methodology, X.L., Y.L. (Yaru Liu) and Z.Z.; software, X.L.; visualization, X.L. and Y.L. (Yaru Liu); writing—original draft, X.L.; writing—review and editing, Y.Y., K.L., Y.L. (Yadong Liu) and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (62006239), the joint funds of the National Natural Science Foundation of China (U19A2083), and the Defense Industrial Technology Development Program (JCKY2020550B003).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Emergency Braking During Simulated Driving Dataset is available under http://bnci-horizon-2020.eu/database/data-sets (accessed on 1 February 2023), and the High Gamma Dataset is available under https://gin.g-node.org/robintibor/high-gamma-dataset/src/master/data (accessed on 1 February 2023).

Acknowledgments

We thank the volunteers who participated in this study.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. The TRM architecture used on the EBDSDD.
Table A1. The TRM architecture used on the EBDSDD.
LayerTRM-(5,5)TRM-(3,3)
Filter/(Kernel Size)OutputFilter/(Kernel Size)Output
Input-----(C = 55, TP = 280)-----(C = 55, TP = 280)
Mapping-----(7, 9, 280)-----(7, 9, 280)
Resize-----(280, 1, 7, 9)-----(280, 1, 7, 9)
Conv3D55/(5, 5)(280, 55, 3, 5)55/(3, 3)(280, 55, 5, 7)
Conv3D55/(3, 5)(280, 55, 1, 1)55/(3, 3)(280, 55, 3, 5)
Conv3D-----(280, 55, 1, 1)55/(3, 3)(280, 55, 1, 3)
Conv3D-----(280, 55, 1, 1)55/(1, 3)(280, 55, 1, 1)
BatchNorm-----(280, 55, 1, 1)-----(280, 55, 1, 1)
Resize-----(55, 280)-----(55, 280)
Note: C: channels, TP: time points, “-----”: not applicable.
Table A2. The TRM architecture used on the HGD.
Table A2. The TRM architecture used on the HGD.
LayerTRM-(5,5)TRM-(3,3)
Filter/(Kernel Size)OutputFilter/(Kernel Size)Output
Input-----(C = 44, TP = 1000)-----(C = 44, TP = 1000)
Mapping-----(7, 7, 1000)-----(7, 7, 1000)
Resize-----(1000, 1, 7, 7)-----(1000, 1, 7, 7)
Conv3D44/(5, 5)(1000, 44, 3, 3)44/(3, 3)(1000, 44, 5, 5)
Conv3D44/(3, 3)(1000, 44, 1, 1)44/(3, 3)(1000, 44, 3, 3)
Conv3D-----(1000, 44, 1, 1)44/(3, 3)(1000, 44, 1, 1)
BatchNorm-----(1000, 44, 1, 1)-----(1000, 44, 1, 1)
Resize-----(44, 1000)-----(44, 1000)
Note: C: channels, TP: time points, “-----”: not applicable.

References

  1. Wolpaw, J.R.; Birbaumer, N.; McFarland, D.J.; Pfurtscheller, G.; Vaughan, T.M. Brain-computer interfaces for communication and control. Clin. Neurophysiol. 2002, 113, 767–791. [Google Scholar] [CrossRef]
  2. Islam, M.S.; Hussain, I.; Rahman, M.M.; Park, S.J.; Hossain, M.A. Explainable Artificial Intelligence Model for Stroke Prediction Using EEG Signal. Sensors 2022, 22, 9859. [Google Scholar] [CrossRef]
  3. Hussain, I.; Park, S.J. HealthSOS: Real-Time Health Monitoring System for Stroke Prognostics. IEEE Access 2020, 8, 213574–213586. [Google Scholar] [CrossRef]
  4. Hussain, I.; Park, S.J. Quantitative Evaluation of Task-Induced Neurological Outcome after Stroke. Brain Sci. 2021, 11, 900. [Google Scholar] [CrossRef] [PubMed]
  5. Hussain, I.; Young, S.; Park, S.J. Driving-Induced Neurological Biomarkers in an Advanced Driver-Assistance System. Sensors 2021, 21, 6985. [Google Scholar] [CrossRef]
  6. Hussain, I.; Hossain, M.A.; Jany, R.; Bari, M.A.; Uddin, M.; Kamal, A.R.M.; Ku, Y.; Kim, J.S. Quantitative Evaluation of EEG-Biomarkers for Prediction of Sleep Stages. Sensors 2022, 22, 3079. [Google Scholar] [CrossRef]
  7. Zhang, X.; Yao, L.; Wang, X.; Monaghan, J.; McAlpine, D.; Zhang, Y. A survey on deep learning-based non-invasive brain signals: Recent advances and new frontiers. J. Neural Eng. 2021, 18, 031002. [Google Scholar] [CrossRef]
  8. Bashashati, A.; Fatourechi, M.; Ward, R.K.; Birch, G.E. A survey of signal processing algorithms in brain-computer interfaces based on electrical brain signals. J. Neural Eng. 2007, 4, R32–R57. [Google Scholar] [CrossRef] [PubMed]
  9. Mcfarland, D.J.; Anderson, C.W.; Muller, K.R.; Schlogl, A.; Krusienski, D.J. BCI meeting 2005-workshop on BCI signal processing: Feature extraction and translation. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 135–138. [Google Scholar] [CrossRef]
  10. Lotte, F.; Bougrain, L.; Cichocki, A.; Clerc, M.; Congedo, M.; Rakotomamonjy, A.; Yger, F. A review of classification algorithms for EEG-based brain-computer interfaces: A 10 year update. J. Neural Eng. 2018, 15, 031005. [Google Scholar] [CrossRef] [Green Version]
  11. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.; Mohamed, A.-r.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.; et al. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  13. Roy, Y.; Banville, H.; Albuquerque, I.; Gramfort, A.; Falk, T.H.; Faubert, J. Deep learning-based electroencephalography analysis: A systematic review. J. Neural Eng. 2019, 16, 051001. [Google Scholar] [CrossRef] [PubMed]
  14. Craik, A.; He, Y.; Contreras-Vidal, J.L. Deep learning for electroencephalogram (EEG) classification tasks: A review. J. Neural Eng. 2019, 16, 031001. [Google Scholar] [CrossRef] [PubMed]
  15. Li, F.; He, F.; Wang, F.; Zhang, D.; Xia, Y.; Li, X. A Novel Simplified Convolutional Neural Network Classification Algorithm of Motor Imagery EEG Signals Based on Deep Learning. Appl. Sci. 2020, 10, 1605. [Google Scholar] [CrossRef]
  16. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain-computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  17. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef] [PubMed]
  18. Bashivan, P.; Rish, I.; Yeasin, M.; Codella, N. Learning representations from EEG with deep recurrent-convolutional neural networks. arXiv 2015, arXiv:1511.06448. [Google Scholar]
  19. Waytowich, N.; Lawhern, V.J.; Garcia, J.O.; Cummings, J.; Faller, J.; Sajda, P.; Vettel, J.M. Compact Convolutional Neural Networks for Classification of Asynchronous Steady-state Visual Evoked Potentials. J. Neural Eng. 2018, 15, 66031. [Google Scholar] [CrossRef]
  20. Liu, M.; Wu, W.; Gu, Z.; Yu, Z.; Qi, F.; Li, Y. Deep learning based on Batch Normalization for P300 signal detection. Neurocomputing 2018, 275, 288–297. [Google Scholar] [CrossRef]
  21. Riyad, M.; Khalil, M.; Adib, A. MI-EEGNET: A novel convolutional neural network for motor imagery classification. J. Neurosci. Methods 2021, 353, 109037. [Google Scholar] [CrossRef]
  22. Li, M.A.; Han, J.F.; Duan, L.J. A Novel MI-EEG Imaging with the Location Information of Electrodes. IEEE Access 2020, 8, 3197–3211. [Google Scholar] [CrossRef]
  23. Tang, Z.; Li, C.; Sun, S. Single-trial EEG classification of motor imagery using deep convolutional neural networks. Optik 2017, 130, 11–18. [Google Scholar] [CrossRef]
  24. Liu, T.; Yang, D. A Densely Connected Multi-Branch 3D Convolutional Neural Network for Motor Imagery EEG Decoding. Brain Sci. 2021, 11, 197. [Google Scholar] [CrossRef]
  25. Liu, X.; Shen, Y.; Liu, J.; Yang, J.; Xiong, P.; Lin, F. Parallel Spatial-Temporal Self-Attention CNN-Based Motor Imagery Classification for BCI. Front. Neurosci. 2020, 14, 587520. [Google Scholar] [CrossRef] [PubMed]
  26. Zhu, X.; Li, P.; Li, C.; Yao, D.; Zhang, R.; Xu, P. Separated channel convolutional neural network to realize the training free motor imagery BCI systems. Biomed. Signal Process. Control 2019, 49, 396–403. [Google Scholar] [CrossRef]
  27. Alwasiti, H.; Yusoff, M.Z.; Raza, K. Motor Imagery Classification for Brain Computer Interface Using Deep Metric Learning. IEEE Access 2020, 8, 109949–109963. [Google Scholar] [CrossRef]
  28. Dai, G.; Zhou, J.; Huang, J.; Wang, N. HS-CNN: A CNN with hybrid convolution scale for EEG motor imagery classification. J. Neural Eng. 2020, 17, 016025. [Google Scholar] [CrossRef]
  29. Lee, B.H.; Jeong, J.H.; Lee, S.W. SessionNet: Feature Similarity-Based Weighted Ensemble Learning for Motor Imagery Classification. IEEE Access 2020, 8, 134524–134535. [Google Scholar] [CrossRef]
  30. Li, J.; Zhang, Z.; He, H. Hierarchical Convolutional Neural Networks for EEG-Based Emotion Recognition. Cogn. Comput. 2018, 10, 368–380. [Google Scholar] [CrossRef]
  31. Altaheri, H.; Muhammad, G.; Alsulaiman, M.; Amin, S.U.; Altuwaijri, G.A.; Abdul, W.; Bencherif, M.A.; Faisal, M. Deep learning techniques for classification of electroencephalogram (EEG) motor imagery (MI) signals: A review. Neural Comput. Appl. 2021, 1–42. [Google Scholar] [CrossRef]
  32. Yang, J.; Ma, Z.; Wang, J.; Fu, Y. A Novel Deep Learning Scheme for Motor Imagery EEG Decoding Based on Spatial Representation Fusion. IEEE Access 2020, 8, 202100–202110. [Google Scholar] [CrossRef]
  33. Jeong, J.-H.; Lee, B.-H.; Lee, D.-H.; Yun, Y.-D.; Lee, S.-W. EEG Classification of Forearm Movement Imagery Using a Hierarchical Flow Convolutional Neural Network. IEEE Access 2020, 8, 66941–66950. [Google Scholar] [CrossRef]
  34. Fan, C.C.; Yang, H.; Hou, Z.G.; Ni, Z.L.; Chen, S.; Fang, Z. Bilinear neural network with 3-D attention for brain decoding of motor imagery movements from the human EEG. Cogn. Neurodyn. 2021, 15, 181–189. [Google Scholar] [CrossRef]
  35. Li, D.; Xu, J.; Wang, J.; Fang, X.; Ji, Y. A Multi-Scale Fusion Convolutional Neural Network Based on Attention Mechanism for the Visualization Analysis of EEG Signals Decoding. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 2615–2626. [Google Scholar] [CrossRef]
  36. Amin, S.U.; Alsulaiman, M.; Muhammad, G.; Mekhtiche, M.A.; Shamim Hossain, M. Deep Learning for EEG motor imagery classification based on multi-layer CNNs feature fusion. Future Gener. Comput. Syst. 2019, 101, 542–554. [Google Scholar] [CrossRef]
  37. Wu, H.; Niu, Y.; Li, F.; Li, Y.; Fu, B.; Shi, G.; Dong, M. A Parallel Multiscale Filter Bank Convolutional Neural Networks for Motor Imagery EEG Classification. Front. Neurosci. 2019, 13, 1275. [Google Scholar] [CrossRef] [PubMed]
  38. Haufe, S.; Treder, M.S.; Gugler, M.F.; Sagebaum, M.; Curio, G.; Blankertz, B. EEG potentials predict upcoming emergency brakings during simulated driving. J. Neural Eng. 2011, 8, 056001. [Google Scholar] [CrossRef] [PubMed]
  39. Avilov, O.; Rimbert, S.; Popov, A.; Bougrain, L. Optimizing Motor Intention Detection with Deep Learning: Towards Management of Intraoperative Awareness. IEEE Trans. Biomed. Eng. 2021, 68, 3087–3097. [Google Scholar] [CrossRef]
  40. Xu, L.; Xu, M.; Ma, Z.; Wang, K.; Jung, T.P.; Ming, D. Enhancing transfer performance across datasets for brain-computer interfaces using a combination of alignment strategies and adaptive batch normalization. J. Neural Eng. 2021, 18, 0460e5. [Google Scholar] [CrossRef] [PubMed]
  41. Liu, Z.; Meng, L.; Zhang, X.; Fang, W.; Wu, D. Universal adversarial perturbations for CNN classifiers in EEG-based BCIs. J. Neural Eng. 2021, 18, 0460a4. [Google Scholar] [CrossRef]
  42. Gemein, L.A.W.; Schirrmeister, R.T.; Chrabaszcz, P.; Wilson, D.; Boedecker, J.; Schulze-Bonhage, A.; Hutter, F.; Ball, T. Machine-learning-based diagnostics of EEG pathology. Neuroimage 2020, 220, 117021. [Google Scholar] [CrossRef] [PubMed]
  43. Yang, D.; Liu, Y.; Zhou, Z.; Yu, Y.; Liang, X. Decoding Visual Motions from EEG Using Attention-Based RNN. Appl. Sci. 2020, 10, 5662. [Google Scholar] [CrossRef]
  44. Zhu, Y.; Li, Y.; Lu, J.; Li, P. EEGNet With Ensemble Learning to Improve the Cross-Session Classification of SSVEP Based BCI From Ear-EEG. IEEE Access 2021, 9, 15295–15303. [Google Scholar] [CrossRef]
  45. Shi, R.; Zhao, Y.; Cao, Z.; Liu, C.; Kang, Y.; Zhang, J. Categorizing objects from MEG signals using EEGNet. Cogn. Neurodyn. 2021, 16, 365–377. [Google Scholar] [CrossRef]
  46. Zhao, H.; Yang, Y.; Karlsson, P.; McEwan, A. Can recurrent neural network enhanced EEGNet improve the accuracy of ERP classification task? An exploration and a discussion. Health Technol. 2020, 10, 979–995. [Google Scholar] [CrossRef]
  47. Tsukahara, A.; Anzai, Y.; Tanaka, K.; Uchikawa, Y. A design of EEGNet-based inference processor for pattern recognition of EEG using FPGA. Electron. Commun. Jpn. 2020, 104, 53–64. [Google Scholar] [CrossRef]
  48. Kostas, D.; Rudzicz, F. Thinker invariance: Enabling deep neural networks for BCI across more people. J. Neural Eng. 2020, 17, 056008. [Google Scholar] [CrossRef] [PubMed]
  49. Topic, A.; Russo, M. Emotion recognition based on EEG feature maps through deep learning network. Eng. Sci. Technol. Int. J. 2021, 24, 1442–1454. [Google Scholar] [CrossRef]
  50. Chao, H.; Dong, L.; Liu, Y.; Lu, B. Emotion Recognition from Multiband EEG Signals Using CapsNet. Sensors 2019, 19, 2212. [Google Scholar] [CrossRef]
  51. Zhao, X.; Zhang, H.; Zhu, G.; You, F.; Kuang, S.; Sun, L. A Multi-Branch 3D Convolutional Neural Network for EEG-Based Motor Imagery Classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 2164–2177. [Google Scholar] [CrossRef] [PubMed]
  52. Liao, J.J.; Luo, J.J.; Yang, T.; So, R.Q.Y.; Chua, M.C.H. Effects of local and global spatial patterns in EEG motor-imagery classification using convolutional neural network. Brain Comput. Interfaces 2020, 7, 47–56. [Google Scholar] [CrossRef]
  53. Li, Y.; Huang, J.; Zhou, H.; Zhong, N. Human Emotion Recognition with Electroencephalographic Multidimensional Features by Hybrid Deep Neural Networks. Appl. Sci. 2017, 7, 1060. [Google Scholar] [CrossRef]
  54. Xu, M.; Yao, J.; Zhang, Z.; Li, R.; Yang, B.; Li, C.; Li, J.; Zhang, J. Learning EEG topographical representation for classification via convolutional neural network. Pattern Recognit. 2020, 105, 107390. [Google Scholar] [CrossRef]
Figure 1. Overall visualization of the TRM. C: channels, TP: time points, H: height, W: width, (M × N): kernel size. The parts framed by green and blue dashed lines are the mapping block and the convolution block, respectively. In the 3-D topographic map, the coordinates of the yellow points correspond to the electrode locations of the raw EEG signals, and the values are the potential values of the electrodes, while the gray points have no corresponding electrodes, and their values are set to 0. The outputs and inputs of the TRM possess the same size.
Figure 1. Overall visualization of the TRM. C: channels, TP: time points, H: height, W: width, (M × N): kernel size. The parts framed by green and blue dashed lines are the mapping block and the convolution block, respectively. In the 3-D topographic map, the coordinates of the yellow points correspond to the electrode locations of the raw EEG signals, and the values are the potential values of the electrodes, while the gray points have no corresponding electrodes, and their values are set to 0. The outputs and inputs of the TRM possess the same size.
Brainsci 13 00268 g001
Figure 2. Correspondence between the electrode locations and matrix coordinates. The upper and lower panels show the correspondence between the electrode locations and matrix coordinates for the EBDSDD and HGD, respectively. The values of the yellow points in the matrix are the potential values of the corresponding electrodes, and values of the gray points are set to 0. For the EBDSDD, a total of 55 electrodes are used, and the size of the corresponding matrix is 7 × 9. For the HGD, a total of 44 electrodes covering the motor cortex (all central electrodes except Cz, which is used as the recording reference electrode) are used, and the size of the corresponding matrix is 7 × 7.
Figure 2. Correspondence between the electrode locations and matrix coordinates. The upper and lower panels show the correspondence between the electrode locations and matrix coordinates for the EBDSDD and HGD, respectively. The values of the yellow points in the matrix are the potential values of the corresponding electrodes, and values of the gray points are set to 0. For the EBDSDD, a total of 55 electrodes are used, and the size of the corresponding matrix is 7 × 9. For the HGD, a total of 44 electrodes covering the motor cortex (all central electrodes except Cz, which is used as the recording reference electrode) are used, and the size of the corresponding matrix is 7 × 7.
Brainsci 13 00268 g002
Figure 3. Training cross-entropy loss curves of different algorithms. The left and right panels show the average training loss curves for 18 subjects on the EBDSDD and 14 subjects on the HGD, respectively. Algorithms with “-TRM” use the TRM.
Figure 3. Training cross-entropy loss curves of different algorithms. The left and right panels show the average training loss curves for 18 subjects on the EBDSDD and 14 subjects on the HGD, respectively. Algorithms with “-TRM” use the TRM.
Brainsci 13 00268 g003
Figure 4. Validation cross-entropy loss curves of different algorithms. The left and right panels show the average validation loss curves on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Figure 4. Validation cross-entropy loss curves of different algorithms. The left and right panels show the average validation loss curves on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Brainsci 13 00268 g004
Figure 5. The average number of training epochs yielding the lowest validation loss for each algorithm. The left and right panels show the average numbers of training epochs yielding the lowest validation loss on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Figure 5. The average number of training epochs yielding the lowest validation loss for each algorithm. The left and right panels show the average numbers of training epochs yielding the lowest validation loss on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Brainsci 13 00268 g005
Figure 6. Average time consumption required for training and validation for 300 epochs of algorithms. The left and right panels show the average time consumption on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Figure 6. Average time consumption required for training and validation for 300 epochs of algorithms. The left and right panels show the average time consumption on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Brainsci 13 00268 g006
Figure 7. The numbers of trainable parameters for each algorithm. The left and right panels show the numbers of trainable parameters of the algorithms when used on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Figure 7. The numbers of trainable parameters for each algorithm. The left and right panels show the numbers of trainable parameters of the algorithms when used on the EBDSDD and HGD, respectively. Algorithms with “-TRM” use the TRM.
Brainsci 13 00268 g007
Table 1. Classification accuracies achieved by different algorithms on the EBDSDD.
Table 1. Classification accuracies achieved by different algorithms on the EBDSDD.
SubjectDeepConvNetEEGNetShallowConvNet
OriginalTRM-(5,5) TRM-(3,3)OriginalTRM-(5,5) TRM-(3,3)OriginalTRM-(5,5) TRM-(3,3)
VPae75.82 ± 4.1084.78 ± 4.7882.61 ± 3.2090.22 ± 2.3592.93 ± 4.2194.02 ± 3.2686.96 ± 3.9791.85 ± 2.5991.85 ± 2.08
VPbad88.51 ± 2.7995.27 ± 2.8896.17 ± 1.3595.05 ± 2.1497.97 ± 1.3597.97 ± 1.5493.02 ± 4.7997.75 ± 2.2798.65 ± 0.52
VPbax83.77 ± 2.0991.89 ± 2.9993.64 ± 1.9590.57 ± 2.5292.32 ± 1.5094.74 ± 1.8994.74 ± 1.6095.39 ± 1.8194.30 ± 1.68
VPbba55.82 ± 9.5878.42 ± 3.2483.22 ± 7.5388.70 ± 4.5391.44 ± 1.7291.44 ± 3.6084.93 ± 4.8890.75 ± 1.7291.10 ± 2.37
VPdx81.19 ± 3.7091.09 ± 3.1396.78 ± 1.6991.83 ± 3.7494.55 ± 0.9993.32 ± 2.2093.32 ± 1.4996.29 ± 3.4797.03 ± 2.91
VPgaa94.70 ± 3.8197.67 ± 1.0798.09 ± 1.0798.94 ± 0.4298.52 ± 0.4298.09 ± 1.6097.25 ± 1.4598.52 ± 0.8198.09 ± 0.81
VPgab88.89 ± 2.5193.98 ± 1.6094.21 ± 2.1994.91 ± 1.9396.30 ± 0.7694.68 ± 2.8796.99 ± 1.5897.92 ± 1.1796.99 ± 1.17
VPgac91.52 ± 3.1495.09 ± 0.8995.54 ± 2.1996.88 ± 1.1597.77 ± 0.8997.77 ± 0.5297.32 ± 1.6397.99 ± 1.6997.99 ± 1.12
VPgae87.28 ± 3.6988.82 ± 4.1489.47 ± 3.4390.57 ± 4.3892.98 ± 2.4891.89 ± 4.0186.84 ± 2.7787.28 ± 2.0991.45 ± 3.68
VPgag89.95 ± 4.1996.81 ± 2.7098.04 ± 0.8095.83 ± 2.5898.04 ± 1.3998.04 ± 1.3995.83 ± 1.4797.06 ± 0.8097.79 ± 0.94
VPgah81.45 ± 4.9286.56 ± 4.1689.52 ± 7.9791.13 ± 2.2292.47 ± 3.0492.74 ± 3.0991.13 ± 2.6993.55 ± 2.1591.94 ± 1.39
VPgal80.69 ± 2.9784.90 ± 1.8789.60 ± 2.4993.81 ± 2.7394.06 ± 0.8192.82 ± 2.0494.31 ± 1.2594.80 ± 1.6995.30 ± 1.69
VPgam84.52 ± 3.0590.95 ± 2.9691.43 ± 3.2192.86 ± 2.9694.76 ± 3.1693.81 ± 2.2792.38 ± 3.0195.71 ± 1.2395.48 ± 0.91
VPih81.37 ± 8.4989.86 ± 3.6491.04 ± 3.5793.63 ± 1.9495.05 ± 2.3696.70 ± 1.8195.05 ± 2.8295.52 ± 1.6196.23 ± 0.77
VPii95.91 ± 1.2997.63 ± 1.2997.84 ± 0.5097.63 ± 0.8398.92 ± 0.4399.35 ± 0.8398.28 ± 0.7098.71 ± 0.5098.49 ± 0.83
VPja89.56 ± 3.9191.99 ± 3.2193.45 ± 3.3093.20 ± 2.3895.87 ± 2.4395.87 ± 0.9395.87 ± 0.9396.60 ± 2.0295.15 ± 2.10
VPsaj94.44 ± 2.6295.14 ± 2.6693.98 ± 1.2095.37 ± 0.7696.99 ± 1.5895.60 ± 1.9196.76 ± 0.9397.45 ± 1.1797.69 ± 1.60
VPsal70.43 ± 7.2182.69 ± 5.0380.77 ± 3.4289.90 ± 2.2991.11 ± 1.9893.03 ± 1.9886.78 ± 5.7991.83 ± 2.2991.35 ± 1.36
Average84.2190.7591.9793.3995.1195.1093.2195.2895.38
p value-----4.54 × 10−55.71 × 10−5-----3.90 × 10−71.71 × 10−4-----2.43 × 10−46.69 × 10−4
Note: The p value is calculated by two-tailed paired t tests, where “-----” means not applicable.
Table 2. Classification accuracies achieved by different algorithms on the HGD.
Table 2. Classification accuracies achieved by different algorithms on the HGD.
SubjectDeepConvNetEEGNetShallowConvNet
OriginalTRM-(5,5) TRM-(3,3)OriginalTRM-(5,5) TRM-(3,3)OriginalTRM-(5,5) TRM-(3,3)
S157.81 ± 3.7062.81 ± 3.7763.91 ± 1.5654.84 ± 6.2858.75 ± 3.3156.72 ± 5.8767.34 ± 1.7272.66 ± 1.7277.81 ± 4.22
S270.78 ± 2.1969.84 ± 2.6273.91 ± 1.7271.72 ± 3.4473.13 ± 2.3477.34 ± 2.4179.53 ± 1.5682.97 ± 1.8081.25 ± 2.22
S383.13 ± 7.0990.16 ± 1.4892.03 ± 3.2092.81 ± 1.5795.31 ± 0.8196.72 ± 1.2995.16 ± 1.0795.31 ± 0.8195.00 ± 1.53
S484.06 ± 4.6490.00 ± 2.7587.34 ± 6.5293.44 ± 1.8895.47 ± 1.2997.50 ± 1.3596.09 ± 0.6097.03 ± 1.2997.81 ± 0.81
S562.97 ± 5.1667.66 ± 5.1482.50 ± 7.1270.16 ± 8.5372.97 ± 4.5581.88 ± 8.5281.25 ± 4.8990.16 ± 1.2989.06 ± 2.31
S657.97 ± 2.1970.63 ± 5.6665.47 ± 4.1974.84 ± 3.8385.47 ± 1.7278.28 ± 9.3986.09 ± 1.3991.88 ± 0.8892.03 ± 1.56
S759.69 ± 3.4867.66 ± 4.6968.28 ± 1.9365.31 ± 6.0569.53 ± 4.0071.41 ± 5.2675.16 ± 0.9481.09 ± 5.7484.38 ± 4.27
S872.66 ± 2.0075.47 ± 2.5774.69 ± 3.3375.16 ± 1.3982.81 ± 5.9283.75 ± 5.4781.25 ± 3.7887.34 ± 3.6992.19 ± 2.58
S947.51 ± 7.9869.84 ± 4.0070.00 ± 3.9976.56 ± 8.5068.75 ± 6.8884.84 ± 8.0778.75 ± 2.1076.88 ± 2.9885.63 ± 2.93
S1078.91 ± 1.3984.22 ± 0.7976.09 ± 3.1284.22 ± 3.8087.34 ± 2.3086.88 ± 3.0285.47 ± 2.9989.69 ± 0.8189.22 ± 0.31
S1162.66 ± 2.9960.00 ± 5.7168.28 ± 6.3873.44 ± 6.3075.66 ± 7.4579.38 ± 8.1278.28 ± 3.9796.72 ± 1.0795.31 ± 1.80
S1280.94 ± 4.6978.91 ± 0.6079.69 ± 4.1382.19 ± 5.2785.16 ± 2.8686.88 ± 3.8588.59 ± 2.8692.03 ± 1.6491.09 ± 0.31
S1359.38 ± 4.5965.63 ± 3.3567.81 ± 3.1775.78 ± 1.1880.94 ± 4.7577.81 ± 6.3079.06 ± 3.7787.66 ± 1.5685.47 ± 3.48
S1460.78 ± 5.9671.09 ± 5.0475.78 ± 7.1580.31 ± 4.1081.72 ± 8.8382.19 ± 6.1373.59 ± 4.9676.09 ± 3.4877.34 ± 2.72
Average67.0973.1474.7076.4879.5081.5481.8386.9788.11
p value-----3.73 × 10−31.75 × 10−3-----0.0151.95 × 10−5-----1.72 × 10−32.04 × 10−4
Note: The p value is calculated by two-tailed paired t tests, where “-----” means not applicable.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liang, X.; Liu, Y.; Yu, Y.; Liu, K.; Liu, Y.; Zhou, Z. Convolutional Neural Network with a Topographic Representation Module for EEG-Based Brain—Computer Interfaces. Brain Sci. 2023, 13, 268. https://doi.org/10.3390/brainsci13020268

AMA Style

Liang X, Liu Y, Yu Y, Liu K, Liu Y, Zhou Z. Convolutional Neural Network with a Topographic Representation Module for EEG-Based Brain—Computer Interfaces. Brain Sciences. 2023; 13(2):268. https://doi.org/10.3390/brainsci13020268

Chicago/Turabian Style

Liang, Xinbin, Yaru Liu, Yang Yu, Kaixuan Liu, Yadong Liu, and Zongtan Zhou. 2023. "Convolutional Neural Network with a Topographic Representation Module for EEG-Based Brain—Computer Interfaces" Brain Sciences 13, no. 2: 268. https://doi.org/10.3390/brainsci13020268

APA Style

Liang, X., Liu, Y., Yu, Y., Liu, K., Liu, Y., & Zhou, Z. (2023). Convolutional Neural Network with a Topographic Representation Module for EEG-Based Brain—Computer Interfaces. Brain Sciences, 13(2), 268. https://doi.org/10.3390/brainsci13020268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop