Next Article in Journal
Data Processing Approaches to Measure Velocity of Electromagnetic Gun on Laser Screen in Complex Environment
Next Article in Special Issue
Comparison of QEEG Findings before and after Onset of Post-COVID-19 Brain Fog Symptoms
Previous Article in Journal
Unsupervised Domain Adaptive Corner Detection in Vehicle Plate Images
Previous Article in Special Issue
Dynamic Connectivity Analysis Using Adaptive Window Size
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distribution Adaptation and Classification Framework Based on Multiple Kernel Learning for Motor Imagery BCI Illiteracy

School of Instrumentation Science and Engineering, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(17), 6572; https://doi.org/10.3390/s22176572
Submission received: 4 August 2022 / Revised: 27 August 2022 / Accepted: 28 August 2022 / Published: 31 August 2022

Abstract

:
A brain-computer interface (BCI) translates a user’s thoughts such as motor imagery (MI) into the control of external devices. However, some people, who are defined as BCI illiteracy, cannot control BCI effectively. The main characteristics of BCI illiterate subjects are low classification rates and poor repeatability. To address the problem of MI-BCI illiteracy, we propose a distribution adaptation method based on multi-kernel learning to make the distribution of features between the source domain and target domain become even closer to each other, while the divisibility of categories is maximized. Inspired by the kernel trick, we adopted a multiple-kernel-based extreme learning machine to train the labeled source-domain data to find a new high-dimensional subspace that maximizes data divisibility, and then use multiple-kernel-based maximum mean discrepancy to conduct distribution adaptation to eliminate the difference in feature distribution between domains in the new subspace. In light of the high dimension of features of MI-BCI illiteracy, random forest, which can effectively handle high-dimensional features without additional cross-validation, was employed as a classifier. The proposed method was validated on an open dataset. The experimental results show that that the method we proposed suits MI-BCI illiteracy and can reduce the inter-domain differences, resulting in a reduction in the performance degradation of both cross-subjects and cross-sessions.

1. Introduction

A brain–computer interface (BCI) based on electroencephalography (EEG) enables a user to control external devices by decoding brain activities that reflect the user’s thoughts [1]. For example, a user’s motor imagery (MI) can be translated into external device control by an MI-BCI. Some subjects cannot effectively control BCI equipment, meaning that they achieve a classification accuracy of less than 70%; such subjects are referred to as BCI illiterate [2,3]. Poor repeatability is obvious with MI-BCI, which can elicit SMR underpinned by neurophysiological processes [4,5]. As shown in Figure 1, the power spectral density (PSD) of Subject 46 was quite different in each of the two sessions. However, it is generally assumed that the training samples and test samples followed the same statistical distribution when a BCI system is based on machine learning. Domain adaptation (DA), as it pertains to transfer learning, has proven to be an effective method to handle inter-domain shift [6]. To make the distribution of features between the source domain and target domain become even closer to each other, we need to adapt both the marginal distribution and the conditional distribution. Furthermore, MI-BCI illiterate subjects do not display typical brain events such as event-related desynchronization (ERD) and event-related synchronization (ERS) [7]. Their divisibility of features was low, as is shown in Figure 2. Therefore, when studying the shared model of the source domain and target domain, the maximum divisibility of features and the impact of a low classification rate should be taken into consideration alongside the inter-domain shift.
The goal of feature-based marginal distribution adaptation (MDA) methods is to have a common feature space in which the marginal distribution of the source domain and the target domain are as close as possible. Certain achievements have been made in the adaptation of marginal distribution in many fields including EEG signal processing with these methods. Liu et al. [9] applied transfer component analysis (TCA) in EEG-based cross-subject mental fatigue recognition. Zhang et al. [10] reduced distribution differences through an inter-domain scatter matrix based on cross-subject mental workload classification. Chai et al. [11] proposed the use of subspace alignment (SA) to transform features into a domain-invariant subspace to solve the adaptation problem of EEG-based emotion recognition. He et al. [12] applied correlation alignment (CORAL) to minimize the spatial offset when solving the problem of different set domain adaptation for BCIs. Hua et al. [13] applied geodesic flow kernel (GFK) in EEG-based cross-subject emotion recognition. With the development of transfer learning, new progress has also been made in MDA. Wei et al. [14] applied the linear weighting method to the four frequently adopted DA methods (TCA, manifold alignment, CORAL, and SA) to determine coefficients through repeated iterations using the principle of neighborhood consistency. Ma et al. [15] identified the center particle position between the two domains and aligned the center position of the source domain and the target domain by translation. A common problem of these approaches is that although they reduce the marginal distribution differences between the source domain and the target domain in the new subspace, the data categories remain indistinguishable, as displayed in Figure 3a. Considering the difficulties in classifying the features extracted by the MI-BCI illiterate subjects, the above works were probably not optimal for BCI illiteracy. We need to find a new subspace in which the divisibility of categories is maximized and the difference between domains is minimized, as displayed in Figure 3b.
Inspired by the kernel trick that data can be mapped to high-dimensional space to increase data divisibility, the kernel method provides a powerful prediction framework for learning nonlinear prediction models. Therefore, we attempted to map features to the Reproducing Kernel Hilbert Space (RKHS) to find latent features of the subjects, especially BCI illiterate subjects, in this multi-dimensional nonlinear space to improve the class divisibility. Meanwhile, to address the problem that a single kernel has relatively more limitations for wide feature distribution, instead, we applied the linear combination of a series of basic kernels. The combined kernel function could still satisfy the Mercer condition, that is, the function satisfies the symmetry and positive definiteness [16,17]. Multiple kernel based maximum mean discrepancy (MK-MMD) put forward in [18] maps both the source domain and target domain data to multiple-kernel-based RKHS and then minimizes the center distance to reduce the marginal distribution difference. Following this idea, we adopted multiple kernel learning (MKL) combined with MK-MMD to build a DA framework. DA combined with MKL was then addressed using the classifier-based DA method in many studies [19,20,21,22,23]. The method involved adding the objective function that minimized the distance between domains in the mapped feature space to the risk function with a kernel-based classifier and applied a weight parameter λ to balance the data distribution differences between the two domains and structural risk functions. This method demonstrated improved classification and generalization capabilities. However, Chen et al. [24] pointed out that the results of minimizing the risk function with the above methods depend on the parameter λ, which may sometimes sacrifice domain similarity to achieve a high classification accuracy for the source only. Zhang et al. [25] put forward a marginal distribution adaptive framework for kernel-based learning machines. This framework first maps the original features to RKHS to improve the divisibility of categories and then transfers the original data to the target domain through linear operators in the result space to make the processed data become close to the covariance of the target data. Inspired by this method, we proposed a distribution adaptation framework based on multiple kernel. Specifically, we used the source domain data to train the multiple-kernel extreme learning machine (MK-ELM) [26] and found that the multiple-kernel induced RKHS, which can maximize the divisibility of source domain feature categories. We then applied MK-MMD to align the source domain and the target domain under this result space. Features after transformation can retain the information of the original data as much as possible [27,28,29]. Therefore, the proposed method can achieve the maximal divisibility of categories and minimal shift between domains.
It is necessary to retain as much information as possible for MI-BCI illiteracy during feature extraction, so the dimension of features will be relatively high. In light of this point, we applied random forest (RF) as the classifier, which can be used without dimensionality reduction and cross validation. RF has been widely used in the field of BCI and has achieved good results [30,31,32].
Considering MI-BCI illiteracy and referring to the existing technology, we proposed a framework combined distribution adaptation with RF based on multiple kernel learning (MK-DA-RF). We then verified this framework using an open dataset containing BCI illiterate subjects [8]. The main contributions of this study are as follows.
  • The source domain data were applied to train kernel-based ELM to find a subspace that could achieve the best classification effect, that is, the separability of features was the best in this new subspace;
  • To overcome the limitations of a single kernel, a linear connection framework using multiple basic kernels was proposed;
  • MK-MMD was applied to align the distribution of the mapped source and target domain data in this subspace.
The rest of this paper is organized as follows. Section 2 introduces the related work and methods used in this study. The experimental results and discussion are provided in Section 3 and Section 4, respectively. Section 5 presents the conclusions of the study.

2. Methodology

Our proposed distribution adaptation and classification framework based on multi-kernel learning is displayed in Figure 4. The extracted features of the EEG-based BCI system used for the training classifier were defined as source domain features, and the features used for testing were defined as target domain features. The source domain features were used to train the multiple-kernel ELM and the weights were then determined. Next, the kernel function that allows for the maximal inter-class divisibility after the data are mapped to the new RKHS is obtained. Then, the features of the source domain and the target domain were aligned based on MK-MMD under the new RKHS. Then, we applied the adapted training features to train the RF to obtain a suitable classifier.
In this study, X S ,   T S is the labeled source domain data, where X S R D × N S is the source domain data with D as the data dimension, N S is the number of data points in the source domain, and T S R 1 × N S is the corresponding label; X T is defined as the unlabeled target domain data, where X T R D × N T is the source domain data with N T as the number of data points in the source domain without available labels; and Class C is contained in both the source domain and target domain.

2.1. Distribution Alignment Based on Multiple Kernel

The goal of DA is to have equal probability densities of distributions in the new subspace between the source and target domains, in other words, P φ X S φ X T .

2.1.1. Multiple Kernel Expression

To overcome the limitations of a single kernel, a linear connection framework using multiple basic kernels is proposed. The mathematical expression is defined as follows:
φ , γ = p = 1 m γ p φ p ,     p = 1 ,   2 , , k  
K , ; γ = p = 1 m γ p k p , ,   p = 1 , 2 , , k
where φ p refers to the basic mapping function; k p , is the corresponding kernel function; and γ p is the coefficient.

2.1.2. Multiple-Kernel Extreme Learning Machine

Assume that there are N training samples, in other words, X ,   T = x i ,   t i ,   i = 1 ,   2 ,   ,   N . According to the research by Huang et al. [33], the output of kernel based ELM for binary classification is:
y = s i g n h x β
where β refers to connection weight, and h x refers to the feature mapping function.
The learning objective of ELM is to minimize the learning error and weight coefficient, which can be expressed as:
Minimize :   L P E L M = 1 2 β 2 + C 1 2 i = 1 N || ξ i || 2
Subject   to : h ( x i ) β = t i ξ i , i = 1 , 2 , , N
where ξ i is the training error and C is a parameter set by the user and provides a tradeoff between the output weights and training error.
Based on the Karush–Kuhn–Tucker (KKT) theory and Bartlett’s theory [34], the Lagrangian function can be written as follows:
L D E L M = 1 2 || β || 2 + C 1 2 i = 1 N || ξ i || 2 i = 1 N α i h ( x i β t i + ξ i )
According to the solution method of the KKT and Mercer’s theorem, we set the derivative of (4) with regard to the parameters, which can be expressed as:
L D E L M β j   =   0 β j   =   i = 1 N α i , j h x i T β = H T α
L D E L M ξ i = 0 α i = C ξ i ,   i = 1 , 2 , , N
L D E L M α i = 0 h ( x i ) β t i T + ξ i T = 0 ,   i = 1 , 2 , , N
Subsequently, by substituting Equation (5a,b) into Equation (5c), it can thus be inferred that
I C + H H T α = T
By substituting Equation (6) in Equation (5a),
β   =   H T ( I C + H H T ) 1 T ,
Combining (2) with (7), the relationship between the input and output of kernel-based ELM can be expressed as:
f x =   K x , x 1 K x , x N ( I C + Ω ELM ) 1 T
where:
Ω E L M = H H T :   Ω E L M i , j   =   K x i , x j =   h x i h x j
On this basis, the single-kernel linear combination is replaced by a multiple-kernel linear combination. The target function of MK-ELM is gained combined with (1) [26]. It can be expressed as:
min γ min β , ξ 1 2 || β || F 2 + C 2 i = 1 n || ξ i || 2
s . t . β T φ x i ; γ = t i     ξ i ,   i
p = 1 m γ p q = 1 ,   γ p     0 ,   p
Herein, q = 2, β ˜ = β ˜ 1 ,   β ˜ 2 , ,   β ˜ m , and β ˜ p = γ p β p ,   p = 1 ,   2 ,   ,   m . The Lagrangian function is:
L β ˜ , ξ , γ = 1 2 p = 1 m || β ˜ p || F 2 γ p + C 2 i = 1 n || ξ i || 2 t = 1 T i = 1 n α i t p = 1 m β ˜ p T φ p x i     t t i + ξ t i + τ ( p = 1 m γ p 2     1 )
According to the KKT theory, it can be concluded that:
|| β ˜ p || F = γ p s , t = 1 T i , j = 1 n α i t α j s K p x i , x j
By taking the derivative of (10) with respect to γ p , we obtain:
1 2 || β ˜ p || F 2 γ p 2 + q τ γ p q     1 = 0 ,   p = 1 , m
Combining (11) with p = 1 m γ p q = 1 , we obtain:
γ p = || β ˜ p || F 2 / 1 + q ( p = 1 m || β ˜ p || F 2 q / 1 + q ) 1 / q ,   p
By gradual iteration coefficient, γ n e w is updated and the optimal coefficient is obtained.

2.1.3. Multiple Kernel Maximum Mean Discrepancy

It is assumed that P X S     P X T ; however, there is a mapping φ so that P φ X S = P φ X T . The MMD put forward in the research by Pan et al. [27] is often used as an indicator for calculating the distribution distance in the RKHS:
D i s t K X s ,   X T = 1 n S i = 1 n S φ x S i     1 n T i = 1 n T φ x T i H 2
In combination with (1), the multiple basic mapping functions can be regarded as one mapping function after linear combination (i.e., the mapping function is defined as φ = φ , γ = p = 1 m γ p φ p ) and a single kernel based MMD can form MK-MMD in this way.
If X = X S ,   X T R D × n S + n T , then the kernel mapping is φ X = { φ x 1 ,   φ x 2 , ,   φ x n S + n T } and the kernel matrix is K = φ X T φ X . According to the Kernel PCA theory [27,35], the transformed features can be expressed as Z = W T φ x T φ x = W T K , where W is the kernel-PCA transformation matrix. By definition, this vector can retain the maximal mapped feature space information, and (14) is then written as:
D i s t K X s ,   X T = 1 n S i = 1 n S W T K i     1 n T j = n S   + 1 n S + n T W T K j H 2 = t r W T K L K W
where
L i j = 1 n s 2 x i , x j X s 1 n T 2 x i , x j X T 1 n S n T o t h e r w i s e
The maximal mean difference is to be minimized in infinite-dimensional RKHS space. Combining kernel-based PCA, the problem for domain adaptation then reduces to:
m i n W   t r W T K L K W + μ t r W T W
s . t .   W T K H K W = I m
where t r W T W is a regular term that controls the model complexity; μ is a trade-off parameter; I R m × m is the identity matrix; H = I n S + n T 1 n S + n T 11 T is the centering matrix, where 1 R n S   + n T is the column vector with all ones; and I n S + n T   R n S + n T × n S + n T is the identity matrix. Defining A as a symmetric matrix, the Lagrangian of (16) is
L = t r W T μ I + K L K W     t r W T K H K W     I A ,
Setting the derivative of (17) with regard to W to zero, we have
μ I + K L K W = K H K W A
Take the former m eigenvector of μ I + K L K   1 K H K as W. The transformed feature is then expressed as:
Z = W T φ x T φ x = W T K
The process of marginal distribution adaption is presented in Algorithm 1.
Algorithm 1. Marginal Distribution Adaptation:
1: Input: labeled source samples  X s ,   T s   and unlabeled target samples X T ; several basic kernel functions   K p p = 1 m ,   q, and C
2: Output:   Z S , Z T
3: Initialize: γ = γ 0 and t = 0
4: repeat
5:   Compute K , ;   γ 0 by solving (1)
6:   Update || β ˜ p || F 2 by solving (11)
7:   Update γ t + 1 by (13)
8: until max { γ t + 1     γ t }   ε
9: Compute K , ;   γ with the obtained γ by solving (1)
10: Compute the eigenvector of μ I + K L K 1 K H K
11: Take the former m eigenvector as W
12: Compute   Z S and Z T with (19).

2.2. Random Forest

Random forest [36] is a type of ensemble learning, which is to combine several base classifiers to obtain a strong classifier with significantly superior classification performance. The principle of random forest is to obtain the final classification result by voting. The generation process of random forest is shown in Figure 5.
Suppose that there is a training set T consisting of N samples (i.e., T = t i ,   i = 1 ,   2 ,   ,   N ) and the corresponding feature vector F with M dimensions (i.e., F = f j ,   j = 1 ,   2 ,   ,   M ). We applied a random forest with k decision trees, and the training steps are as follows:
  • Resample randomly from the training set based on bootstrap to form a training subset T k ;
  • Randomly extract m features from F of T k without replacement ( m = log 2 M is set in this paper) to generate a complete decision tree S k without pruning;
  • Repeat the above two steps k times to generate k decision trees, and then combine all of the decision trees to form a random forest;
  • Take the test sample as the input of the random forest, and then vote on the result of each decision tree based on majority voting algorithm to obtain the classification result.

3. Results

We validated our method by an open-access dataset, namely the BMI dataset (http://gigadb.org/dataset/view/id/100542/ (accessed on 16 May 2021)). The research was provided by Lee et al. [8].

3.1. Experiment Materials and Preprocessing

The BCI system analyzed was based on Brain Amp, which utilizes 62 Ag/AgCI electrodes [8]. The EEG signals were collected at a frequency of 1000 Hz, and electrodes were placed in accordance with the international 10/20 system standard. Fifty-four subjects participated in this experiment, and none had a history of mental illness or psychoactive drugs that would affect the results of the study.
MI-BCI was tested with a dichotomous experiment in which subjects imagine their left and right hands in accordance with the directions of arrows, as shown in Figure 6. The EEG signals were recorded in two different sessions on different days. For all blocks of a session, black fixation was displayed on the screen for 3 s before each trial task began. The subjects then imagined they were performing the hand-grabbing action (grasping) in the direction specified by the visual cue. After the task, the screen display was blank for 6 s to allow the subjects to rest. There were 200 trials in one experiment per subject, half on the left and half on the right.
To retain the EEG information of the subjects as much as possible, as shown in Figure 7, 20-channel EEG data of the motor cortex region were selected: {FC5, FC3, FC1, FC2, FC4, FC6, C5, C3, C1, Cz, C2, C4, C6, CP5, CP3, CP1, CPz, CP2, CP4, CP6}. The EEG signals were downsampled to 100 Hz, and the 5th order Butterworth digital filter was utilized to obtain 8–30 Hz signals. Then, the range of 500–3500 s after the task started was selected. A common spatial pattern (CSP) was applied to maximize the difference between the two types of tasks. The first five dimensions of the feature vector were selected, and after that, the log-variance feature was calculated. Therefore, the CSP feature of a single trial was 1 × 10.

3.2. Model Generation

We verified the effect of the proposed domain adaption framework that combined MK-ELM and MK-MMD (denoted as MK-DA) on the aforementioned open dataset. First, the following three base kernel functions were chosen to form multiple-kernel ELM:
  • Polynomial kernel function
K x , y = x y + a d , d = 1 , 2 , , N .
  • Gaussian kernel function
K x , y = exp ( x     y 2 2 σ 2 ) ,
  • Translation-invariant of wavelet kernel function
K x , y = h ( x y w a ) h x = c o s w b x e x p ( x 2 w c )
The relaxation coefficient C of the classifier was selected from C 0.001 ,   0.01 ,   0.1 ,   1 ,   10 ,   50 ,   100 . The optimal parameters p a and p d of the poly-kernel function were selected from p a 0.001 ,   0.01 ,   0.1 ,   1 ,   10 ,   50 ,   100 , and p d 2 ,   3 ,   4 , respectively. The optimal parameter of the Gaussian kernel function was determined from σ 0.001 ,   0.01 ,   0.1 ,   1 ,   10 ,   50 ,   100 . The optimal parameters w a , w b , and w c of the wavelet kernel function were searched from w { 0.001 ,   0.01 ,   0.1 ,   1 ,   10 ,   50 ,   100 } , respectively.
Then, random forest was used as the classifier. The number of decision trees was selected from k 10 ,   20 ,   50 .

3.3. Experimental Results

3.3.1. Methods for Comparison

This study primarily addressed the domain shift problem of BCI illiteracy by applying an open dataset containing BCI illiterate subjects [8] for validation. The classic method in which CSP was applied to extract features and linear discriminant analysis (LDA) was applied to classify, which was used as the reference framework.
The performance of the proposed DA method was compared with the performance of the DA methods that are widely used and known to achieve good results. At the same time, RF was employed as the classifier, which was the proposed method in this paper. To ensure fairness in the comparison, we gave the same parameter to all parts using the same operation, and the other parameters were given the optimal value according to the suggestions in the literature. The comparison DA methods were as follows:
  • SA: We set the parameters referring to the research by Xiao et al. [37]. Considering the poor classification effect of BCI illiteracy, we set the subspace dimension of principal component analysis (PCA) to all to avoid information loss;
  • GFK: We referred to the research by Wei et al. [38]. We determined the optimal dimension of the subspace by adopting the subspace disagreement measure (SDM) after the source domain and target domain data were determined;
  • CORAL: Referring to the research by He et al. [12], we conducted a distributed computation on the feature covariance matrix of each domain and then minimized the distance between the covariance matrices of different domains;
  • TCA: We referred to the research by Jayaram et al. [39]. In this experiment, when carrying out a multiple-kernel linear combination, the weight of the Gaussian kernel was generally the largest. Therefore, we chose the Gaussian kernel function and set its parameters to be the same as those of the Gaussian kernel function in MK-ELM;
  • MKL: Referring to the research by Sun et al. [19] and Dai et al. [20], we combined Gaussian-kernel-based support vector machine (SVM) with MKL and applied the classifier-based DA method to optimize the target function of SVM, while minimizing the inter-domain offset based on MKL. MKL uses the three kernels above-mentioned and applied the second-order Newton method recommended by Sun et al. [18] to obtain the combination coefficients. The balance parameter was λ = 0.5. Note that the combined coefficients obtained by this method can be different from those obtained by the method proposed in this paper.
Then, the performance of the proposed classifier was compared with the performance of classifiers that are widely used in MI-BCI. The comparison classifiers were as follows:
  • LDA: The reference method proposed by Lee et al. [8].
  • SVM: We referred to the research by Lotte et al. [40]. We chose the Gaussian kernel function and set its parameters to be the same as those of the Gaussian kernel function in MK-ELM;
  • KNN: We referred to the research by Lotte et al. [40]. We set the number k = 5 .
  • EEGnet: We referred to the research by Lawhern et al. [41]. We set the number of channels as 20.
  • FBCNet: We referred to the research by Mane et al. [42]. We set C as 20.

3.3.2. Performance of the Domain Adaption and Classification Framework

We set the threshold value as 0.05, so p ≤ 0.05 indicates the statistical significance. During the experiment, according to the classification results obtained from the literature and the definition of BCI illiteracy, the subjects were divided into the following two groups:
  • BCI (the classification result was greater than 70% in both sessions), denoted as BNI;
  • BCI illiteracy (the classification result was less than 70% in both sessions), denoted as BI.
We performed experiments from two perspectives (i.e., cross-subject experiment and cross-session experiment). For preciseness, the Kruskal-Wallis test was adopted to display the statistical significance of the differences between methods.
1. 
Results of Cross-Subject Experiments
To ensure the simplicity of the comparison factors, both the source and target domains in this part were of the same session. Based on NBI and BI grouping, we randomly selected one subject as the source domain and another subject in the same session as the target domain in the following two ways. The first method was random sampling limited to NBI, and the second method was random sampling limited to BI. We used the proposed method DA and the control method to align the marginal distribution and employed RF as the classifier. Then, we applied different classifiers to the features adapted by MK-DA for the classification and comparison. The experiment was repeated 30 times, and the average accuracy was taken as the result. The results including the average classification accuracies (mean), standard deviation (Std), and confidence interval under 95% signification level (CI) are shown in Table 1 and Table 2.
2. 
Results of the Cross-Session Experiments
The experiments of the two sessions of each subject were taken as the source domain and the target domain, respectively. The results were divided into the BI group and the NBI group, and the average value of each group was taken as the result of that group. Then, the data of the source domain and the target domain were switched. The experimental verification was conducted in the same way as the cross-subject experiments, and the results including the average classification accuracies (mean), standard deviation (Std), and confidence interval under 95% signification level (CI) are shown in Table 3 and Table 4.

4. Discussion

Based on the above experimental results, the rationality of the proposed method is discussed from two perspectives: cross-subject experiment and cross-session experiment.
  • Cross-Subject Experiments
We randomly selected one subject as the subject of the training field and another subject in the same session as the target domain based on two methods, namely, the first subject was selected from those whose target domain was specified as the NBI group, and the second subject was selected from those whose target domain was specified as the BI group. When applying the proposed MK-ELM for classification, compared with LDA as the reference method, the average classification accuracies were improved, among which the biggest gain was 1.26% and 1.18%, respectively, which proved that the MK-ELM adopted in this paper increased the data divisibility in the new RKHS. Then, the competitive marginal distribution method was applied to the feature distribution adaptation. As can be seen from Figure 8, the classification accuracies of MK-DA-RF improved by 2.65% and 3.72%, respectively, compared with LDA as the reference method and by 0.78% and 0.46% compared with TCA-RF, which was the best-performing control method.
  • Cross-Session Experiments
The two sessions of the same subject were chosen as the source domain and target domain, respectively. From the results of the average classification accuracy of all subjects, the classification accuracies of the proposed MK-DA-RF improved by 6.24% and 5.74%, respectively, compared with LDA as the reference method and by 0.72% and 1.31% compared with TCA-RF, which was the best-performing control method. Then, we averaged the subjects according to the NBI and BI groups. The results of group NBI and group BI are displayed in Figure 9 and Figure 10, respectively. It can be seen that the classification accuracy of MK-ELM gained an average increase of 3.93% in the two tasks compared with the reference method for the BI group, but an average increase of 2.06% for the NBI group. Since the subjects in the BI group could not effectively control the BCI, the extracted features were difficult to distinguish. The divisibility was increased after features were mapped to multiple-kernel-based RHKS by MK-ELM, which was consistent with the experimental phenomenon from the results. Then, under the adjustment of DA, the combined method of MK-ELM and MK-MMD proposed in this study significantly improved the classification accuracy. In particular, in the BI group, the average classification accuracies of MK-DA-RF were 5.9% and 6.3% higher than those of the reference method and 0.62% and 1.34% higher than those of TCA-RF, which was the best-performing control method.
The feature distribution of the same subject before and after MDA was observed, and the first two dimensions of the feature value were taken as the X-axis and Y-axis, respectively, to check the feature distribution. Figure 11 shows the original feature distribution of Subject 19 and the feature distribution obtained with the method proposed in this study. Specifically, Figure 11a,b is the feature space distribution of class 1 (left-hand motor imagery) and class 2 (right-hand motor imagery), respectively. It can be seen that the class divisibility of the feature distribution was increased with MK-DA. Figure 11c,d refers to the feature distribution before and after MK-DA was employed. It can be seen that the feature space of the source domain and the target domain were further closer by the method proposed in this paper.
  • Performance of Random Forest
In this section, we analyzed the performance of different classifiers separately in two experiments. The performance evaluation metrics were calculated referring to the research by Giannakakis et al. [43].
In the cross-subject experiment, as shown in Figure 12, RF achieved relatively better results in all experiments. The classification accuracies of the proposed RF improved by 0.17% and 0.17%, respectively, compared with LDA, which was the best-performing control method. The results of the performance evaluation metrics are shown in Table 5. The confusion matrices and receiver operating characteristic curves are shown in Figure 13 and Figure 14.
In the cross-session experiment, the results of all subjects are displayed in Figure 15. It can be seen that the classification accuracy of RF gained an average increase of 0.24% and 0.58%, respectively, in the two tasks compared with the control method with the best performance. The results of the performance evaluation metrics are shown in Table 6. The confusion matrices and receiver operating characteristic curve are shown in Figure 16 and Figure 17.
In particular, in all experiments, the performance of EEGnet was worse than that of the LDA without domain adaptation, so we believe that it is related to the fact that 20 channels were used, but the training data were so small that the model overfitted.
  • Computational Complexity
Let l S and l T denote the number of training samples and testing samples, respectively, and each sample x i R d . Suppose we grow   k trees for RF. The computational complexity of each step is shown in Table 7, which is based on Liu et al. [26], Pan et al. [27], and Biau [36], where t γ is the maximum number of iterations and m is the number of base kernels. Therefore, the computational complexity of MK-DA is t γ O 1 + m l S 2 O d + O ( d l S + l T 2 ) . Then, the computational complexity of the proposed framework is t γ O 1 + m l S 2 O d + O ( d l S + l T 2 ) + O k d l S log l S .
  • Limitations
However, there are also problems with the proposed method. First, after the features were adapted, the classification accuracies applying RF in both types of experiments were higher than the LDA after domain adaption, that is, the average classification accuracy of RF was 0.17% higher than that of LDA in the cross-subject experiment, and was 0.41% higher than that of the LDA in the cross-session experiment. However, the computational complexity of RF was significantly higher than that of the LDA. Therefore, it is necessary to combine classification accuracy with the computational complexity in selecting the suitable classifier. Second, parameters involved in the proposed framework (i.e., the relaxation coefficient C of ELM, the initial parameters of the kernels) were all selected from among a limited number of values, but the choice of the parameters would affect the classification results. Therefore, optimization methods will be suggested to solve this problem in our subsequent studies.

5. Conclusions

The method proposed in this paper aimed to address the inter-domain differences of EEG-based motor imagery BCI, especially for BCI illiteracy. It was found that BCI illiterate subjects could not effectively control the BCI due to two major problems: difficulties in classifying and poor repeatability. We proposed a domain adaption method that combines MK-ELM and MK-MMD. To demonstrate the effectiveness of the method, we performed experiments from two perspectives (i.e., cross-subject experiment and cross-session experiment). The MK-ELM achieved relatively better results than the LDA in all experiments. Meanwhile, it can be seen from the results of MK-DA that the MK-DA with each classifier achieved relatively better results in all combination forms. The average accuracies of all experiments of MK-DA combined with LDA was 3% higher than that of LDA in the cross-subject experiments, and was 5.6% higher in the cross-session experiments. Therefore, the divisibility was increased after the features were mapped to multiple-kernel-based RHKS by MK-ELM, and the domain shift decreased by MK-MMD, which was consistent with the experimental phenomenon from the results. At the same time, RF that could effectively handle high-dimensional features was employed as a classifier. It can be seen from the results of MK-DA-RF in the cross-subject experiments that the average classification accuracy of all the experiments could reach 70.4%, which was 2.7% higher than that of the reference method “CSP + LDA” and 0.3% higher than that of the best-performing control method. In the cross-session experiments, the average classification accuracy of the proposed method for all experiments could reach 73.6%, 6.1% higher than that of the reference method, and 0.4% more than that of the best-performing control method. Particularly for the BCI illiterate subjects, the average classification accuracy of all the experiments with target subjects showed that BCI illiteracy could reach 63.4% with the proposed method, which was 5.3% higher than the reference method without domain adaption. Therefore, the method proposed in this paper could achieve a significant effect in the BCI illiteracy group.

Author Contributions

Conceptualization, L.T. and D.L.; Methodology, L.T.; Software, L.T. and T.C.; Validation, L.T., T.C. and D.L.; Formal analysis, L.T.; Investigation, L.T.; Resources, Q.W.; Data curation, L.T.; Writing—original draft preparation, L.T.; Writing—review and editing, J.S.; Visualization, L.T.; Supervision, J.S. and D.L.; Project administration, J.S. and D.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (grant number 61471140); the Fundamental Research Funds for the Central Universities (grant number IR2021222), the Sci-tech Innovation Foundation of Harbin (grant number 2016RALGJ001), and the China Scholarship Council and the Future Science and Technology Innovation Team Project of HIT (grant number 216506).

Data Availability Statement

We validated our method by an open-access dataset, namely, the BMI dataset (http://gigadb.org/dataset/view/id/100542/ (accessed on 16 May 2021)). Thus, our study did not involve any self-recorded datasets from human participants or animals. The authors declare that they have no conflicts of interest.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pasqualotto, E.; Federici, S.; Belardinelli, M.O. Toward functioning and usable brain-computer interfaces (BCIs): A literature review. Disabil. Rehabil. Assist. Technol. 2012, 7, 89–103. [Google Scholar] [CrossRef] [PubMed]
  2. Blankertz, B.; Vidaurre, C. Towards a cure for BCI illiteracy: Machine learning based co-adaptive learning. BMC Neurosci. 2009, 10, P85. [Google Scholar] [CrossRef]
  3. Blankertz, B.; Sannelli, C.; Halder, S.; Hammer, E.M.; Kübler, A.; Müller, K.R.; Curio, G.; Dickhaus, T. Neurophysiological predictor of SMR-based BCI performance. Neuroimage 2010, 51, 1303–1309. [Google Scholar] [CrossRef] [PubMed]
  4. Saha, S.; Baumert, M. Intra- and Inter-subject Variability in EEG-Based Sensorimotor Brain Computer Interface: A Review. Front. Comput. Neurosci. 2020, 13, 87. [Google Scholar] [CrossRef] [PubMed]
  5. Kragel, P.A.; Knodt, A.R.; Hariri, A.R.; Labar, K.S. Decoding Spontaneous Emotional States in the Human Brain. PLoS Biol. 2016, 14, e2000106. [Google Scholar] [CrossRef]
  6. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A Survey of Transfer Learning; Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar] [CrossRef]
  7. Shu, X.; Yao, L.; Sheng, X.; Zhang, D.; Zhu, X. Enhanced motor imagery-based BCI performance via tactile stimulation on unilateral hand. Front. Hum. Neurosci. 2017, 11, 585. [Google Scholar] [CrossRef]
  8. Lee, M.H.; Kwon, O.Y.; Kim, Y.J.; Kim, H.K.; Lee, Y.E.; Williamson, J.; Fazli, S.; Lee, S.W. EEG dataset and OpenBMI toolbox for three BCI paradigms: An investigation into BCI illiteracy. Gigascience 2019, 8, giz002. [Google Scholar] [CrossRef]
  9. Liu, Y.; Lan, Z.; Cui, J.; Sourina, O.; Muller-Wittig, W. EEG-Based cross-subject mental fatigue recognition. In Proceedings of the 2019 International Conference on Cyberworlds (CW), Kyoto, Japan, 2–4 October 2019; pp. 247–252. [Google Scholar] [CrossRef]
  10. Zhang, J.; Wang, Y.; Li, S. Cross-subject mental workload classification using kernel spectral regression and transfer learning techniques. Cogn. Technol. Work 2017, 19, 587–605. [Google Scholar] [CrossRef]
  11. Chai, X.; Wang, Q.; Zhao, Y.; Li, Y.; Liu, D.; Liu, X.; Bai, O. A fast, efficient domain adaptation technique for cross-domain electroencephalography(EEG)-based emotion recognition. Sensors 2017, 17, 1014. [Google Scholar] [CrossRef]
  12. He, H.; Wu, D. Different Set Domain Adaptation for Brain-Computer Interfaces: A Label Alignment Approach. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 1091–1108. [Google Scholar] [CrossRef]
  13. Hua, Y.; Zhong, X.; Zhang, B.; Yin, Z.; Zhang, J. Manifold feature fusion with dynamical feature selection for cross-subject emotion recognition. Brain Sci. 2021, 11, 1392. [Google Scholar] [CrossRef] [PubMed]
  14. Wei, H.; Ma, L.; Liu, Y.; Du, Q. Combining Multiple Classifiers for Domain Adaptation of Remote Sensing Image Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1832–1847. [Google Scholar] [CrossRef]
  15. Ma, L.; Crawford, M.M.; Zhu, L.; Liu, Y. Centroid and Covariance Alignment-Based Domain Adaptation for Unsupervised Classification of Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 2305–2323. [Google Scholar] [CrossRef]
  16. Liu, X.; Wang, L.; Zhu, X.; Li, M.; Zhu, E.; Liu, T.; Liu, L.; Dou, Y.; Yin, J. Multiple Kernel Learning Algorithms. J. Mach. Learn. Res. 2011, 42, 1303–1316. [Google Scholar] [CrossRef]
  17. Bucak, S.S.; Jin, R.; Jain, A.K. Multiple kernel learning for visual object recognition: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1354–1369. [Google Scholar] [CrossRef] [PubMed]
  18. Long, M.; Cao, Y.; Wang, J.; Jordan, M.I. Learning transferable features with deep adaptation networks. In Proceedings of the 32nd International Conference on Machine Learning, ICML, Lille, France, 7–9 July 2015; Volume 1, pp. 97–105. [Google Scholar]
  19. Sun, Z.; Wang, C.; Wang, H.; Li, J. Learn multiple-kernel SVMs for domain adaptation in hyperspectral data. IEEE Geosci. Remote Sens. Lett. 2013, 10, 1224–1228. [Google Scholar] [CrossRef]
  20. Dai, M.; Wang, S.; Zheng, D.; Na, R.; Zhang, S. Domain Transfer Multiple Kernel Boosting for Classification of EEG Motor Imagery Signals. IEEE Access 2019, 7, 49951–49960. [Google Scholar] [CrossRef]
  21. Deng, C.; Liu, X.; Li, C.; Tao, D. Active multi-kernel domain adaptation for hyperspectral image classification. Pattern Recognit. 2018, 77, 306–315. [Google Scholar] [CrossRef] [Green Version]
  22. Zheng, Y.; Wang, X.; Zhang, G.; Xiao, B.; Xiao, F.; Zhang, J. Multi-Kernel Coupled Projections for Domain Adaptive Dictionary Learning. IEEE Trans. Multimed. 2019, 21, 2292–2304. [Google Scholar] [CrossRef]
  23. Wang, W.; Wang, H.; Zhang, Z.; Zhang, C.; Gao, Y. Semi-supervised domain adaptation via Fredholm integral based kernel methods. Pattern Recognit. 2019, 85, 185–197. [Google Scholar] [CrossRef]
  24. Chen, X.; Lengelĺe, Ŕ. Domain adaptation transfer learning by SVM subject to a maximum-mean-discrepancy-like constraint. In Proceedings of the ICPRAM 2017—6th International Conference on Pattern Recognition Applications and Methods, Porto, Portugal, 24–26 February 2017; pp. 89–95. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Wang, M.; Huang, Y.; Nehorai, A. Aligning Infinite-Dimensional Covariance Matrices in Reproducing Kernel Hilbert Spaces for Domain Adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3437–3445. [Google Scholar] [CrossRef]
  26. Liu, X.; Wang, L.; Huang, G.; Zhang, J.; Yin, J. Multiple kernel extreme learning machine. Neurocomputing 2015, 149, 253–264. [Google Scholar] [CrossRef]
  27. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain adaptation via transfer component analysis. IEEE Trans. Neural Netw. 2011, 22, 199–210. [Google Scholar] [CrossRef]
  28. Sch, B. Nonlinear Component Analysis as a Kernel Eigenvalue Problem. Neural Comput. 1998, 1319, 1299–1319. [Google Scholar]
  29. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer feature learning with joint distribution adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar] [CrossRef]
  30. Kaeseler, R.L.; Johansson, T.W.; Struijk, L.N.S.A.; Jochumsen, M. Feature and Classification Analysis for Detection and Classification of Tongue Movements from Single-Trial Pre-Movement EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 678–687. [Google Scholar] [CrossRef]
  31. Torres, E.P.; Torres, E.A.; Hernández-Álvarez, M.; Yoo, S.G. Emotion Recognition Related to Stock Trading Using Machine Learning Algorithms with Feature Selection. IEEE Access 2020, 8, 199719–199732. [Google Scholar] [CrossRef]
  32. Bentlemsan, M.; Zemouri, E.T.; Bouchaffra, D.; Yahya-Zoubir, B.; Ferroudji, K. Random forest and filter bank common spatial patterns for EEG-based motor imagery classification. In Proceedings of the International Conference on Intelligent Systems, Modelling and Simulation, ISMS, Kuala Lumpur, Malaysia, 9–12 February 2015; pp. 235–238. [Google Scholar] [CrossRef]
  33. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2012, 42, 513–529. [Google Scholar] [CrossRef] [PubMed]
  34. Huang, G.B.; Ding, X.; Zhou, H. Optimization method based extreme learning machine for classification. Neurocomputing 2010, 74, 155–163. [Google Scholar] [CrossRef]
  35. Lee, J.M.; Yoo, C.K.; Choi, S.W.; Vanrolleghem, P.A.; Lee, I.B. Nonlinear process monitoring using kernel principal component analysis. Chem. Eng. Sci. 2004, 59, 223–234. [Google Scholar] [CrossRef]
  36. Biau, G. Analysis of a random forests model. J. Mach. Learn. Res. 2012, 13, 1063–1095. [Google Scholar]
  37. Xiao, T.; Liu, P.; Zhao, W.; Tang, X. Iterative landmark selection and subspace alignment for unsupervised domain adaptation. J. Electron. Imaging 2018, 27, 1. [Google Scholar] [CrossRef]
  38. Wei, J. Learning Discriminative Geodesic Flowkernel For Unsupervised Domain Adaption. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), San Diego, CA, USA, 23–27 July 2018. [Google Scholar]
  39. Jayaram, V.; Alamgir, M.; Altun, Y.; Grosse-wentrup, M. Transfer Learning in Brain-Computer Interfaces. IEEE Comput. Intell. Mag. 2016, 11, 20–31. [Google Scholar] [CrossRef]
  40. Lotte, F.; Congedo, M.; Lécuyer, A.; Lamarche, F.; Arnaldi, B. A review of classification algorithms for EEG-based brain-computer interfaces. J. Neural Eng. 2007, 4, R1. [Google Scholar] [CrossRef] [PubMed]
  41. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain-computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef] [PubMed]
  42. Mane, R.; Chew, E.; Chua, K.; Ang, K.K.; Robinson, N.; Vinod, A.P.; Lee, S.-W.; Guan, C. FBCNet: A Multi-view Convolutional Neural Network for Brain-Computer Interface. arXiv 2021, arXiv:2104.01233. [Google Scholar]
  43. Giannakakis, G.; Trivizakis, E.; Tsiknakis, M.; Marias, K. A novel multi-kernel 1D convolutional neural network for stress recognition from ECG. In Proceedings of the 2019 8th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos, ACIIW, Cambridge, UK, 3–6 September 2019; pp. 273–276. [Google Scholar] [CrossRef]
Figure 1. The power spectral density (PSD) of Subject 46 performing motor imagery at different times: the charts are the power spectra (blue represents negative, and red represents positive). The EEG signals were provided by an open dataset with BCI illiterate subjects [8] and they were recorded in two different sessions on different days. (a) and (b) are the PSD diagrams of Subject 46 in session 1 and session 2, respectively. The classification accuracies of Subject 46 were 53% and 58% in the two sessions, respectively, so the subject was classified as BCI illiterate.
Figure 1. The power spectral density (PSD) of Subject 46 performing motor imagery at different times: the charts are the power spectra (blue represents negative, and red represents positive). The EEG signals were provided by an open dataset with BCI illiterate subjects [8] and they were recorded in two different sessions on different days. (a) and (b) are the PSD diagrams of Subject 46 in session 1 and session 2, respectively. The classification accuracies of Subject 46 were 53% and 58% in the two sessions, respectively, so the subject was classified as BCI illiterate.
Sensors 22 06572 g001
Figure 2. The feature distribution comparison of Subject 46. The features were extracted by a common spatial pattern (CSP). (a) The distribution of features for session 1; (b) The distribution comparison of features for Sessions 1 and 2.
Figure 2. The feature distribution comparison of Subject 46. The features were extracted by a common spatial pattern (CSP). (a) The distribution of features for session 1; (b) The distribution comparison of features for Sessions 1 and 2.
Sensors 22 06572 g002
Figure 3. The distribution adaptation. (a) Purely for marginal distribution adaptation; (b) With the features being mapped into the new space, the inter-class maximal divisibility was achieved and the distribution difference between the source domain and target domain was reduced.
Figure 3. The distribution adaptation. (a) Purely for marginal distribution adaptation; (b) With the features being mapped into the new space, the inter-class maximal divisibility was achieved and the distribution difference between the source domain and target domain was reduced.
Sensors 22 06572 g003
Figure 4. The joint distribution adaptation framework.
Figure 4. The joint distribution adaptation framework.
Sensors 22 06572 g004
Figure 5. The generation process of random forest. (a) The generation of the forest. (b) The implementation of decisions.
Figure 5. The generation process of random forest. (a) The generation of the forest. (b) The implementation of decisions.
Sensors 22 06572 g005
Figure 6. The experimental design for binary class MI. The experimental description was provided by Lee et al. [8].
Figure 6. The experimental design for binary class MI. The experimental description was provided by Lee et al. [8].
Sensors 22 06572 g006
Figure 7. The electrode position.
Figure 7. The electrode position.
Sensors 22 06572 g007
Figure 8. The results of the different distribution adaptation methods in the cross-subject experiments.
Figure 8. The results of the different distribution adaptation methods in the cross-subject experiments.
Sensors 22 06572 g008
Figure 9. The results of the different distribution adaptation methods for the NBI group.
Figure 9. The results of the different distribution adaptation methods for the NBI group.
Sensors 22 06572 g009
Figure 10. The results of the different distribution adaptation methods for the BI group.
Figure 10. The results of the different distribution adaptation methods for the BI group.
Sensors 22 06572 g010
Figure 11. The feature space distribution of Subject 22. With the reference method, Subject 22 had a classification accuracy of 68% from session 1 to session 2. With the improved method, the classification accuracy was 83%. (a) The original feature space distribution of class 1 (left-hand imagery) and class 2 (right-hand imagery) in session 1; (b) The feature space distribution of class 1 and class 2 obtained after marginal distribution adaptation session 1; (c) The original feature space distribution in session 1 (source domain) and session 2 (target domain) for all categories; (d) The feature space distribution obtained after marginal distribution adaptation in session 1 (source domain) and session 2 (target domain) for all categories.
Figure 11. The feature space distribution of Subject 22. With the reference method, Subject 22 had a classification accuracy of 68% from session 1 to session 2. With the improved method, the classification accuracy was 83%. (a) The original feature space distribution of class 1 (left-hand imagery) and class 2 (right-hand imagery) in session 1; (b) The feature space distribution of class 1 and class 2 obtained after marginal distribution adaptation session 1; (c) The original feature space distribution in session 1 (source domain) and session 2 (target domain) for all categories; (d) The feature space distribution obtained after marginal distribution adaptation in session 1 (source domain) and session 2 (target domain) for all categories.
Sensors 22 06572 g011
Figure 12. The results of different classifiers in the cross-subject experiment.
Figure 12. The results of different classifiers in the cross-subject experiment.
Sensors 22 06572 g012
Figure 13. The confusion matrices of RF in the cross-subject experiment. (a) Random-NBI; (b) Random-BI.
Figure 13. The confusion matrices of RF in the cross-subject experiment. (a) Random-NBI; (b) Random-BI.
Sensors 22 06572 g013
Figure 14. Receiver operating characteristic curve of RF in the cross-subject experiment. (a) Random-NBI; (b) Random-BI.
Figure 14. Receiver operating characteristic curve of RF in the cross-subject experiment. (a) Random-NBI; (b) Random-BI.
Sensors 22 06572 g014
Figure 15. The results of the different classifiers in the cross-session experiment.
Figure 15. The results of the different classifiers in the cross-session experiment.
Sensors 22 06572 g015
Figure 16. The confusion matrices of RF in the cross-session experiment. (a) S1-S2; (b) S2-S1.
Figure 16. The confusion matrices of RF in the cross-session experiment. (a) S1-S2; (b) S2-S1.
Sensors 22 06572 g016
Figure 17. The receiver operating characteristic curve of RF in the cross-session experiment. (a) S1-S2; (b) S2-S1.
Figure 17. The receiver operating characteristic curve of RF in the cross-session experiment. (a) S1-S2; (b) S2-S1.
Sensors 22 06572 g017
Table 1. A comparison of the different distribution adaptation methods for the source and target data from different subjects.
Table 1. A comparison of the different distribution adaptation methods for the source and target data from different subjects.
TasksResultsLDA **MK-ELM *SA-RF *GFK-RF **CORAL-RF *TCA-RF *MKL *MK-DA-RF
Random-NBIMean75.8177.0775.0575.1277.077.6875.2278.46
Std8.259.3211.9812.3110.7311.149.149.5
CI2.953.334.284.403.843.983.273.40
Random-BIMean58.6059.7859.1158.4759.8661.8660.3662.32
Std7.478.747.646.647.175.048.678.01
CI2.673.132.732.372.561.083.102.87
Note: * indicates p < 0.05, ** indicates p < 0.01, and there is no * when p > 0.05. The best results are indicated in bold text; SA-RF indicates the DA method is SA and the classifier is RF, Other interpretations are the same.
Table 2. A comparison of the different classifiers in the cross-subject experiments.
Table 2. A comparison of the different classifiers in the cross-subject experiments.
TasksResultsLDA *SVM **KNN **EEGnet **FBCNet **RF
Random-NBIMean78.2977.8178.2664.6078.2378.46
Std8.487.408.639.777.609.50
CI3.032.653.093.492.723.40
Random-BIMean62.1561.6762.1361.7262.0962.32
Std7.496.626.918.364.806.95
CI2.682.372.472.991.722.49
Note: * indicates p < 0.05, ** indicates p < 0.01, and there is no * when p > 0.05. The best results are indicated with bold text.
Table 3. A comparison of the different distribution adaptation methods for the source and target data from different sessions.
Table 3. A comparison of the different distribution adaptation methods for the source and target data from different sessions.
GroupTaskResultsLDA **MK-ELM **SA-RF *GFK-RF **CORAL-RF *TCA-RF *MKL *MK-DA-RF
NBIS1-S2Mean79.0282.0683.2178.9882.9884.8383.7585.69
S2-S1Mean80.7781.8580.8580.8582.8384.5484.4485.81
BIS1-S2Mean58.9062.8560.0860.0861.3764.1863.7564.80
S2-S1Mean56.7760.6858.6558.6559.1561.7361.5063.07
ALLS1-S2Mean67.8471.3968.4868.4870.9773.3672.6474.08
Std15.0414.3715.0914.9014.8715.6414.8014.43
CI4.123.794.064.114.153.934.023.90
S2-S1Mean67.4470.0968.5268.5269.6871.8771.6973.18
Std15.4314.2215.2115.4015.5714.7215.0514.64
CI4.013.834.033.973.974.173.953.85
Note: * indicates p < 0.05, ** indicates p < 0.01, and there is no * when p > 0.05. The best results are indicated with bold text; SA-RF indicates the DA method is SA and the classifier is RF, Other interpretations are the same.
Table 4. A comparison of the different classifiers in the cross-session experiments.
Table 4. A comparison of the different classifiers in the cross-session experiments.
TasksResultsLDA *SVM *KNN **EEGnet **FBCNet **RF
S1-S2Mean73.8472.1872.6166.5673.8474.08
Std14.5014.1915.1314.4914.1014.43
CI3.873.784.043.863.763.85
S2-S1Mean72.6072.1272.2265.8472.3673.18
Std14.2314.6215.2015.4314.9114.64
CI3.873.904.064.123.983.91
Note: * indicates p < 0.05, ** indicates p < 0.01, and there is no * when p > 0.05. The best results are indicated with bold text.
Table 5. The performance matrices of RF in the cross-subject experiments.
Table 5. The performance matrices of RF in the cross-subject experiments.
TaskKappaRecallF1-ScorePrecisionAUC
S1-S20.7030.8160.7910.7680.762
S2-S10.4510.6130.6260.6190.640
Table 6. The performance matrices of RF in the cross-session experiments.
Table 6. The performance matrices of RF in the cross-session experiments.
TaskKappaRecallF1-ScorePrecisionAUC
S1-S20.6310.7430.7410.7390.736
S2-S10.6230.7600.7380.7170.732
Table 7. The computational complexity of each step.
Table 7. The computational complexity of each step.
StepComputational Complexity
MK-ELM t γ O 1 + m l S 2 O d
TCA O ( d l S + l T 2 )
RF O k d l S log l S
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tao, L.; Cao, T.; Wang, Q.; Liu, D.; Sun, J. Distribution Adaptation and Classification Framework Based on Multiple Kernel Learning for Motor Imagery BCI Illiteracy. Sensors 2022, 22, 6572. https://doi.org/10.3390/s22176572

AMA Style

Tao L, Cao T, Wang Q, Liu D, Sun J. Distribution Adaptation and Classification Framework Based on Multiple Kernel Learning for Motor Imagery BCI Illiteracy. Sensors. 2022; 22(17):6572. https://doi.org/10.3390/s22176572

Chicago/Turabian Style

Tao, Lin, Tianao Cao, Qisong Wang, Dan Liu, and Jinwei Sun. 2022. "Distribution Adaptation and Classification Framework Based on Multiple Kernel Learning for Motor Imagery BCI Illiteracy" Sensors 22, no. 17: 6572. https://doi.org/10.3390/s22176572

APA Style

Tao, L., Cao, T., Wang, Q., Liu, D., & Sun, J. (2022). Distribution Adaptation and Classification Framework Based on Multiple Kernel Learning for Motor Imagery BCI Illiteracy. Sensors, 22(17), 6572. https://doi.org/10.3390/s22176572

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop