Next Article in Journal
EMG-Based 3D Hand Motor Intention Prediction for Information Transfer from Human to Robot
Next Article in Special Issue
Self-Relative Evaluation Framework for EEG-Based Biometric Systems
Previous Article in Journal
Leveraging Deep Learning for Visual Odometry Using Optical Flow
Previous Article in Special Issue
Induction of Neural Plasticity Using a Low-Cost Open Source Brain-Computer Interface and a 3D-Printed Wrist Exoskeleton
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing SSVEP-Based Brain-Computer Interface with Two-Step Task-Related Component Analysis

Department of Electronics and Communication Engineering, Kwangwoon University, Seoul 01897, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(4), 1315; https://doi.org/10.3390/s21041315
Submission received: 12 January 2021 / Revised: 6 February 2021 / Accepted: 9 February 2021 / Published: 12 February 2021
(This article belongs to the Collection EEG-Based Brain–Computer Interface for a Real-Life Appliance)

Abstract

:
Among various methods for frequency recognition of the steady-state visual evoked potential (SSVEP)-based brain-computer interface (BCI) study, a task-related component analysis (TRCA), which extracts discriminative spatial filters for classifying electroencephalogram (EEG) signals, has gathered much interest. The TRCA-based SSVEP method yields lower computational cost and higher classification performance compared to existing SSVEP methods. In spite of its utility, the TRCA-based SSVEP method still suffers from the degradation of the frequency recognition rate in cases where EEG signals with a short length window are used. To address this issue, here, we propose an improved strategy for decoding SSVEPs, which is insensitive to a window length by carrying out two-step TRCA. The proposed method reuses the spatial filters corresponding to target frequencies generated by the TRCA. Followingly, the proposed method accentuates features for target frequencies by correlating individual template and test data. For the evaluation of the performance of the proposed method, we used a benchmark dataset with 35 subjects and confirmed significantly improved performance comparing with other existing SSVEP methods. These results imply the suitability as an efficient frequency recognition strategy for SSVEP-based BCI applications.

1. Introduction

The brain-computer interface (BCI) provides a bidirectional system between the human brain and external devices by decoding electrical brain waves measured in specific environments. Among various measurements of brain activities, electroencephalography (EEG) is the most common tool in BCI systems due to inexpensive cost, portability, usability, and so forth [1]. EEG-based BCI may help severely disabled people, which is used in rehabilitative applications and the internet of medical things (IoMT) [2,3]. Typically, in years past, real-time BCI applications such as brain-controlled vehicles (BCVs) [4] and brain-controlled wheelchairs (BCWs) [5] that can be facilitated in daily life have received enormous attention. To control these applications, in the BCI study, EEG signals can be divided into different forms depending on the purpose of use, its type, and so on. Among those forms, steady-state visual evoked potential (SSVEP) has attracted much attention due to the high communication rate, classification accuracy, and high signal-to-noise ratio (SNR) [6,7]. Driven by these advantages, the number of SSVEP-based real-time BCI applications have resulted in remarkable achievements [4,5,8,9].
In SSVEP-based BCIs, in terms of detection target frequency among visual stimuli with specific frequencies and spatial filtering techniques have been widely developed due to high SNR by removing the external noise of EEG signals caused by artifacts and eye blinks. Recently, canonical correlation analysis (CCA) has been presented to identify target frequency based on the use of sinusoidal signals as reference signals [10]. Due to its high efficiency and easy implementation, CCA has been widely utilized in SSVEP-based BCI research. Moreover, according to other studies [11,12], CCA with a high information transfer rate (ITR) has shown great potential in online BCI applications. However, due to the interference in spontaneous EEG signals, CCA may suffer from its degradation of detection performance. To address this issue, in much of the literature, a number of variants of CCA have been proposed to achieve higher frequency recognition performance. For example, individual CCA (IT-CCA) [12], L1-regularized multi-way CCA (L1-MCCA) [13], multiset CCA (MsetCCA) [14], and latent common source extraction (LCSE) [15] have gained interest among BCI communities. Among those, the combination of the standard CCA and IT-CCA has led to the highest performance [16].
As another approach, several spatial filtering methods have shed light on frequency recognition to extract task-specific source activities from EEG signals. Among them, the task-related component analysis (TRCA) [17] has been developed to extract the spatial filter closely related to task-specific by finding the maximum correlation of the internal component between trials. Based on this approach, Nakanishi et al. introduced TRCA into the SSVEP-based BCI, leading to the best performance among existing methods [18]. The TRCA method achieved ITR of 325.33 ± 38.17 bits/min, implying practicality in real-life BCI application. More recently, an ensemble approach for incorporating a generated spatial filter has shown significant improvement to frequency recognition regardless of time window length [16]. In this line of thought, fusing all the canonical correlation coefficient of CCA yielded robust results and improved performance in terms of classification accuracy and ITR compared to CCA [19]. However, in spite of several advantages of the aforementioned methods, SSVEP-based BCI still suffers from the degradation of performance in cases where a short time window (TW) of EEG signals is used. In order to be available in online SSVEP-based BCI applications, robustness regarding the TW is an essential property of frequency recognition in SSVEP.
In this study, we present a novel frequency recognition method for SSVEP-based BCI by expanding the concept of the standard TRCA. The proposed method consists of two subsequent steps, which is referred to as a two-step TRCA (TSTRCA). First, we generate the subject-specific spatial filter using the standard TRCA. Then, motivated by an ensemble approach, the target frequency recognition is carried out by ensembling and emphasizing discriminative information from the correlation between the individual templates and test data. Thus, the proposed TSTRCA method can improve performance in a short TW by reflecting the correlation of inter-subjects as well as inter-sessions and accentuating features as ensemble classifiers. We validated the frequency recognition performance of the proposed method using the SSVEP benchmark dataset, comprised of 35 subjects [20]. In addition, we compared the average accuracy and ITR of the proposed TSTRCA with CCA, extended CCA (ExtCCA), and TRCA.
The rest of this paper is organized as follows: Section 2 presents an introduction of the benchmark SSVEP dataset and describes existing methods and the proposed method. In Section 3, the experimental results are exhibited. Section 4 provides the conclusion of this study.

2. Materials and Algorithms

2.1. Benchmark SSVEP EEG Dataset and Preprocessing

In this study, the benchmark SSVEP dataset provided by Wang et al. [20] was utilized to evaluate the proposed method. Thirty-five healthy subjects, consisting of seventeen females and eighteen males, participated in an SSVEP experiment by staring at an offline 40-target BCI speller (5 × 8 character matrix), each with a different frequency. The 40-target BCI speller has a range between 8 Hz and 15.8 Hz with an interval of 0.2 Hz. The SSVEP EEG signals were recorded with 64 channels, sampled at 1000 Hz, and band-pass filtered between 0.15 Hz and 200 Hz. A notch filter at 50 Hz was employed in order to remove power-line interference. For each subject, the dataset was made up of a total of 6 s per trial, and one trial was repeated six times. In each trial, a visual cue indicating the beginning of the experiment appeared for 0.5 s.
All 40 target frequencies were presented randomly to all subjects. After the end of the visual stimulation, it was blanked for 0.5 s before the next experiment was presented. During the experiment, the subject was asked to avoid blinking and a suitable rest was also provided between two consecutive trials.
To facilitate the signal-processing analytics, the SSVEP datasets were further preprocessed. Firstly, the band-pass filter with an IIR filter was applied to all data epochs. The frequency range was considered from 7 Hz to 90 Hz. Then, as shown in [20], considering a latency delay in the visual system, the SSVEP data were extracted between 0.64 s and 0.64 + d s from each epoch, where d s is the length of TW for frequency recognition. The supplementary information for this dataset was elucidated in [20].

2.2. Conventional SSVEP Frequency Recognition Methods

In this section, we first provide a brief introduction of the conventional SSVEP frequency recognition methods such as CCA, ExtCCA, and TRCA, which is followed by the proposed method. Then, the frequency recognition using a filter bank approach, which is known as its capability to improve the performance of standard SSVEP methods, is provided.

2.2.1. Standard Canonical Correlation Analysis

CCA is a conventional statistical method to explore the underlying correlation between two sets of multidimensional variables. Assume that multidimensional signals are given as X D 1 × T and Y D 2 × T ; CCA aims at finding a pair of weight vectors, w x D 1 × 1 and w y D 2 × 1 , which maximize the correlation between their linear combinations x = w x T X and y = w y T Y . Formally, the correlation coefficient of CCA is given by
ρ =   E x y T E x x T E y y T =   w x T X Y T w y w x T X X T w x w y T Y Y T w y
where ρ is the Pearson correlation coefficient between x and y . Then, the weight vectors w x and w y based on CCA is obtained by maximizing the correlation ρ in Equation (1). Formally, this problem for finding the weight vectors w x and w y can be represented by
arg   max w x ,   w y   ρ =   w x T X Y T w y s . t .   w x T X X T w x = 1 , w y T Y Y T w y = 1
Then, the optimal weight vectors are obtained through a generalized eigenvalue problem [21]. Here, the maximum of ρ regarding the weight vectors is referred to as the maximum canonical correlation.
In the SSVEP-based BCI, CCA has been widely used for frequency recognition by obtaining the maximum canonical correlation between test signals of SSVEP EEGs and reference signals [10]. In addition, the reference signals are composed of sinusoidal signals, denoted as Z i 2 N h × N s ,   i = 1 ,   2 ,   ,   N f , which is given by
Z i =   sin 2 π f i t cos 2 π f i t sin 2 π N h f i t cos 2 π N h f i t , t =   1 F s ,   2 F s ,   ,   N s F s
where, N h is the number of harmonics, N s is the number of sample points, N f is the number of target frequencies, and F s is the sampling rate. However, for the reference signals, it is difficult to determine the appropriate number of harmonics. Therefore, in the current study, we employed the individual template proposed by IT-CCA as the reference signals [12]. The test signal consists of a single trial of multichannel EEG signals, written as X ¯ N c × N s , where N c denotes the number of channels.
Finally, the target frequency is identified in the cases where the correlation coefficient ρ i —calculated by CCA between a test signal and each reference signal—is at its maximum, as follows:
f t a r g e t =   max i ρ i , i =   1 ,   2 ,   ,   N f

2.2.2. Extended Canonical Correlation Analysis

In [12,22], ExtCCA enhanced the signal to noise (SNR) of SSVEP by combining two frequency recognition methods, i.e., the standard CCA and IT-CCA. IT-CCA is a variant of the standard CCA in that the individual templates are used as the reference signals. The individual template, denoted as Y i N c × N s , i = 1 ,   2 ,   ,   N f , is constructed by averaging across multiple EEG trials acquired from the same subjects. Furthermore, ExtCCA makes use of three weight vectors generated by three kinds of EEG signals, i.e., test signal and two reference signals, as the spatial filters. Specifically, three spatial filters are as follows: (1) W X ¯ X ¯ Y i between the test signal X ¯ and the individual template Y i , (2) W X ¯ X ¯ Z i between the test signal X ¯ and a set of sinusoidal signals Z i , (3) W X ¯ Y i Z i between the individual template and a set of sinusoidal signals. A correlation vector r i   ,   i = 1 ,   2 ,   ,   N f , for the i -th template signal can be obtained using the designed spatial filters as follows:
r i =   r i ,   1 r i ,   2 r i ,   3 r i ,   4 = ρ X ¯ T W X ¯ X ¯ Z i , Z i T W Z X ¯ Z i ρ X ¯ T W X ¯ X ¯ Y i , Y i T W X ¯ X ¯ Y i ρ X ¯ T W X ¯ X ¯ Z i , Y i T W X ¯ X ¯ Z i ρ X ¯ T W X ¯ Y i Z i , Y i T W X ¯ Y i Z i
where ρ ·   ,   · is the Pearson correlation coefficient between two multidimensional SSVEP EEG signals. For each target frequency, the four correlation values in Equation (5) are combined as a weighted correlation coefficient ρ i , which is given by
ρ i =   k = 1 4 sign r i , k · r i , k 2
where sign · denotes the signum function and is used to reflect discriminative information from the negative value of r i , k . Then, the target frequency of each test signal is identified by the aforementioned Equation (4).

2.2.3. Standard Task-Related Component Analysis

In TRCA, maximizing reproducibility between time-locked task trials leads to the spatial filters, which are capable of reflecting task-specific components. Assume SSVEP EEG signals of l -th trial X t N c × N s , t = 1 ,   2 ,   ,   N t , where N t is the number of trials. Then, a linear combination of X t is written as Y t = w T X t .
The TRCA method aims at designing the weight vector w which is obtained by maximizing the sum of covariance between available combinations of all trials. The covariance between t 1 -th and t 2 -th trials are computed as
t 1   ,   t 2 = 1 t 1   t 2 N t Cov Y t 1   , Y t 2     =   t 1   ,   t 2 = 1 t 1   t 2 N t c 1   , c 2 = 1 N c w c 1 w c 2 Cov X C 1 t 1   , X C 2 t 2 =     w T S w
where Cov ( ·   , · ) denotes the covariance between two variables, X C 1 t 1 and X C 2 t 2 denote C 1 -th and C 2 -th channels of EEG signals of X t 1 and X t 2 , respectively.
Denote a concatenated matrix of all trials X t as X ^ = X 1 X 2 X N t . Constraining the variance of Y t by normalizing to one leads to
Var Y t =   c 1   ,     c 2 = 1 N c w c 1 w c 2 Cov X ^ C 1   , X ^ C 2 = w T Q w = 1
where Var ( · ) denotes the variance of a variable, X ^ C 1 and X ^ C 2 denote the C 1 -th and C 2 -th channels of X ^ , respectively.
Finally, the optimal weight vector w ^ can be obtained through a constrained optimization problem as follows:
w ^ =   arg   max w w T S w w T Q w
In Equation (9), the optimal weight vector w ^ is the eigenvector of Q 1 S , which corresponds to the largest eigenvalue. In the SSVEP-based BCI, a spatial filter approach such as TRCA has the effect of eliminating background activities by filtering out the principal components inherent in EEG signals [17,18]. In TRCA, for each target frequency, the correlation coefficient between the test signal and the individual template is determined from the training signal with the given optimal spatial filters as follows:
ρ ^ i =   ρ w i T X ¯   , w i T Y i , i = 1 , 2 ,   ,   N f
Subsequently, target identification is performed as follows:
f t a r g e t =   max i ρ ^ i , i = 1 ,   2 ,   ,   N f

2.3. The Proposed Two-Step TRCA

As mentioned earlier, advanced versions based on the spatial filter accomplished improved performance of frequency identification of SSVEP. In [16,18], using an ensemble approach yielded better robustness and superior performance than standard TRCA by integrating the spatial filters of all target frequency.
With this regard, we propose a novel SSVEP frequency recognition method by utilizing the relationship between all spatial filters and individual templates, which is referred to as two-step TRCA (TSTRCA). Figure 1 shows the flowchart of the proposed method. The proposed TSTRCA method consists of two steps: (1) First-step: construction of the spatial filters using standard TRCA and individual templates by averaging SSVEP EEG signals except test data corresponding to the target frequencies; (2) Second-step: target identification based on an ensemble approach.
The first step aims to obtain spatial filters and individual templates from training data for each target frequency as done in standard TRCA. In the second step, we emphasize the relationships between the test data and individual template to yield the informative features for frequency recognition.
We newly formulate the parameter β i ,   k to further intensify the correlation coefficient with the feature corresponding to target frequency by redefining the relationship between test data and individual template. Specifically, the parameter β i ,   k , k = 1 ,   2 ,   ,   N f is defined as the correlation coefficient between the i -th individual template and the spatial filter for the k -th target frequency with the test data, which is given by
β i     =   β i , 0 β i ,     1 β i ,     2 β i ,     N f   =   ρ Y i , X ¯   ρ w 1 Y i ,   w 1 X ¯ ρ w 2 Y i ,   w 2 X ¯ ρ w N f Y i ,   w N f X ¯
where β i is the correlation vector which consists of β i ,   k , k = 1 ,   2 ,   ,   N f , and β i , 0 denotes the correlation coefficient between Y i and X ¯ without a spatial filter. Then, the correlation coefficient ρ ˜ i is obtained as a weighted sum of squares of β i ,   k as an ensemble.
ρ ˜ i =   k = 0 N f sign β i ,     k · β i ,     k 2
where sign · is used to reflect discriminative information from the negative value of β i ,   k .
Finally, the target identification of the proposed TSTRCA method is calculated as follows:
f t a r g e t =   max i ρ ˜ i , i = 1 ,   2 ,   ,   N f

2.4. Frequency Recognition Based on Filter Bank Approach

Recently, the filter bank approach, which extracts independent components by decomposing the frequency band of the input signal into multiple sub-bands using band-pass filters, has considerably contributed to improving the classification performance of BCI models [23,24]. For instance, the filter bank common spatial pattern (FBCSP) integrated the filter bank and the standard CSP, thus yielding an improved classification accuracy by correctly extracting the frequency bands that have prominently feature in the motor imagery BCI [23]. Similarly, the filter bank CCA (FBCCA) provided an improved frequency recognition performance compared to the conventional CCA [24]. Inspired by these results, we further adopted the filter bank approach to the proposed TSTRCA and compared the standard TRCA with the filter bank.
As introduced in [18], the filter bank approach in SSVEP-based BCI can effectively separate sub-band components, including independent information embedded in the harmonic frequency bands. In [24], depending on the type of sub-band components, the filter bank approach consists of three categories. Here, we use the third one, which is referred to as the M3 method. By using the M3 method, we can obtain multiple harmonic frequency bands with a high cut-off frequency. In the M3 method, the cut-off frequency range of sub-band is set between b × 8 Hz and 90 Hz, where b indicates the sub-band index. The zero-phase Chebyshev Type Ⅰ infinite impulse response (IIR) is used as a band-pass filter. After that, the arranged b th sub-band is applied to SSVEP EEG signals and learned spatial filters for each target frequency to generate a set of correlation vectors between test data and individual template. Finally, in order to recognize the target frequency, a set of correlation vectors are combined into a single metric using the linear combination method presented by [25] and is given by
ρ i =   s b · r i b , i = 1 ,   2 ,   ,   N f
where r i b is a set of correlation vectors according to b th sub-band and s b =   b 1.25 + 0.25 . Here, s b plays a role in compensating for the reduction in the SNR of SSVEP harmonic as the response frequency increases [24]. Then, the target identification is carried out using a given Equation (14).

3. Results

3.1. Performance Evaluation

This work was performed in the MATLAB environment on an Intel 3.60 GHz Core i7 with 64GB of RAM. In addition, we used the MATLAB codes, such as TRCA and filter bank method provided on the website (https://github.com/mnakanishi/TRCA-SSVEP (accessed on 12 February 2021)).
To evaluate the proposed TSTRCA method compared to the existing SSVEP frequency recognition methods such as CCA, ExtCCA, and TRCA, we used the classification accuracy and the information transfer rate (ITR) as two metrics to measure the frequency detection performance.
The ITR is described as the amount of information transmitted by a system’s output and given by [26]
ITR   =   log 2 N f + P log 2 P + 1 P log 2 1 P N f 1 × 60 T
where N f indicates the number of target frequency, P is the classification accuracy, and T is the selected TW for visual stimulation including gaze shifting time. In this work, we predetermined the gaze shifting time of 0.5 s as presented in [20] and evaluated ITR for TWs with an interval of 0.1 s from 0.2 to 1.0 s. In addition, we used one-way repeated measure ANOVA as a statistical analysis to determine the significant difference.
A leave-one-out cross validation was applied to estimate the performance of various SSVEP frequency recognition methods for test data. Among the six trials in SSVEP EEG signals as described in Section 2.1, the five trials were comprised of the training data and the remaining trial was used as the test data. This process was repeated six times and the average values of all six accuracies and ITRs were represented as an average accuracy and ITR corresponding to the target frequency.

3.2. Target Identification Performance

Figure 2a,b show the average accuracies and simulated ITR of CCA, ExtCCA, TRCA, and the proposed TSTRCA across all subjects for different TWs, respectively. Among 64 channels of EEG signals, we used 9 channels (Pz, PO5, PO3, POz, PO4, PO6, O1, Oz, O2) as in [11]. As can be seen in these figures, the standard TRCA was superior to CCA and ExtCCA and the proposed TSTRCA showed the best performance compared with other methods in terms of both average accuracy and ITR. Note that, for the short TWs from 0.2 to 0.5 s, the superiority of the proposed TSTRCA was clearer than the cases of TWs above 0.6 s. For example, the differences of average accuracy between TRCA and TSTRCA for the time window of 0.2 and 1.0 s are approximately 17% and 3%, respectively. In the figures, one-way repeated measures ANOVA analysis indicated that there were significant difference between four methods across all TWs in terms of average accuracy (TW = 0.2: F(3, 102) = 35.46, p < 0.001; TW = 0.3: F(3, 102) = 30.55, p < 0.001; TW = 0.4: F(3, 102) = 26.34, p < 0.001; TW = 0.5: F(3, 102) = 23.79, p < 0.001; TW = 0.6: F(3, 102) = 23.42, p < 0.001; TW = 0.7: F(3, 102) = 21.28, p < 0.001; TW = 0.8: F(3, 102) = 18.6, p < 0.001; TW = 0.9: F(3, 102) = 16.44, p < 0.001; and TW = 1.0: F(3, 102) = 16.13, p < 0.001). In addition, for simulated ITR, significant difference between four methods were similarly observed (TW = 0.2: F(3, 102) = 25.56, p < 0.001; TW = 0.3: F(3, 102) = 25.42, p < 0.001; TW = 0.4: F(3, 102) = 22.98, p < 0.001; TW = 0.5: F(3, 102) = 21.25, p < 0.001; TW = 0.6: F(3, 102) = 21.3, p < 0.001; TW = 0.7: F(3, 102) = 19.92, p < 0.001; TW = 0.8: F(3, 102) = 18.68, p < 0.001; TW = 0.9: F(3, 102) = 16.59, p < 0.001; and TW = 1.0: F(3, 102) = 16.65, p < 0.001).
Table 1 summarizes the statistical analysis results of the performance of each method between the different number of channels. ExtCCA, TRCA, and TSTRCA showed statistical difference as the number of channels increased. In addition, Table 2 exhibits the statistical analysis results of the performance between four methods when a different number of channels was used. It can be observed that except for the channels of 3, there was a statistically significant difference between the four methods.
To further verify the performance comparison of the four methods, we examined the average performance for the number of channels. Here, the TW was set to 0.8 s. Figure 3 illustrates the average accuracy of the four methods in cases where a different number of channels was used. In Figure 3a, we can observe that the average accuracy increased as more channels were used for all methods. The average accuracy comparison results for all methods in terms of the number of channels are shown in Figure 3b. As can be seen in the figure, across all cases of the number of channels, the proposed TSTRCA achieved the best average accuracy among the four methods. Note that the average accuracy of TSTRCA with a lower number of channels was comparable to ExCCA and TRCA or outperformed CCA, ExCCA, and TRCA.
Figure 4 and Figure 5 illustrate the average accuracy and ITR of two methods—TRCA and TSTRCA—across all subjects for a TW of 0.3 s. As shown in the figures, the average accuracy and ITR were improved for most subject. However, the amount of improvement of accuracy can be biased by several subjects. Thus, to avoid the impact of a specific subject on average performance in the comparison analysis, we further investigated median accuracy and ITR analysis, as depicted in Figure 6 and Figure 7, respectively.
In Figure 6 and Figure 7, we verified that the proposed TSTRCA outperformed other methods in terms of median accuracy and median ITR for TWs with an interval of 0.2 s from 0.3 to 0.9 s. It implies that the performance improvement of TSTRCA came from improvement on most subjects, not due to specific subjects.
Finally, to validate the effect of the filter bank approach, we examined the performance of TRCA and the proposed TSTRCA with filter banks at different TWs. Figure 8a,b indicate the average accuracy and simulated ITR of TRCA and TSTRCA with a filter bank, which are referred to as FBTRCA and FBTSTRCA, respectively. As in TRCA and TSTRCA without a filter bank in Figure 2, we observed that the TSTRCA with a filter bank—FBTSTRCA—was superior to the TRCA with a filter bank—FBTRCA—across all TWs. Through one-way repeated measure ANONA analysis, we confirmed a significant difference between the two methods.

4. Discussions

Due to simplicity and improved performance, the standard CCA and its variants, such as L1-MCCA and MsetCCA, have contributed to enhanced SSVEP-BCI. Followingly, the use of spatial filters in SSVEP-BCI research has enhanced the performance of target frequency recognition significantly [18,25].
Recently, the TRCA approach, which extracts the spatial filers with task-specific components, has yielded notable improvement compared to conventional SSVEP-BCI methods [18,27]. However, the TRCA approach is beneficial especially for a sufficiently long length of SSVEP EEG recordings. In this regard, the proposed TSTRCA has provided improved performance for the recognition of target frequencies for short and long time windows.
The proposed TSTRCA consists of two steps. First, it aims to develop the spatial filters and individual templates using training data. Second, the target frequencies are identified by applying an ensemble classifier. In the second step, all spatial filters are utilized to accentuate the features corresponding to the target frequencies. The first and second steps correspond to the training stage and test stage of TRCA, respectively. In Figure 2, Figure 6, and Figure 7 of Section 4, the results demonstrate that the proposed TSTRCA yields enhanced accuracy and ITR compared to conventional SSVEP methods. While ExtCCA utilizes two reference signals, TSTRCA uses a single reference, thus implying its simplicity.
Furthermore, we carried out a performance comparison of the proposed TSTRCA and conventional SSVEP methods in terms of precision, recall, and F1-score, shown in Table 3. These metrics are obtained by averaging each metric over all subjects and all trials with a TW of 0.5 s. As can be seen, the TSTRCA shows more robust performance than other methods in short time windows.
Recently, several fusion-based SSVEP-BCI studies have shown the remarkable performance of frequency recognition. Liu et al. [19] developed FoCCA, which fuses all correlation coefficients of the standard CCA. While FoCCA represented approximately 80% accuracy with a TW of 2 s, the proposed TSTRCA shows an 83.71% accuracy with a TW of 0.7 s. Besides, the average accuracy and simulated ITR of fusion of maximum signal fraction analysis (FoMSFA) [28] were less than 31% and 100 bits/min for a TW of 0.2 s, respectively. Compared to these results, TSTRCA demonstrates 42.06% accuracy and 120.51 bits/min ITR using the same length of TW. This comparison suggests the superiority of TSTRCA over fusion-based SSVEP-BCI methods.
For practical use, the improved performance of the proposed TSTRCA with a short time window suggests its promising capability as a new communication tool for both healthy and disabled people. Thus, SSVEP EEG signals would play a role in daily life, such as the use of photoplethysmography (PPG) and electrocardiogram (ECG) [29,30,31].
The proposed TSTRCA was designed and evaluated on the offline experiment. Thus, future works should be conducted to (1) establish the real-time SSVEP-BCI system using TSTRCA, (2) explore how different spatial filtering mechanisms address the trade-off between computational complexity and the performance for SSVEP frequency recognition, and (3) construct the SSVEP-BCI EEG dataset extracted from the various subjects and environments to pursue a general-purpose SSVEP-BCI framework by extending our work.

5. Conclusions

We presented a novel frequency recognition method for SSVEP-based BCI based on the TRCA method. The proposed TSTRCA accentuates the features corresponding to target frequencies by: (1) redefining a correlation vector based on the spatial filters of all target frequencies, (2) emphasizing the relationship between test data and individual templates using an ensemble classifier. Through validation using the SSVEP benchmark dataset, we confirmed that the proposed TSTRCA outperforms the existing SSVEP frequency recognition methods including the standard TRCA in terms of average accuracy and simulated ITR. Furthermore, we introduced the proposed TSTRCA with a filter bank, which is called FBTSTRCA, and validated its superior performance over the standard TRCA with a filter bank. The experimental results suggest that the proposed TSTRCA can play an important role for SSVEP frequency recognition since it possesses efficacy in frequency recognition in case of a short time window. These properties of TSTRCA imply the suitability as a promising frequency recognition strategy for SSVEP-based BCI applications.

Author Contributions

H.K.L. and Y.-S.C. conceived and designed the methodology, and were responsible for analyzing and writing the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT) (No. NRF-2019R1F1A1045607) and Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government(MSIT) (No. 2021-0-00013, Development of AI Technology for Detecting Korean Violence Situation using Multimodal Data), and the Research Grant of Kwangwoon University in 2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nicolas-Alonso, L.F.; Gomez-Gil, J. Brain Computer Interfaces, a Review. Sensors 2012, 12, 1211–1279. [Google Scholar] [CrossRef]
  2. Mak, J.N.; Wolpaw, J.R. Clinical Applications of Brain-Computer Interfaces: Current State and Future Prospects. IEEE Rev. Biomed. Eng. 2009, 2, 187–199. [Google Scholar] [CrossRef] [Green Version]
  3. Chatterjee, R.; Maitra, T.; Hafizul Islam, S.; Hassan, M.M.; Alamri, A.; Fortino, G. A novel machine learning based feature selection for motor imagery EEG signal classification in Internet of medical things environment. Future Gener. Comput. Syst. 2019, 98, 419–434. [Google Scholar] [CrossRef]
  4. Lu, Y.; Bi, L. EEG Signals-Based Longitudinal Control System for a Brain-Controlled Vehicle. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 323–332. [Google Scholar] [CrossRef]
  5. Liu, M.; Wang, K.; Chen, X.; Zhao, J.; Chen, Y.; Wang, H.; Wang, J.; Xu, S. Indoor Simulated Training Environment for Brain-Controlled Wheelchair Based on Steady-State Visual Evoked Potentials. Front. Neurorobot. 2020, 13, 101. [Google Scholar] [CrossRef] [PubMed]
  6. Chen, X.; Chen, Z.; Gao, S.; Gao, X. A high-ITR SSVEP-based BCI speller. Brain-Comput. Interfaces 2014, 1, 181–191. [Google Scholar] [CrossRef]
  7. İşcan, Z.; Nikulin, V.V. Steady state visual evoked potential (SSVEP) based brain-computer interface (BCI) performance under different perturbations. PLoS ONE 2018, 13, e0191673. [Google Scholar] [CrossRef] [Green Version]
  8. Martišius, I.; Damaševičius, R. A Prototype SSVEP Based Real Time BCI Gaming System. Comput. Intell. Neurosci. 2016, 2016, 3861425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Shao, L.; Zhang, L.; Belkacem, A.N.; Zhang, Y.; Chen, X.; Li, J.; Liu, H. EEG-Controlled Wall-Crawling Cleaning Robot Using SSVEP-Based Brain-Computer Interface. J. Healthc. Eng. 2020, 2020, 6968713. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Lin, Z.; Zhang, C.; Wu, W.; Gao, X. Frequency recognition based on canonical correlation analysis for SSVEP-based BCIs. IEEE Trans. Biomed. Eng. 2007, 54, 1172–1176. [Google Scholar] [CrossRef]
  11. Bin, G.; Gao, X.; Yan, Z.; Hong, B.; Gao, S. An online multi-channel SSVEP-based brain-computer interface using a canonical correlation analysis method. J. Neural Eng. 2009, 6, 046002. [Google Scholar] [CrossRef] [PubMed]
  12. Nakanishi, M.; Wang, Y.; Wang, Y.-T.; Mitsukura, Y.; Jung, T.-P. A high-speed brain speller using steady-state visual evoked potentials. Int. J. Neural Syst. 2014, 24, 1450019. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Zhou, G.; Jin, J.; Wang, M.; Wang, X.; Cichocki, A. L1-Regularized Multiway Canonical Correlation Analysis for SSVEP-Based BCI. IEEE Trans. Neural Syst. Rehabil. Eng. 2013, 21, 887–896. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Zhou, G.; Jin, J.; Wang, X.; Cichocki, A. Frequency recognition in ssvep-based bci using multiset canonical correlation analysis. Int. J. Neural Syst. 2013, 24, 1450013. [Google Scholar] [CrossRef]
  15. Kumar, G.R.K.; Reddy, M.R. Latent common source extraction via a generalized canonical correlation framework for frequency recognition in SSVEP based brain–computer interfaces. J. Neural Eng. 2019, 16, 046004. [Google Scholar] [CrossRef] [PubMed]
  16. Nakanishi, M.; Wang, Y.; Wang, Y.-T.; Jung, T.-P. A Comparison Study of Canonical Correlation Analysis Based Methods for Detecting Steady-State Visual Evoked Potentials. PLoS ONE 2015, 10, e0140703. [Google Scholar] [CrossRef] [Green Version]
  17. Tanaka, H.; Katura, T.; Sato, H. Task-related component analysis for functional neuroimaging and application to near-infrared spectroscopy data. NeuroImage 2013, 64, 308–327. [Google Scholar] [CrossRef]
  18. Nakanishi, M.; Wang, Y.; Chen, X.; Wang, Y.; Gao, X.; Jung, T. Enhancing Detection of SSVEPs for a High-Speed Brain Speller Using Task-Related Component Analysis. IEEE Trans. Biomed. Eng. 2018, 65, 104–112. [Google Scholar] [CrossRef] [PubMed]
  19. Liu, T.; Zhang, Y.; Wang, L.; Li, J.; Xu, P.; Yao, D. Fusing Canonical Coefficients for Frequency Recognition in SSVEP-Based BCI. IEEE Access 2019, 7, 52467–52472. [Google Scholar] [CrossRef]
  20. Wang, Y.; Chen, X.; Gao, X.; Gao, S. A Benchmark Dataset for SSVEP-Based Brain–Computer Interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1746–1752. [Google Scholar] [CrossRef] [PubMed]
  21. Yang, X.; Weifeng, L.; Liu, W.; Tao, D. A Survey on Canonical Correlation Analysis. IEEE Trans. Knowl. Data Eng. 2019. [Google Scholar] [CrossRef]
  22. Wang, Y.; Nakanishi, M.; Wang, Y.-T.; Jung, T.-P. Enhancing detection of steady-state visual evoked potentials using individual training data. In Proceedings of the 2014 36th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Chicago, IL, USA, 26–30 August 2014; pp. 3037–3040. [Google Scholar]
  23. Ang, K.K.; Chin, Z.Y.; Zhang, H.; Guan, C. Filter Bank Common Spatial Pattern (FBCSP). In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Hong Kong, China, 1–8 June 2008; pp. 2390–2397. [Google Scholar]
  24. Chen, X.; Wang, Y.; Gao, S.; Jung, T.-P.; Gao, X. Filter bank canonical correlation analysis for implementing a high-speed SSVEP-based brain–computer interface. J. Neural Eng. 2015, 12, 046008. [Google Scholar] [CrossRef]
  25. Zhang, Y.; Guo, D.; Li, F.; Yin, E.; Zhang, Y.; Li, P.; Zhao, Q.; Tanaka, T.; Yao, D.; Xu, P. Correlated Component Analysis for Enhancing the Performance of SSVEP-Based Brain-Computer Interface. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 948–956. [Google Scholar] [CrossRef] [PubMed]
  26. McFarland, D.J.; Sarnacki, W.A.; Wolpaw, J.R. Brain–computer interface (BCI) operation: Optimizing information transfer rates. Biol. Psychol. 2003, 63, 237–251. [Google Scholar] [CrossRef]
  27. Mary Judith, A.; Baghavathi Priya, S. Multiset Task Related Component Analysis (M-TRCA) for SSVEP Frequency Recognition in BCI. J. Ambient Intell. Human. Comput. 2020. [Google Scholar] [CrossRef]
  28. Li, Z.; Liu, K.; Deng, X.; Wang, G. Spatial Fusion of Maximum Signal Fraction Analysis for Frequency Recognition in SSVEP-Based BCI. Biomed. Signal Process. Control 2020, 61, 102042. [Google Scholar] [CrossRef]
  29. Cvetkovic, D.; Holland, G.; Cosic, I. The Relationship between EEG, PPG and Oxymetric Signal Responses During the OSAH Events: A Pilot Sleep Study. In Proceedings of the 5th World Sleep Congress of the WFSRSMS (WorldSleep07), Cairns, QLD, Australia, 2–6 September 2007. [Google Scholar]
  30. Baldoumas, G.; Peschos, D.; Tatsis, G.; Chronopoulos, S.K.; Christofilakis, V.; Kostarakis, P.; Varotsos, P.; Sarlis, N.V.; Skordas, E.S.; Bechlioulis, A.; et al. A Prototype Photoplethysmography Electronic Device that Distinguishes Congestive Heart Failure from Healthy Individuals by Applying Natural Time Analysis. Electronics 2019, 8, 1288. [Google Scholar] [CrossRef] [Green Version]
  31. Baldoumas, G.; Peschos, D.; Tatsis, G.; Christofilakis, V.; Chronopoulos, S.K.; Kostarakis, P.; Varotsos, P.A.; Sarlis, N.V.; Skordas, E.S.; Bechlioulis, A.; et al. Remote Sensing Natural Time Analysis of Heartbeat Data by Means of a Portable Photoplethysmography Device. Int. J. Remote Sens. 2021, 42, 2292–2302. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed TSTRCA method. In the first step, the standard TRCA is employed to compute the spatial filters for each target frequency in the training data X i   N c × N s × N t ,   i = 1 ,   2 ,   ,   N f and obtain the individual templates, i.e., Y i ,   i = 1 ,   2 ,   ,   N f , by group averaging across multiple training blocks. Here, the remaining blocks are stored as test data. Followingly, in the second step, the obtained spatial filter w i for each target frequency is used in yielding the correlation coefficients between the test data X ¯ and the individual template Y i . We repeat this procedure for test data and all individual templates to compute the parameters β i ,   k ,   k = 1 ,   2 ,   ,   N f . β i ,   0 denotes the correlation coefficient without the spatial filter.
Figure 1. Flowchart of the proposed TSTRCA method. In the first step, the standard TRCA is employed to compute the spatial filters for each target frequency in the training data X i   N c × N s × N t ,   i = 1 ,   2 ,   ,   N f and obtain the individual templates, i.e., Y i ,   i = 1 ,   2 ,   ,   N f , by group averaging across multiple training blocks. Here, the remaining blocks are stored as test data. Followingly, in the second step, the obtained spatial filter w i for each target frequency is used in yielding the correlation coefficients between the test data X ¯ and the individual template Y i . We repeat this procedure for test data and all individual templates to compute the parameters β i ,   k ,   k = 1 ,   2 ,   ,   N f . β i ,   0 denotes the correlation coefficient without the spatial filter.
Sensors 21 01315 g001
Figure 2. Average results of four methods without filter bank. (a) Average accuracy and (b) simulated ITR across 35 subjects for different time windows (TWs). Error bars represent standard errors. The asterisks indicate significant difference between four methods obtained by one-way repeated measures ANOVA (p < 0.001).
Figure 2. Average results of four methods without filter bank. (a) Average accuracy and (b) simulated ITR across 35 subjects for different time windows (TWs). Error bars represent standard errors. The asterisks indicate significant difference between four methods obtained by one-way repeated measures ANOVA (p < 0.001).
Sensors 21 01315 g002
Figure 3. Average accuracy across 35 subjects for four methods in cases of the different number of channels with TW of 0.8 s. (a) Methods; (b) The number of channels. Error bars represent standard errors.
Figure 3. Average accuracy across 35 subjects for four methods in cases of the different number of channels with TW of 0.8 s. (a) Methods; (b) The number of channels. Error bars represent standard errors.
Sensors 21 01315 g003aSensors 21 01315 g003b
Figure 4. Average accuracy across all subjects for TW of 0.3 s using TRCA and TSTRCA. The error bars represent standard errors. Green and blue bars represent the average accuracies of TRCA and TSTRCA, respectively.
Figure 4. Average accuracy across all subjects for TW of 0.3 s using TRCA and TSTRCA. The error bars represent standard errors. Green and blue bars represent the average accuracies of TRCA and TSTRCA, respectively.
Sensors 21 01315 g004
Figure 5. Average ITR across all subjects for TW of 0.3 s using TRCA and TSTRCA. The error bars represent standard errors. Green and blue bars represent the average ITRs of TRCA and TSTRCA, respectively.
Figure 5. Average ITR across all subjects for TW of 0.3 s using TRCA and TSTRCA. The error bars represent standard errors. Green and blue bars represent the average ITRs of TRCA and TSTRCA, respectively.
Sensors 21 01315 g005
Figure 6. Median accuracy for TWs with an interval of 0.2 s from 0.3 s to 0.9 s. (a) 0.3 s time window. (b) 0.5 s time window. (c) 0.7 s time window. (d) 0.9 s time window. Here, on each box, the central red line indicates the median, and the bottom and top edges of the box refer to the 25th and 75th percentiles, respectively. Whiskers extend to the maximum or minimum performance not considered by outliers, and outliers are denoted by ‘+’.
Figure 6. Median accuracy for TWs with an interval of 0.2 s from 0.3 s to 0.9 s. (a) 0.3 s time window. (b) 0.5 s time window. (c) 0.7 s time window. (d) 0.9 s time window. Here, on each box, the central red line indicates the median, and the bottom and top edges of the box refer to the 25th and 75th percentiles, respectively. Whiskers extend to the maximum or minimum performance not considered by outliers, and outliers are denoted by ‘+’.
Sensors 21 01315 g006
Figure 7. Median ITR for TWs with an interval of 0.2 s from 0.3 s to 0.9 s. (a) 0.3 s time window. (b) 0.5 s time window. (c) 0.7 s time window. (d) 0.9 s time window. Here, on each box, the central red line indicates the median, and the bottom and top edges of the box refer to the 25th and 75th percentiles, respectively. Whiskers extend to the maximum or minimum performance not considered by outliers, and outliers are denoted by ‘+’.
Figure 7. Median ITR for TWs with an interval of 0.2 s from 0.3 s to 0.9 s. (a) 0.3 s time window. (b) 0.5 s time window. (c) 0.7 s time window. (d) 0.9 s time window. Here, on each box, the central red line indicates the median, and the bottom and top edges of the box refer to the 25th and 75th percentiles, respectively. Whiskers extend to the maximum or minimum performance not considered by outliers, and outliers are denoted by ‘+’.
Sensors 21 01315 g007
Figure 8. Average results of TRCA and TSTRCA with filter bank. (a) Averaged accuracy and (b) simulated ITR across 35 subjects for different TWs. Error bars represent standard errors. The asterisks indicate significant difference between four methods obtained by one-way repeated measures ANOVA (p < 0.001).
Figure 8. Average results of TRCA and TSTRCA with filter bank. (a) Averaged accuracy and (b) simulated ITR across 35 subjects for different TWs. Error bars represent standard errors. The asterisks indicate significant difference between four methods obtained by one-way repeated measures ANOVA (p < 0.001).
Sensors 21 01315 g008
Table 1. Statistical analysis results of average accuracy and ITR between the different number of channels for each method.
Table 1. Statistical analysis results of average accuracy and ITR between the different number of channels for each method.
Method
CCAExtCCATRCATSTRCA
AccuracyF(6, 204)1.395.594.227.43
p0.22<0.001<0.001<0.001
ITRF(6, 204)1.295.574.17.74
p0.26<0.001<0.001<0.001
Table 2. Statistical analysis results of average accuracy and ITR between four methods for the different number of channels.
Table 2. Statistical analysis results of average accuracy and ITR between four methods for the different number of channels.
Channels
3456789
AccuracyF(3, 102)3.265.7310.613.6315.0717.218.6
p0.0020.001<0.001<0.001<0.001<0.001<0.001
ITRF(3, 102)3.666.2110.5512.814.5316.8318.68
p0.001<0.001<0.001<0.001<0.001<0.001<0.001
Table 3. Comparison of performance (Precision, Recall, and F1-score) of SSVEP frequency recognition for TW of 0.5 s.
Table 3. Comparison of performance (Precision, Recall, and F1-score) of SSVEP frequency recognition for TW of 0.5 s.
Method
(Average ± std. dev. in %)
CCAExtCCATRCATSTRCA
Precision22.19 ± 21.6342.73 ± 19.8648.57 ± 25.6264.33 ± 22.81
Recall28.94 ± 23.2452.43 ± 19.8456.32 ± 24.8971.92 ± 20.59
F1-score25.14 ± 22.4546.94 ± 20.0252 ± 25.4667.78 ± 21.94
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, H.K.; Choi, Y.-S. Enhancing SSVEP-Based Brain-Computer Interface with Two-Step Task-Related Component Analysis. Sensors 2021, 21, 1315. https://doi.org/10.3390/s21041315

AMA Style

Lee HK, Choi Y-S. Enhancing SSVEP-Based Brain-Computer Interface with Two-Step Task-Related Component Analysis. Sensors. 2021; 21(4):1315. https://doi.org/10.3390/s21041315

Chicago/Turabian Style

Lee, Hyeon Kyu, and Young-Seok Choi. 2021. "Enhancing SSVEP-Based Brain-Computer Interface with Two-Step Task-Related Component Analysis" Sensors 21, no. 4: 1315. https://doi.org/10.3390/s21041315

APA Style

Lee, H. K., & Choi, Y. -S. (2021). Enhancing SSVEP-Based Brain-Computer Interface with Two-Step Task-Related Component Analysis. Sensors, 21(4), 1315. https://doi.org/10.3390/s21041315

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop