Next Article in Journal
Few-Shot Image Classification of Crop Diseases Based on Vision–Language Models
Previous Article in Journal
Design of an FPGA-Based Controller for Fast Scanning Probe Microscopy
Previous Article in Special Issue
Algorithms for Liver Segmentation in Computed Tomography Scans: A Historical Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing EEG-Based MI-BCIs with Class-Specific and Subject-Specific Features Detected by Neural Manifold Analysis

by
Mirco Frosolone
1,
Roberto Prevete
2,
Lorenzo Ognibeni
1,3,
Salvatore Giugliano
2,
Andrea Apicella
2,
Giovanni Pezzulo
1 and
Francesco Donnarumma
1,*
1
Institute of Cognitive Sciences and Technologies, National Research Council, Via Gian Domenico Romagnosi, 00196 Rome, Italy
2
Department of Electrical Engineering and Information Technology (DIETI), University of Naples Federico II, 80125 Naples, Italy
3
Department of Computer, Control and Management Engineering ‘Antonio Ruberti’ (DIAG), Sapienza University of Rome, 00185 Rome, Italy
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 6110; https://doi.org/10.3390/s24186110
Submission received: 3 August 2024 / Revised: 12 September 2024 / Accepted: 18 September 2024 / Published: 21 September 2024
(This article belongs to the Special Issue Biomedical Sensing and Bioinformatics Processing)

Abstract

:
This paper presents an innovative approach leveraging Neuronal Manifold Analysis of EEG data to identify specific time intervals for feature extraction, effectively capturing both class-specific and subject-specific characteristics. Different pipelines were constructed and employed to extract distinctive features within these intervals, specifically for motor imagery (MI) tasks. The methodology was validated using the Graz Competition IV datasets 2A (four-class) and 2B (two-class) motor imagery classification, demonstrating an improvement in classification accuracy that surpasses state-of-the-art algorithms designed for MI tasks. A multi-dimensional feature space, constructed using NMA, was built to detect intervals that capture these critical characteristics, which led to significantly enhanced classification accuracy, especially for individuals with initially poor classification performance. These findings highlight the robustness of this method and its potential to improve classification performance in EEG-based MI-BCI systems.

1. Introduction

The field of Brain–Computer Interfaces (BCIs) has witnessed significant advancements over the past decades, forecasting exceptional achievements in bioengineering applications [1,2,3]. BCIs rely on the measurement of neural activity as a method to determine the user’s intentions. Suitable command sequences can be produced using BCI applications, including those based on non-invasive methods such as functional magnetic resonance imaging (fMRI), magnetoencephalography (MEG), and electroencephalography (EEG). Among these, the most widely used BCI systems leverage EEG signals due to their simplicity, affordability, and ability to support effective real-time implementations thanks to their intrinsic high temporal resolution [4,5]. BCIs are particularly effective in motor imagery (MI) tasks, where users imagine performing motor movements to control external devices. Typically, in an MI-based BCI, the user is required to imagine the motor action related to their body parts (e.g., left hand, right hand, feet, or tongue). It is widely accepted that the mental imagination of movements involves brain regions similar to those engaged in the actual execution of these movements [6,7]. Differently from the physical execution, in MI the movement is blocked at a corticospinal level. However, functional brain imaging studies demonstrated that patterns of brain activity during motor imagery are similar to those observed during actual movement execution. Specifically, an activation is observed in various structures involved in the early stages of motor control, such as motor programming and planning. For example, imagining right-hand or left-hand movements activates the contralateral hand area, the top-central for feet, and the parietofrontal for tongue MI [8,9,10,11]. This neural activation can be readily detected through the EEG signals [12]. MI-BCI systems prominently depend on sensorimotor rhythms (SMR), event-related potentials (ERPs), visually evoked potentials (VEPs), and slow cortical potentials (SCPs). Among these, SMR-based BCI systems offer significant freedom in real-time control and motor imagery activities, such as movements of the tongue, hand, arm, and feet [13]. Decoding the recorded EEG signals and mapping the corresponding MI to a command for an external device is the primary challenge in EEG-based MI-BCI systems. The EEG-based MI-BCIs are crucial for designing systems that enable specific activities such as controlling and governing wheelchairs [14], home appliances, speech synthesizers, robotic prostheses [15], post-stroke rehabilitation [16,17], digital computers, and competitive or collaborative games [18,19,20,21].
An EEG-based MI-BCI system encompasses a pipeline involving various steps [18]: (a) Signal acquisition: EEG signals are collected from the scalp using specialized hardware while the user performs MI tasks. (b) Signal processing: the goal is to increase the signal-to-noise ratio of the weak EEG signals, which are often contaminated by artifacts and interferences such as muscle movements, eye blinks, heartbeats, and powerline noise. (c) Feature extraction and selection: this involves extracting relevant properties in the time domain, frequency domain, or time–frequency domain and selecting those that are most successful in representing the task. (d) Classification: the extracted features are used to decode the EEG signals. (e) Control: proper commands are then sent to an external device, such as a wheelchair, based on the decoded EEG signals.
In this work, the focus is on the crucial step in the BCI pipeline: feature extraction. Extracted suitable features correspond to vital information encapsulated in the signal. Various techniques can be employed to extract the most informative parts of the input signals, including fast Fourier transform (FFT) [22], autoregressive model (AR) [23], Common Spatial Pattern (CSP) [24], and Wavelet Transform (WT) [25]. This step corresponds to a Neural Manifold Analysis (NMA) [26], where EEG signals are denoised and reorganized, reducing the high-dimensional input signal space to a more manageable lower-dimensional space [27]. The classification process then converts the features encoded in the manifold generated by the feature extractor into commands. The more readable this manifold is, the easier the classification becomes. BCI classification techniques translate these discriminatory features into decoded motor activities such as tongue movement, left–right movement, and foot movement. Several classification methods like artificial neural networks (ANN), linear discriminant analysis (LDA), k-nearest neighbors (k-NN), support vector machines (SVM), Gaussian Naive Bayes (GNB), and deep learning (DL) have been used for MI-BCI systems [28]. A major challenge in BCIs is that different users have varying neuronal responses to the same stimulus, and even the same user can exhibit different neuronal responses to the same stimulus at different times or conditions. Additionally, calibrating the BCI system requires acquiring a large number of subject-specific labeled training examples for each new subject, which is both time-consuming and expensive.
To address these issues, different approaches involving Transfer Learning have been explored [29,30,31,32]. Those studies use data extracted from one or more source domains to help construct the representation manifold in the target domain, effectively addressing these problems. In [33], the authors employ DL models for feature extraction in EEG signals, achieving notable improvements in classification performance through the use of convolutional neural networks. In [34], a DL approach is used to select samples in the source neuronal domain that are closest to the data in the target manifold domain, assigning them high correlation weights. In [35], the authors adapt a DL approach to utilize pre-trained features from different datasets. Additionally, in [36], a method to extract cross-channel specific-mutual features is proposed, further enhancing cross-subject generalization.
Nevertheless, several recent studies have explored various methods to improve EEG-based BCI systems, focusing on feature extraction and classification while highlighting the importance of addressing cross-subject variability [4,37]. For instance, in [38], the authors proposed a method for regularizing CSP features to enhance the robustness of BCI systems against subject-specific variations. This is crucial since EEG signals often exhibit large inter-subject variability, making consistent classification challenging. Importantly, studies utilized NMA techniques, such as Principal Component Analysis (PCA), to detect spatio-temporal features in EEG inputs, which are useful for increasing task classification accuracy [39,40]. Additionally, works like [41] focused on time–frequency feature extraction for MI tasks, emphasizing the importance of selecting optimal time intervals to boost classification accuracy. Furthermore, ref. [42] introduced a subject-independent approach for MI classification, demonstrating the potential of using data from multiple subjects to improve overall classification accuracy. CSP is a widely used technique for spatial filtering in EEG signals, aiming to maximize the variance difference between two classes. Despite its success, CSP’s reliance on data from the same subject and specific time intervals often limits its generalizability across different subjects or sessions.
Consequently, feature extraction in motor imagery has evolved from single-domain (time, frequency, or spatial) approaches to multi-domain fusion, especially combining spatial and frequency domain information. Thus, methods for NMA, are employed to determine patterns of covariance in participants’ responses, extracting separate features for each class and subject.
In this study, it is proposed a novel approach leveraging NMA to identify optimal time intervals for feature extraction, which are critical for improving classification performance [37]. NMA involves analyzing the EEG signal in a multi-dimensional feature space to detect intervals that capture class-specific and subject-specific characteristics. By applying state-of-the-art feature extraction algorithms within these identified ranges, the goal is to improve the discriminative power of the extracted features. Furthermore, this work addresses the challenge of subjects with poor classification performance by cross-validating the extracted features across different subjects. By incorporating features from subjects with high classification accuracy, significant improvements are achieved for subjects that initially exhibit poor performance. Traditionally, NMA is used to derive a reduced manifold, enabling more effective feature extraction and, consequently, better classification. In the present paper, however, an innovative use of NMA is proposed: using a separability measure on the neural manifolds, it is possible to identify specific temporal segments of the EEG signal, where, based on manifold analysis, we can extract more relevant features. This method allows for the construction of multiple manifolds over specific temporal segments, which can then be combined into a more complex overall manifold. The resulting manifold demonstrates enhanced discriminability compared with traditional approaches, making it more effective for classification tasks. To demonstrate the effectiveness of the proposed approach, particularly suited for the classification of oscillatory components of SMR during MI tasks, results from two datasets are presented [43,44,45]: Graz Dataset 2b, which highlights the robustness of the NMA pipelines in binary classification, and Graz Dataset 2a, a key benchmark for motor imagery, including a more complex four-class classification problem. This work develops NMA processing pipelines by building on the insights gained from the winners of BCI Competition IV [46,47], while also integrating recent advances in deep learning methods [33]. The performance results in both two-class and four-class motor imagery tasks in the selected datasets highlight the significant potential of NMA to improve cross-class and cross-subject classification.
The present paper is organized as follows: Section 2, Materials and Methods, provides a detailed explanation of the MNA pipelines used in the experiments. This is followed by the results section (Section 3), which is divided into two parts: Section 3.1 presents results from Graz Dataset 2b, while Section 3.2 focuses on Graz Dataset 2a. Finally, Section 4 summarizes the findings achieved with the proposed method and discusses potential avenues for future research and development.

2. Materials and Methods

Our starting point is the FBCSP Algorithm, which was the winner of the Graz BCI Competition [44,45] and has since become a gold standard in the BCI community. As previously mentioned, the core components of an EEG BCI pipeline include several critical steps. In this paper, the following FBCSP pipeline is employed:
  • Signal processing: Filter Bank.
  • Feature extraction: common spatial pattern algorithm.
  • Feature selection: mutual information-based best individual feature.
  • Classification: quadratic discriminant analysis classifier.
In Figure 1, a depiction of the process flow of a BCI system, including its key modules, is presented. Additionally, in this paper, the standard pipeline was enhanced with further NMA modules, leading to notable improvements in classification accuracy. In the following subsections, the modules of the system used to test the methodology will be described.

2.1. Signal Processing: Filter Bank

After acquisition, the input EEG signal can be represented as a time series X R N c h × T , where N c h is the number of channels and T is the number of time samples acquired. The signal undergoes the first of a series of preprocessing steps. A filter bank comprising multiple ( B = 9 ) Chebyshev Type II band-pass filters is utilized, i.e., nine Chebyshev Type II filters, each one designed for a specific band, as described in [45]. Each filter has a 4 Hz-wide pass band, resulting in nine non-overlapping bands from 4 Hz to 40 Hz, denoted as { ( 4 b , 4 b + 4 ) } b = 1 B . The best bands are selected on a subject-by-subject basis during feature selection. The attenuation in the stop band is set to −20 dB, and the filter order is 4, achieving a sharp roll-off in the frequency response.
Thus, starting with an input signal X , the result of the filtering is an output χ R N c h × T × B . The chosen band-pass frequency ranges allow for a stable frequency response and coverage of the 4–40 Hz range.

2.2. Feature Extraction: Common Spatial Pattern Algorithm

The CSP algorithm creates a reduced space to maximize the separability of labeled samples. Typically, CSP approaches are applied when two classes are present, but multi-class extensions are possible [48]. In this work, the one-versus-rest (OVR) approach is adopted [45], which allows for discrimination among an arbitrary number K of classes. In the OVR–CSP method, selecting m CSP components enables the construction of projection matrices W k b R N c h × 2 m , one for each band b and for each class k, which can then be applied to the filtered data χ . These projection matrices are derived from a sample set of labeled data (training set) during the training phase. Consequently, each filter band and class has an associated projection matrix W k b . In the transformed space, the first m CSP components have maximum variance associated with class k and minimum variance for the remaining classes, while the last m components have minimum variance for class k and maximum variance for others; see [45,48] for details.
The N samples in the input set can be partitioned into K sets { Π k } k = 1 K , where Π k denotes the sample set of the k-th class containing N k data points. For each band, selecting the samples χ n belonging to class k, the class covariance matrix can be computed as
S k = 1 N k χ n Π k χ n χ n T
and the corresponding composite covariance matrix is S = k S k . The complete projection matrix W k is obtained by solving the eigenvalue decomposition problem:
S k W k = S W k Λ k
where Λ k is a diagonal matrix of the eigenvalues, sorted in ascending order, and W k consists of the corresponding eigenvectors. The final projection matrix W k b is obtained by selecting the first m and the last m columns of W k (in the reported tests m = 2 , see [45]).
Each time series χ n R N c h × T for a single trial and band is projected into a new space using W k b , resulting in V = ( W k b ) T χ n R 2 m × T . From the projected trial V, a covariance matrix A = V V T is computed. From this matrix A, a vector is obtained by selecting the elements of the diagonal (the variances) a = A 1 1 , A 2 2 , , A 2 m 2 m . Renormalizing this vector and tacking the logarithmic values a ˜ k = l o g a / a , a feature vector a k of 2 m elements for each class k is obtained. Concatenating features from each class results in an element of the feature space:
z ˜ b = a ˜ 1 , , a ˜ k
with z ˜ b R m × 2 K , i.e., a CSP feature with 2 K values. Notice that these features are related to one band b, thus the total features are z ˜ = z ˜ 1 , , z ˜ B , organized in a space Z ˜ R 2 m B × K . These F = 2 m B features are subjected to the feature selection phase, described in the next subsection.

2.3. Feature Selection: Mutual Information-Based Best Individual Feature

The feature selection algorithm is crucial for identifying discriminative features in Z ˜ for the subject’s task. The mutual information-based best individual feature (MIBIF) was the winning algorithm of the BCI Competition [45,49]. Feature selection is executed on the training set by selecting the most discriminative CSP features based on the mutual information computed between each feature and the corresponding motor imagery classes. A parameter is chosen to select a number of D features. The N samples in the manifold can be partitioned into K sets { Π k } k = 1 K , where Π k denotes the sample set of the k-th class containing N k data points.
A set of F = 2 m · B · K features, z ˜ n = ( z ˜ n 1 , , z ˜ n F ) R F , is associated with an input trial x n belonging to a specific class k. For each j [ 1 , F ] , the mutual information M z ˜ n j ; k with each class label k can be computed. It is possible to define M z ˜ n j ; k = H k H k | z ˜ n j , where H k = k = 1 K P ( k ) log 2 P ( k ) is the entropy over the choice of k, and the conditional entropy H k | z ˜ n j = k = 1 K p k | z ˜ n j log 2 p k | z ˜ n j .
The conditional probability p k | z ˜ n j of class k given the j-th feature z ˜ n j is estimated. Initially, the probability is constructed such that, given a class k, the j-th feature z ˜ n j is found:
p z ˜ n j | k = 1 N k z ˜ k Π k ϕ z ˜ n j z ˜ k j , h
where ϕ x , h = 1 2 π e x 2 / 2 h 2 is a Gaussian kernel with an attenuation parameter h. Then the probability is computed using Bayes’ theorem, p k | z ˜ n j p z ˜ n j | k p ( k ) , where p ( k ) = N k / N . The resulting selected features z lie in the manifold Z R D . In the current implementation, D = 4 · K . To keep a conservative approach, the twin CSP feature is preserved if it is not included in this set. Thus, in the case where none of the twin CSP features were already included in the set D could raise to D = 8 · K (see [49]).

2.4. Classification: Quadratic Discriminant Analysis

Quadratic discriminant analysis (QDA) is a widely used approach for classification [50]. Given a manifold Z R D with N data samples, each data sample z n Z belongs to one of K classes and is represented by a one-hot encoded label vector d n such that if z n belongs to the k-th class, then d n ( k ) = 1 , where d n { 0 , 1 } K . Once again, the N samples in the manifold can be partitioned into K sets { Π k } k = 1 K , where Π k denotes the sample set of the k-th class containing N k data points. QDA models each class with a multivariate Gaussian distribution:
P z n | d n ( k ) = 1 = N k z n ; μ k , Σ k
where μ k and Σ k are the mean vector and covariance matrix for each class, respectively. The decision boundaries in QDA are designed to enhance class separability by maximizing the within-class scatter:
D w i t h i n = k = 1 K z n Π k ( z n μ k ) ( z n μ k )
and the between-class scatter:
D b e t w e e n = k = 1 K N k ( μ k μ ) ( μ k μ )
where μ is the global mean of the N input samples. The quadratic discriminant functions can be written as
log P d ( k ) = 1 | z δ k ( z ) = 1 2 log | Σ k | 1 2 ( z μ k ) Σ k 1 ( z μ k ) + log N k N
from the relation P d ( k ) = 1 | z P z | d ( k ) = 1 · P ( k ) . This allows computation of the score of a new signal time series z in the manifold for belonging to the k-th class.

2.5. Neural Manifold Analysis

In the previous subsections, a procedure for each EEG trial x R T , considered a time series, was formalized to extract points in a reduced manifold z Z in such a way that MI classes can be handled with a multi-class classifier.
In this subsection, an approach utilizing a NMA on EEG data is presented (see Figure 1) to reorganize time series by identifying specific time intervals, capturing class-specific and subject-specific characteristics, and improving class discriminability. This method involves analyzing the EEG signal in a multi-dimensional feature space to detect intervals that best capture the relevant features corresponding to the different MI tasks. Additionally, a cross-validation of the extracted time features across subjects is performed, significantly improving classification accuracy for challenging subjects. This underscores the reliability and potential of the presented method for enhancing cross-subject classification in EEG-based BCI systems.
Different approaches to analyzing acquired neuronal signals involve PCA, PPCA, GPFA [51], demixed PCA [52], pi-Variational Auto Envoders (VAEs) [53], UMAP [54], or frameworks like MIND [55], LFADS [56], and CEBRA [57,58]. NMA aims to uncover the underlying structure of high-dimensional neuronal data by projecting it into a lower-dimensional space where the data’s intrinsic properties are more apparent. Techniques such as PCA [26,40] are applied to reduce the dimensionality of the EEG data while preserving its most significant features. This step transforms the high-dimensional EEG signals into a more manageable lower-dimensional space, capturing the essential patterns and structures inherent in the data.
Movement planning functions in the brain are hypothesized to occur in a low-dimensional subspace of movements called movement primitives, often corresponding to a reduced neuronal manifold. These neuronal primitives enable the control of multiple degrees of freedom of movement with fewer control signals [59,60]. To quantitatively compare the differences between the encoded variables (such as direction and task), is considered an H-dimensional neuronal manifold, formed by H sub-manifolds identified by a set of elements { s 1 , , s H } , resulting in a neural set of sub-manifolds or dictionary [60,61,62], emerging from the space reduction with PCA. Each EEG trial, i.e., neural trajectories, can be approximated as
x ( t ) = h = 1 H c h ( t ) · s h
with t T , where c h ( t ) are the coefficients of the decomposition with respect to the sub-manifold h. This can be viewed as an H-dimensional trajectory in R H . For each trajectory, is computed a separability measure in each direction of R H in a supervised manner with respect to the classes of the dataset. The N samples in the input set can be partitioned into K sets { Π k } k = 1 K , where Π k denotes the sample set of the k-th class containing N k data points. For each class, can be selected the coefficient c k h ( t ) relative to the trials of the corresponding class x Π k .
Given these trajectories in each sub-manifold, a separability measure reflecting the probability of the trajectories being separated is computed, and it is used as a probability of separation among classes. For simplicity, a one-way ANOVA is performed at each time step among the values c k ( t ) , obtaining a p-value that measures the separability of the classes over time. The more this p-value approaches zero, the more separable the classes are. Additionally, using a post hoc Tukey test [63], the separability measure p i j between two distinct classes can also be obtained. By studying the trend of these measures, the minimum values where the classes are most separated can be identified. Corresponding to these values, the time s of maximum separability among classes ( arg min t T p ( t ) ) and the time s i j of maximum separability between class i and class j ( arg min t T and p ( t ) < 0.05 p i j ( t ) ) can be determined.
Whenever s and s i j are computed, these times are used to organize new trials x { x [ s Δ t , s + Δ t ] } or x { x [ s i j Δ t , s i j + Δ t ] } , respectively. On these new trials, the FBCSP procedure is performed to obtain better features and improve classification [64]. Seven selected viable pipelines for trial pre-processing are presented, followed by an explanation of how these pipelines are applied to generate new trials. In the results section, the effectiveness of these pipelines in improving classification accuracy is demonstrated.

2.6. NMA Pipelines

In the FBCSP approach [44], the analysis interval corresponds to the entire period considered informative for the task (e.g., in the presented experiments, T corresponds to [ 0.5 , 2.5 ] , see Section 3). However, using CSP to extract a unique multi-dimensional point over this entire interval may discard valuable information within the noise. Moreover, splitting intervals and aggregating them afterwards can be computationally expensive due to the numerous potential choices, and arbitrary splits may introduce additional problems, discarding critical information. In this framework, a guided procedure that extracts crucial information, enhancing classification, is proposed.
The starting point is to use the entire interval of interest, as done in standard FBCSP. Here, various pipelines are explored to form new trials x in different intervals, aiming to improve the procedure. Once defined T as the original motor imagery interval, NMA is performed over all T to identify interval of interests for the pipelines: T s = [ s Δ T , s + Δ T ] , where s is the time of maximum separability found with the NMA procedure; T s i j = [ s i j Δ T , s i j + Δ T ] , where s i j is the maximum time of separability between classes i and j identified by the NMA procedure, where i and j are chosen looking at the confusion matrix of a classification with FBCSP procedure; T s i j 0 = [ s i j Δ T , s i j + Δ T ] , where s i j is the maximum time of separability between classes i and j identified by the NMA procedure, where i and j are chosen looking at the confusion matrix of a classification with Pipeline 0 procedure (see below).
Thus, the following pipelines are realized in this study with NMA:
FBCSP: 
The entire EEG signal interval T, during which the motor imagery task is performed, resulting in trials x R N c h × T .
Pipeline 0:
Reduced EEG trials x R N c h × T 0 centered on the maximum separability time point among classes.
Pipeline 1:
For each trial, two time series are obtained: one corresponding to standard FBCSP x 1 R N c h × T and the other to Pipeline 0 x 2 R N c h × T s . The features obtained from each FBCSP procedure are concatenated and sent to the classifier.
Pipeline 2:
For each trial, two time series are obtained: x 1 R N c h × T and x 2 R N c h × T s . These signals are concatenated to form a new time series x = [ x 1 , x 2 ] R N c h × T + T s , on which the feature extraction procedure is applied.
Pipeline 3:
For each trial, two time series are obtained: one corresponding to standard FBCSP x 1 R N c h × T and the other to x 2 R N c h × T s i j , corresponding to the maximum separability time point between classes i and j. The features obtained from each FBCSP procedure are concatenated and sent to the classifier.
Pipeline 4:
For each trial, two time series x 1 R N c h × T and x 2 R N c h × T s i j are obtained. These signals are concatenated to form a new time series x = [ x 1 , x 2 ] R N c h × T + T s i j , on which the feature extraction procedure is applied.
Pipeline 5:
For each trial, two time series are obtained: one corresponding to standard FBCSP x 1 R N c h × T and the other to x 2 R N c h × T s i j 0 , corresponding to the maximum separability time point between classes i and j. The features obtained from each FBCSP procedure are concatenated and sent to the classifier.
Pipeline 6:
For each trial, two time series x 1 R N c h × T and x 2 R N c h × T s i j 0 are obtained. These signals are concatenated to form a new time series x = [ x 1 , x 2 ] R N c h × T + T s i j , on which the feature extraction procedure is applied.
Note that the intervals T, T s , T s i j , and T s i j 0 are all subject-specific. However, combining this information from different subjects can further enhance the accuracy of a BCI system, as demonstrated in Section 3.

3. Experimental Results

This section is divided into two subsections, each corresponding to a different dataset. The first subsection covers tests on a two-class dataset, Graz Dataset 2b, while the second subsection focuses on a four-class dataset, Graz Dataset 2a. Results from both datasets are presented and thoroughly discussed.

3.1. Tests on Graz Dataset 2b

This part provides a brief description of the Graz Dataset 2b, followed by a report and discussion on the improvements introduced by NMA. Comparisons are made with the approaches that demonstrated the best overall performance across all subjects among the algorithms submitted to the BCI Competition IV [46,47].

3.1.1. Graz Dataset 2b Description

The Graz Dataset 2b [65] consists of EEG recordings from nine right-handed subjects who participated in a two-class motor imagery study: left hand (class 1) and right hand (class 2). EEG data were recorded using three bipolar electrodes placed at positions C3, Cz, and C4. The experiment consisted of five sessions, with the first two dedicated to training without feedback (screening sessions) and the last three incorporating real-time feedback. During the screening sessions, subjects sat in front of a computer screen. After 3 s, a cue in the form of an arrow (pointing left or right) indicating the motor imagery task to perform appeared for 1.25 s. Subjects were instructed to continue the motor imagery task (which involved imagining the movement of either their right or left hand) until the fixation cross disappeared at 7 s. A short pause followed, with a black screen (see Figure 2a). During the feedback session, at the beginning of each trial (second 0), the feedback (a gray smiley) was centered on the screen. At second 2, a short warning beep (1 kHz, 70 ms) was given. From second 3 to 7.5, a visual cue was presented, and depending on the cue, the subjects were required to move the smiley towards the left or right side by imagining the corresponding hand movement. During this feedback period, the smiley turned green when moved in the correct direction and red if incorrect. The smiley’s distance from the origin was adjusted based on the integrated classification output over the past two seconds. Additionally, the classifier output influenced the curvature of the smiley’s mouth, making it appear happy (corners of the mouth upward) or sad (corners downward). At second 7.5, the screen went blank, followed by a random interval between 1.0 and 2.0 s. Subjects were instructed to keep the smiley on the correct side for as long as possible, continuing the MI task throughout the trial (see Figure 2b). The training data comprised the first two sessions (screening) and the third session (with feedback), totaling 240 trials without visual feedback and 160 trials with feedback (named BT). The evaluation data were drawn from the remaining two sessions, consisting of 320 trials (named BE). The dataset is open and freely available at www.bbci.de/competition/iv/ (accessed on 2 February 2024).

3.1.2. Performance Comparison on Graz Dataset 2b

This section illustrates how the MNA approach can be effectively applied in a binary classification task with a limited number of channels (3), showing its potential for improving performance. Table 1 presents the results of 10-fold cross-validation conducted on the BT session, comparing NMA and FBCSP. Apart from Pipeline 0, every pipeline utilizing NMA demonstrates notable improvements across different subjects, consistently achieving higher accuracy compared with the FBCSP. The parameters learned during the BT session were subsequently used to evaluate performance on the BE session, which served as a test set. Table 2 presents the results obtained from the BE session on the Graz Dataset 2b, with the trained models assessing generalization capabilities in a new session. A new model is added for comparison, a ShallowConvNet (SCN) architecture [33]. SCN consists of two convolutional layers (temporal, then spatial), a squaring nonlinearity ( f ( x ) = x 2 ), an average pooling layer, and l o g nonlinearity ( f ( x ) = l o g ( x ) ). The SCN architecture was specifically designed for oscillatory signal classification (by extracting features related to log-band power). The accuracy results on the BE session show that the pipelines introduced in this paper outperform both FBCSP and SCN in the majority of subjects, demonstrating superior classification accuracy in distinguishing between left and right hand movements. This dataset effectively serves as a testbed, illustrating that even with only two classes, the NMA approach can successfully identify intervals that enhance feature separability. The following section extends this approach to a multi-class scenario, where NMA is used to isolate intervals that improve the separability of specific classes that are otherwise poorly distinguishable.

3.2. Tests on Graz Dataset 2a

This subsection provides a brief overview of the Graz Dataset 2a and compares the performance improvements introduced by NMA in a multi-class problem to other approaches, such as FBCSP and SCN [33], both of which have demonstrated effectiveness on this dataset. Additionally, this section demonstrates how features extracted from one subject can be effectively leveraged to improve classification performance in another subject.

3.2.1. Graz Dataset 2a Description

The Graz Dataset 2a [43] for the BCI Competition IV is specifically designed to systematically study EEG responses associated with various motor imagery tasks, thereby facilitating the analysis of brain activity patterns. This dataset is particularly valuable due to its complexity, involving four distinct motor imagery classes.
The dataset includes EEG data from nine subjects participating in a cue-based BCI paradigm involving four motor imagery tasks: imagining the movement of the left hand (class 1), right hand (class 2), both feet (class 3), and tongue (class 4). During the experiment, subjects sat in front of a computer screen. Each subject completed two sessions, named A Training (AT) and A Evaluation (AE), recorded on different days. Each session includes recordings from three EOG channels and twenty-two EEG channels. However, only the EEG channels were considered. The registered session consists of six runs separated by short breaks. Each run included 48 trials (12 for each class), resulting in a total of 288 trials per session.
At the start of each trial ( t = 0 s), a fixation cross appeared on the black screen, accompanied by a short acoustic warning tone. After two seconds ( t = 2 s), a cue in the form of an arrow (left, right, down, or up) appeared for 1.25 s, indicating the motor imagery task to perform. Subjects were instructed to continue the MI task until the fixation cross disappeared at t = 6 s, followed by a short break with a black screen (see Figure 3 for a detailed depiction of the task recording). The dataset is open and is freely available at www.bbci.de/competition/iv/ (accessed on 02 February 2024).

3.2.2. Enhancing Class Separation via NMA

NMA allows for the detection of features that are more discriminative concerning motor imagery tasks. By applying the NMA procedure to each subject, subject-specific intervals such as T, T s , T s i j , and T s i j 0 can be identified, enabling the refinement of features for the classifier. A series of figures is provided to clearly illustrate the method. In Figure 4, the results for a sample subject (Subject 9) are reported, depicting the computation of the separability measure of the trajectories among all classes. For this subject, two directions explain more than 95 % of the variance. Panels a and b in Figure 4 display the separability measure for all classes across two directions. Lower values indicate greater separation, allowing us to identify optimal times (e.g., T and T s ; see Section 2) for centering the time series analysis.
Similarly, Figure 5 illustrates the separability measure of the trajectories between pairs of specific classes for the same subject. This is again shown for two directions, where lower values signify greater separation, enabling the identification of times (e.g., T s i j , T s i j 0 , see Section 2) for focusing the time series analysis.
From the classification phase, confusion matrices for the four classes can be constructed for each subject using both FBCSP and Pipeline 0. Figure 6 shows the confusion matrix obtained with FBCSP for a sample Subject 9. These confusion matrices are utilized to detect the intervals T s i j and T s i j o , respectively, based on the minimum separability measure between two classes detected by NMA.
To further clarify the analysis for Subject 9, 2D trajectories of the coefficients c ¯ k 1 ( t ) against c ¯ k 2 ( t ) , averaged for each class, are plotted (see Figure 7), where k corresponds to one of the four motor imagery classes: left hand, right hand, feet, and tongue. These plots highlight the points of maximum separation among the classes (crosses in blue rectangles) and between right hand and tongue (crosses in red rectangles). The same trajectories are also depicted as 2D ellipsoids (see Figure 8), illustrating c ¯ k 1 ( t ) and c ¯ k 2 ( t ) against time, with the ellipsoid dimensions based on the standard deviation for each manifold sub-dimension. This further illustrates the points of maximum separation among classes, specifically between right hand and tongue.
Table 3 shows the results from the AT session of the Graz Dataset, performing 10-fold cross-validation. Each pipeline involving NMA demonstrates improvements in different subjects. Apart from Pipeline 0, each pipeline shows improvements in accuracy compared with the FBCSP winner of the BCI competition. Moreover, by choosing the most successful pipeline for each subject, it is evident that the presented method brings significant improvements over the state-of-the-art algorithm.
The same comparison was conducted on the AE session, utilizing the model trained on the AT session to evaluate its generalization capabilities in a new session. This aligns with the spirit of the BCI competition, where the AE session served as the test set for comparing methods, and the AT session was used for system tuning. As shown in Table 4, the accuracy results are lower in the AE session compared with the AT session, as expected. However, the presented method consistently defines pipelines that outperform the competition winner and demonstrate better accuracy than the ones of FBCSP and SCN. Interestingly, in the case of four-class classification, the performance gap between NMA-based pipelines and other approaches appears to widen. This suggests that as the number of classes increases, our separability properties allow the NMA-based approach to construct manifolds that more effectively distinguish among different conditions, leading to improved discrimination. By selecting the best pipeline for each subject, the proposed method achieves superior accuracy compared with the FBCSP and SCN approaches, which have been used as benchmarks for SMR classification.

3.2.3. Cross-Subjects Manifold Sharing

Until now, each result was obtained by analyzing the manifolds of specific subjects and attempting to improve classification by enhancing specific features whenever NMA detected them. However, this section shows that this analysis can be extended further. By leveraging the capabilities of certain subjects who excel in class discrimination, it is possible to augment the performance of subjects with poorer classification abilities. Specifically, this approach imports results of MNA from high-performing subjects and projects them onto other subjects. This approach combines previous BCI systems (explained in the previous subsection) with new features detected with the aid of other successful subjects.
Figure 9 illustrates the results of this hybridization approach. This combined confusion matrix shown presents the classification performance when subject i uses an NMA insight j from another subject. Each matrix element A i j indicates the percentage improvement in accuracy compared with the FBCSP and SCN benchmarks. The diagonal elements represent pipelines using only NMA from the same subject, while the off-diagonal elements show results using pipelines augmented with the best features obtained through NMA analysis from other successful subjects.
This figure demonstrates that augmenting features by NMA from other subjects significantly enhances the BCI system’s classification capabilities. This approach highlights the potential of cross-subject NMA information sharing in improving overall system performance, showcasing how knowledge transfer among subjects can lead to better generalization and more robust BCI systems. Table 5 shows the accuracies on the AE session when using the best NMA result per subject (computed in the previous subsection) and the best NMA result across subjects. The results demonstrate that NMA information shared across subjects can further improve classification performance.
Finally, to provide insight into the learned features, the frequency bands (see Figure 10) and topographic maps (Figure 11) are presented, both obtained from the CSP projection matrix of the first selected features. These figures illustrate the characteristic patterns for different motor imagery tasks: a contralateral pattern for left hand and right hand, top-central activation for feet [8,9], and parietofrontal activation for tongue [44]. Notably, several features are “borrowed” from other subjects, demonstrating that cross-subject features can significantly enhance accuracy. This cross-subject feature borrowing is particularly beneficial in improving the robustness and generalization of the BCI system, making it more adaptable to various users. The integration of these features across subjects highlights the potential of NMA to uncover critical patterns that are not only subject-specific but also generalizable across different individuals, further strengthening the classification performance and reliability of EEG-based BCI systems.

4. Conclusions

The presented study demonstrates the significant potential of integrating NMA with traditional EEG-based MI-BCI systems to enhance feature extraction and classification accuracy. By identifying specific time intervals that capture class-specific and subject-specific characteristics, the presented approach has demonstrated the ability to enhance the performance of MI-BCI systems, particularly for challenging subjects. This method not only refines the features for individual subjects but also leverages cross-subject information to further boost classification accuracy. The primary objective of this paper was to develop a method specifically designed for classifying the oscillatory components of sensorimotor rhythms during MI tasks. The Graz Datasets 2a and 2b, widely recognized benchmarks in this field, were utilized to validate the proposed approach. This work builds on the most efficient existing methods tailored for MI tasks (such as the FBCSP algorithm, the winner of the BCI competition, and SCN, specifically designed for MI tasks). By introducing NMA-based preprocessing to create novel MI-BCI pipelines, the discriminability of trials is significantly improved, as features become more separable with respect to distinct motor imagery classes. The presented results demonstrate that incorporating NMA in the preprocessing stage enhances the performance of established algorithms. As detailed in Section 3, the results underscore the robustness and adaptability of the presented approach, paving the way for more reliable and efficient MI-BCI systems.
Future research directions could involve incorporating successful neural network approaches, such as SCN, which has been shown to be effective for SMR classification, by completely replacing the CSP modules within the pipeline. While DL techniques typically automate feature extraction and selection, they could potentially benefit from a preprocessing phase that begins with manifold analysis. Furthermore, an intriguing possibility would be to develop a conditional VAE [66]—a deep VAE model conditioned on the separability properties in NMA outlined in our Methods section. This approach could further optimize the feature selection process, contributing to more refined and user-friendly BCI applications.

Author Contributions

Conceptualization, M.F., R.P., G.P. and F.D.; Methodology, M.F., R.P., L.O., S.G., A.A., G.P. and F.D.; Software, M.F., L.O., S.G., A.A. and F.D.; Validation, M.F. and F.D.; Formal analysis, M.F., G.P. and F.D.; Investigation, M.F., G.P. and F.D.; Resources, M.F. and F.D.; Data curation, M.F. and F.D.; Writing—original draft, M.F., R.P. and F.D.; Writing—review & editing, M.F., R.P., L.O., S.G., A.A., G.P. and F.D.; Visualization, F.D.; Supervision, F.D.; Funding acquisition, G.P. and F.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received funding from the European Research Council under the Grant Agreement No. 820213 (ThinkAhead) to G.P.; the Italian National Recovery and Resilience Plan (NRRP), M4C2, funded by the European Union—NextGenerationEU (Project IR0000011, CUP B51E22000150006, “EBRAINS-Italy”; Project PE0000013, CUP B53C22003630006, “FAIR”; Project PE0000006, CUP J33C22002970002 “MNESYS”) to G.P. and F.D.; PRIN PNRR P20224FESY to G.P. and F.D.; the MUR PRIN2020 project Free energy principle and the brain: neuronal and phylogenetic mechanisms of Bayesian inference—Grant No. 2020529PCP to FD. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are available at: www.bbci.de/competition/iv/.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hosseini, M.P.; Hosseini, A.; Ahi, K. A review on machine learning for EEG signal processing in bioengineering. IEEE Rev. Biomed. Eng. 2020, 14, 204–218. [Google Scholar] [CrossRef] [PubMed]
  2. Dhiman, R. Machine learning techniques for electroencephalogram based brain-computer interface: A systematic literature review. Meas. Sens. 2023, 28, 100823. [Google Scholar]
  3. Pahuja, S.; Veer, K. Recent approaches on classification and feature extraction of EEG signal: A review. Robotica 2022, 40, 77–101. [Google Scholar]
  4. Lotte, F.; Bougrain, L.; Cichocki, A.; Clerc, M.; Congedo, M.; Rakotomamonjy, A.; Yger, F. A review of classification algorithms for EEG-based brain–computer interfaces: A 10 year update. J. Neural Eng. 2018, 15, 031005. [Google Scholar] [CrossRef] [PubMed]
  5. Apicella, A.; Arpaia, P.; Frosolone, M.; Improta, G.; Moccaldi, N.; Pollastro, A. EEG-based measurement system for monitoring student engagement in learning 4.0. Sci. Rep. 2022, 12, 5857. [Google Scholar] [CrossRef]
  6. Decety, J. The neurophysiological basis of motor imagery. Behav. Brain Res. 1996, 77, 45–52. [Google Scholar] [CrossRef] [PubMed]
  7. Jeannerod, M. Mental imagery in the motor context. Neuropsychologia 1995, 33, 1419–1432. [Google Scholar] [CrossRef] [PubMed]
  8. Pfurtscheller, G.; Neuper, C. Motor imagery and direct brain-computer communication. Proc. IEEE 2001, 89, 1123–1134. [Google Scholar] [CrossRef]
  9. Pfurtscheller, G. Functional brain imaging based on ERD/ERS. Vis. Res. 2001, 41, 1257–1260. [Google Scholar] [CrossRef]
  10. Giannopulu, I.; Mizutani, H. Neural kinesthetic contribution to motor imagery of body parts: Tongue, hands, and feet. Front. Hum. Neurosci. 2021, 15, 602723. [Google Scholar] [CrossRef]
  11. Ehrsson, H.H.; Geyer, S.; Naito, E. Imagery of voluntary movement of fingers, toes, and tongue activates corresponding body-part-specific motor representations. J. Neurophysiol. 2003, 90, 3304–3316. [Google Scholar] [CrossRef] [PubMed]
  12. Praamstra, P.; Boutsen, L.; Humphreys, G.W. Frontoparietal control of spatial attention and motor intention in human EEG. J. Neurophysiol. 2005, 94, 764–774. [Google Scholar] [CrossRef] [PubMed]
  13. He, B.; Baxter, B.; Edelman, B.J.; Cline, C.C.; Wenjing, W.Y. Noninvasive brain-computer interfaces based on sensorimotor rhythms. Proc. IEEE 2015, 103, 907–925. [Google Scholar] [CrossRef]
  14. Edelman, B.J.; Meng, J.; Suma, D.; Zurn, C.; Nagarajan, E.; Baxter, B.S.; Cline, C.C.; He, B. Noninvasive neuroimaging enhances continuous neural tracking for robotic device control. Sci. Robot. 2019, 4, eaaw6844. [Google Scholar] [CrossRef]
  15. Pfurtscheller, G.; Guger, C.; Müller, G.; Krausz, G.; Neuper, C. Brain oscillations control hand orthosis in a tetraplegic. Neurosci. Lett. 2000, 292, 211–214. [Google Scholar] [CrossRef]
  16. Zimmermann-Schlatter, A.; Schuster, C.; Puhan, M.A.; Siekierka, E.; Steurer, J. Efficacy of motor imagery in post-stroke rehabilitation: A systematic review. J. Neuroeng. Rehabil. 2008, 5, 8. [Google Scholar] [CrossRef]
  17. Apicella, A.; Arpaia, P.; Frosolone, M.; Moccaldi, N. High-wearable EEG-based distraction detection in motor rehabilitation. Sci. Rep. 2021, 11, 5297. [Google Scholar] [CrossRef]
  18. Wu, D.; Xu, Y.; Lu, B.L. Transfer learning for EEG-based brain—Computer interfaces: A review of progress made since 2016. IEEE Trans. Cogn. Dev. Syst. 2020, 14, 4–19. [Google Scholar] [CrossRef]
  19. Bonnet, L.; Lotte, F.; Lécuyer, A. Two brains, one game: Design and evaluation of a multiuser BCI video game based on motor imagery. IEEE Trans. Comput. Intell. AI Games 2013, 5, 185–198. [Google Scholar] [CrossRef]
  20. Mladenović, J.; Frey, J.; Pramij, S.; Mattout, J.; Lotte, F. Towards identifying optimal biased feedback for various user states and traits in motor imagery BCI. IEEE Trans. Biomed. Eng. 2021, 69, 1101–1110. [Google Scholar] [CrossRef]
  21. Gómez, C.M.; Arjona, A.; Donnarumma, F.; Maisto, D.; Rodríguez-Martínez, E.I.; Pezzulo, G. Tracking the time course of Bayesian inference with event-related potentials: A study using the central Cue Posner Paradigm. Front. Psychol. 2019, 10, 1424. [Google Scholar] [CrossRef] [PubMed]
  22. Subasi, A. EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Syst. Appl. 2007, 32, 1084–1093. [Google Scholar] [CrossRef]
  23. Atyabi, A.; Shic, F.; Naples, A. Mixture of autoregressive modeling orders and its implication on single trial EEG classification. Expert Syst. Appl. 2016, 65, 164–180. [Google Scholar] [CrossRef] [PubMed]
  24. Blankertz, B.; Tomioka, R.; Lemm, S.; Kawanabe, M.; Muller, K.R. Optimizing spatial filters for robust EEG single-trial analysis. IEEE Signal Process. Mag. 2007, 25, 41–56. [Google Scholar] [CrossRef]
  25. Qin, L.; He, B. A wavelet-based time–frequency analysis approach for classification of motor imagery for brain–computer interface applications. J. Neural Eng. 2005, 2, 65. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, W.; Wu, D. Manifold embedded knowledge transfer for brain-computer interfaces. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 1117–1127. [Google Scholar] [CrossRef]
  27. Arpaia, P.; Donnarumma, F.; Esposito, A.; Parvis, M. Channel selection for optimal EEG measurement in motor imagery-based brain-computer interfaces. Int. J. Neural Syst. 2021, 31, 2150003. [Google Scholar] [CrossRef]
  28. Gu, X.; Cao, Z.; Jolfaei, A.; Xu, P.; Wu, D.; Jung, T.P.; Lin, C.T. EEG-based brain-computer interfaces (BCIs): A survey of recent studies on signal sensing technologies and computational intelligence approaches and their applications. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 1645–1666. [Google Scholar] [CrossRef]
  29. Jayaram, V.; Alamgir, M.; Altun, Y.; Scholkopf, B.; Grosse-Wentrup, M. Transfer learning in brain-computer interfaces. IEEE Comput. Intell. Mag. 2016, 11, 20–31. [Google Scholar] [CrossRef]
  30. Wu, D.; Lawhern, V.J.; Hairston, W.D.; Lance, B.J. Switching EEG headsets made easy: Reducing offline calibration effort using active weighted adaptation regularization. IEEE Trans. Neural Syst. Rehabil. Eng. 2016, 24, 1125–1137. [Google Scholar] [CrossRef]
  31. Kang, H.; Nam, Y.; Choi, S. Composite common spatial pattern for subject-to-subject transfer. IEEE Signal Process. Lett. 2009, 16, 683–686. [Google Scholar] [CrossRef]
  32. Zanini, P.; Congedo, M.; Jutten, C.; Said, S.; Berthoumieu, Y. Transfer learning: A Riemannian geometry framework with applications to brain—Computer interfaces. IEEE Trans. Biomed. Eng. 2017, 65, 1107–1116. [Google Scholar] [CrossRef] [PubMed]
  33. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain—Computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  34. Zheng, M.; Lin, Y. A deep transfer learning network with two classifiers based on sample selection for motor imagery brain-computer interface. Biomed. Signal Process. Control 2024, 89, 105786. [Google Scholar] [CrossRef]
  35. Xie, Y.; Wang, K.; Meng, J.; Yue, J.; Meng, L.; Yi, W.; Jung, T.P.; Xu, M.; Ming, D. Cross-dataset transfer learning for motor imagery signal classification via multi-task learning and pre-training. J. Neural Eng. 2023, 20, 056037. [Google Scholar] [CrossRef]
  36. Li, D.; Wang, J.; Xu, J.; Fang, X.; Ji, Y. Cross-channel specific-mutual feature transfer learning for motor imagery EEG signals decoding. IEEE Trans. Neural Netw. Learn. Syst. 2023. [Google Scholar] [CrossRef] [PubMed]
  37. Arpaia, P.; Covino, A.; Cristaldi, L.; Frosolone, M.; Gargiulo, L.; Mancino, F.; Mantile, F.; Moccaldi, N. A systematic review on feature extraction in electroencephalography-based diagnostics and therapy in attention deficit hyperactivity disorder. Sensors 2022, 22, 4934. [Google Scholar] [CrossRef]
  38. Lotte, F.; Guan, C. Regularizing common spatial patterns to improve BCI designs: Unified theory and new algorithms. IEEE Trans. Biomed. Eng. 2010, 58, 355–362. [Google Scholar] [CrossRef]
  39. Vallabhaneni, A.; He, B. Motor imagery task classification for brain computer interface applications using spatiotemporal principle component analysis. Neurol. Res. 2004, 26, 282–287. [Google Scholar] [CrossRef]
  40. Simola, J.; Silander, T.; Harju, M.; Lahti, O.; Makkonen, E.; Pätsi, L.M.; Smallwood, J. Context independent reductions in external processing during self-generated episodic social cognition. Cortex 2023, 159, 39–53. [Google Scholar] [CrossRef]
  41. Kwon, O.Y.; Lee, M.H.; Guan, C.; Lee, S.W. Subject-independent brain–computer interfaces based on deep convolutional neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 3839–3852. [Google Scholar] [CrossRef] [PubMed]
  42. Dong, Y.; Wen, X.; Gao, F.; Gao, C.; Cao, R.; Xiang, J.; Cao, R. Subject-independent EEG classification of motor imagery based on dual-branch feature fusion. Brain Sci. 2023, 13, 1109. [Google Scholar] [CrossRef] [PubMed]
  43. Brunner, C.; Leeb, R.; Müller-Putz, G.; Schlögl, A.; Pfurtscheller, G. BCI Competition 2008–Graz data set A. Inst. Knowl. Discov. (Laboratory-Brain-Comput. Interfaces) Graz Univ. Technol. 2008, 16, 1–6. [Google Scholar]
  44. Chin, Z.Y.; Ang, K.K.; Wang, C.; Guan, C.; Zhang, H. Multi-class filter bank common spatial pattern for four-class motor imagery BCI. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 571–574. [Google Scholar]
  45. Ang, K.K.; Chin, Z.Y.; Wang, C.; Guan, C.; Zhang, H. Filter bank common spatial pattern algorithm on BCI competition IV datasets 2a and 2b. Front. Neurosci. 2012, 6, 21002. [Google Scholar] [CrossRef]
  46. Tangermann, M.; Müller, K.R.; Aertsen, A.; Birbaumer, N.; Braun, C.; Brunner, C.; Leeb, R.; Mehring, C.; Miller, K.J.; Müller-Putz, G.R.; et al. Review of the BCI competition IV. Front. Neurosci. 2012, 6, 55. [Google Scholar] [CrossRef]
  47. Ang, K.K.; Chin, Z.Y.; Zhang, H.; Guan, C. Filter Bank Common Spatial Pattern (FBCSP) in Brain-Computer Interface. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 2390–2397. [Google Scholar]
  48. Thiyam, D.; Rajkumar, E. Common Spatial Pattern Algorithm Based Signal Processing Techniques for Classification of Motor Imagery Movements: A Mini Review. Int. J. Circuit Theory Appl. 2016, 9, 53–65. [Google Scholar]
  49. Ang, K.K.; Chin, Z.Y.; Zhang, H.; Guan, C. Mutual information-based selection of optimal spatial–temporal patterns for single-trial EEG-based BCIs. Pattern Recognit. 2012, 45, 2137–2144. [Google Scholar] [CrossRef]
  50. Hastie, T.; Tibshirani, R.; Friedman, J.H.; Friedman, J.H. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer: New York, NY, USA, 2009; Volune 2. [Google Scholar]
  51. Yu, B.M.; Cunningham, J.P.; Santhanam, G.; Ryu, S.; Shenoy, K.V.; Sahani, M. Gaussian-process factor analysis for low-dimensional single-trial analysis of neural population activity. In Proceedings of the NIPS’08: 21st International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–10 December 2008. [Google Scholar]
  52. Kobak, D.; Brendel, W.; Constantinidis, C.; Feierstein, C.E.; Kepecs, A.; Mainen, Z.F.; Qi, X.L.; Romo, R.; Uchida, N.; Machens, C.K. Demixed principal component analysis of neural population data. elife 2016, 5, e10989. [Google Scholar] [CrossRef]
  53. Zhou, D.; Wei, X.X. Learning identifiable and interpretable latent models of high-dimensional neural activity using pi-VAE. Adv. Neural Inf. Process. Syst. 2020, 33, 7234–7247. [Google Scholar]
  54. McInnes, L.; Healy, J.; Melville, J. Umap: Uniform manifold approximation and projection for dimension reduction. arXiv 2018, arXiv:1802.03426. [Google Scholar]
  55. Low, R.J.; Lewallen, S.; Aronov, D.; Nevers, R.; Tank, D.W. Probing variability in a cognitive map using manifold inference from neural dynamics. BioRxiv 2018, 418939. [Google Scholar] [CrossRef]
  56. Pandarinath, C.; O’Shea, D.J.; Collins, J.; Jozefowicz, R.; Stavisky, S.D.; Kao, J.C.; Trautmann, E.M.; Kaufman, M.T.; Ryu, S.I.; Hochberg, L.R.; et al. Inferring single-trial neural population dynamics using sequential auto-encoders. Nat. Methods 2018, 15, 805–815. [Google Scholar] [CrossRef] [PubMed]
  57. Schneider, S.; Lee, J.H.; Mathis, M.W. Learnable latent embeddings for joint behavioural and neural analysis. Nature 2023, 617, 360–368. [Google Scholar] [CrossRef]
  58. Mathis, M.W. Brain dynamics uncovered using a machine-learning algorithm. Nature 2023. [Google Scholar] [CrossRef]
  59. Vinjamuri, R.; Weber, D.J.; Mao, Z.H.; Collinger, J.L.; Degenhart, A.D.; Kelly, J.W.; Boninger, M.L.; Tyler-Kabara, E.C.; Wang, W. Toward synergy-based brain-machine interfaces. IEEE Trans. Inf. Technol. Biomed. 2011, 15, 726–736. [Google Scholar] [CrossRef]
  60. Prevete, R.; Donnarumma, F.; d’Avella, A.; Pezzulo, G. Evidence for sparse synergies in grasping actions. Sci. Rep. 2018, 8, 616. [Google Scholar] [CrossRef] [PubMed]
  61. Donnarumma, F.; Prevete, R.; Maisto, D.; Fuscone, S.; Irvine, E.M.; van der Meer, M.A.; Kemere, C.; Pezzulo, G. A framework to identify structured behavioral patterns within rodent spatial trajectories. Sci. Rep. 2021, 11, 468. [Google Scholar] [CrossRef]
  62. Jenatton, R.; Mairal, J.; Obozinski, G.; Bach, F.R. Proximal methods for sparse hierarchical dictionary learning. In Proceedings of the ICML, Haifa, Israel, 21–24 June 2010; Citeseer: Gaithersburg, MD, USA, 2010; Volume 1, p. 2. [Google Scholar]
  63. Montgomery, D.C. Design and Analysis of Experiments; John Wiley & Sons: Hoboken, NJ, USA, 2017. [Google Scholar]
  64. Arpaia, P.; Frosolone, M.; Gargiulo, L.; Moccaldi, N.; Nalin, M.; Perin, A.; Puttilli, C. Specific feature selection in wearable EEG-based transducers for monitoring high cognitive load in neurosurgeons. Comput. Stand. Interfaces 2025, 92, 103896. [Google Scholar] [CrossRef]
  65. Leeb, R.; Brunner, C.; Müller-Putz, G.; Schlögl, A.; Pfurtscheller, G. BCI Competition 2008–Graz data set B. Graz Univ. Technol. Austria 2008, 16, 1–6. [Google Scholar]
  66. Bethge, D.; Hallgarten, P.; Grosse-Puppendahl, T.; Kari, M.; Chuang, L.L.; Özdenizci, O.; Schmidt, A. EEG2Vec: Learning affective EEG representations via variational autoencoders. In Proceedings of the 2022 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Prague, Czech Republic, 9–12 October 2022; pp. 3150–3157. [Google Scholar]
Figure 1. Depiction of a BCI system, including modules for Signal Acquisition, Feature Extraction, Feature Selection, and Classification. The system’s ability to discriminate between classes enables it to send commands to a control system, such as a wheelchair, prosthetic hand, or other BCI-controlled devices. In the presented approach, this process is augmented with NMA performed on the EEG signal to design better features that enhance the classification system. Furthermore, this information, especially for poorly performing subjects, can be improved by incorporating analysis from other subjects.
Figure 1. Depiction of a BCI system, including modules for Signal Acquisition, Feature Extraction, Feature Selection, and Classification. The system’s ability to discriminate between classes enables it to send commands to a control system, such as a wheelchair, prosthetic hand, or other BCI-controlled devices. In the presented approach, this process is augmented with NMA performed on the EEG signal to design better features that enhance the classification system. Furthermore, this information, especially for poorly performing subjects, can be improved by incorporating analysis from other subjects.
Sensors 24 06110 g001
Figure 2. Graz Dataset 2b MI timing scheme of the paradigm. (a) Screening session: Subjects sat in front of a computer screen. After t = 3 s, a cue in the form of an arrow (left or right) appeared for t = 1.25 s, indicating the MI task to perform. Subjects were instructed to continue the MI task until the fixation cross disappeared at t = 7 s, followed by a short break with a black screen. (b) Feedback session: At the beginning ( t = 0 s), a gray smiley appeared on the screen. From t = 3 s to t = 7.5 s, a visual cue was presented, and the subject started the MI hand movement. During this feedback period, the smiley turned green when the motor imagery moved in the correct direction and red if incorrect. After this period, the screen went blank, and a new trial began.
Figure 2. Graz Dataset 2b MI timing scheme of the paradigm. (a) Screening session: Subjects sat in front of a computer screen. After t = 3 s, a cue in the form of an arrow (left or right) appeared for t = 1.25 s, indicating the MI task to perform. Subjects were instructed to continue the MI task until the fixation cross disappeared at t = 7 s, followed by a short break with a black screen. (b) Feedback session: At the beginning ( t = 0 s), a gray smiley appeared on the screen. From t = 3 s to t = 7.5 s, a visual cue was presented, and the subject started the MI hand movement. During this feedback period, the smiley turned green when the motor imagery moved in the correct direction and red if incorrect. After this period, the screen went blank, and a new trial began.
Sensors 24 06110 g002aSensors 24 06110 g002b
Figure 3. Graz Dataset 2a MI Task: timing scheme of the paradigm: subjects sat in front of a computer screen. After t = 2 s, a cue in the form of an arrow (left, right, down, or up) appeared for t = 1.25 s, indicating the MI task to perform. Subjects were instructed to continue the motor imagery task until the fixation cross disappeared at t = 6 s, followed by a short break with a black screen.
Figure 3. Graz Dataset 2a MI Task: timing scheme of the paradigm: subjects sat in front of a computer screen. After t = 2 s, a cue in the form of an arrow (left, right, down, or up) appeared for t = 1.25 s, indicating the MI task to perform. Subjects were instructed to continue the motor imagery task until the fixation cross disappeared at t = 6 s, followed by a short break with a black screen.
Sensors 24 06110 g003
Figure 4. Results for Subject 9 showing the computation of the separability measure of the manifold trajectories among all classes. We present the probability p S e p ( t ) of trajectory separation under the null hypothesis of no effect among classes. Above each plot, the percentage of explained variance (EV) is reported for each manifold direction, with two manifold directions accounting together for over 95 % of the variance. A one-way ANOVA [63] was performed at each time step among the values c k ( t ) , yielding a p-value that quantifies the separability of the classes over time. The closer this p-value is to zero, the more separable the classes are. Based on these values, the time s of maximum separability among classes ( arg min t T p S e p ( t ) ) can be identified.
Figure 4. Results for Subject 9 showing the computation of the separability measure of the manifold trajectories among all classes. We present the probability p S e p ( t ) of trajectory separation under the null hypothesis of no effect among classes. Above each plot, the percentage of explained variance (EV) is reported for each manifold direction, with two manifold directions accounting together for over 95 % of the variance. A one-way ANOVA [63] was performed at each time step among the values c k ( t ) , yielding a p-value that quantifies the separability of the classes over time. The closer this p-value is to zero, the more separable the classes are. Based on these values, the time s of maximum separability among classes ( arg min t T p S e p ( t ) ) can be identified.
Sensors 24 06110 g004
Figure 5. Results for Subject 9, showing the computation of the separability measure of the manifold trajectories one class versus the other, following the one-way ANOVA test (see Figure 4). A post hoc Tukey test was used to obtain the separability measure p i j S e p ( t ) between two distinct classes. By analyzing the trend of these measures, the minimum values where the classes are most separated can be identified. Based on these values, the time s i j of maximum separability between class i and class j ( arg min t T and p S e p ( t ) < 0.05 p i j S e p ( t ) ) can be determined.
Figure 5. Results for Subject 9, showing the computation of the separability measure of the manifold trajectories one class versus the other, following the one-way ANOVA test (see Figure 4). A post hoc Tukey test was used to obtain the separability measure p i j S e p ( t ) between two distinct classes. By analyzing the trend of these measures, the minimum values where the classes are most separated can be identified. Based on these values, the time s i j of maximum separability between class i and class j ( arg min t T and p S e p ( t ) < 0.05 p i j S e p ( t ) ) can be determined.
Sensors 24 06110 g005
Figure 6. Confusion Matrix A i j obtained with FBCSP for a sample Subject 9 for the four classes: (1) left hand, (2) right hand, (3) feet, and (4) tongue. Each row represents the true instances in class i, while each column represents the instances in the predicted class j. Thus, the diagonal elements A i i represent correctly predicted instances, while the off-diagonal elements correspond to misclassifications. In the figure, each distinct color highlights the reciprocal misclassification between pairs of classes, visually illustrating the degree of confusion in classification results. Consequently, arg max i , i ( A i j + A j i ) gives the value of the worst class pair, in this case, 2 vs. 4 (right hand vs. tongue). Consequently, it is possible to select the interval T s i j based on the separability between these two classes detected by NMA.
Figure 6. Confusion Matrix A i j obtained with FBCSP for a sample Subject 9 for the four classes: (1) left hand, (2) right hand, (3) feet, and (4) tongue. Each row represents the true instances in class i, while each column represents the instances in the predicted class j. Thus, the diagonal elements A i i represent correctly predicted instances, while the off-diagonal elements correspond to misclassifications. In the figure, each distinct color highlights the reciprocal misclassification between pairs of classes, visually illustrating the degree of confusion in classification results. Consequently, arg max i , i ( A i j + A j i ) gives the value of the worst class pair, in this case, 2 vs. 4 (right hand vs. tongue). Consequently, it is possible to select the interval T s i j based on the separability between these two classes detected by NMA.
Sensors 24 06110 g006
Figure 7. Two−dimensional trajectories of the coefficients c ¯ k 1 ( t ) against c ¯ k 2 ( t ) for the manifold directions of Subject 9, averaged for each class, where k corresponds to one of the four motor imagery classes: left hand, right hand, feet, and tongue. These plots highlight the points of maximum separation among the classes, indicated by crosses in blue rectangles, and specifically between right hand and tongue, indicated by crosses in red rectangles (note that black borders are drawn around the right hand and tongue points).
Figure 7. Two−dimensional trajectories of the coefficients c ¯ k 1 ( t ) against c ¯ k 2 ( t ) for the manifold directions of Subject 9, averaged for each class, where k corresponds to one of the four motor imagery classes: left hand, right hand, feet, and tongue. These plots highlight the points of maximum separation among the classes, indicated by crosses in blue rectangles, and specifically between right hand and tongue, indicated by crosses in red rectangles (note that black borders are drawn around the right hand and tongue points).
Sensors 24 06110 g007
Figure 8. Two−dimensional ellipsoids for manifold directions of Subject 9, showing coefficients c ¯ k 1 ( t ) and c ¯ k 2 ( t ) against time, where k corresponds to one of the four motor imagery classes: left hand, right hand, feet, and tongue. The dimensions of the ellipsoids are based on the standard deviation for each manifold sub-dimension. This illustration further highlights the points of maximum separation among the classes, and specifically between the right hand and tongue classes.
Figure 8. Two−dimensional ellipsoids for manifold directions of Subject 9, showing coefficients c ¯ k 1 ( t ) and c ¯ k 2 ( t ) against time, where k corresponds to one of the four motor imagery classes: left hand, right hand, feet, and tongue. The dimensions of the ellipsoids are based on the standard deviation for each manifold sub-dimension. This illustration further highlights the points of maximum separation among the classes, and specifically between the right hand and tongue classes.
Sensors 24 06110 g008
Figure 9. The figure illustrates the results of the cross−subject NMA approach. The confusion matrix presents the classification performance when subject i uses NMA insights from subject j. Each matrix element A i j indicates the percentage improvement in accuracy compared with the FBCSP benchmark. Diagonal elements represent pipelines using only NMA from the same subject, while off-diagonal elements show results using pipelines augmented with the best features obtained through NMA analysis from other successful subjects.
Figure 9. The figure illustrates the results of the cross−subject NMA approach. The confusion matrix presents the classification performance when subject i uses NMA insights from subject j. Each matrix element A i j indicates the percentage improvement in accuracy compared with the FBCSP benchmark. Diagonal elements represent pipelines using only NMA from the same subject, while off-diagonal elements show results using pipelines augmented with the best features obtained through NMA analysis from other successful subjects.
Sensors 24 06110 g009
Figure 10. Statistics of the successful frequency bands for each subject, sorted and obtained from the CSP projection matrix of the first selected feature per subject. The figure presents the distribution of the most effective frequency bands that contribute to the classification accuracy across different subjects. Each rectangle represents the frequency band that was most frequently selected for optimal feature extraction, highlighting the variability and commonality of effective frequency ranges among subjects. This analysis underscores the significance of individual–specific frequency bands in enhancing the performance of motor imagery tasks in BCI systems. The detailed examination of these bands provides valuable insights into the neural oscillatory patterns that are critical for accurate classification. For further details, refer to the main text.
Figure 10. Statistics of the successful frequency bands for each subject, sorted and obtained from the CSP projection matrix of the first selected feature per subject. The figure presents the distribution of the most effective frequency bands that contribute to the classification accuracy across different subjects. Each rectangle represents the frequency band that was most frequently selected for optimal feature extraction, highlighting the variability and commonality of effective frequency ranges among subjects. This analysis underscores the significance of individual–specific frequency bands in enhancing the performance of motor imagery tasks in BCI systems. The detailed examination of these bands provides valuable insights into the neural oscillatory patterns that are critical for accurate classification. For further details, refer to the main text.
Sensors 24 06110 g010
Figure 11. Topographic maps obtained from the CSP projection matrix for the first selected feature per subject. These maps illustrate the best feature per class, ordered by classification accuracy per class. The topographic maps highlight characteristic spatial patterns for different motor imagery tasks: contralateral patterns for left and right hand movements, top–central activation for foot movements, and parietofrontal activation for tongue movements. In the accompanying table, three values are indicated: the subject from which the feature is derived (with features borrowed from another subject highlighted in bold), the Pipeline P l from which the feature is extracted, and the reference interval to which the feature corresponds (see Section 2.6). It is notable that several features are borrowed from other subjects, and certain subjects contribute more frequently to these feature–lending scenarios. This suggests that features from high–performing subjects can be effectively used to prototype efficient BCI systems.
Figure 11. Topographic maps obtained from the CSP projection matrix for the first selected feature per subject. These maps illustrate the best feature per class, ordered by classification accuracy per class. The topographic maps highlight characteristic spatial patterns for different motor imagery tasks: contralateral patterns for left and right hand movements, top–central activation for foot movements, and parietofrontal activation for tongue movements. In the accompanying table, three values are indicated: the subject from which the feature is derived (with features borrowed from another subject highlighted in bold), the Pipeline P l from which the feature is extracted, and the reference interval to which the feature corresponds (see Section 2.6). It is notable that several features are borrowed from other subjects, and certain subjects contribute more frequently to these feature–lending scenarios. This suggests that features from high–performing subjects can be effectively used to prototype efficient BCI systems.
Sensors 24 06110 g011
Table 1. Accuracy results from the BT sessions of the Graz 2b Dataset, performing 10-fold cross-validation. Each pipeline involving NMA demonstrates improvements in different subjects. Apart from Pipeline 0, each pipeline shows increased accuracy compared with the FBCSP. Values for pipelines that outperform FBCSP are in bold. The best pipeline per subject is highlighted in green.
Table 1. Accuracy results from the BT sessions of the Graz 2b Dataset, performing 10-fold cross-validation. Each pipeline involving NMA demonstrates improvements in different subjects. Apart from Pipeline 0, each pipeline shows increased accuracy compared with the FBCSP. Values for pipelines that outperform FBCSP are in bold. The best pipeline per subject is highlighted in green.
BT Session
Accuracy [%]
Subject [#]FBCSPPipeline 0Pipeline 1Pipeline 2Best Pipelines
171.665.874.073.074.0
255.261.558.060.360.3
361.353.358.862.062.0
493.489.392.494.094.0
583.381.484.083.884.0
672.268.073.374.374.3
773.173.878.076.578.0
865.168.268.466.868.4
970.969.072.871.072.8
mean ± SE71.8 ± 3.870.0 ± 3.573.3 ± 3.773.5 ± 3.574.2 ± 3.5
Table 2. Results from the BE session of Graz Dataset 2b, using models trained on the BT session to assess generalization capabilities in a new session, are presented. Accuracy results higher than those of FBCSP and SCN are shown in bold. The best approach for each subject is highlighted in red.
Table 2. Results from the BE session of Graz Dataset 2b, using models trained on the BT session to assess generalization capabilities in a new session, are presented. Accuracy results higher than those of FBCSP and SCN are shown in bold. The best approach for each subject is highlighted in red.
BE Session
Accuracy [%]
Subject [#]FBCSPSCNPipeline 0Pipeline 1Pipeline 2Best Pipelines
166.376.264.165.065.365.3
256.151.055.458.657.958.6
351.353.455.359.158.859.1
494.495.795.996.695.996.6
587.287.283.188.886.388.8
676.377.665.677.278.178.1
775.676.386.378.874.786.3
886.375.686.389.487.589.4
982.886.377.582.880.682.8
mean ± SE75.1 ± 4.976.8 ± 5.174.4 ± 4.977.3 ± 4.476.1 ± 4.478.3 ± 4.7
Table 3. Accuracy results from the AT session of the Graz Dataset 2a, performing 10-fold cross-validation. Each pipeline involving NMA demonstrates improvements in different subjects. Apart from Pipeline 0, each pipeline shows increased accuracy compared with the FBCSP winner of the BCI competition (considered as a benchmark). Values for pipelines that outperform FBCSP are in bold. The best pipeline per subject is highlighted in green. Notice that apart from Pipeline 0, each pipeline achieves higher average accuracy across subjects compared with FBCSP.
Table 3. Accuracy results from the AT session of the Graz Dataset 2a, performing 10-fold cross-validation. Each pipeline involving NMA demonstrates improvements in different subjects. Apart from Pipeline 0, each pipeline shows increased accuracy compared with the FBCSP winner of the BCI competition (considered as a benchmark). Values for pipelines that outperform FBCSP are in bold. The best pipeline per subject is highlighted in green. Notice that apart from Pipeline 0, each pipeline achieves higher average accuracy across subjects compared with FBCSP.
AT Session
Accuracy [%]
Subject [#]FBCSPPipeline 0Pipeline 1Pipeline 2Pipeline 3Pipeline 4Pipeline 5Pipeline 6Best Pipelines
178.166.377.476.080.979.575.379.980.9
246.545.149.746.249.350.749.353.153.1
381.975.084.783.779.586.881.987.287.2
449.349.054.551.453.552.852.154.554.5
558.056.355.665.554.960.153.859.065.5
650.052.458.056.352.853.555.652.858.0
778.877.880.684.479.276.776.779.584.4
885.484.788.983.083.086.883.785.188.9
983.380.283.079.578.885.186.583.786.5
mean ± SE67.9 ± 5.565.2 ± 5.070.3 ± 5.269.5 ± 5.068.0 ± 4.970.2 ± 5.268.3 ± 5.170.5 ± 5.073.2 ± 5.1
Table 4. Results on the AE session of Graz Dataset 2a, using the models trained on the AT session to assess generalization capabilities in a new session. Accuracy results higher than those of FBCSP and SCN are shown in bold. The best approach for each subject is highlighted in red.
Table 4. Results on the AE session of Graz Dataset 2a, using the models trained on the AT session to assess generalization capabilities in a new session. Accuracy results higher than those of FBCSP and SCN are shown in bold. The best approach for each subject is highlighted in red.
AE Session
Accuracy [%]
Subject [#]FBCSPSCNPipeline 0Pipeline 1Pipeline 2Pipeline 3Pipeline 4Pipeline 5Pipeline 6Best Pipelines
167.871.457.367.766.767.771.271.271.971.9
247.539.246.949.048.345.149.044.146.249.0
383.382.577.184.482.383.783.782.681.384.4
453.558.753.157.658.355.663.254.557.663.2
532.444.338.236.843.135.839.236.538.543.1
642.646.941.342.743.847.942.442.043.447.9
780.676.974.778.183.378.178.878.178.183.3
882.274.877.179.280.681.380.980.980.280.9
971.775.970.171.577.171.573.377.177.177.1
mean ± SE62.4 ± 6.363.4 ± 1.959.5 ± 5.263.0 ± 5.764.8 ± 5.663.0 ± 5.864.6 ± 5.763.0 ± 6.263.8 ± 5.866.7 ± 5.5
Table 5. Comparison of accuracies on the AE session of Graz Dataset 2a between the FBCSP and SCN benchmarks, the best NMA result per subject, and the best NMA result across subjects. The values clearly demonstrate that sharing NMA information across subjects can further improve classification performance. The highest accuracy for each subject is highlighted in bold.
Table 5. Comparison of accuracies on the AE session of Graz Dataset 2a between the FBCSP and SCN benchmarks, the best NMA result per subject, and the best NMA result across subjects. The values clearly demonstrate that sharing NMA information across subjects can further improve classification performance. The highest accuracy for each subject is highlighted in bold.
AE Session
FBCSPSCNNMA per SubjectNMA Cross Subjects
Subject [#]Accuracy [%]
167.871.471.972.6
247.539.249.051.7
383.382.584.484.4
453.558.763.264.2
532.444.343.143.1
642.646.947.948.6
780.676.983.384.4
882.274.880.981.3
971.775.977.178.5
mean ± SE62.4 ± 6.363.4 ± 5.566.7 ± 5.567.6 ± 5.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Frosolone, M.; Prevete, R.; Ognibeni, L.; Giugliano, S.; Apicella, A.; Pezzulo, G.; Donnarumma, F. Enhancing EEG-Based MI-BCIs with Class-Specific and Subject-Specific Features Detected by Neural Manifold Analysis. Sensors 2024, 24, 6110. https://doi.org/10.3390/s24186110

AMA Style

Frosolone M, Prevete R, Ognibeni L, Giugliano S, Apicella A, Pezzulo G, Donnarumma F. Enhancing EEG-Based MI-BCIs with Class-Specific and Subject-Specific Features Detected by Neural Manifold Analysis. Sensors. 2024; 24(18):6110. https://doi.org/10.3390/s24186110

Chicago/Turabian Style

Frosolone, Mirco, Roberto Prevete, Lorenzo Ognibeni, Salvatore Giugliano, Andrea Apicella, Giovanni Pezzulo, and Francesco Donnarumma. 2024. "Enhancing EEG-Based MI-BCIs with Class-Specific and Subject-Specific Features Detected by Neural Manifold Analysis" Sensors 24, no. 18: 6110. https://doi.org/10.3390/s24186110

APA Style

Frosolone, M., Prevete, R., Ognibeni, L., Giugliano, S., Apicella, A., Pezzulo, G., & Donnarumma, F. (2024). Enhancing EEG-Based MI-BCIs with Class-Specific and Subject-Specific Features Detected by Neural Manifold Analysis. Sensors, 24(18), 6110. https://doi.org/10.3390/s24186110

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop