Next Article in Journal
Event-Triggered Asynchronous Filter of Nonlinear Switched Positive Systems with Output Quantization
Next Article in Special Issue
Infrared Small Target Detection Based on Partial Sum Minimization and Total Variation
Previous Article in Journal
Multi-Stage Multi-Product Production and Inventory Planning for Cold Rolling under Random Yield
Previous Article in Special Issue
Enhanced Convolutional Neural Network Model for Cassava Leaf Disease Identification and Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Class-Incremental Learning Method Based on Preserving the Learned Feature Space for EEG-Based Emotion Recognition

by
Magdiel Jiménez-Guarneros
* and
Roberto Alejo-Eleuterio
Division of Postgraduate Studies and Research, National Technological of Mexico, Technological Institute of Toluca, Metepec 52149, Mexico
*
Author to whom correspondence should be addressed.
Mathematics 2022, 10(4), 598; https://doi.org/10.3390/math10040598
Submission received: 13 January 2022 / Revised: 8 February 2022 / Accepted: 11 February 2022 / Published: 15 February 2022

Abstract

:
Deep learning-based models have shown to be one of the main active research topics in emotion recognition systems from Electroencephalogram (EEG) signals. However, a significant challenge is to effectively recognize new emotions that are incorporated sequentially, as current models must perform retraining from scratch. In this paper, we propose a Class-Incremental Learning (CIL) method, named Incremental Learning preserving the Learned Feature Space (IL2FS), in order to enable deep learning models to incorporate new emotions (classes) into the already known. IL2FS performs a weight aligning to correct the bias on new classes, while it incorporates margin ranking loss and triplet loss to preserve the inter-class separation and feature space alignment on known classes. We evaluated IL2FS over two public datasets (DREAMER and DEAP) for emotion recognition and compared it with other recent and popular CIL methods reported in computer vision. Experimental results show that IL2FS outperforms other CIL methods by obtaining an average accuracy of 59.08 ± 08.26% and 79.36 ± 04.68% on DREAMER and DEAP, recognizing data from new emotions that are incorporated sequentially.

1. Introduction

Emotion analysis has shown to be an important part of research fields such as human–computer interaction and health care, in order to improve the interactive experience and understand the behavior of patients [1,2]. Existing approaches in emotion recognition characterize the responses of emotions in two main modalities [3,4]: behavioral and physiological signals. The first type of modality includes those approaches based on facial expression [5,6], speech emotion recognition [7] and body language. Unlike this type of modality, the physiological signals provide a reliable way to recognize emotions since these signals are produced by the human body that may not be susceptible to subjective approaches based on behavioral signals [8]. In this sense, Electrocardiogram (ECG) [9], Electromyography (EMG) [10], Electroencephalogram (EEG) [4] or even a combination of them [11,12], have been used for emotion recognition. Among these physiological-signal-based approaches, EEG has provided a reliable and promising indicator to identify different emotional states, as it directly reflects brain activity [12]. Furthermore, EEG is a non-invasive device, easy to use, and has a low cost [4,13]. Thus, EEG has been widely used in emotion recognition systems in the last years [3,8,13,14,15,16,17,18].
Reported works have been mainly focused on extracting discriminative EEG emotional features and building more effective emotion recognition systems. The collected EEG signals are usually analyzed in three categories to extract discriminative features: time domain (e.g., statistics of signal), frequency domain (e.g., differential entropy), and time-frequency domain (e.g., Fourier transform). In this direction, many methods have been proposed via machine learning to leverage the features extracted from EEG signals [17,18,19]. Recently, several methods are gradually moving towards the deep learning-based approaches, becoming dominant in EEG-based emotion recognition [3,8,13,14,15,16,20,21]. For example, different deep learning methods have been proposed to consider the spatial information, such as convolutional neural networks (CNNs) [3,14,16], capsule networks (CapsNets) [21] and graph neural networks (GNN) [8,13]. Likewise, attention mechanisms and recurrent networks [15] have been used to extract spatial and temporal information as emotion features.
Although remarkable progress has been achieved, there is a growing demand for adaptive, scalable, and responsive deep learning methods for emotion recognition tasks. Reported works are focused on recognizing emotions with fixed models while being unable to incorporate other emotions into their knowledge. New emotions may be recorded over time so that devices with pre-installed emotion recognition models may fail to recognize this new knowledge. Whenever samples from a new emotion become available, deep neural network models require retraining the whole model from scratch. This issue may be infeasible both in time or storage while using all training data or when the size of the main memory is limited [22]. Instead, the knowledge learned by a trained model should only be modified by using samples from a new emotion. In this sense, Class-incremental learning (CIL) provides a solution when new samples emerge, updating the knowledge of the model according to samples from new classes, avoiding re-configure the entire system [23].
CIL methods have been widely studied in computer vision [22] since several works have shown that deep learning models suffer from catastrophic forgetting when they are trained incrementally [24]. The catastrophic forgetting is the performance degradation of a neural network model affecting previously learned concepts whenever new ones are incorporated sequentially [25]. Different approaches have rapidly emerged to alleviate catastrophic forgetting. A first approach extends the model capacity to accommodate the latest knowledge as new data are integrated [26,27]. Although no sample is retained during incremental stages, these works may not scale well in specific scenarios since new weights are added each time. A second approach [28,29,30,31] uses a fixed model to generate feature representations across different incremental stages while multiple classifiers are trained for new classes. Although the retraining of the entire model is avoided, the performance of these methods depends on the quality of an initial representation, producing sub-optimal classification results in some cases [22]. Moreover, a third approach [25,32,33,34,35,36,37,38,39,40], named memory replay, stores a small set of representative samples from old classes and updates deep learning models via Fine-tuning (FT) across different incremental stages. The memory replay-based approach has shown better performance than previous approaches [35], but certainly the catastrophic forgetting is still under-studied. Mainly, in EEG-based signal recognition, Lee et al. [41] explored CIL for the imagined speech recognition task, but the authors used one of the most straightforward memory replay-based methods under an undemanding evaluation, as only a single incremental stage was tested for CIL. On the other hand, no work has been reported to study the dynamic changes in class for the EEG-based emotion recognition task. Thus, this research focuses on studying CIL for emotion recognition from EEG signals to enable deep learning models to incorporate new emotions into already known.
In this paper, we introduce Incremental Learning preserving the Learned Feature Space (IL2FS), a CIL method to address the catastrophic forgetting in EEG-based emotion recognition. The proposed method aims to preserve the feature space learned over past incremental stages, performing a bias correction on new classes, as well as encouraging the inter-class separation and feature space alignment over old classes. Firstly, we use Weighting Aligning (WA) [36] for bias correction on the weights at the output layer since class imbalance is present. Secondly, we use margin ranking loss to set a margin between scores of the ground-truth from old classes and their nearest score from any class (old or new), instead of only ensuring a separation between old and new classes, as reported in [33]. Finally, unlike previous CIL works for embedding networks [42,43,44], we propose to use triplet loss [45] to maintain the feature space alignment of old classes. IL2FS was implemented on a Capsule Network (CapsNet) architecture, which presents one of the best performances in terms of accuracy for emotion recognition [21]. We evaluate and validate our proposal on incremental learning tasks over two public datasets, DREAMER [46] and DEAP [11], using a reduced set of samples from old classes and the maximum number of incremental stages that may be built for each dataset.
The main contributions of this work are:
  • We present a Class-incremental Learning method, named IL2FS, for emotion recognition from EEG signals, addressing the catastrophic forgetting problem.
  • IL2FS incorporates a strategy based on bias correction of the new classes while ensuring an inter-class separation and feature alignment of the old classes. This strategy allows better preservation of the learned knowledge for a greater number of incremental stages and a reduced number of reserved samples in memory.
  • We conduct experiments on two benchmarks, DEAP and DREAMER, for emotion recognition research. The proposed method achieves a significant improvement when compared with existing CIL methods.
The rest of this paper is organized as follows: in Section 2, we review previous works on class-incremental learning. Section 3 describes the proposed method in detail. Section 4 presents datasets, preprocessing procedure, neural network architecture and experimental setup. The corresponding results are reported in Section 5. Finally, the discussion and conclusions are reported in Section 6 and Section 7.

2. Related Work

Existing works in EEG-based emotion recognition have focused on dynamic data distribution changes, but dynamic changes in class have not been studied yet. In [41], the authors explored CIL using a memory replay-based approach for the imagined speech task. Even so, a simple method [47] based on fine-tuning and the nearest neighbor classifier was adopted. Likewise, an undemanding evaluation was performed since only a single new class was tested, while a considerable percentage of data from old classes is reserved in memory when a new class is added. On the other hand, several CIL methods are available in computer vision to address the catastrophic forgetting problem. Among different approaches, reported in [22], we are interested in memory replay-based methods since they have shown superior performance in terms of accuracy. Thus, we describe several methods based on memory replay to deal with the catastrophic forgetting problem. We group these methods according to the problem they address.
Less forgetting. Knowledge distillation [48] was introduced as a regularizer on the outputs of a reference network and a new network in [49], in order to preserve the predictions of classes learned at previous CIL stages. For this, knowledge distillation aims to keep the new network weights close to the weights of the reference network. Moreover, Hou et al. [33] presented Learning a Unified Classifier Incrementally via Rebalancing (LUCIR), which introduces a less-forget constraint through the cosine distance, considering the local geometric structures of old classes in their feature space. More recently, Simon et al. [25] proposed a distillation loss, named Geodesic, by adopting the concept of geodesic flow between two tasks, that is, the gradual changes between tasks projected in intermediate subspaces.
Bias correction. In this group, CIL methods focus on updating the neural network weights in order to calibrate the bias produced by the class imbalance of representative samples. Wu et al. [50] proposed Bias correction (BiC) to rectify the weights of the model output, but a validation set is still required. In [33], authors observed that magnitudes of the weight vectors for new classes are higher than those of old classes, then, cosine normalization is used over the output layer to reduce the impact of imbalanced data. In this sense, Incremental Learning with Dual Memory (LI2M) [47] corrects scores of old classes storing their statistical information in an additional memory. Classifier Weights Scaling for Class Incremental Learning (ScaIL) [51] rectifies the weights of old classes to make them more comparable to those of new classes. Zhao et al. [36] proposed Weight Aligning (WA) to correct the biased weights at the output layer once the training process has ended. For this, only weight vectors of new classes are aligned to those of old classes using normalization.
Inter-class separation. The knowledge distillation loss has proven to be useful while producing more discriminative results within old classes when a bias correction is performed [36]. However, distillation loss may not be sufficient to ensure an inter-class separation between old and new classes since decision boundaries are re-configured during training over new classes. Thus, authors in [33] introduced margin ranking loss to encourage a margin that separates old and new classes. Chaudhry et al. [37] used bilevel optimization to update the model with new classes, keeping predictions intact on anchor points of old classes that lie close to the class decision boundaries.
Representative samples. Some strategies have been reported to select representative samples of old classes in order to avoid the model from overfitting to new classes. The baseline method, named Herding [32,52], selects the closest samples as most representative of a class, based on a histogram of the distances to the mean sample of that class. Authors in [53] introduced a more complex solution, named Mnemonics, which uses a strategy based on meta-learning to update the memory via gradient descent, selecting those samples located on boundary decisions. Generative solutions may also be found in [54,55], where artificial samples are drawn from each incremental stage, using generative adversarial networks (GANs). However, since GANs have proven to be difficult to optimize, they present scalability issues.

3. Proposed Method

In this section, we introduce the proposed method in detail. First, the Class-incremental learning setting is described. Then, we introduce an overview of the proposed method and its components. Finally, the training algorithm of the proposed method is presented.

3.1. Class Incremental Learning Setting

This research is focused on Class-Incremental Learning (CIL) based on the memory replay approach [22,32], where the neural network model complexity is maintained constant through S incremental stages, while new emotions are sequentially incorporated. In each incremental stage, samples from new emotions and a few samples from old emotions are available to retrain an existing neural network model.
Let X be a feature space with a label space Y belonging to classes (emotions) in C . A labeled dataset is defined as D = { ( x , y ) | x X , y Y } . We assume one initial stage and S incremental stages, where C is split into S + 1 sets C 0 , C 1 , C S with C = C 0 C 1 C S and C i C j = for i j . A budget is determined for the memory M = { ( x , y ) | x X , y Y } , which is used to store a limited amount of representative samples from old classes. In the initial stage, a deep neural network model is trained on a labeled dataset D 0 . Next, a representative set of samples E 0 is selected and stored in memory M as a replacement of D 0 , with | E 0 | | D 0 | . In the incremental stage s, a deep network model is updated using the labeled dataset D s and memory M , that is, D s M . Notice that M now contains representative samples of old classes E 0 : s 1 from incremental stage 0 to s 1 . We assume all training samples in D s are available to train a neural network. In CIL, the main objective is to use a deep network model and D s M to accurately classify samples belonging to old and new classes in each incremental stage s, avoiding catastrophic forgetting.
A deep neural network model is usually denoted as a labeling function f with trainable weights Φ , such that y ^ = f ( x ; Φ ) . The function f may be represented as composite of two functions, f enc f cls . Here, f enc represents the part of network that encodes an input x into a latent feature representation z , that is, z = f enc ( x ; θ ) ; θ is the set of trainable weights. Then, latent features z are fed to a feature labeling function f cls with weights ϕ , in order to produce a classification score y ^ , i.e., y ^ = f cls ( z ; ϕ ) . In CIL, the number of classes of the model output increases at each incremental stage. Thus, the network model f is expected to classify | C s | more classes at incremental stage s than at stage s 1 .

3.2. Overview of the Proposed Method

The proposed method, named Incremental Learning preserving the Learned Feature Space (IL2FS), faces the catastrophic forgetting problem aiming to preserve the learned feature space from old classes. For this, IL2FS performs a bias correction of new classes, while the inter-class separation and feature space alignment of old classes are ensured. Firstly, a bias correction is performed on the weights at the output layer via Weight Aligning [36], as imbalanced data are present when trained over a reduced set of representative samples of old classes. Then, an inter-class separation is encouraged between scores from old classes and their nearest class (old or new) via margin ranking loss, instead of only encouraging a separation between old and new classes, as reported in [33]. Finally, since that new knowledge may modify the learned feature space at previous CIL stages, we propose to use triplet loss [45] to preserve the feature space alignment of old classes.
The complete flowchart is shown in Figure 1 and the overall objective can be written as follows
L inc ( D s , M , f ( s 1 ) ; Φ ) = β · L tri + L cls + α · L mr ,
where L tri is the triplet loss, L cls is a classification loss, and L mr is the margin ranking loss. λ , α , β are the trade-off hyper-parameters.
As shown in Figure 1, the network model f 0 is trained at stage 0 on D 0 , using the classification loss. Next, the Herding method [32,52] is employed to select m representative samples to be stored in memory M . At the incremental stage s, weights Φ are initialized using those learned at stage s 1 . Then, the network model f s is retrained on D s M , using loss function L inc . Exponential Moving Average (EMA) [56] is also incorporated into IL2FS in order to stabilize the training of f s over n training steps:
Φ EMA ( n ) = ( 1 λ EMA ) · Φ EMA ( n 1 ) + λ EMA · Φ ( n ) ,
where Φ EMA ( n ) is the EMA of successive Φ weights over n and λ EMA is the decay rate or momentum. Then, at the end of the model’s training, Weighting Aligning is used to align the norms of the weight vectors between old and new classes at the output layer. Likewise, m representative samples are selected on D s M , considering a balanced selection. This procedure is repeated every time new classes emerge, which must be incorporated into an existing model.

3.3. Bias Correction

Weight Aligning (WA) [36] has been used for bias correction, given that a class imbalance is produced by using a reduced set of representative samples of old classes in new incremental stages. Thus, WA rectifies the weight vectors at the output layer of a network model, aligning the norms of the weight vectors between old and new classes.
The output layer is rewritten as
W = ( W old , W new ) ,
where
W old = ( w 1 , w 2 , , w C old ) R d × C old , W new = ( w C old + 1 , , w C new ) R d × C ,
while the norms of the weight vectors are expressed as follows
W old = ( w 1 , w 2 , , w C old ) , W new = ( w C old + 1 , , w C new ) .
Then, the weights of new classes are normalized by using
W ¯ new = γ · W new ,
where
γ = M ( W old ) M ( W new ) .
Here, M ( · ) computes the mean value using these weight vectors.

3.4. Inter-Class Separation

We assume that decision regions of old classes may change during model retraining, as representative samples of old classes are used for this process. Then, an inter-class separation is ensured by setting a margin over class scores throughout the different incremental learning stages.
Margin ranking loss was introduced in [33] to ensure a separation between old and new classes (see Section 2). Unlike previous work, we use a variant of the margin ranking loss to encourage an inter-class separation between the ground-truth score of an old class and its nearest score coming from any class, old or new.
For each sample x in memory M , a separation is encouraged between the ground-truth old classes and their nearest class (old or new). For each sample x in memory M , the score ω ¯ ( x ) of the ground-truth old class is considered positive, while the maximum score ω ¯ k ( x ) among the remaining classes is considered hard negative. We have:
L mr = x { M } k = 1 K max ( b ω ¯ ( x ) + ω ¯ k ( x ) , 0 ) ,
where b is the margin, ω ¯ ( x ) is the score of the ground-truth class for the sample x , and ω ¯ k ( x ) is the nearest class score for x .

3.5. Feature Space Alignment

We incorporate triplet loss [45] to leverage the less forgetting, preserving the alignment of the feature space of classes learned at previous incremental stages. Note that existing strategies are mainly focused on maintaining the same output predictions of old classes (see Section 2). On the other hand, previous works in CIL [42,43,44] have mainly used triplet loss to train embedding networks and ensure an inter-class separation. However, unlike previous works, we incorporate triplet loss to preserve the feature space alignment of old samples, producing near feature representations from f enc ( S 1 ) and f enc ( S ) for the same processed sample. Here, f enc ( S 1 ) is the model learned at the last incremental stage ( s 1 ) and f enc ( S ) is the new model to train in the current stage s. Representations from different samples, processed by f enc ( S 1 ) and f enc ( S ) , are pushed away from each other by a small margin. Note that class labels for the processed samples are not used in our proposal, as an inter-class separation is not pursued.
More specifically, we use triplet loss to push latent feature representations z s 1 = f enc ( S 1 ) ( x 1 ) and z s = f enc ( S ) ( x 1 ) close to each other for the same sample x 1 . Meanwhile, the latent features z s 1 and z s , produced by f enc ( S 1 ) and f enc ( S ) , but coming from samples x 1 and x 2 , are pushed away from each other by a margin.
The triplet loss is defined as follows
L tri = x { M } max ( d ( z a , z p ) d ( z a , z n ) + a , 0 ) ,
where z a is the anchor input, z p is a positive input of the same label as z a , while z n is a negative input of a different label as z a ; a is the margin and d is the cosine dissimilarity measure. Anchor-positive pairs are formed by latent features generated by f enc ( S 1 ) and f enc ( S ) for the same sample, while anchor-negative pairs are formed by latent features generated by f enc ( s 1 ) and f enc ( S ) for a pair of different samples. f enc ( S 1 ) processes all samples within the current batch to generate their respective latent feature representations. After, each featured sample is labeled according to its index into the batch of samples. This procedure is repeated for all samples but using f enc ( S ) ; later, featured samples are concatenated with those obtained by f enc ( S 1 ) . Then, the multi-similarity miner [57] is used to generate anchor-positive pairs ( z a , z p ) and anchor-negative pairs ( z a , z n ) over labeled feature representations in order to preserve the feature alignment of old classes.

3.6. Training of IL2FS

Algorithm 1 presents the training procedure of IL2FS at incremental stage s. First, the set of weights Φ is initialized using weights Φ ( s 1 ) (line 1). Next, we compute latent features for x using the reference model and current model (lines 7–8). Featured samples are labeled according to their indices into the dataset (lines 9–10). Anchor-positive and anchor-negative pairs are generated using the Multi-similarity miner (line 11) to be employed in triplet loss L tri . Then, scores for ground-truth old classes and their nearest classes are computed in order to be used in margin ranking loss L mr (lines 12–14). After, neural network model f s is trained using the loss function L inc (line 15). Note that L inc is composed of classification loss L cls , triplet loss L tri and margin ranking loss L mr . The EMA weights Φ EMA ( n ) are computed from Φ ( n ) (line 16). After training f s , weight vectors of the output layer are rectified employing the Weighting Aligning method (line 18). Finally, the memory M is updated by selecting m representative samples on D s M by means of the Herding method (line 19).
Algorithm 1 Training algorithm of IL2FS at incremental stage s.
Inputs: D s – training labeled dataset from new classes; M – memory containing representative samples from old classes; f ( s 1 ) ( · ; Φ s 1 ) – reference model trained at incremental stage s 1 ; λ mr , α , β – trade-off hyperparameters; λ EMA – decay rate; η – learning rate; n – number of epochs.
Output: f s ( · ; Φ EMA ) – a trained neural network model; M – updated memory with representative samples from old classes.
1: Initialize Φ s with Φ ( s 1 ) .
2:  x o l d , y o l d M
3:  x n e w , y n e w D s
4:  x x o l d x n e w
5:  y y o l d y n e w
6: repeat
7:   z r e f f enc ( S 1 ) ( x )   ▹ Compute features for samples using the reference model
8:   z c u r f enc ( S ) ( x )   ▹ Compute features for samples using the current model
9:   v r e f GenerateLabels ( z r e f )   ▹ Assign labels based on indices into the dataset
10:   v c u r GenerateLabels ( z c u r )
11:   z a , z p , z n MultiSimilartyMiner ( z r e f z c u r , v r e f v c u r )   ▹ Generate anchor-positive and anchor-negative pairs
12:   ω ¯ ( x ) f ( s 1 ) ( x o l d )   ▹ Compute scores for samples from old classes using the reference model
13:   ω ¯ a ( x ) f s ( x )   ▹ Compute scores for all samples using the current model
14:   ω ¯ k ( x ) NearestClass ( ω ¯ ( x ) , ω ¯ a ( x ) )   ▹ Obtain scores from the nearest classes to old classes
15:   Φ ( i ) Φ ( i 1 ) η · [ β · L tri ( z a , z p , z n ; Φ ( i 1 ) ) + L cls ( f s ( x ) , y ; Φ ( i 1 ) ) + α · L mr ( ω ¯ ( x ) , ω ¯ k ( x ) ; Φ ( i 1 ) ) ]
16:   Φ EMA ( i ) = ( 1 λ EMA ) · Φ EMA ( i 1 ) + λ EMA · Φ ( i )
17: until n epochs are reached
18:  Φ EMA WeightingAligning ( Φ EMA )   ▹ Bias correction
19:  M Herding ( D s M ) .   ▹ memory is updated using the Herding method
20: return f s ( · ; Φ EMA ) , M

4. Experimental Design

This section first describes two public datasets used in our experiments. Then, the neural network architecture, comparison methods and implementation details are introduced. (Code is available at https://github.com/mjmnzg/IL2FS. Accessed on 11 January 2022).

4.1. Datasets

Experiments were performed on two public datasets, DREAMER [46] and DEAP [11], since they are benchmarks for emotion recognition research [3,14,15,21]. DREAMER is a multi-channel dataset containing records of nine emotions from EEG signals per subject. Likewise, DEAP is a large-scale dataset containing EEG signals with different emotional evaluations. More importantly, both datasets were selected since a high number of classes may be obtained from EEG data, making it useful for the analysis of the catastrophic forgetting problem in emotion recognition.
The DREAMER dataset comprises EEG data from 23 subjects (14 male and nine female). EEG data were collected while the subjects watched 18 film clips, which contain cut-out scenes to evoke nine emotions: calmness, surprise, amusement, fear, excitement, disgust, happiness, anger, and sadness. The length of each film clip is between 65 to 393 s (M = 199 s). EEG signals were recorded at a sampling rate of 128 Hz using an Emotiv EPOC system that uses 16 electrodes, following locations according to the International 10–20 systems: AF3, F7, F3, FC5, T7, P7, O1, O2, P8, T8, FC6, F4, F8, AF4, M1, and M2. Sensor M1 acts as a ground reference, while M2 is a feed-forward reference; then, the remaining 14 electrodes were recorded and used for feature extraction. EEG data from all subjects have 18 experimental EEG trials, two per elicited emotion. Each EEG trial begins with a neural film to help the subjects return to the neutral emotion state, while data serve as a baseline. EEG signals of each trial were filtered with Hamming bandpass linear phase FIR filters to extract frequencies inside the ranges of interest (4–30 Hz). Likewise, artifacts were removed by using artifact subspace reconstruction (ASR) [58]. At the final step, the Common Average Reference (CAR) method [59] was applied to compute the average value over all electrodes and subtracts it from each sample of each electrode. In our experiments, we adopt a discrete categorization instead of a dimensional categorization, with nine classes available.
The DEAP dataset contains EEG and peripheral physiological signals from 32 subjects while watching 40 music videos. EEG signals were collected using a cap of 32 electrodes, placed according to the international 10–20 system [60]. For this, a sampling rate of 512 Hz was used, then downsampled to 128 Hz. We used the pre-processed data (https://www.eecs.qmul.ac.uk/mmv/datasets/deap/readme.html. Accessed on 1 June 2021), where each trial contains 60 s of recorded signals under stimulation and 3 s of baseline signals in a relaxed state. A bandpass filter from 4.0–45.0 Hz was applied over EEG signals, and eye artifacts were removed as in [11] using independent component analysis (ICA). EEG data were averaged to the common reference. Subjects rate their levels of arousal, valence, linking, and dominance from 1 to 9 for each music video. In our experiments, we adopt a multi-class categorization scheme, combining discrete ratings of valence, arousal and dominance. Firstly, we divide each emotion dimension into two categories using a rating of 5 as threshold: low/high valence, low/high arousal and low/high dominance. Secondly, we label each EEG trial used as a combination of binary categorization in three dimensions. For instance, its label is 0 when the rating is low for the three dimensions, while the label is 1 when the rating for valence and arousal is low, but the rating for dominance is high. Finally, the recognition task is a multi-class classification composed of a maximum of 8 classes, given that not all subjects rate for every level of arousal, valence and dominance.

4.2. Preprocessing

We applied the preprocessing procedure of baseline removal on EEG signals as in the works reported by [3,15,21,61] since this method highlights the effects of stimulated emotions. We begin by using a non-overlapping window to slice baseline signals into N segments of 1 s for each trial and C electrodes. From the set of N segments, we obtain the mean segment, which represents the base emotional state without stimulation. Next, the mean segment is subtracted from the EEG signals under stimulation. The obtained differences represent the electrical changes in the brain under stimulation. Following this pre-processing, 1080 EEG samples are obtained for each subject in DREAMER, where 60 segments are obtained from each experimental trial; 18 experimental trials per subject. In this direction, each trial in DEAP is divided into 60 segments, each one containing 128 sampling points. Then, we obtain 2400 EEG samples for each subject since there are 40 trials per subject. Finally, each EEG sample in DREAMER and DEAP is a 32 × 128 matrix and 14 × 128 matrix, composed of the number of electrodes and sampling points, respectively.

4.3. Neural Network Architecture

We adopted a Capsule Network (CapsNet) architecture [21], which showed one of the best accuracy performances for EEG-based emotion recognition research. Figure 2 presents the CapsNet architecture and Table 1 describes the implementation details. Unlike the original CapsNet architecture, we add a module based on an attention mechanism, which includes a Channel-Attention block [62] into the modules from Convolutional to PrimaryCaps. In addition, the bottleneck layer proposed in [21] was removed since it dramatically increases the resources used in memory. To train CapsNet, the classification loss L cls uses the margin and reconstruction losses, as suggested in [63]. For this purpose, CapsNet employs a separated margin loss L k for each class k. On the other hand, reconstruction loss L rec uses the sum of squared differences between the outputs of a decoder and the input EEG signal values. This decoder consists of 3 fully-connected layers that model the EEG signals.

4.4. Comparison Methods

We compared IL2FS with eight popular and recent CIL methods based on memory replay: Fine-tuning (FT) [51], Fine-tuning+Nearest Centroid Classifier (FT+NCC) [41,51], Less without Forgetting (LwF) [49], Incremental Classifier and Representation Learning (iCARL) [32], Mnemonics [53], ScaIL [51], Weighting Aligning (WA) [36], and Geodesic+LUCIR [25]. We selected such CIL methods in our comparison since they arise as promising solutions to address the catastrophic forgetting problem in emotion recognition. All comparison methods were downloaded from repositories of original authors and then adapted for our experiments, except FT and FT+NCC, which do not represent a challenge to implement as they are basic methods. Note that all CIL methods use the same preprocessing procedure and the CapsNet architecture described in the previous sections.

4.5. Implementation Details

We first configured the hyper-parameters for the classification loss of the CapsNet architecture. Thus, the margins m + and m for the separated margin loss L cls were set to 0.9 and 0.1, as suggested in [21,63]. Likewise, the reconstruction loss L rec was scaled by 0.3 during training; this value was selected from {0.01, 0.1, 0.2, 0.3, and 0.5}.
Concerning the specific configuration of our proposed method, we adopted a mean layer instead of a normalization layer (N14) in the CapsNet architecture. For L tri , we used a margin a = 0.1 since a feature space alignment is pursued between extracted features from a reference network model and a new network model; a larger margin showed to affect the classification results negatively. To ensure an inter-class separation via margin ranking loss L mr , we used a margin b equal to 5, which was selected from {1, 3, 5, 8, and 10}. Finally, for trade-off hyper-parameters, we used α = 1 and β = 0.1 , which were selected from {0.01, 0.1, 1, and 2}. We use a momentum λ EMA = 0.995 to place a greater significance on the most recent values.
Table 2 describes the specific hyper-parameters of CIL methods used in our comparison. Similar to our proposal, all hyperparameters were selected via grid search in combination with coordinated descent [64] in order to ensure the best configuration. Specifically, we select a small finite list of values for each hyper-parameter and each value is changed at a time while the rest of the hyper-parameters remains fixed.
Regarding the training algorithm of the CIL methods, we used Adam optimizer employing a mini-batch size of 10; a larger size showed to reduce the classification results using an incremental learning evaluation. For DREAMER, we set a learning rate of 0.001 up to epoch 30, when it decays to 0.0001, keeping this value until epoch 50 when the training concludes. For DEAP, we set an initial learning rate of 0.001 up to epoch 15, when it also decays by a factor of 10, and then holds this value until the end of epoch 20. Other learning rates (0.1, 0.01, 0.001, 0.0001) were evaluated, but they did not improve the accuracy performance. An L1 regularizer was incorporated to CapsNet with a weight decay of 0.0004 for the Adam algorithm.
Our proposal and the comparison methods were implemented with PyTorch and trained on an Intel(R) Core (TM) i7 PC with an Nvidia GTX 1080 graphics card and Ubuntu v20.04 LTS.

4.6. Evaluation

As reported in [22,32], we follow the standard evaluation protocol used for the CIL setting based on the memory replay approach. The Holdout method is applied for a given dataset to build the training and testing data for each available class. Likewise, classes are arranged in a fixed random order. Each method is trained in a class-incremental way on available training data, as described in Section 3.1. At the end of each incremental stage, the resulting classifier is evaluated on testing data for already trained classes. Note that the testing dataset is not revealed to the CIL methods during training in each incremental stage to avoid overfitting. At the end of S incremental stages, we obtain S classification accuracies, averaged and reported as the final result.
We adopted an instantiation of the above protocol for each subject’s data on the DREAMER and DEAP datasets, considering the most challenging scenario possible. Firstly, we start from a model trained on two classes, while remaining classes in DREAMER and DEAP come in 7 and at most six incremental stages, respectively. Secondly, we set the memory size M to approximately 1% of the full training set from each subject in order to store representative samples from old classes. We used 90% of the data of each class for training, while the rest of the data was used for testing. Thus, about ten samples can be stored in memory for DREAMER through 7 incremental stages, while at least 28 samples can be stored for DEAP during six stages. Note that not all subjects in DEAP rate the same levels of arousal, valence and dominance, producing an imbalanced dataset; an oversampling was applied using a random selection. Classes from incremental stages are arranged in sequence with a fixed random order. We performed five repetitions with different partitions of data and different classes, using different random seeds; a stratified sampling was performed with respect to the classes. From accuracy results by training in a class-incremental way, we compute the average and standard deviation over the incremental stages as final results. We assumed that training and testing datasets are independent and identically distributed, i.e., both datasets were drawn from the same distribution. Thus, we did not consider any change of distribution.

5. Results

Table 3 shows the average accuracy and standard deviation for all methods over DREAMER. We observed that IL2FS achieved the best average accuracy (59.08%) with one of the lowest standard deviations (8.26). Notice that IL2FS outperformed the second-best method (Mnemonics) by 8.96 percentage points (pp). Statistical differences were computed among the evaluated methods on the average accuracy of the 23 subjects. The Friedman test was applied, followed by Wilcoxon signed-rank as post hoc with the Finner correction. Friedman’s test showed significant differences among CIL methods ( χ 2 ( 8 ) = 169.99 , p = 0.0 ). The Wilcoxon test indicated that the difference between IL2FS and CIL methods was statistically significant ( p < 0.05 ).
Table 4 shows the average accuracy and standard deviation for all methods on DEAP. We can see that IL2FS achieved the best average accuracy (79.36%) with the lowest standard deviation (4.68). The second-best method was Geodesic+LUCIR, obtaining an average accuracy of 8.94 percentage points shorter than IL2FS. Friedman’s test indicated significant differences among compared methods ( χ 2 ( 8 ) = 230.03 , p = 0.0 ). Wilcoxon signed-rank test revealed that differences between IL2FS and CIL methods are statistically significant ( p < 0.05 ) for the given dataset.
Comparison with baseline.Figure 3 presents a comparison of IL2FS and existing CIL methods concerning the baseline approach (CapsNet-wo-memory), that is, when CapsNet did not include any data from old classes in a CIL training. In addition, we also included the average accuracy when CapsNet is trained using all training samples from old classes (CapsNet-Full) in each incremental stage. We observed that CapsNet-wo-memory obtained the worst accuracy results when samples of old classes are not available in the memory, suggesting the presence of catastrophic forgetting. However, CapsNet improved its accuracy performance when samples of old classes were employed during Fine-tuning (FT). Note that IL2FS and advanced CIL methods improved the average accuracy of FT by incorporating a specific strategy to address the catastrophic forgetting problem, except ScaIL and LwF. Finally, we observed that IL2FS is still exposed to catastrophic forgetting as CapsNet-Full achieved 90.63% and 98.17% on DEAP and DREAMER.

5.1. Ablation Studies

In this section, we present an analysis with respect to the number of reserved samples from old classes. After, we study the impact of the number of new emotions incorporated into the neural network model. Finally, we analyze the impact of each component of IL2FS.

5.1.1. Effect of the Number of Reserved Samples

Figure 4 shows the comparison of IL2FS with CIL methods, when the memory of old samples has a size close to 1%, 2%, and 5% of the size of the full training set for each subject on DREAMER and DEAP. As expected, CIL methods improved their accuracy performance when more samples were stored in the memory. However, we can see that IL2FS still maintains the best average accuracy for different sizes of the reserved samples. For DREAMER, 66.73% and 75.06% were obtained by IL2FS when the memory is close to 2% and 5% of the size of the full training set. For DEAP, IL2FS achieved average accuracies of 85.35% and 90.73% using memory sizes of 2% and 5%, respectively. Note that our proposal obtained a greater gain in average accuracy than the comparison methods when a smaller number of samples from old classes is reserved in the memory.

5.1.2. Effect of the Number of Incremental Stages

Figure 5 shows the average accuracy of IL2FS for each incremental stage in comparison to CIL methods on DREAMER. We reported the average accuracy of CIL methods over all subjects, employing memory sizes of 1% and 5% of the size of the full training dataset. Accuracy results for CapsNet-wo-memory and CapsNet-Full were also included as baselines. We observed that a CIL strategy helps reduce catastrophic forgetting by improving the accuracy performance of CapsNet-wo-memory. However, note that CIL methods decrease their accuracy performance when the number of stages is increased. It is worth mentioning that IL2FS achieved the best average accuracies throughout different incremental stages. In addition, IL2FS obtained a greater gain than existing methods during the last incremental stages because fewer samples from old classes can be stored in the memory.

5.1.3. Effect of Each Component of IL2FS

The proposed method comprises three main strategies: weight aligning for bias correction, margin ranking loss for inter-class separation and triplet loss for a feature space alignment of old classes. Table 5 shows the average accuracy over all subjects for each evaluated dataset. Note that Fine-tuning achieved an average accuracy of 41.69% and 58.69% over DREAMER and DEAP, respectively. By using weight aligning for bias correction, the average accuracy is improved by 4.86 and 3.99 percentage points over the DREAMER and DEAP datasets. A variant of margin ranking loss was incorporated to encourage a separation between each ground-truth old class and its nearest class (old or new). This modification allowed IL2FS to outperform Fine-tuning+Weight aligning by 10.22 and 15.33 percentage points over DREAMER and DEAP, respectively. In addition, triplet loss was used to keep a similar alignment of the feature space of old classes. From experiments, we found that by encouraging such alignment, an improvement of 2.31 and 1.35 percentage points is observed in average accuracy on DREAMER and DEAP, respectively.

6. Discussion

Experiments showed that a standard deep learning model for emotion recognition (CapsNet) degrades its accuracy performance when trained in a class-incremental way over only samples from new emotions. This problem, known as catastrophic forgetting, is presented because previously learned emotions are negatively affected when new ones are incorporated into the classifier model. Thus, unlike previous works as reported in [3,8,13,14,15,16,20,21], this research is focused on studying the catastrophic forgetting problem in EEG-based emotion recognition.
By incorporating existing CIL methods to CapsNet, we showed that classification results of the baseline approach (CapsNet-wo-memory) can be improved, suggesting that CIL methods can help mitigate the catastrophic forgetting in EEG-based emotion recognition. However, experimental results on two public datasets showed that existing CIL methods do not ensure high average accuracies. Thus, a CIL method was developed and validated to address the catastrophic forgetting problem.
Previously, Lee et al. [41] studied the CIL over the imagined speech task from EEG signals. Authors used fine-tuning and the nearest neighbor classifier to address the catastrophic forgetting, however, they stored 20% of the full data of every old class in each incremental stage. Furthermore, only one incremental stage was used for CIL evaluation, while more stages are needed to observe the negative impact of catastrophic forgetting. On the other hand, our experiments consider a rigorous evaluation over two datasets for emotion recognition, including popular and recent CIL methods in our comparison. Based on our results, we found that IL2FS outperformed existing CIL methods on two public datasets: DREAMER and DEAP. Note that we integrated a weighting aligning as the WA method for bias correction, but an inter-class separation and a feature space alignment were also considered by IL2FS, outperforming WA by 14.28 pp and 17.35 pp on DREAMER and DEAP, respectively. Like IL2FS, the Mnemonics and Geodesic+LUCIR methods ensure an inter-class separation via margin ranking loss, but IL2FS encourages the separation between old classes and their nearest one, including old or new, instead of only ensuring a separation between old and new classes. Although Mnemonics and Geodesic+LUCIR also consider strategies for bias correction and an alignment of output predictions, our proposal outperformed Mnemonics by 8.96 pp and 9.95 pp on DREAMER and DEAP, while Geodesic+LUCIR was outperformed by 9 pp and 8.94 pp, respectively. In addition, note that comparison methods, such as LwF, iCARL, WA, Mnemonics, and Geodesic+LUCIR, use different strategies to align the output predictions of old classes to leverage the less forgetting. Unlike these works, IL2FS incorporates triplet loss to preserve the feature space alignment of old classes instead of the output predictions.
Regarding the evaluation which varies the number of reserved samples and the number of incremental stages, IL2FS showed a clear advantage compared to existing methods when the number of reserved samples in the memory is reduced. This issue is also observed when a greater number of CIL stages are achieved since a lower number of samples per class may be stored in memory. The above indicates that IL2FS preserves the learned knowledge better than compared methods throughout different incremental stages on the most challenging scenario possible for the evaluated datasets. On the other hand, as expected, every evaluated method improved its accuracy performance whenever the number of reserved samples in the memory is increased. However, by using a memory size near 5%, IL2FS still obtained the best average accuracy on the DREAMER, while it is similar with respect to the existing CIL methods for the DEAP dataset.
Concerning the effect of each component of IL2FS, weight aligning improved the average accuracy over the Fine-tuning method, which indicates that performing a bias correction is important to reduce the catastrophic forgetting problem in EEG-based emotion recognition. Then, margin ranking loss was incorporated to ensure an inter-class separation between each old class and its nearest class (old or new). Previous work in [33] showed that a separation between old and new classes might be sufficient to help reduce the catastrophic forgetting. However, we found that this strategy [33] on IL2FS obtained a similar average accuracy on DREAMER (58.55 ± 7.33% vs. 58.74 ± 07.56%), but the accuracy performance is drastically reduced on DEAP (53.36 ± 08.84% vs. 79.36 ± 04.68%). These results suggest that it is preferable to encourage an inter-class separation between each old class and its nearest class (old or new) instead of only ensuring a separation between old and new classes. Finally, unlike previous CIL works [42,43,44] where triplet loss is mainly used to train embedding networks and provide an inter-class separation, we used such loss function to maintain the same aligning of the feature space learned at previous incremental stages. For this, IL2FS aims to produce near feature representations coming from a reference model and a new model for the same processed sample, while features from different samples are pushed away from each other by a small margin. This strategy showed to be beneficial for the CIL task in two emotion recognition datasets.
The presented study may contribute to designing and building more adaptive and scalable classifiers, as our study showed a first Class-incremental Learning solution to avoid reconfiguring the entire system when new emotions are incorporated sequentially. For this, we consider an evaluation of the most challenging scenario that may be built over the two public datasets for emotion recognition. However, our study did not consider other CIL settings and evaluation protocols. Furthermore, other preprocessing procedures and neural network architectures were also not explored.

7. Conclusions

In this paper, we presented IL2FS, a CIL method to address the catastrophic forgetting in EEG-based emotion recognition from EEG signals. IL2FS aims to preserve the feature space learned over previous incremental stages, performing a bias correction of new classes and ensuring the inter-class separation and feature space alignment from classes learned at previous incremental stages. The proposed method was incorporated into a Capsule Network architecture for EEG-based emotion recognition. Our experiments showed that IL2FS achieved the best average accuracy over two public emotion datasets, outperforming popular and recent CIL methods under different memory sizes. Furthermore, Friedman and Wilcoxon’s tests showed that IL2FS significantly outperformed existing CIL methods over the evaluated datasets, using the standard protocol for CIL methods based on memory replay. By using IL2FS, better preservation of the learned knowledge is possible when presented with a greater number of incremental stages and a reduced number of reserved samples in memory. In this direction, new emotions may be incorporated into an existing deep neural network classifier without retraining from scratch, employing a set of representative samples of emotions previously learned in a sequential way. However, the presented results suggest that the proposed solution is still exposed to catastrophic forgetting for a high number of incremental stages and limited memory size.
As future work, we are interested in studying the negative effect of batch normalization layers since a bias may be produced over learned statistics from old classes by training over imbalanced data.

Author Contributions

Conceptualization, M.J.-G.; formal analysis, M.J.-G.; investigation, M.J.-G.; methodology, M.J.-G.; project administration, M.J.-G. and R.A.-E.; software, M.J.-G.; supervision, R.A.-E.; writing—original draft preparation, M.J.-G. and R.A.-E.; writing—review and editing, M.J.-G. and R.A.-E. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Mexican Council of Science and Technology (COMECYT), grant number CAT2021-0193.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BiCBias correction
CapsNetCapsule Network
CILClass-incremental Learning
CNNConvolutional Neural Network
DBNDeep Belief Network
EEGElectroencephalogram
FTFine-tuning
NCCNearest Centroid Classifier
GANGenerative adversarial network
GNNGraph Neural Network
iCARLIncremental Classifier and Representation Learning
IL2FSIncremental Learning preserving the Learned Feature Space
LI2MIncremental Learning with Dual Memory
LUCIRLearning a Unified Classifier Incrementally via Rebalancing
LwFLearning without Forgetting
ScaILClassifier Weights Scaling for Class Incremental Learning
WAWeight Aligning

References

  1. Deng, J.J.; Leung, C.H.C.; Milani, A.; Chen, L. Emotional States Associated with Music: Classification, Prediction of Changes, and Consideration in Recommendation. ACM Trans. Interact. Intell. Syst. 2015, 5, 1–36. [Google Scholar] [CrossRef]
  2. Ali, M.; Mosa, A.H.; Al Machot, F.; Kyamakya, K. EEG-based emotion recognition approach for e-healthcare applications. In Proceedings of the 2016 Eighth International Conference on Ubiquitous and Future Networks (ICUFN), Vienna, Austria, 5–8 July 2016; pp. 946–950. [Google Scholar]
  3. Huang, D.; Chen, S.; Liu, C.; Zheng, L.; Tian, Z.; Jiang, D. Differences first in asymmetric brain: A bi-hemisphere discrepancy convolutional neural network for EEG emotion recognition. Neurocomputing 2021, 448, 140–151. [Google Scholar] [CrossRef]
  4. Alarcão, S.M.; Fonseca, M.J. Emotions Recognition Using EEG Signals: A Survey. IEEE Trans. Affect. Comput. 2019, 10, 374–393. [Google Scholar] [CrossRef]
  5. Kim, J.H.; Poulose, A.; Han, D.S. The Extensive Usage of the Facial Image Threshing Machine for Facial Emotion Recognition Performance. Sensors 2021, 21, 2026. [Google Scholar] [CrossRef]
  6. Poulose, A.; Reddy, C.S.; Kim, J.H.; Han, D.S. Foreground Extraction Based Facial Emotion Recognition Using Deep Learning Xception Model. In Proceedings of the 2021 Twelfth International Conference on Ubiquitous and Future Networks (ICUFN), Barcelona, Spain, 5–8 July 2021; pp. 356–360. [Google Scholar]
  7. Han, K.; Yu, D.; Tashev, I. Speech Emotion Recognition Using Deep Neural Network and Extreme Learning Machine. Interspeech 2014, 223–227. [Google Scholar]
  8. Song, T.; Liu, S.; Zheng, W.; Zong, Y.; Cui, Z. Instance-adaptive graph for EEG emotion recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 2–12 February 2020; Volume 34, pp. 2701–2708. [Google Scholar]
  9. Agrafioti, F.; Hatzinakos, D.; Anderson, A.K. ECG pattern analysis for emotion detection. IEEE Trans. Affect. Comput. 2011, 3, 102–115. [Google Scholar] [CrossRef]
  10. Cheng, B.; Liu, G. Emotion Recognition from Surface EMG Signal Using Wavelet Transform and Neural Network. In Proceedings of the 2008 2nd International Conference on Bioinformatics and Biomedical Engineering, Shanghai, China, 16–18 May 2008; pp. 1363–1366. [Google Scholar] [CrossRef]
  11. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis; Using Physiological Signals. IEEE Trans. Affect. Comput. 2012, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  12. Lan, Y.T.; Liu, W.; Lu, B.L. Multimodal Emotion Recognition Using Deep Generalized Canonical Correlation Analysis with an Attention Mechanism. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 19–24 July 2020; pp. 1–6. [Google Scholar]
  13. Zhong, P.; Wang, D.; Miao, C. EEG-based emotion recognition using regularized graph neural networks. IEEE Trans. Affect. Comput. 2020, 1–12. [Google Scholar] [CrossRef]
  14. Shen, L.; Zhao, W.; Shi, Y.; Qin, T.; Liu, B. Parallel Sequence-Channel Projection Convolutional Neural Network for EEG-Based Emotion Recognition. IEEE Access 2020, 8, 222966–222976. [Google Scholar] [CrossRef]
  15. Tao, W.; Li, C.; Song, R.; Cheng, J.; Liu, Y.; Wan, F.; Chen, X. EEG-based Emotion Recognition via Channel-wise Attention and Self Attention. IEEE Trans. Affect. Comput. 2020, 1–12. [Google Scholar] [CrossRef]
  16. Topic, A.; Russo, M. Emotion recognition based on EEG feature maps through deep learning network. Eng. Sci. Technol. Int. J. 2021, 24, 1442–1454. [Google Scholar] [CrossRef]
  17. Salankar, N.; Mishra, P.; Garg, L. Emotion recognition from EEG signals using empirical mode decomposition and second-order difference plot. Biomed. Signal Process. Control 2021, 65, 102389. [Google Scholar] [CrossRef]
  18. Shen, F.; Peng, Y.; Kong, W.; Dai, G. Multi-Scale Frequency Bands Ensemble Learning for EEG-Based Emotion Recognition. Sensors 2021, 21, 1262. [Google Scholar] [CrossRef] [PubMed]
  19. Xu, X.; Zhang, Y.; Tang, M.; Gu, H.; Yan, S.; Yang, J. Emotion Recognition Based on Double Tree Complex Wavelet Transform and Machine Learning in Internet of Things. IEEE Access 2019, 7, 154114–154120. [Google Scholar] [CrossRef]
  20. Chao, H.; Liu, Y. Emotion Recognition From Multi-Channel EEG Signals by Exploiting the Deep Belief-Conditional Random Field Framework. IEEE Access 2020, 8, 33002–33012. [Google Scholar] [CrossRef]
  21. Liu, Y.; Ding, Y.; Li, C.; Cheng, J.; Song, R.; Wan, F.; Chen, X. Multi-channel EEG-based emotion recognition via a multi-level features guided capsule network. Comput. Biol. Med. 2020, 123, 103927. [Google Scholar] [CrossRef] [PubMed]
  22. Belouadah, E.; Popescu, A.; Kanellos, I. A comprehensive study of class incremental learning algorithms for visual tasks. Neural Netw. 2021, 135, 38–54. [Google Scholar] [CrossRef]
  23. Geng, X.; Smith-Miles, K. Incremental Learning; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  24. Kirkpatrick, J.; Pascanu, R.; Rabinowitz, N.; Veness, J.; Desjardins, G.; Rusu, A.A.; Milan, K.; Quan, J.; Ramalho, T.; Grabska-Barwinska, A.; et al. Overcoming catastrophic forgetting in neural networks. Proc. Natl. Acad. Sci. USA 2017, 114, 3521–3526. [Google Scholar] [CrossRef] [Green Version]
  25. Simon, C.; Koniusz, P.; Harandi, M. On learning the geodesic path for incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1591–1600. [Google Scholar]
  26. Liu, Y.; Schiele, B.; Sun, Q. Adaptive aggregation networks for class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2544–2553. [Google Scholar]
  27. Yan, S.; Xie, J.; He, X. DER: Dynamically Expandable Representation for Class Incremental Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3014–3023. [Google Scholar]
  28. Hayes, T.L.; Kafle, K.; Shrestha, R.; Acharya, M.; Kanan, C. Remind your neural network to prevent catastrophic forgetting. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 466–483. [Google Scholar]
  29. Liu, Y.; Parisot, S.; Slabaugh, G.; Jia, X.; Leonardis, A.; Tuytelaars, T. More classifiers, less forgetting: A generic multi-classifier paradigm for incremental learning. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 699–716. [Google Scholar]
  30. Rajasegaran, J.; Khan, S.; Hayat, M.; Khan, F.S.; Shah, M. Itaml: An incremental task-agnostic meta-learning approach. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 13588–13597. [Google Scholar]
  31. Hu, W.; Qin, Q.; Wang, M.; Ma, J.; Liu, B. Continual Learning by Using Information of Each Class Holistically. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event. 2–9 February 2021; Volume 35, pp. 7797–7805. Available online: https://ojs.aaai.org/index.php/AAAI/article/view/16952 (accessed on 12 January 2022).
  32. Rebuffi, S.A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. Icarl: Incremental classifier and representation learning. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2001–2010. [Google Scholar]
  33. Hou, S.; Pan, X.; Loy, C.C.; Wang, Z.; Lin, D. Learning a unified classifier incrementally via rebalancing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 831–839. [Google Scholar]
  34. Iscen, A.; Zhang, J.; Lazebnik, S.; Schmid, C. Memory-efficient incremental learning through feature adaptation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 699–715. [Google Scholar]
  35. Prabhu, A.; Torr, P.H.; Dokania, P.K. Gdumb: A simple approach that questions our progress in continual learning. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 524–540. [Google Scholar]
  36. Zhao, B.; Xiao, X.; Gan, G.; Zhang, B.; Xia, S.T. Maintaining discrimination and fairness in class incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13208–13217. [Google Scholar]
  37. Chaudhry, A.; Gordo, A.; Dokania, P.K.; Torr, P.H.S.; Lopez-Paz, D. Using Hindsight to Anchor Past Knowledge in Continual Learning. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, AAAI 2021, Virtual Event. 2–9 February 2021; pp. 6993–7001. Available online: https://ojs.aaai.org/index.php/AAAI/article/view/16861 (accessed on 12 January 2022).
  38. Bang, J.; Kim, H.; Yoo, Y.; Ha, J.W.; Choi, J. Rainbow Memory: Continual Learning with a Memory of Diverse Samples. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8218–8227. [Google Scholar]
  39. Kurmi, V.K.; Patro, B.N.; Subramanian, V.K.; Namboodiri, V.P. Do not Forget to Attend to Uncertainty while Mitigating Catastrophic Forgetting. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikola, HI, USA, 5–9 January 2021; pp. 736–745. [Google Scholar]
  40. Tang, S.; Chen, D.; Zhu, J.; Yu, S.; Ouyang, W. Layerwise optimization by gradient decomposition for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9634–9643. [Google Scholar]
  41. Lee, D.Y.; Lee, M.; Lee, S.W. Decoding Imagined Speech Based on Deep Metric Learning for Intuitive BCI Communication. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 1363–1374. [Google Scholar] [CrossRef]
  42. Huo, J.; Zyl, T.L.v. Comparative Analysis of Catastrophic Forgetting in Metric Learning. In Proceedings of the 2020 7th International Conference on Soft Computing Machine Intelligence (ISCMI), Stockholm, Sweden, 14–15 November 2020; pp. 68–72. [Google Scholar] [CrossRef]
  43. Yu, L.; Twardowski, B.; Liu, X.; Herranz, L.; Wang, K.; Cheng, Y.; Jui, S.; Weijer, J.V.d. Semantic drift compensation for class-incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 6982–6991. [Google Scholar]
  44. Zhao, H.; Fu, Y.; Kang, M.; Tian, Q.; Wu, F.; Li, X. Mgsvf: Multi-grained slow vs. fast framework for few-shot class-incremental learning. arXiv 2021, arXiv:2006.15524. [Google Scholar] [CrossRef]
  45. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  46. Katsigiannis, S.; Ramzan, N. DREAMER: A Database for Emotion Recognition Through EEG and ECG Signals From Wireless Low-cost Off-the-Shelf Devices. IEEE J. Biomed. Health Inform. 2018, 22, 98–107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Belouadah, E.; Popescu, A. Il2m: Class incremental learning with dual memory. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 583–592. [Google Scholar]
  48. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  49. Li, Z.; Hoiem, D. Learning without forgetting. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 2935–2947. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Wu, Y.; Chen, Y.; Wang, L.; Ye, Y.; Liu, Z.; Guo, Y.; Fu, Y. Large scale incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 374–382. [Google Scholar]
  51. Belouadah, E.; Popescu, A. ScaIL: Classifier Weights Scaling for Class Incremental Learning. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020. [Google Scholar]
  52. Welling, M. Herding Dynamic Weights for Partially Observed Random Field Models. In Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence, UAI 2009, Montreal, QC, Canada, 18–21 June 2009; pp. 599–606. [Google Scholar]
  53. Liu, Y.; Su, Y.; Liu, A.A.; Schiele, B.; Sun, Q. Mnemonics training: Multi-class incremental learning without forgetting. In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 12245–12254. [Google Scholar]
  54. Wu, C.; Herranz, L.; Liu, X.; van de Weijer, J.; Raducanu, B. Memory replay gans: Learning to generate new categories without forgetting. Adv. Neural Inf. Process. Syst. 2018, 31, 5962–5972. [Google Scholar]
  55. van de Ven, G.M.; Siegelmann, H.T.; Tolias, A.S. Brain-inspired replay for continual learning with artificial neural networks. Nat. Commun. 2020, 11, 4069. [Google Scholar] [CrossRef]
  56. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Proceedings of the Advances in Neural Information Processing Systems 30, Long Beach, CA, USA, 4–9 December 2017; pp. 1195–1204. [Google Scholar]
  57. Wang, X.; Han, X.; Huang, W.; Dong, D.; Scott, M.R. Multi-Similarity Loss with General Pair Weighting for Deep Metric Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 5022–5030. [Google Scholar]
  58. Mullen, T.R.; Kothe, C.A.; Chi, Y.M.; Ojeda, A.; Kerth, T.; Makeig, S.; Jung, T.P.; Cauwenberghs, G. Real-time neuroimaging and cognitive monitoring using wearable dry EEG. IEEE Trans. Biomed. Eng. 2015, 62, 2553–2567. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  59. Davidson, R.J. Affective neuroscience and psychophysiology: Toward a synthesis. Psychophysiology 2003, 40, 655–665. [Google Scholar] [CrossRef]
  60. Sanei, S.; Chambers, J.A. EEG Signal Processing; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  61. Yang, Y.; Wu, Q.; Fu, Y.; Chen, X. Continuous Convolutional Neural Network with 3D Input for EEG-Based Emotion Recognition. In Information Processing; Cheng, L., Leung, A.C.S., Ozawa, S., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 433–443. [Google Scholar]
  62. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision–ECCV 2018, Munich, Germany, 8–14 September 2018; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar]
  63. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic Routing Between Capsules. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2017; Volume 30. [Google Scholar]
  64. Montavon, G.; Orr, G.B.; Müller, K. (Eds.) Neural Networks: Tricks of the Trade, 2nd ed.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7700. [Google Scholar]
Figure 1. Flowchart of the proposed method throughout S incremental stages.
Figure 1. Flowchart of the proposed method throughout S incremental stages.
Mathematics 10 00598 g001
Figure 2. Diagram of the Capsule Network architecture.
Figure 2. Diagram of the Capsule Network architecture.
Mathematics 10 00598 g002
Figure 3. Comparison of CIL methods with baseline approaches on (a) DREAMER and (b) DEAP. CapsNet-wo-memory represents the average accuracy obtained by CapsNet when data from old classes are not included in the memory during the CIL training. CapsNet-Full indicates the average accuracy using all training samples from old classes in each incremental stage. Mnem and Geo+LUC indicate Mnemonics and Geodesic+LUCIR, respectively.
Figure 3. Comparison of CIL methods with baseline approaches on (a) DREAMER and (b) DEAP. CapsNet-wo-memory represents the average accuracy obtained by CapsNet when data from old classes are not included in the memory during the CIL training. CapsNet-Full indicates the average accuracy using all training samples from old classes in each incremental stage. Mnem and Geo+LUC indicate Mnemonics and Geodesic+LUCIR, respectively.
Mathematics 10 00598 g003
Figure 4. Effect of the number of reserved samples in memory: (a) DREAMER and (b) DEAP.
Figure 4. Effect of the number of reserved samples in memory: (a) DREAMER and (b) DEAP.
Mathematics 10 00598 g004
Figure 5. Effect of the number of incremental stages on the DREAMER dataset, using memory sizes of (a) 1% and (b) 5% of the full training dataset of each subject. We reported average accuracy in each incremental stage over all subjects.
Figure 5. Effect of the number of incremental stages on the DREAMER dataset, using memory sizes of (a) 1% and (b) 5% of the full training dataset of each subject. We reported average accuracy in each incremental stage over all subjects.
Mathematics 10 00598 g005
Table 1. Specifications of the Capsule Network architecture. We include a Channel-Attention block before the PrimaryCaps module. The decoder setting for reconstruction loss is shown at the bottom.
Table 1. Specifications of the Capsule Network architecture. We include a Channel-Attention block before the PrimaryCaps module. The decoder setting for reconstruction loss is shown at the bottom.
IdModulesLayers (Input ID)HyperparametersOutput Shape
I1InputDREAMER: 14 × 128
DEAP: 32  ×  128
C2ConvolutionalConvolution-2D (I1)DREAMER: 64 filters, size = 6 , stride = 1 , activation = ReLU DREAMER: 64 × 123 × 9
DEAP: 64 filters, size = 9 , stride = 2 , activation = ReLU DEAP: 64 × 60 × 12
A3Channel-AttentionAverage pooling (C2)Size = 1, stride = 1
C4Convolution-2D (A3)32 filters, size = 1 , stride = 1 , activation = ReLU
C5Convolution-2D (C4)64 filters, size = 1 , stride = 1
M6Maxpooling (C2)size=1, stride=1
C7Convolution-2D (M6)32 filters, size = 1 , stride = 1 , activation = ReLU
C8Convolution-2D (C7)64 filters, size = 1 , stride = 1
S9Sum (C5, C8)
A10Activation (S9)SigmoidDREAMER: 64 × 123 × 9
DEAP: 64 × 60 × 12
C11PrimaryCapsConvolution-2D (A10)DREAMER: 8 × 16 filters, size = 6 , stride = 2
DEAP: 8 × 16 filters, size = 9 , stride = 2
R12Reshape (C11)DREAMER: 1088 × 8
DEAP: 832 × 8
E13EmotionCapsDynamic routing (R12)16 units16 × 16
N14NormNormalization (E13)16
O15FCFully connected (N14)Dynamic outputsDREAMER: 9
DEAP: 8
Decoder
F1FC1Fully connected (O15)256 units256
F2FC2Fully connected (F1)512 units512
F3FC3Fully connected (F2)DREAMER: 14 × 128 unitsDREAMER: 14 × 128
DEAP: 32 × 128 unitsDEAP: 32 × 128
Table 2. Hyper-parameter setting. a, b are margins; λ , λ o , λ dis , λ mr , are the trade-off hyperparameters for each CIL method.
Table 2. Hyper-parameter setting. a, b are margins; λ , λ o , λ dis , λ mr , are the trade-off hyperparameters for each CIL method.
MethodsHyper-Parameters
FT, FT+NCC, ScaIL-
LwF, iCARL T = 2 , λ o = 1
Mnemonics λ dis = 0.5 , λ mr = 1 , b = 5
WA T = 2 , λ = 0.4
Geodesic+LUCIR λ dis = 1 , λ mr = 1 , b = 5
Table 3. Accuracy and standard deviation for CIL methods on the DREAMER dataset using approximately 1% of the training size. The best results are in bold.
Table 3. Accuracy and standard deviation for CIL methods on the DREAMER dataset using approximately 1% of the training size. The best results are in bold.
Geodesic
Subj.FTFT+NCCLwFiCARLMnemonicsScaILWA+LUCIRIL2FS
144.08 ± 13.8648.63 ± 13.3744.78 ± 10.2548.26 ± 09.9551.94 ± 14.0040.09 ± 11.2147.65 ± 14.9253.29 ± 16.2964.58 ± 12.02
235.82 ± 13.8738.31 ± 14.3137.27 ± 15.8439.95 ± 16.1947.86 ± 17.6430.56 ± 11.6439.83 ± 19.9442.54 ± 14.5053.64 ± 11.78
332.51 ± 11.5934.93 ± 09.8733.32 ± 12.9137.31 ± 13.3442.54 ± 12.5029.96 ± 11.9635.33 ± 14.4544.37 ± 14.4654.43 ± 11.12
453.96 ± 11.8856.55 ± 11.4154.23 ± 14.2356.84 ± 14.4760.72 ± 15.3040.09 ± 08.6455.13 ± 16.9360.16 ± 13.1064.08 ± 09.19
538.76 ± 14.3541.65 ± 13.5235.65 ± 15.6737.73 ± 16.5251.42 ± 16.9529.24 ± 10.6239.87 ± 17.6551.13 ± 17.0960.74 ± 12.52
635.80 ± 13.3938.47 ± 13.3836.38 ± 15.3140.12 ± 15.0644.35 ± 14.4234.98 ± 09.5937.01 ± 16.3542.08 ± 12.7453.31 ± 12.94
732.77 ± 11.3335.89 ± 11.1431.71 ± 11.9835.87 ± 13.1542.01 ± 13.4927.76 ± 08.1334.81 ± 14.2945.17 ± 14.0656.34 ± 10.35
833.77 ± 11.4136.90 ± 11.6331.03 ± 11.2933.39 ± 11.8441.25 ± 12.3327.18 ± 10.7134.57 ± 12.0643.31 ± 13.6549.98 ± 10.66
928.69 ± 08.1232.11 ± 09.0928.75 ± 11.8932.67 ± 11.0337.97 ± 13.0725.95 ± 09.1030.72 ± 14.5039.73 ± 13.1049.07 ± 12.59
1035.62 ± 13.8138.06 ± 14.2436.46 ± 15.8538.56 ± 15.5440.70 ± 13.6128.70 ± 08.4638.47 ± 13.6740.54 ± 12.5351.48 ± 11.21
1135.91 ± 10.1939.53 ± 10.2033.02 ± 13.7036.61 ± 12.8641.62 ± 12.3234.08 ± 13.0436.83 ± 12.1046.29 ± 13.2456.08 ± 14.38
1245.95 ± 09.9550.54 ± 10.4246.08 ± 12.0549.81 ± 12.1056.21 ± 14.5037.23 ± 09.5551.48 ± 15.6957.70 ± 13.6566.81 ± 09.14
1346.41 ± 15.8848.24 ± 16.2545.25 ± 16.2648.39 ± 15.2257.09 ± 14.8439.42 ± 09.5752.48 ± 16.7855.51 ± 15.1068.07 ± 11.11
1445.27 ± 12.8448.52 ± 11.6446.56 ± 18.0249.50 ± 17.3854.18 ± 16.5938.57 ± 14.7050.10 ± 21.2151.96 ± 17.2460.94 ± 13.67
1564.27 ± 11.4766.17 ± 11.1366.48 ± 10.6469.18 ± 09.8572.71 ± 12.8156.49 ± 09.0869.19 ± 11.4671.30 ± 11.1681.16 ± 07.11
1634.37 ± 14.0337.87 ± 13.6732.26 ± 11.4335.40 ± 09.3842.87 ± 12.6729.04 ± 08.5935.41 ± 14.7243.04 ± 15.0650.80 ± 12.11
1737.99 ± 12.4041.80 ± 11.5038.42 ± 12.1642.11 ± 11.8845.49 ± 14.3231.50 ± 11.7840.99 ± 13.4946.73 ± 13.1653.91 ± 09.47
1850.14 ± 16.9252.06 ± 17.3148.56 ± 15.2851.88 ± 15.4557.97 ± 17.4141.17 ± 13.1352.66 ± 17.8156.76 ± 15.6362.78 ± 11.35
1957.42 ± 12.9561.90 ± 11.3056.79 ± 15.5259.20 ± 15.4065.67 ± 11.7250.43 ± 10.2361.13 ± 16.2566.94 ± 10.9871.56 ± 09.73
2027.09 ± 13.4428.79 ± 12.8729.47 ± 15.6331.08 ± 14.6235.81 ± 13.1523.91 ± 09.1132.62 ± 15.2135.83 ± 13.9849.75 ± 14.18
2141.34 ± 10.8743.63 ± 10.4439.95 ± 13.0442.68 ± 12.7448.10 ± 12.2031.82 ± 07.0242.34 ± 15.8748.22 ± 10.9954.25 ± 12.23
2257.39 ± 12.7161.14 ± 11.3455.15 ± 14.1157.58 ± 13.7166.89 ± 10.4047.36 ± 08.7257.42 ± 16.1664.07 ± 10.8868.49 ± 13.69
2343.66 ± 15.6345.99 ± 14.3338.93 ± 16.6241.39 ± 16.6747.38 ± 16.0832.77 ± 10.2140.98 ± 17.7645.17 ± 16.2956.66 ± 12.17
Avg.41.69 ± 09.8044.68 ± 09.9141.15 ± 09.8544.15 ± 09.8550.12 ± 09.3135.14 ± 08.2044.22 ± 10.2250.08 ± 09.3159.08 ± 08.26
Table 4. Accuracy and standard deviation for CIL methods on the DEAP dataset using approximately 1% of the training size. The best results are in bold.
Table 4. Accuracy and standard deviation for CIL methods on the DEAP dataset using approximately 1% of the training size. The best results are in bold.
Geodesic
Subj.FTFT+NCCLwFiCARLMnemonicsScaILWA+LUCIRIL2FS
157.65 ± 13.8464.62 ± 11.9555.14 ± 12.1262.57 ± 12.4874.60 ± 17.0453.14 ± 11.9461.80 ± 17.9375.06 ± 13.4286.47 ± 06.85
256.83 ± 16.6861.35 ± 15.7556.30 ± 18.1258.82 ± 18.0864.09 ± 17.7754.10 ± 05.8159.24 ± 19.9262.53 ± 16.9074.33 ± 10.94
356.23 ± 09.2760.40 ± 08.7759.29 ± 12.8463.56 ± 12.3363.78 ± 15.9052.40 ± 06.2362.50 ± 12.5867.24 ± 14.6674.31 ± 09.64
458.44 ± 15.7763.74 ± 15.6060.47 ± 15.1465.21 ± 16.4662.51 ± 18.6957.31 ± 03.4460.78 ± 17.6069.60 ± 14.8875.91 ± 12.26
557.34 ± 13.3963.42 ± 12.0555.10 ± 11.9059.98 ± 11.6866.71 ± 17.3151.64 ± 09.1460.28 ± 14.0369.12 ± 12.8678.55 ± 08.14
655.84 ± 14.1063.04 ± 12.1356.09 ± 15.0863.04 ± 14.7172.35 ± 15.7651.28 ± 08.9860.66 ± 17.9472.20 ± 16.0183.16 ± 06.61
764.44 ± 16.0470.54 ± 14.4662.59 ± 17.6969.65 ± 14.6875.40 ± 15.9759.38 ± 04.1967.31 ± 19.1079.58 ± 14.9785.28 ± 06.72
860.46 ± 14.5667.00 ± 12.3960.82 ± 14.3566.61 ± 13.2570.96 ± 16.6653.71 ± 03.9964.64 ± 15.4871.79 ± 12.4182.59 ± 06.95
952.65 ± 15.7759.46 ± 16.0152.85 ± 15.6259.02 ± 15.6065.73 ± 17.8948.87 ± 04.2455.72 ± 16.7464.45 ± 13.7678.22 ± 10.15
1072.54 ± 11.8978.08 ± 10.0670.43 ± 11.9574.33 ± 11.3072.70 ± 13.2262.91 ± 04.4274.19 ± 12.9482.04 ± 09.1185.71 ± 05.15
1152.94 ± 19.0359.58 ± 17.3555.42 ± 18.4963.22 ± 18.3664.73 ± 21.2052.12 ± 07.2357.57 ± 20.2066.36 ± 19.7377.96 ± 10.11
1268.06 ± 16.9872.76 ± 15.1668.35 ± 16.7774.88 ± 15.3371.02 ± 17.6569.16 ± 09.7670.10 ± 15.4676.61 ± 14.5279.11 ± 11.29
1346.32 ± 12.1952.48 ± 11.5345.36 ± 13.1350.33 ± 12.7564.25 ± 16.5938.55 ± 02.5848.49 ± 17.1359.71 ± 15.8774.74 ± 12.02
1451.04 ± 15.1158.49 ± 13.7149.41 ± 16.7455.65 ± 16.9664.15 ± 14.5745.08 ± 03.8751.73 ± 15.9363.48 ± 15.6875.37 ± 10.65
1551.02 ± 13.9657.96 ± 13.5753.10 ± 15.4159.35 ± 15.5171.36 ± 15.6953.49 ± 04.2254.30 ± 16.3463.75 ± 15.1580.50 ± 08.37
1661.86 ± 09.8167.57 ± 09.2062.10 ± 09.9667.35 ± 09.1161.78 ± 19.8854.16 ± 05.1263.79 ± 11.4670.24 ± 13.3582.10 ± 10.33
1749.18 ± 10.9755.58 ± 11.1549.45 ± 13.8355.88 ± 12.9361.75 ± 16.9042.52 ± 04.6150.79 ± 13.5762.45 ± 12.5868.84 ± 10.02
1863.45 ± 09.9769.54 ± 09.2761.68 ± 08.8267.09 ± 08.0276.78 ± 15.3166.03 ± 07.8567.34 ± 14.3074.56 ± 13.6879.82 ± 06.34
1961.31 ± 12.8766.52 ± 12.2762.42 ± 13.4567.94 ± 10.4871.57 ± 17.2859.35 ± 04.8166.83 ± 13.8771.54 ± 16.7282.06 ± 08.84
2068.44 ± 15.6973.53 ± 13.5266.13 ± 17.8270.82 ± 15.5367.01 ± 19.2061.15 ± 05.1870.73 ± 16.4474.44 ± 15.0882.96 ± 09.52
2169.70 ± 14.8874.58 ± 13.0769.10 ± 13.6273.16 ± 13.3970.00 ± 21.4367.61 ± 07.9072.44 ± 16.9077.39 ± 15.3579.11 ± 10.35
2266.38 ± 09.9873.02 ± 10.0764.10 ± 12.2569.00 ± 11.3870.31 ± 13.6858.79 ± 06.3169.03 ± 14.7772.83 ± 09.9680.14 ± 08.90
2347.61 ± 12.3455.60 ± 11.6047.39 ± 09.8856.89 ± 12.9577.79 ± 15.1348.54 ± 05.0555.51 ± 17.8569.97 ± 09.8483.30 ± 06.33
2466.44 ± 15.9571.95 ± 13.3568.02 ± 18.0473.78 ± 16.0272.32 ± 18.8666.30 ± 09.0670.66 ± 18.6274.67 ± 15.4684.28 ± 07.59
2549.94 ± 12.3057.68 ± 10.6352.73 ± 10.5360.48 ± 09.4163.30 ± 16.3947.68 ± 04.9552.64 ± 13.2062.91 ± 12.4176.20 ± 10.29
2648.79 ± 08.9354.62 ± 09.7447.65 ± 08.7152.85 ± 08.8660.39 ± 15.6248.20 ± 08.4551.76 ± 09.7859.99 ± 11.2366.78 ± 10.25
2754.70 ± 11.1864.47 ± 09.5649.45 ± 07.8266.25 ± 08.3080.64 ± 12.1554.49 ± 10.9460.36 ± 15.1874.16 ± 11.7280.90 ± 06.87
2857.35 ± 10.5064.41 ± 09.2557.00 ± 11.4664.69 ± 10.1469.88 ± 14.8955.19 ± 07.7160.71 ± 16.0571.67 ± 11.6377.46 ± 09.02
2973.05 ± 07.9977.94 ± 06.1072.27 ± 07.8877.25 ± 08.1078.84 ± 11.2873.64 ± 08.8574.40 ± 08.6782.36 ± 09.3885.87 ± 05.34
3069.05 ± 13.6974.11 ± 11.4869.88 ± 13.7774.64 ± 12.9075.45 ± 15.1167.46 ± 03.9172.50 ± 12.9680.69 ± 11.5683.45 ± 08.12
3161.55 ± 14.2766.49 ± 13.0062.38 ± 12.1766.14 ± 11.6967.71 ± 17.9957.06 ± 04.0764.61 ± 14.7969.30 ± 13.9877.63 ± 09.48
3247.53 ± 09.3454.37 ± 09.1646.70 ± 10.9055.44 ± 12.8871.31 ± 14.2946.57 ± 06.3350.94 ± 15.0260.83 ± 10.5076.49 ± 09.43
Avg.58.69 ± 07.8364.84 ± 07.2458.41 ± 07.7664.56 ± 07.0069.41 ± 05.4855.56 ± 08.2062.01 ± 07.6070.42 ± 06.5079.36 ± 04.68
Table 5. Effect of each component of IL2FS on DREAMER and DEAP. The best results are in bold.
Table 5. Effect of each component of IL2FS on DREAMER and DEAP. The best results are in bold.
MethodDREAMERDEAP
Fine-tuning (FT)41.69 ± 09.8058.69 ± 07.83
FT+Weight Aligning46.55 ± 09.0162.68 ± 07.29
FT+Weight Aligning+Margin ranking loss56.77 ± 08.4778.01 ± 04.25
FT+Weight Aligning+Margin ranking loss+Triplet loss (IL2FS)59.08 ± 08.2679.36 ± 04.68
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiménez-Guarneros, M.; Alejo-Eleuterio, R. A Class-Incremental Learning Method Based on Preserving the Learned Feature Space for EEG-Based Emotion Recognition. Mathematics 2022, 10, 598. https://doi.org/10.3390/math10040598

AMA Style

Jiménez-Guarneros M, Alejo-Eleuterio R. A Class-Incremental Learning Method Based on Preserving the Learned Feature Space for EEG-Based Emotion Recognition. Mathematics. 2022; 10(4):598. https://doi.org/10.3390/math10040598

Chicago/Turabian Style

Jiménez-Guarneros, Magdiel, and Roberto Alejo-Eleuterio. 2022. "A Class-Incremental Learning Method Based on Preserving the Learned Feature Space for EEG-Based Emotion Recognition" Mathematics 10, no. 4: 598. https://doi.org/10.3390/math10040598

APA Style

Jiménez-Guarneros, M., & Alejo-Eleuterio, R. (2022). A Class-Incremental Learning Method Based on Preserving the Learned Feature Space for EEG-Based Emotion Recognition. Mathematics, 10(4), 598. https://doi.org/10.3390/math10040598

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop