Next Article in Journal
Deposition of Different Metallic Coatings as Repair Materials for Concrete by Using a Twin-Wire Arc Thermal Spray Process
Next Article in Special Issue
Advances in Machine Learning for Sensing and Condition Monitoring
Previous Article in Journal
Laboratory Experimental Study on Polymer Flooding in High-Temperature and High-Salinity Heavy Oil Reservoir
Previous Article in Special Issue
Bibliometric Analysis of the Application of Artificial Intelligence Techniques to the Management of Innovation Projects
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Imaginary Speech Recognition Using a Convolutional Network with Long-Short Memory

Department of Applied Electronics and Information Engineering, Faculty of Electronics, Telecommunications and Information Technology, Polytechnic University of Bucharest, 060042 Bucharest, Romania
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(22), 11873; https://doi.org/10.3390/app122211873
Submission received: 10 October 2022 / Revised: 10 November 2022 / Accepted: 18 November 2022 / Published: 21 November 2022
(This article belongs to the Special Issue Applied Artificial Intelligence (AI))

Abstract

:
In recent years, a lot of researchers’ attentions were concentrating on imaginary speech understanding, decoding, and even recognition. Speech is a complex mechanism, which involves multiple brain areas in the process of production, planning, and precise control of a large number of muscles and articulation involved in the actual utterance. This paper proposes an intelligent imaginary speech recognition system of eleven different utterances, seven phonemes, and four words from the Kara One database. We showed, during our research, that the feature space of the cross-covariance in frequency domain offers a better perspective of the imaginary speech by computing LDA for 2D representation of the feature space, in comparison to cross-covariance in the time domain and the raw signals without any processing. In the classification stage, we used a CNNLSTM neural network and obtained a performance of 43% accuracy for all eleven different utterances. The developed system was meant to be a subject’s shared system. We also showed that, using the channels corresponding to the anatomical structures of the brain involved in speech production, i.e., Broca area, primary motor cortex, and secondary motor cortex, 93% of information is preserved, obtaining 40% accuracy by using 29 electrodes out of the initial 62.

1. Introduction

Imaginary speech recognition (ISR) systems have grown in popularity in recent years. These types of systems usually aim to acquire signals from the brain using invasive or non-invasive methods, preprocess them to improve the quality of the signal, extracts the features with the main purpose of concentrating the significant information into smaller data, and finally, classifies the features in order to offer feedback to the user. This processing chain that leads to obtaining the final system must work in real time on portable devices to be actually used in everyday activities by the people who need it. The main purpose of ISR is to come to the aid of persons who, after suffering some disorders (afflictions,) such as the lock-down syndrome, cerebral palsy, etc., lost the ability to speak.
An important part of the ISR systems is represented by the signals acquisition stage. There are several methods of collecting the signals for a brain computer interface (BCI) system, such as electroencephalography (EEG), electrocorticography (ECoG), magnetoencephalography (MEG), positron emission tomography (PET), and functional magnetic resonance imaging (fMRI). The most common method used is EEG. The main advantage of the EEG signals is the fact that the acquisition is non-invasive and low-cost, compared to MEG, PET, and fMRI. However, since the signals are acquired from the surface of the scalp, there are multiple layers between the source and the electrode, which leads to a high attenuation of the signal, and it is more sensitive to noises. The second most used acquisition method for imaginary speech decoding is ECoG, due to the higher quality of the collected data, since the signals are acquired directly from the cortex (invasive). Nevertheless, this advantage also brings the disadvantage of being an invasive method, leading to difficulties in data acquisition.

2. State of the Art

In recent years, recognizing silent speech from cortical signals attracted more and more attention of the researchers. Different approaches were tried over the years, aiming to achieve the best performances.
The first attempts consisted of creating the desired words from letters using different methods of choosing the letters, such as moving a cursor on a monitor [1] or following a matrix of ASCII characters as the lines and columns are highlighted [2]. A successful real-time communication system, based on creating the words letter-by-letter, was developed by Ujwal Chaudhary et al. [3] on a patient with amyotrophic lateral sclerosis (ASL). Due to the degradation of the motor functions, the only way of communication for the patient was by using brain signals. The researchers implanted a 64-microelectrode array, which allowed the subject to modulate neural firing rates after receiving audio feedback with a block of letters first and each letter from the block after to create words. After days of training, on the 245th day after implantation, the patient was able to create complex sentences, such as, “I would like to listen to the album by Tool loud”, with a mean of one character per minute.
These methods offered significant results and have the advantage of improving over time as the subjects practice mastering the system. However, this type of communication is difficult, unnatural, and takes relatively a lot of time to form a word (one character per minute [3]).
Other approaches targeted decoding the speech directly from the imagined words or phonemes. These methods assume that, during imaginary speech, the brain has different marks of electric activity, according to different pronunciations of each word.
Trying to find different marks of imaginary speech, Thimotee Proix et al. [4] investigated the ECoG signals acquired during overt and imagined speech. In the conducted study, the researchers first tried to find similarities between overt and imagined speech. They computed the power spectrum of four frequency bands, theta (4–8 Hz), low-beta (12–18 Hz), low-gamma (25–35 Hz), and BHA (80–150 HZ), and they observed that the BHA power for both overt and imagined speech increased in the sensory and motor regions, while beta power decreased over the same regions. Nevertheless, when trying to differentiate different syllables in binary classes, articulatory, phonetic, and vocalic, the results for imaginary speech significantly dropped, compared to overt speech in the BHA band. Better performance was recorded when using power in the beta band, but with room for improvement. Significant improvements of the results were obtained when the trials were aligned on recorded speech onset, which was impossible to do for the imaginary speech.
Other study based on ECoG signals managed to obtain important results when classifying five different words in a patient-specific system [5]. The researchers used high-gamma time features, which were aligned using dynamic time warping (DTW), and finally obtained a matrix feature by computing the DTW-distance between the realigned trials. These features, combined with support vector machine (SVM), offered encouraging results for a five-class classification task, obtaining a 58% mean accuracy for all five subjects. In 2019, Miguel Angrick et al. [6] managed to synthesize speech from ECoG signals by also using the speech signal acquired, together with the ECoG signal. The obtained results registered a correlation between the reconstructed speech signal and the actual speech, higher than the chance for all the analyzed subjects. Moreover, one of the subjects had a correlation of 0.69, which allowed for a good reconstruction of the vocal signal.
Recent studies of stereoelectroencephalography (sEEG) signals regarding decoding the imaginary speech reported the development of a real-time system with speech synthesis for only one patient, due to the difficulty of data acquisition [7]. The researchers processed the neural signals by extracting the multichannel high-gamma band as features that were decoded by assigning the neural activity to a mel-scaled audio spectrogram. Further, a regularized LDA-classifier was trained to predict the energy level for each spectral bin. The results reported an average 0.62 ± 0.15 Pearson correlation between the reconstructed signal and the real speech signal, but the reconstructed speech was not yet intelligible.
However, even if ECoG and sEEG offered a good representation of the neural activity with a high signal-to-noise ratio, they were still invasive methods of acquiring signals from the brain, which leads to great limitations regarding collecting the signals and being easily accepted by the patients. The alternative to these invasive methods can be functional magnetic resonance imaging (fMRI), functional near-infrared spectroscopy (fNIRS), magnetoencephalography (MEG), and electroencephalography (EEG). All these methods have their limitations: fMRI is expensive, non-portable, and has a poor temporal resolution, fNIRS takes time to collect the information from the brain (aprox. 2 to 5 s), which makes it harder to aim a real-time system, and MEG is expensive and non-portable. Finally, EEG remains the best option for a real-time noninvasive wearable device, while having the limitation of small amplitude data with a higher signal-to-noise ratio, comparing to ECoG.
In recent years, several studies have used EEG signals for imaginary speech recognition. In study [8], the researchers analyzed six different words: “could”, “yard”, “give”, “him”, “there”, and “toe”, acquired from 15 subjects. In the feature extraction stage, the signals were decomposed using the Daubechies-4 (db4) mother wavelet to eight levels corresponding to the delta, theta, alpha, beta, gamma bands, and additionally, three other bands, <2 Hz, 64–128 Hz and 128–256 Hz, where features such as root mean square, standard deviation, and relative wavelet energy were computed. The features were fed into a random forest (RF) classifier and support vector machine (SVM). The registered results rose above the chance for both classifiers, having a mean accuracy for all 15 subjects of 25.26% RF and 28.61% SVM.
A great breakthrough in the field came along with the open access datasets, such as the Kara One Dataset [9] and Nguyen et al. [10] research dataset. This helped the community to develop research in the imaginary speech field, without having to go through the complicated process of data acquisition, and made it possible to compare the results with other studies developed based on the same environment.
Using the Kara One Dataset, Panachakel et al. [11] managed to develop a patient-specific system based on computing a set of statistical features, root mean square, variance, kurtosis, skewness, and the 3rd order moment, over the obtained signals after decomposition using db4 mother wavelet into seven levels. Afterwards, the signals were fed into a deep learning architecture with two layers of 40 neurons. The results offered a 57.15% mean accuracy for all subjects, significantly higher compared to the results obtained without using deep learning. With a more complex deep learning architecture, the researchers from University of British Columbia managed, in their paper [12], to present a generalized system for imaginary speech binary classification, achieving the best accuracy of 85.23% for discrimination between consonant and vocal phonemes. The results were achieved using the covariance matrix between the channels as a feature matrix and a combination between convolutional neural networks (CNN), long-short term memory (LSTM) neural networks, and a deep autoencoder (DAE).
In 2021, a larger dataset was collected in Moscow, Russia, containing signals acquired from 270 healthy subjects during the silent speech of eight Russian words: “forward”, “backward”, “up”, “down”, “help”, “take”, “stop”, and “release”. The patient-specific developed system offered an accuracy of 84.5% for all nine words classification and 87.9% for binary classification using the deep learning architecture of ResNet18, in combination with two layers of gated recurrent units (GRUs). Finally, the researchers claimed that there are strong differences between the signals from different subjects, and it is more plausible to develop a patient-specific system with a high accuracy than a generalized one.
Deep learning architectures showed a better performance of imaginary speech signals and gained popularity over the years. Recently, EEG studies started to concentrate on the LSTM neural network, due to the advantages it has for a continuous series. The LSTM neural network offered significant results on applications such as epilepsy prediction [13] and imaginary speech recognition [14]. The results obtained by the researchers in [14] registered a maximum accuracy of 73.56% for a subject-independent system and a four-class problem: “sos”, ”stop”, ”medicine”, and ”washroom”.
In this paper, we focused on developing a subject’s shared system for imaginary speech recognition. By the subject’s shared system, we mean that one system was developed for all registered users in the database. However, this system was different from a patient-specific system in such a matter that, when introducing a new subject in the database, it will only need a fine-tuning of the neural network and not an entire subject training. During our research, we used Kara One database, which was preprocessed for further usage. Cross-covariance in frequency domain was computed in the feature extraction stage, and all the signals were further introduced into a CNNLSTM neural network. We showed that CNNLSTM performed better than the CNN neural network by increasing the accuracy from 37% to 43%. In the paper, the system behavior after reducing the number of channels and selecting the channels from the main brain areas and the areas corresponding to the anatomical structures involved in imaginary speech production was also tested. We showed that the channels corresponded to the anatomical structures involved in speech production concentrates 93% of information, obtaining a maximum accuracy of approx. 40%. Even if the accuracy dropped by 3%, a lot more was gained in terms of time of execution, comfort, portability, and costs.
The system was developed using python as programming language, and the CNNLSTM neural network was developed using TensorFlow, an end-to-end open-source platform for machine learning [15].

3. Materials and Methods

3.1. Preparing Database

The Kara One Database was developed following a partnership between the University of Toronto and Toronto Rehabilitation Institute [9]. It contains signals acquired from eight male and four female participants (a total of twelve subjects), all right-handed, with no visual, hearing, or motor problems. All signals were collected using a standard protocol consisting of a (1) rest state, (2) stimulus state, in which the prompt with the stimulus appeared, (3) a five-second of stimulus imagined speaking state, and (4) an actual articulation of the stimulus. An important aspect of this database consisted of randomly presenting the stimulus to the subject at each epoch, avoiding the accidental creation of temporal correlations in EEG signals, as presented in [16]. The prompts contained seven phonemes: /iy/, /uw/, /piy/, /tiy/, /diy/, /m/, /n/ and four words chosen from the Kent’s list of phonetically similar pairs: “pat”, “pot”, “knew”, and “gnaw”. The stimulus was chosen to cover a wide range of pronunciation mechanisms, such as nasal: /m/, /n/, “knew”, and “gnaw”, bilabial: /piy/, “pat”, “pot”, and /m/, closed vocal tract: /piy/, /tiy/, /diy/, vocals: /iy/, /uw/, and consonants: /m/, /n/. Further, in the study, we only used the signals corresponding to the imaginary speech.
All signals from the database were visually analyzed by a specialist. This process led to elimination of the epochs containing high noises. Additionally, during the visual analysis of the signal, four of the twelve subjects were eliminated from the study, due to the very high noise signals or the presence of unattached ground wires. The number of signals that passed on in the study were of 993 for all subjects and all prompts.

3.2. Preprocessing

In the preprocessing stage, the signals were filtered using a notch filter to remove the 60 Hz power line artifact and all multiples of 60 Hz smaller than the Nyquist frequency.
Afterwards, from the 5 s epoch of imaginary speech, corresponding to each utterance, we removed the first and the last 0.5 s corresponding to the transition from one state to another. Further, we segmented the 4 s remaining signal into windows of 0.25 s without overlapping. The window dimension was chosen following a previous study focused, among others, on the analysis of different windows length: 0.25 s, 0.5 s, and 1 s [17]. Finally, half of the windows were randomly assigned to the training set and the other half to the testing set for the subject’s shared system.

3.3. Feature Extraction

The feature extraction stage aims to decode the information hidden in the signal and pass the result to a neural network capable of classifying the words and phonemes. In this paper, we used as features the inter-channel covariation in frequency domain. Let F X c h be the fast Fourier transform (FFT) corresponding to the channel c h of the input signal X c h , the FFT can be described by the following equation:
F X c h ( f ) = t = 0 n 1 X t c h e j 2 π f t n
The cross-covariation between the channels in frequency domain is given by:
C o v ( X c 1 ( t ) , X c 2 ( t ) ) = E [ [   X c 1 ( t ) E ( X c 1 ( t ) ] [   X c 2 ( t ) E ( X c 2 ( t ) ] ]
where X c 1 ( t ) represents the EEG signal acquired for channel c1, X c 2 ( t ) is the EEG signal acquired for channel c2, and E[ X ch ( t ) ] represents the mean value (where ch is corresponding to the specific channel c1 or c2) and is computed as:
E ( X c h ( t ) ) = 1 W i = 0 W 1 x i c h
The W value of Equation (3) corresponds to the window dimension for which the features are computed. After extracting the features, the resulting matrix for a signal will be of size [ N c h × N c h ], where N c h is the number of channels.
The 0.25 s segmented channels were further cropped into 0.1 s window with 50% overlap to preserve the time evolution of the signal, which was passed to the CNN-LSTM architecture. An example of the computed features over 0.1 s window length for each phoneme and word is presented in Figure 1.
Further, we investigated the features by computing the linear discriminant analysis (LDA) to reduce the features’ dimensionality to only two components for visual inspection of the computed features. Figure 2 presents a comparison between the results obtained by applying LDA to the computed features cross-covariance in frequency domain (a), the raw signal in time (b), and computed features cross-covariance in time domain (c). It can be easily seen that, when using the frequency domain cross-covariance, the features are very well-structured into separate clusters, compared to the results obtained using the time domain signal or the time domain cross-covariance as features, cases where the classes overlap and the vectors are mixed in the same region of feature space. Another observation regarding the LDA results is the separability of the phonemes versus words. It can be seen that words and phonemes occupy different regions of the feature space.
In Table 1, the mean and standard deviation of the LDA components for each class were computed to quantify the results observed after the visual analyses. The computed means and the confidence ellipse for each class are pointed out in Figure 3 for all analyses features: cross-covariance in frequency domain (a), raw-signal features (b), and covariance in time domain (c). In time domain, the means were very close to each other, and the standard deviation was high for all classes, each one incorporating a large part of the other classes for phonemes and for words. However, there was a clear differentiation between the words and the phonemes in the feature space. When it came to the cross-covariance in time domain, the distances between the means were getting higher, but the separability between the phonemes and words was lost, too. In the case of cross-covariance in frequency domain, the separability between the phonemes and words was being preserved, and the distances between the means were bigger. However, the classes were still not fully separable.
After the visual and quantitative analysis of the LDA feature space, we concluded that the cross-covariance in frequency domain offer the best understanding of the imaginary speech signals and was further used in the classification stage.

3.4. Classification

LSTM neural networks started to gain more and more popularity when it comes to developing intelligent systems meant to work with time series, such as EEG signals. A wide range of EEG applications exploit the advantages of LSTM consisting of learning long-term dependencies in time series and making connections between different windows of signals. Among these applications, we can list a seizure prediction systems [13], error detections in a musician’s performances [18], prediction of customers decisions for brand extension scenarios [19], classification of hand movements [20], and imaginary speech recognition systems [14].
In this paper, we used a 2D convolutional long-short term memory neural network (CNNLSTM) in the classification stage. This neural network added to the advantages of LSTM, i.e., the spatial correlation information between the channels. In other words, this neural network is connecting both the spatial and temporal information of the EEG signals conducting to a better decoding of the brain activity. The detailed information regarding the CNNLSTM layers of the neural networks are explained in [21].
The architecture of the neural network is based on the previous study of the CNN architectures [17], in which the best results were obtained using two convolutional layers with 64 and 128 filters, respectively, connected to a dense layer with 64 neurons. The final layer corresponds to the decision and consists of a dense layer with 11 neurons, the equivalent number of the seven phonemes and four words. During this study, more complex architectures were also tested, but no significant improvements in accuracy were registered.
The size of the input layer corresponds to the size of the features, a matrix of N × N c h × N c h × N w , where N is the number of observations of the training set, N c h is the number of channels used to compute the cross-covariance matrix, and N w is the number of windows for the LSTM time analysis. The diagram block of the CNNLSTM architecture used in this paper is presented in Figure 4.
Finally, the neural network was trained using the optimizer Adam with a learning rate of 0.0001, having as loss function the categorical cross-entropy.
The neural network was developed using TensorFlow with the implemented CNNLSTM2D layer using the algorithm described in [21].

3.5. Performance Metrics

The performance metrics used to describe the system’s capability of recognizing imaginary speech are implemented in the open-source scikit-learn toolkit detailed in [22].
Let CM be the confusion matrix of the system, described in Table 2, as:
The accuracy is computed as the overall performance of the system:
A c c u r a c y = 1 N k = 1 N c C k k
where N is the number of input vectors, N c is the number of classes, and C k k corresponds to the diagonal of the confusion matrix.
The balanced accuracy is used to determine the performance of the system when the imbalanced data is used:
B a l a n c e d   a c c u r a c y = 1 N c k = 1 N c C k k t k
where t k is the number of times that class k occurs in the input dataset and is computed as:
t k = j = 1 N c C j k
Kappa coefficient (Cohen’s Kappa) is generally used to compute the degree of agreement between independent observations and was computed as follows:
K a p p a = C   x   N k N c p k × t k N 2 k N c p k × t k
where C is the total number of correctly predicted elements, and p k is the number of times that class k was predicted, and they are described by the following equations:
C = k = 1 N c C k k
p k = j = 1 N c C k j
Recall measurement offers information regarding the capability of the system to correctly predict the phonemes in the dataset and is computed using the following equation:
R e c a l l = C C + k N c f n k
where f n k is the number of times the class k was predicted as one of the other classes:
f n k = i = 1 ,   i k N c C k i

4. Results

This paper aimed to develop an intelligent subject’s shared ISR system for differentiating seven phonemes and four words of the Kara One database. The signals from the database were preprocessed, in order to obtain a better quality of the data and were further introduced to a feature extraction stage, with the purpose of extracting the main information hidden in the EEG signals regarding imaginary speech. The features used were based on the inter-channel covariation in frequency domain, a method of feature extraction first introduced in the ISR domain in the study [17], which has the main advantage of encoding the variability of the electrodes when a stimulus is sent by the brain to the motor neurons involved in the process. Another advantage of the method consists of eliminating the possible delays of brain impulse over the channels, due to the computation in frequency domain, with respect to the time domain, first introduced to imaginary speech by Pramit Saha and Sidney Fels in their study [12]. In the classification stage, we used a 2D CNNLSTM neural network in order to connect both the spatial and temporal correlations between the different windows and the electrodes.
During the development of the system, we also studied the system performances for different regions of the brain, frontal (F), central (C), occipital (O), and different areas, left (L) and right (R) and combinations between them:
  • FC = frontal and central;
  • CO = central and occipital;
  • FCO = frontal, central and occipital;
  • Frontal left and right;
  • Central left and right;
  • Occipital left and right.
To these regions, we added an ASBA (anatomical brain areas involved in speech) region. Considering the nowadays inclination to develop systems that can be easily transposed to a portable device, reducing the number of channels can be a big breakthrough towards this aspiration. By using only the electrodes that are directly responsible for the speech production, the device is more likely to be accepted by the users. Another advantage of selecting the channels consists of reducing the memory and the execution time of the system.

4.1. CNNLSTM vs CNN

This study focused on highlighting the advantages of the 2D CNNLSTM neural network in recognizing imaginary phonemes and words using the inter-channel cross-covariance method in the feature extraction stage. The results achieved were compared with the ones recorded in study [17], where the same feature extraction method was used with a CNN neural network as classifier. The results comparison can be observed in Table 3.
The CNNLSTM neural network managed to increase the accuracy from 0.3758 to 0.4398, according to Table 3, when the same preprocessing and feature extraction processing chain was followed. This difference in accuracy is due to the ability of the LSTM neural network to learn the long-term dependencies of time series in combination with the 2D CNN neural network, which encodes the spatial correlation between the electrodes too.
The mean confusion matrix of the k-folds is presented in Figure 5 and reveals that there is hardly any confusion between the phonemes and words, as expected after the visualization of the LDA dimension reduction results in Figure 2.
An analysis of the distances between the mean of each class for every phoneme and word was conducted in Table 4 and Table 5, with the cross-covariance computed in frequency domain features (Table 4) and for the features computed by the CNNLSTM neural network after the first CNNLSTM layer (Table 5). This analysis aims to highlight the role of the CNNLSTM layers in the feature extraction adaptive process, designed to transform the input space and to increase the separability between the classes.
To compute the Euclidean distance between the mean of each class, first the input matrix was reshaped into a matrix with the size N × Nfeatures, where N is the number of input vectors, and Nfeatures is the total number of features corresponding to Nwindow × Nchannels × Ndepth for the cross-covariance matrix in frequency domain and Nwindow × Nchannels × Nchannels × Ndepth for the features computed by the first layer of the CNNLSTM neural network. All features were normalized in the range [0, 1], using Equation (12), before computing the Euclidean distances (13).
X i j n o r m = X i j min i X i j max i X min i X i j
where X i j n o r m is the input vector X i j , i = 0 ,   N 1 ¯ , j = 0 ,   N f e a t u r e s 1 ¯ , min i X i j is the minim value of the vector i for the feature j, and max i X i j is the maximum value of the vector i for the feature j.
The Euclidean distance between the mean of utterance u1, and utterance u2 was computed as:
D = j = 0 N f e a t u r e s 1 ( μ u 1 , j μ u 2 , j ) 2
where μ u , j is equivalent to the mean of all utterance u, computed as:
μ u = 1 N u i = 0 ,   X i u N u 1 X i
It can be seen from Table 4 and Table 5 that the Euclidean distances of the means increased significantly after the first convolutional layer. This growth highlights the important role of the convolutional layer in transforming the feature space into a space with a better separability between classes.
Table 6 and Table 7 summarizes the minimum and maximum of the computed Euclidean distance for the phonemes and words, together with the minimum and maximum of the standard deviation of the features to emphasize that, even if the averages move away from each other, the standard deviation has not undergone significant changes. The standard deviation of each feature for each class was computed using the following equation:
σ u = 1 N u i = 1 ,     X i u   N u 1 ( X i μ u ) 2
where σ u is the standard deviation for each feature for the utterance u.

4.2. Brain Areas Analysis

The second study conducted in this paper was based on evaluating the performance of the system when using a smaller number of electrodes from different regions of the brain. Table 8 details the region selected for the study, along with the electrodes used for that region.
The regions were selected at first based on the major brain areas corresponding to the EEG electrodes, i.e., frontal, central, and occipital areas, and afterwards based on the anatomical brain areas involved in speech conceptualization and articulatory plans, initiation, and coordination of the motor stimulus to be sent to the effectors. Regarding the anatomical brain areas involved in speech, is well-known that the Broca area has an important role in speech production [23]: the primary motor cortex generates the signals that control the execution of the movements and was discovered to also relate with the motor imagination [24], and the secondary motor area is responsible for motor planning [25]. The next step was to identify the spatial position of each anatomical structure identified to be involved in speech, regarding to the electrodes positioned in the 10-20 system. The Broca area is positioned in the posterior half of the inferior frontal gyrus; the primary motor cortex corresponds to the precentral gyrus, and the secondary motor cortex is mostly identified as Brodmann area 6, which is positioned in the precentral gyrus, the caudal superior frontal gyrus, and the caudal middle central gyrus. These anatomical spatial structures were identified as electrodes in the 10-20 positioning system by the researcher L. Koessler et al. in [26] as the channels selected for the anatomical speech brain areas detailed in Table 8.
For a better perspective of the analyzed brain areas, Figure 6 presents the 10-20 system for the electrode positioning used to acquire the signals from Kara One database having the selected regions colored differently.
The obtained results for each brain area and the combinations of the brain areas studied in this paper are presented in Table 9.
As expected, after selecting fewer channels corresponding to different brain areas of interest, the system classification accuracy decreased. The best accuracy was achieved when using the electrodes from the identified anatomical speech brain areas, according to the specialized literature, from both the left and right hemispheres, reaching a value of approx. 40%. Although there was a decrease in performance, compared to using all the available channels, and there were advantages regarding the computation, memory, costs, and even the comfort of using a smaller number of electrodes.

4.3. Complexity and Memory Analysis

The tendency of an ISR system is to perform with the highest possible accuracy on a portable device with limited resources. There will always be a tension between the performance and the complexity when developing an ISR system. In this section of the paper, we focused on a study regarding the complexity and memory of the developed system.
When it comes to the complexity and memory of an intelligent system, the primary resource consumer is generally the neural network. In the case of a CNN neural network with long-term memory, the complexity is given by the long-term memory convolutional layer, due to the need of the implied gates output computation, i.e., the input gate, the forgetting gate, and the output gate, considering the 2D spatial correlation in this process, too. For the 3D convolution corresponding to the input tensor (Nlines × Ncolumns × Nchannels) and the kernel (klines × kcolumns × kchannels), the complexity can be measured as:
O ( ( N l i n e s + k l i n e s ) ( N c o l u m n s + k c o l u m n s ) ( N c h a n n e l s + k c h a n n e l s ) log ( ( N l i n e s + k l i n e s ) ( N c o l u m n s + k c o l u m n s ) ( N c h a n n e l s + k c h a n n e l s ) ) )
when using the fast Fourier transform, as described in [27].
In our case, the N l i n e s = N c o l u m n s ,  k l i n e s = k c o l u m n s , and N c h a n n e l s = k c h a n n e l s , so the Equation (16) becomes:
O ( 2 ( N + k ) 2 N c h log ( ( 2 ( N + k ) 2 N c h ) ) )
where N is the input lines/columns, k is the kernel dimension, and N c h is the number of input channels.
However, for a ConvLSTM2D layer, this computation (described in Equation (17)), will be made for all the filters of the layer ( N f ), for all windows ( N w ), and for all gates, plus the computation of the last cell output (a total of four computations). Finally, the complexity can be approximated to:
O ( 4 N w ( 2 N 2 N c h ) log ( ( 2 N 2 N c h ) ) N f )
The details regarding the complexity for all system stages and for all specific layers of the used neural network are presented in Table 10, along with the memory consumption and the execution time measured using an AMD Ryzen 7 4800HS CPU.

5. Discussion

This paper aimed to develop an intelligent system for the imaginary speech recognition of seven phonemes and four words from the Kara One database. To achieve our goal, we passed the signals from the database through a preprocessing stage, in which an expert analyzed the imaginary speech epochs and eliminated the ones with high noises. Afterwards, in the feature extraction stage, we computed the cross-covariance between the channels in the frequency domain to conserve the channels connections in a compact matrix. In this stage, we showed, by using the LDA algorithm for feature visualization in a 2D feature space, that this method is better for decoding the features than using the signals in time or the cross-covariance applied over the channels in time. We used, as a classifier, an CNNLSTM neural network and concluded that it performed better than a CNN (in comparison to paper [17]), due to the consideration of not only the space correlations, but of the time variations, too. We were looking to develop a system that can easily be passed to a low-cost portable device, so we also studied the possibility of using a smaller number of electrodes divided into different brain regions. Finally, we also studied the complexity of the algorithm and the time required for the system to decide the membership of an input.

5.1. LDA for Feature Extraction Analysis

In the feature extraction stage, the signals were passed through a feature reduction algorithm used for the visual analysis of the three different features: the signal in time without processing, the cross-covariance in time, and the cross-covariance in the frequency domain. We observed that, when using the cross-covariance in the frequency domain, the features were clustered into the desired classes, however overlapped, but obtaining a good distinction between the phonemes and the words. We also observed that the distances between the means got higher using this type of feature extraction. When using the cross-covariance in time domain, the phonemes and the words were distributed in clusters, but almost completely overlapped, which made the classification harder. The computed means of each class was higher, compared to the raw feature signal, but the separability between the phonemes and words was lost. We observed that the means of the raw signals were close to each other for phonemes and for words, but a distinction between the two was clearly made.

5.2. CNNLSTM vs. CNN

The main advantage of a LSTM neural network is its capability of memorizing long-term dependencies. This neural network ability comes in handy when non-stationary time variant signals, such as EEG, are analyzed. In addition to the LSTM long-term memory, the spatial dependencies of the EEG signals were taken into consideration by adding CNN to the LSTM layer. This junction helped the neural network to learn both spatial and time-variant features, increasing the accuracy of the system from 0.37, obtained with a CNN architecture with similar parameters up to 0.43.
The mean confusion matrix of the k-folds presented in Figure 5 reveals that there is hardly any confusion between the phonemes and words, as expected after the visualization of the LDA dimension reduction results in Figure 2. Confusions were made between the phonemes /tiy/ and /diy/, with a relatively higher percent, compared to the rest of the phonemes. This behavior can be explained by the similar mechanisms of the pronunciation of the sound “t” and the sound “d”. For both utterances, the vocal tract maintains almost the same position, and the sound is being produced by presence or absence of vocal cord vibration. Major confusions of approx. 19% were also made between the words “pat” and “pot”, for the same reason of similar mechanism of pronunciation.
In Table 4, we computed the Euclidean distance between the mean of each phoneme and word analyzed for the cross-covariance in the frequency domain. We observed that these distances are smaller than the ones computed for the extracted features obtained after the first CNNLSTM layer (Table 5), meaning that, after the signal was processed by the first neural network layer, the feature space of the utterances was modified, and the features became more separable.

5.3. Brain Area Analysis

The final goal of an automatic ISR system is to obtain the best possible accuracy using a portable device with limited resources. Considering this, we further studied the system behavior by using a smaller number of electrodes, located in specific areas of the brain, for classifications. This approach enhanced the portability of the device and decreased the needed resources used for development, but with the cost of accuracy, as can be seen in Table 9.
When the number of channels was reduced, the accuracy dropped, as well. However, when using the electrodes positioned on the anatomical regions of the brain responsible for speech intention and production, the accuracy of the system reached 0.40, a value with a drop of only 3%, in comparison to all channel accuracy. This means that 93% of speech information is concentrated in these channels, and only 7% of information is distributed to the parietal and occipital regions. The main advantage of using only the channels corresponding to the anatomical structures of the brain regarding speech is that the number of electrodes is reduced, in this case by more than half, from 62 to 29, which is a big computational and cost gain.
Important information obtained after the study of the brain areas involved in imaginary speech recognition is that the data acquisition using the visual stimulus appearing on a prompt does not affect the study because the occipital area is less involved in the decision-making of the system. This is due to the two-second period introduced in the protocol design, the time between the prompt display and the actual imagination of the phoneme.

5.4. Complexity and Memory Analysis

An important aspect when developing an ISR system is considering the complexity of the algorithm and the memory used. Usually, the main consummator of the resources is the neural network, as can also be seen in Table 10. The maximum number of operations is given by the second layer of the CNNLSTM neural network and is to the order of approx. O (6.3 × 109). However, the execution time for a decision is under 100 ms, even when using all the channels in the computation, which means that it can still be implemented in a real-time device.
When reducing the number of electrodes, we can see in Table 11 an important decrease in the time execution, as well. An alternative to an all-electrode system can be the ASBA electrodes system, in which the time of execution considerably dropped from 80 ms to 20 ms, with a drop in accuracy of only 3%. This alternative is better from other points of view, too, such as increasing the user comfort when using the device and decreasing the cost of the final product.
When it comes to the memory usage, the system has its limitation because it needs at least 2 GB of memory only to retain the weights, due to the gates and the long-term memory implication of the LSTM network. However, the LSTM addition to the system improved the overall classification and added great value to the final system.

6. Conclusions

This paper aimed to develop a subject’s shared system for recognizing seven phonemes and four words collected during imaginary speech from the Kara One Database. To achieve the proposed goal, the database was preprocessed and features were computed, in order to decode the hidden information regarding imaginary speech. The features were based on computing the cross-covariance in the frequency domain and were introduced to a CNNLSTM neural network for final classification in the desired classes.
During our research, we observed that, by computing the cross-covariance in frequency domain, the feature space turned into a space where the utterances could be separated easily. This conclusion was drawn after analyzing the feature space in two dimensions by computing the LDA algorithm for feature reduction.
Another feature comparison was made between the extracted features and the features obtained after the first CNNLSTM layer. Table 4 and Table 5 represent the Euclidean distances between the mean of each class, and it can be easily seen that these distances increased after the CNNLSTM layer. This suggests that the feature space of the input vectors changed, so that the classes became more separable. This is one of the advantages of deep neural networks, especially CNN networks, which use the first layers for feature extraction before the actual classification of the data.
This paper also showed an improvement for the system performance when using a CNNLSTM neural network, with respect to CNN. The accuracy increased from 37% to 43% when using CNNLSTM and the same processing chain of the database. The advantage of CNNLSTM is the consideration of both the spatial and temporal connections. CNN uses the convolution for the spatial connection between the channels, while LSTM brings the long-term memory that is vital for the non-stationary time variant signals, such as EEG.
The proposed system considers the portability of a possible real-time portable device. Therefore, we also studied the system’s behavior when reducing the number of channels for classification. We separated the electrodes into main areas, i.e., frontal, central, and occipital, studied the areas involved in imaginary speech production, and selected the channels accordingly for the anatomical speech brain areas study. We concluded that 93% of the information is concentrated in the anatomical speech brain areas, obtaining an accuracy of 40% for 29 channels used, in comparison to the 62 used in the beginning. Even if the accuracy dropped by 3% using only 29 electrodes, the system brings more advantages in a matter of time for execution, portability, comfort, and cost.

Author Contributions

Conceptualization, A.-L.R. and O.G.; Methodology, A.-L.R. and O.G.; Software, A.-L.R.; Data curation, A.-L.R.; Writing – original draft, A.-L.R.; Writing – review & editing, O.G.; Supervision, O.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical approval was obtained from both the University of Toronto and the University Health Network, of which Toronto Rehab is a member.

Informed Consent Statement

Consent was obtained from all subjects involved in the study.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kennedy, P.R.; Bakay, R.A.E.; Moore, M.M.; Adams, K.; Goldwaithe, J. Direct control of a computer from the human central nervous system. IEEE Trans. Rehab. Eng. 2000, 8, 198–202. [Google Scholar] [CrossRef] [PubMed]
  2. Jayabhavani, G.N.; Rajaan, N.R. Brain enabled mechanized speech synthesizer using Brain Mobile Interface. Int. J. Eng. Technol. 2013, 5, 333–339. [Google Scholar]
  3. Chaudhary, U.; Vlachos, I.; Zimmermann, J.B.; Espinosa, A.; Tonin, A.; Jaramillo-Gonzalez, A.; Khalili-Ardali, M.; Topka, H.; Lehmberg, J.; Friehs, G.M.; et al. Spelling interface using intracortical signals in a completely locked-in patient enabled via auditory neurofeedback training. Nat. Commun. 2022, 13, 1236. [Google Scholar] [CrossRef] [PubMed]
  4. Proix, T.; Delgado Saa, J.; Christen, A.; Martin, S.; Pasley, B.N.; Knight, R.T.; Tian, X.; Poeppel, D.; Doyle, W.K.; Devinsky, O.; et al. Imagined speech can be decoded from low- and cross-frequency intracranial EEG features. Nat. Commun. 2022, 13, 48. [Google Scholar] [CrossRef] [PubMed]
  5. Martin, S.; Brunner, P.; Iturrate, I.; Millán, J.D.R.; Schalk, G.; Knight, R.T.; Pasley, B.N. Word pair classification during imagined speech using direct brain recordings. Sci. Rep. 2016, 6, 25803. [Google Scholar] [CrossRef] [Green Version]
  6. Angrick, M.; Herff, C.; Mugler, E.; Tate, M.C.; Slutzky, M.W.; Krusienski, D.J.; Schultz, T. Speech synthesis from ECoG using densely connected 3D convolutional neural networks. J. Neural Eng. 2019, 16, 036019. [Google Scholar] [CrossRef]
  7. Angrick, M.; Ottenhoff, M.C.; Diener, L.; Ivucic, D.; Ivucic, G.; Goulis, S.; Saal, J.; Colon, A.J.; Wagner, L.; Krusienski, D.J.; et al. Real-time synthesis of imagined speech processes from minimally invasive recordings of neural activity. Commun. Biol. 2021, 4, 1055. [Google Scholar] [CrossRef]
  8. Varshney, Y.V.; Khan, A. Imagined Speech Classification Using Six Phonetically Distributed Words. Front. Signal Process. 2022, 2, 760643. [Google Scholar] [CrossRef]
  9. Zhao, S.; Rudzicz, F. Classifying phonological categories in imagined and articulated speech. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 992–996. [Google Scholar] [CrossRef]
  10. Nguyen, C.H.; Karavas, G.K.; Artemiadis, P. Inferring imagined speech using EEG signals: A new approach using Riemannian manifold features. J. Neural Eng. 2018, 15, 016002. [Google Scholar] [CrossRef]
  11. Panachakel, J.T.; Ramakrishnan, A.G.; Ananthapadmanabha, T.V. Decoding Imagined Speech using Wavelet Features and Deep Neural Networks. In Proceedings of the 2019 IEEE 16th India Council International Conference (INDICON), Rajkot, India, 13–15 December 2019; pp. 1–4. [Google Scholar] [CrossRef]
  12. Saha, P.; Fels, S.; Abdul-Mageed, M. Deep Learning the EEG Manifold for Phonological Categorization from Active Thoughts. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 13–17 May 2019; pp. 2762–2766. [Google Scholar] [CrossRef] [Green Version]
  13. Tsiouris, Κ.Μ.; Pezoulas, V.C.; Zervakis, M.; Konitsiotis, S.; Koutsouris, D.D.; Fotiadis, D.I. A Long Short-Term Memory deep learning network for the prediction of epileptic seizures using EEG signals. Comput. Biol. Med. 2018, 99, 24–37, Aug. [Google Scholar] [CrossRef]
  14. Agarwal, P.; Kumar, S. Electroencephalography-based imagined speech recognition using deep long short-term memory network. ETRI J. 2022, 44, 672–685. [Google Scholar] [CrossRef]
  15. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th Symposium on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  16. Porbadnigk, A.; Wester, M.; Calliess, T.S.J.P. EEG-based speech recognition—Impact of Temporal Effects. In Proceedings of the International Conference on Bio-inspired Systems and Signal Processing, Porto, Portugal, 14–17 January 2009; pp. 376–381. [Google Scholar] [CrossRef] [Green Version]
  17. Rusnac, A.-L.; Grigore, O. CNN Architectures and Feature Extraction Methods for EEG Imaginary Speech Recognition. Sensors 2022, 22, 4679. [Google Scholar] [CrossRef]
  18. Ariza, I.; Tardón, L.J.; Barbancho, A.M.; De-Torres, I.; Barbancho, I. Bi-LSTM neural network for EEG-based error detection in musicians’ performance. Biomed. Signal Process. Control. 2022, 78, 103885. [Google Scholar] [CrossRef]
  19. Ma, Q.; Wang, M.; Hu, L.; Zhang, L.; Hua, Z. A Novel Recurrent Neural Network to Classify EEG Signals for Customers’ Decision-Making Behavior Prediction in Brand Extension Scenario. Front. Hum. Neurosci. 2021, 15, 610890. [Google Scholar] [CrossRef]
  20. Zhang, G.; Davoodnia, V.; Sepas-Moghaddam, A.; Zhang, Y.; Etemad, A. Classification of Hand Movements From EEG Using a Deep Attention-Based LSTM Network. IEEE Sensors J. 2020, 20, 3113–3122. [Google Scholar] [CrossRef] [Green Version]
  21. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.-Y.; Wong, W.; Woo, W. Convolutional LSTM Network: A Machine Learning Approach for Precipitation Nowcasting. arXiv 2015, arXiv:1506.04214v2. [Google Scholar]
  22. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Müller, A.; Nothman, J.; Louppe, G.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  23. Lorca-Puls, D.L.; Gajardo-Vidal, A.; Oberhuber, M.; Prejawa, S.; Hope, T.M.H.; Leff, A.P.; Green, D.W.; Price, C.J.; Team, P. Brain regions that support accurate speech production after damage to Broca’s area. Brain Commun. 2021, 3, fcab230. [Google Scholar] [CrossRef]
  24. Bhattacharjee, S.; Kashyap, R.; Abualait, T.; Chen, S.-H.A.; Yoo, W.-K.; Bashir, S. The Role of Primary Motor Cortex: More Than Movement Execution. J. Mot. Behav. 2021, 53, 258–274. [Google Scholar] [CrossRef]
  25. Olson, J.M.; Li, J.K.; Montgomery, S.E.; Nitz, D.A. Secondary Motor Cortex Transforms Spatial Information into Planned Action during Navigation. Curr. Biol. 2020, 30, 1845–1854.e4. [Google Scholar] [CrossRef]
  26. Koessler, L.; Maillard, L.; Benhadid, A.; Vignal, J.P.; Felblinger, J.; Vespignani, H.; Braun, M. Automated cortical projection of EEG sensors: Anatomical correlation via the international 10–10 system. NeuroImage 2009, 46, 64–72. [Google Scholar] [CrossRef] [PubMed]
  27. Notchenko, A.; Kapushev, E.; Burnaev, E. Large-Scale Shape Retrieval with Sparse 3D Convolutional Neural Networks. arXiv 2017, arXiv:1611.09159. [Google Scholar]
Figure 1. Cross-covariance in frequency domain for 0.1 s window for all phonemes and words from the KODB.
Figure 1. Cross-covariance in frequency domain for 0.1 s window for all phonemes and words from the KODB.
Applsci 12 11873 g001
Figure 2. Feature reduction using LDA algorithm with eigenvalue decomposition and shrinkage value 0.2 for (a) cross-covariance features in frequency domain, (b) raw signal in time domain, and (c) cross-covariance features in time domain.
Figure 2. Feature reduction using LDA algorithm with eigenvalue decomposition and shrinkage value 0.2 for (a) cross-covariance features in frequency domain, (b) raw signal in time domain, and (c) cross-covariance features in time domain.
Applsci 12 11873 g002
Figure 3. The mean of each LDA feature space of KODB classes with the confidence ellipse for feature space analysis of: (a) cross-covariance in frequency domain; (b) raw-time features; (c) cross-covariance in time domain.
Figure 3. The mean of each LDA feature space of KODB classes with the confidence ellipse for feature space analysis of: (a) cross-covariance in frequency domain; (b) raw-time features; (c) cross-covariance in time domain.
Applsci 12 11873 g003
Figure 4. Diagram block for the CNNLSTM neural network used in the classification stage.
Figure 4. Diagram block for the CNNLSTM neural network used in the classification stage.
Applsci 12 11873 g004
Figure 5. The confusion matrix obtained for all eleven classes of KODB when using CNNSLTM as neural network.
Figure 5. The confusion matrix obtained for all eleven classes of KODB when using CNNSLTM as neural network.
Applsci 12 11873 g005
Figure 6. The representation of the different brain areas selected for the channels dimension reduction corresponding to the (a) major areas of the EEG represented with different colors and (b) ASBA.
Figure 6. The representation of the different brain areas selected for the channels dimension reduction corresponding to the (a) major areas of the EEG represented with different colors and (b) ASBA.
Applsci 12 11873 g006
Table 1. The mean and standard deviation of the LDA components for the analyzed features: raw time features, covariance in time domain features, and covariance in frequency domain features.
Table 1. The mean and standard deviation of the LDA components for the analyzed features: raw time features, covariance in time domain features, and covariance in frequency domain features.
LDA–Time FeaturesLDA–COV-Time FeaturesLDA–COV-Frequency Features
μ L D A 1 μ L D A 2 σ L D A 1 σ L D A 2 μ L D A 1 μ L D A 2 σ L D A 1 σ L D A 2 μ L D A 1 μ L D A 2 σ L D A 1 σ L D A 2
/iy/0.74080.62910.09690.13620.71570.35450.08750.08710.17730.56340.05140.0584
/uw/0.76560.62280.09030.11260.67500.44140.07480.08250.20360.81490.05640.0606
/piy/0.77480.60840.09030.11790.67050.33510.07860.08960.18910.22190.05720.0604
/tiy/0.77230.54750.07370.13250.69630.41300.07800.07770.17390.49010.05300.0586
/diy/0.75780.60890.10690.12830.68910.41930.08050.07740.20720.68970.05660.0611
/m/0.77070.51090.08670.14820.67800.38770.07260.07300.16860.31350.05390.0655
/n/0.77130.57440.07970.11990.71120.51870.08760.10160.16410.45600.05420.0614
pat0.18930.59640.10350.05420.48960.58550.09720.13260.76960.51930.05710.0569
pot0.21410.58340.11080.05450.48820.37160.09900.08010.77550.33950.05640.0606
knew0.18290.57650.08890.04390.47610.30420.10030.10050.76410.69950.05740.0599
gnaw0.20920.58550.10370.04810.54700.40730.09780.07540.78180.46350.05810.0599
Table 2. The confusion matrix of a multiclass dataset; y i —the desired output of the system, y i ^ —the output of the system. Nc—number of classes.
Table 2. The confusion matrix of a multiclass dataset; y i —the desired output of the system, y i ^ —the output of the system. Nc—number of classes.
Predicted Label
Classes c 1 c 2 c N c Total
True label c 1  
y i     c 1
y i ^     c 1 y i ^     c 2 y i ^     c N c j N c C c 1 j
c 2  
y i     c 2
y i ^     c 1 y i ^     c 1 y i ^     c N c j N c C c 2 j
c N c  
y i     c N c
y i ^     c 1 y i ^     c 1 y i ^     c N c j N c C N c j
Total j N c C j c 1 j N c C j c 2 j N c C j N c k N c C k k
Table 3. The CNNLSTM neural network performances presented in comparison to CNN neural network, described in [17].
Table 3. The CNNLSTM neural network performances presented in comparison to CNN neural network, described in [17].
AccuracyBalanced AccuracyKappaRecall
CNN [17]0.3758 ± 0.0040.3749 ± 0.0040.3132 ± 0.0040.3750 ± 0.004
CNNLSTM
[This study]
0.4398 ± 0.0040.4392 ± 0.0040.3837 ± 0.0040.4398 ± 0.004
Table 4. The computed Euclidean distances between the mean of each class and every phoneme and word of the KODB for the cross-covariance features in frequency domain (the features were normalized before the computation of the Euclidean distance).
Table 4. The computed Euclidean distances between the mean of each class and every phoneme and word of the KODB for the cross-covariance features in frequency domain (the features were normalized before the computation of the Euclidean distance).
/iy//uw//piy//tiy//diy//m//n/patpotknewgnaw
/iy/00.961.200.971.021.461.213.403.453.193.00
/uw/0.9601.441.161.161.611.183.343.493.172.97
/piy/1.201.4400.971.261.411.153.443.383.233.05
/tiy/0.971.160.9700.941.261.033.233.273.042.84
/diy/1.021.161.260.9401.181.213.083.172.882.70
/m/1.461.611.411.261.1801.122.812.802.672.43
/n/1.211.181.151.031.211.1202.952.962.832.57
pat3.403.343.443.233.082.812.9501.100.650.93
pot3.453.493.383.273.172.802.961.1001.070.98
knew3.193.173.233.042.882.672.830.651.0700.93
gnaw3.002.973.052.842.702.432.570.930.980.930
Table 5. The computed Euclidean distances between the mean of each class and every phoneme and word of the KODB for the features computed by the CNNLSTM first layer (the features were normalized before the computation of the Euclidean distance).
Table 5. The computed Euclidean distances between the mean of each class and every phoneme and word of the KODB for the features computed by the CNNLSTM first layer (the features were normalized before the computation of the Euclidean distance).
/iy//uw//piy//tiy//diy//m//n/patpotknewgnaw
/iy/08.749.918.038.1110.819.6739.1233.6435.0628.91
/uw/8.74011.079.898.5412.109.3336.4531.4832.6226.71
/piy/9.9111.0708.079.0910.669.2838.7132.9734.7328.92
/tiy/8.039.898.0707.678.978.3338.5432.9634.5128.35
/diy/8.118.549.097.6708.909.1337.1031.5333.0726.95
/m/10.8112.1010.668.978.9009.2037.1030.9933.2826.76
/n/9.679.339.288.339.139.20037.3731.4233.4626.93
Pat39.1236.4538.7138.5437.1037.1037.37011.247.1413.92
Pot33.6431.4832.9732.9631.5330.9931.4211.2408.798.88
knew35.0632.6234.7334.5133.0733.2833.467.148.79010.36
gnaw28.9126.7128.9228.3526.9526.7626.9313.928.8810.360
Table 6. The minimum and maximum values of the Euclidian distances and the minimum and maximum of standard deviation over the features for the cross-covariance in frequency domain.
Table 6. The minimum and maximum values of the Euclidian distances and the minimum and maximum of standard deviation over the features for the cross-covariance in frequency domain.
Euclidean DistanceMeasured between ClassesStd. Dev. of Feature Components Limits
σ m i n σ m a x
Min.0.6543pat0.07540.2448
knew0.0640.2367
Max.3.4864/uw/0.04910.2166
pot0.06710.236
Table 7. The minimum and maximum values of the Euclidian distances and the minimum and maximum of standard deviation over the features computed after the first CNNLSTM layer.
Table 7. The minimum and maximum values of the Euclidian distances and the minimum and maximum of standard deviation over the features computed after the first CNNLSTM layer.
Euclidean DistanceMeasured between ClassesStd. Dev. of Feature Components Limits
σ m i n σ m a x
Min.7.1444pat0.01450.3034
knew0.01460.3075
Max.39.1189/uw/0.01260.2769
pot0.01450.3034
Table 8. The analyzed brain areas for channels reduction with the corresponding electrodes from the 10-20 system for electrode positioning.
Table 8. The analyzed brain areas for channels reduction with the corresponding electrodes from the 10-20 system for electrode positioning.
Frontal (F)Central (C)Occipital (O)Anatomical Speech Brain Areas (ASBA)
Left (L)FP1, FPZ, AF3, F7, F5, F3, F1, FZ, FCZ, FC1, FC3, FC5, FT7FT7, FC5, FC3, FC1, FCZ, CZ, C1, C3, C5, T7, TP7, CP5, CP3, CP1, CPZP7, P5, P3, P1, PZ, POZ, PO3, PO5, PO7, O1, OZAF3, F7, F5, F3, F1, FZ, FCZ, FC1, FC3, FC5, FT7, T7, C5, C3, C1, CZ
Right (R)FP2, FPZ, AF4, FZ, F2, F4, F6, F8, FT8, FC6, FC4, FC2, FCZFCZ, FC2, FC4, FC6, FC8, T8, C6, C4, C2, CZ, CPZ, CP2, CP4, CP6, TP8PZ, P2, P4, P6, P8, PO8, PO6, PO4, POZ, OZ, O2AF4, FZ, F2, F4, F6, F8, FT8, FC6, FC4, FC2, FCZ, CZ, C2, C4, C6, T8
Table 9. The mean of k-folds (k = 4) accuracy of the classification results obtained using different areas and combinations of different areas as CNNLSTM input; F—frontal, C—central, O—occipital, FC—frontal and central, CO—central and occipital, FCO—frontal, central, and occipital, ASBA—anatomical speech brain areas, L—left, and R—right.
Table 9. The mean of k-folds (k = 4) accuracy of the classification results obtained using different areas and combinations of different areas as CNNLSTM input; F—frontal, C—central, O—occipital, FC—frontal and central, CO—central and occipital, FCO—frontal, central, and occipital, ASBA—anatomical speech brain areas, L—left, and R—right.
FCOFCCOFCOASBAFCOASBA
L0.2846 ± 0.0050.2892 ± 0.0050.2297 ± 0.0020.3372 ± 0.0020.3132 ± 0.0020.3518 ± 0.0010.3096 ± 0.0060.3582 ± 0.0060.3678 ± 0.0050.2606 ± 0.0060.4027 ± 0.002
R0.3028 ± 0.0030.3032 ± 0.0030.2272 ± 0.0040.3386 ± 0.0070.3189 ± 0.0010.3566 ± 0.0030.3305 ± 0.001
Table 10. The complexity and memory of the system detailed for each step taken for system development.
Table 10. The complexity and memory of the system detailed for each step taken for system development.
System StagesComplexityMemoryTime [s]
Feature ExtractionFFTO(Nf × N × MlogM)~23 MB<0.001
COVO(Nf × N2)~24 MB<0.001
CNNLSTMConvLSTM2D-64O(4 × 4 × (2N2) × log(2N2) × 64)~2.12 GB0.0819
ConvLSTM2D-128O(4 × 4 × (2N2 × 64) × log(2N2 × 64) × 128)
Dense-64O(4 × N2 × 128 × 64)
Dense-11O(64 × 11)
Table 11. The mean execution time needed to obtain a response from the system for all groups of channels analyzed in the paper; L—left, R—right, ASBA—anatomical speech brain areas.
Table 11. The mean execution time needed to obtain a response from the system for all groups of channels analyzed in the paper; L—left, R—right, ASBA—anatomical speech brain areas.
No. ChannelsTime [s]Accuracy [%]
All channels620.082643.98
L/R Frontal130.004628.46/30.28
L/R Central150.005728.92/30.32
L/R Occipital110.003522.97/22.72
L/R Frontal + Central230.012433.72/33.86
L/R Central + Occipital260.015531.32/31.89
L/R Frontal + Central + Occipital340.025835.18/35.66
L/R ASBA160.006840.27
L + R Frontal230.016735.82
L + R Central270.015636.78
L + R Occipital160.008926.06
L + R ASBA290.020940.27
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rusnac, A.-L.; Grigore, O. Imaginary Speech Recognition Using a Convolutional Network with Long-Short Memory. Appl. Sci. 2022, 12, 11873. https://doi.org/10.3390/app122211873

AMA Style

Rusnac A-L, Grigore O. Imaginary Speech Recognition Using a Convolutional Network with Long-Short Memory. Applied Sciences. 2022; 12(22):11873. https://doi.org/10.3390/app122211873

Chicago/Turabian Style

Rusnac, Ana-Luiza, and Ovidiu Grigore. 2022. "Imaginary Speech Recognition Using a Convolutional Network with Long-Short Memory" Applied Sciences 12, no. 22: 11873. https://doi.org/10.3390/app122211873

APA Style

Rusnac, A. -L., & Grigore, O. (2022). Imaginary Speech Recognition Using a Convolutional Network with Long-Short Memory. Applied Sciences, 12(22), 11873. https://doi.org/10.3390/app122211873

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop