1. Introduction
Tinnitus is characterized by the perception of sound in the absence of an external sound source. Although the effects of tinnitus vary from person to person, most patients suffer from varying degrees of fatigue, stress, sleep problems, difficult in concentrating, memory loss, anxiety, and irritability [
1]. The most common cause of tinnitus is exposure to loud noise, while other causes include ototoxic drugs (e.g., aspirin), head injury, ear infection, or other diseases, such as diabetes [
2].
The acute effect is injury to the cochlea and auditory nerve, following auditory plasticity, tinnitus-related effects centralise over time. Due to the heterogeneity of tinnitus, treatment response varies. Clinical judgements about the likelihood of treatment response are currently difficult, and patients are often switched through several treatments before finding an effective intervention. AI may help in personalising the prescription approach to tinnitus by indicating those interventions most likely to be effective for any given patient. This would improve clinical efficiency, reducing patient distress and optimising benefits. Although several AI methods have been applied to brain data in order to classify (diagnosis) tinnitus (e.g., support vector machine, SVM; multilayer perceptron, MLP; logistic regression; and naive Bayes learning) [
3,
4,
5,
6], there is an absence of research on the early prediction of symptom outcomes (prognosis). Prediction of response might be possible by applying machine learning methods to diverse spatio-temporal brain data (e.g., electroencephalogram, EEG [
7]; magnetic resonance imaging, MRI [
8]; functional MRI [
9]; functional near-infrared spectroscopy, fNIRS [
10]; and magnetoencephalography, MEG [
11]), clinical and behavioural measures (e.g., hearing thresholds, sleep quality), neuropsychological tests (e.g., memory); and/or cardiovascular measurements (e.g., heart rate variability) [
12].
The current study focuses is on EEG, which allows for high temporal-resolution recording of cortical electrical activity via electrodes positioned over the scalp [
13,
14]. By using machine learning to model high-dimensional EEG data, feature selection methods can be applied to remove irrelevant data, hence, a smaller number of data features which can reduce the model complexity. This helps ro prevent overfitting and improves learning performance by promoting generalization [
15,
16,
17,
18]. Sub-sampling in the time domain or frequency domain of EEG signals is a common method for feature extraction which is broadly utilized in neurological diagnosis. Combining EEG features from the temporal and frequency domains leads to increased accuracy in pattern classification [
13]. Reference [
19] proposed converting EEG features from the temporal domain to the frequency domain using Fast Fourier Transform (FFT) and illustrated them as a series of multispectral images of brain topology. These images were then used to train machine learning algorithms, including deep learning neural networks, to learn from the robust representation of image sequences. One of the most popular methods in machine learning is based on deep neural networks. Deep neural networks (DNN) are computational models that simulate the way neurons process information. They can identify patterns in data and make predictions based on those patterns. A variety of architectures of DNN have been proposed so far, including the Convolutional Neural Network (CNN) [
20]. The CNN is a feed-forward neural network with artificial neurons that respond to surrounding units, and it is excellent for image processing. The CNN consists of one fully-connected layer (first layer), one or more convolutional layers, as well as associative weights and a pooling layers. CNNs can also be trained using back-propagation algorithms. Compared to other deep neural networks, convolutional neural networks require fewer parameters to be considered, making them an attractive structure for deep learning [
21]. Modelling of high-dimensional EEG data using machine learning methods has been conducted for the diagnosis of various neurological diseases, including epilepsy and seizures [
22]. Reference [
23] proposed a fuzzy decision tree (FDT) classifier for epileptic seizure detection which achieved 99.5% accuracy. Despite the existence of several successful EEG applications in the health field, there are still limited studies on modelling EEG for the prediction of responses to tinnitus treatment. Most of the EEG studies are based on the classification of raw EEG signals using machine learning. Limited investigation has been conducted on using dynamic functional connectivity and dynamic frequency images as deep learning inputs for the prediction of tinnitus treatment outcomes.
In this paper we explain the methodology of our predictive models as well as the feature selection methods used for improving the performance of the models. We report the predicted results of treatment outcome classification (i.e., responder, non-responder). A comparative analysis using different feature selection methods is reported on EEG data in the frequency and functional connectivity domains.
2. Methodology
This research aims to develop an AI-based system for predicting the change in patients’ severity of tinnitus over a period of treatment using computational models of brain data collected prior to and following a tinnitus treatment (the EEG data is explained in [
24]). The primary measure of tinnitus severity used was the Tinnitus Functional Index (TFI), a questionnaire designed to measure the impact of tinnitus on various aspects of life (e.g., sleep, communication, and quality of life) [
25]. The methods are illustrated in
Figure 1 and include modelling EEG data in the frequency domain and functional connectivity networks applied as inputs to artificial intelligence algorithms (neural networks) for extracting patterns and performing prediction of the tinnitus treatment outcomes.
The clinical impact of the proposed methodology is applying AI for the early prediction of response to tinnitus treatment when only baseline data is used. This would allow for an optimal selection of treatment options for patients through the prediction of response to each treatment. We also identified a subset of EEG sensors as informative features that increase the prediction accuracy and make the further development of wearable diagnosis and prognosis AI tools more effective. Sensors have seldom been used clinically in relation to sensory disorders. The application of sensor technology to tinnitus is novel, and the algorithm and its application are innovative approaches to this complex sensory disorder. The literature has clearly identified the need for biomarkers, and this is amongst the first studies looking towards effective wearable solutions in our field.
2.1. Datasets
We used two datasets (EEG, behavioural) from patients with tinnitus (n = 8 shown in
Figure 1a) at baseline (pre-treatment) and after the treatment (a follow up after 3 months). The behavioural data recorded 19 features, including cognitive, psychological, and the TFI scores collected at each follow-up, as listed in
Table 1. The TFI score in behavioural data was used to categorise patients into two groups (responders and non-responders), according to the level of change in their the TFI scores between the pre- and post-treatment phases. The responsiveness of the TFI to treatment-related change was evaluated by examining the change in baseline test–retest scores calculated as
. The smallest detectable change was determined from the variance in these measures and was found to be 4.8. Therefore, the responder group encompasses individuals with TFI changes ≥ 4.8, while the non-responder group encapsulates individuals with TFI changes < 4.8. This labelling criterion was also validated in the literature for New Zealand data [
26]. Among 8 patients, 4 of them were labelled as responders, while the other 4 patients were labelled as non-responder according to their TFI score changes.
EEG was recorded in an electrically shielded and sound-treated booth (ISO 8253–1:2010) from 64 BioSemi active Ag/AgCl recording electrodes. Electrode locations corresponded to the extended international 10/20 system. Electrodes were attached to a fitted BioSemi head cap. Parker Signa gel was applied at each electrode site to ensure reliable conductivity between electrode and scalp. Continuous EEG signals were recorded on a Dell Optiplex 7040 desktop computer at an 8192 Hz sample rate with a 64-channel BioSemi ActiveTwo system referenced to the common mode sense active electrode and grounded to the driven right leg passive electrode. The EEG signals were down-sampled to 256 Hz. The temporal lengths of these EEG signals were segmented into multiple intervals of 256 time-points (i.e., 1 s duration). This resulted in 6642 EEG samples, which were used for training deep neural networks to predict the patient’s tinnitus treatment outcomes (class labels: responder vs. non-responder defined with respect to the changes in their TFI score after 3 months of treatment). The class labels of the EEG samples correspond to the TFI changes in the behavioural data of the same patients.
Tinnitus severity numerical scales (TSNS) have been widely used to assess tinnitus severity and have demonstrated good test–retest reliability and concordance with other participative measures of tinnitus. Participants were asked how much a problem their tinnitus was (0 not a problem−5 very big problem). Numeric rating scales were used to measure tinnitus perception along five dimensions: how strong, intrusive, uncomfortable, unpleasant the tinnitus signal was, and how easy it was to ignore the tinnitus signal (0–10 rating, 0 not a problem−10 extreme problem) [
27]. The literature suggested statistical results that proved the validity of this scale [
28,
29].
2.2. Psychological Function
Depression, Anxiety, and Scale (DASS) [
30] is a validated, widely used self-reporting instrument that measures the dimensions of depression, anxiety, and stress for use in both clinical and nonclinical populations. Participants rate 21 items (7 for each subscale) on a 4-point scale of how much each statement applies to them. The Positive and Negative Affect Schedule (PANAS) [
31] is a self-reporting questionnaire that consists of a list of 10 positive and 10 negative affective adjectives, each rated on a 5-point scale (not at all–very much).
2.3. Prediction of Treatment Outcomes Using EEG Data and Neural Networks
In the current research, we applied neural networks for the classification of EEG data (only baseline EEG) in order to predict patients’ treatment outcomes (responder and non-responder groups) labelled after 3 months of tinnitus sound treatment. To this end, a tinnitus avatar was first generated that was identical to the individual with tinnitus. This sound was then morphed slowly over time until it was identical to an environmental sound. Within the field of virtual reality, one study has attempted to synthesize an auditory replica of tinnitus. However, this relied purely on auditory thresholds, tinnitus pitch, and participant preference judgments [
32].
In the current study, the computational AI model has the potential to distinguish the baseline EEG patterns of patients who are likely to respond to the tinnitus treatment over time. In order to identify different types of predictive patterns from EEG data, we considered two forms of EEG-driven information as inputs to the neural networks: (1) frequency domain (presented in
Section 2.3.1) and (2) EEG functional connectivity (presented in
Section 2.3.2).
2.3.1. EEG Class Prediction on Frequency Domain
In this section, we demonstrate EEG data modelling via deep neural networks for the prediction of treatment outcomes. Here, the EEG modelling is based on the frequency domain. A FFT (fast Fourier transform) was applied to transform each of the EEG samples into the frequency domain image. As described in Equation (1) and Reference [
33], we first created EEG samples with a 1 s duration. As explained in
Section 2.1, 6642 frequency images were generated. To make use of meaningful data, we selected features on the most prominent frequency bands (beta 14–30 HZ, alpha 8–14 HZ, and theta 4–8 HZ) as the analysis objects. The mean (across each 1 s epoch) of the absolute values for each of these frequencies was calculated at each electrode site.
where
is the FFT coefficients, N is the total number of input EEG samples, and n is total number of points in FFT. The frequency samples were then spatially mapped onto 64 electrodes, based on the international 10/20 system co-ordinates, and shown in 2D surface frequency images. Specifically, a grid data function based on Delaunay triangulation was used to interpolate the data points and generate frequency domain images. After the triangulation was completed, any points in the triangle area were interpolated according to the value of each triangle vertex. This process was repeated for all frequency bands of interest, resulting in three topographic maps corresponding to each frequency band (
Figure 2). The amplitudes of the three frequency bands were then averaged and passed to a single topographically based frequency image which dynamically changes over time when streaming temporal EEG data [
34,
35].
Data from the sequence of 1 s epochs was used as a temporal input into the neural networks for training. In this experiment, because the EEG datasets were presented in the form of images, we employed a CNN in deep learning [
36], commonly used in image classification. Such CNNs are inspired by the visual cortex, where the firing rate of every sensory neuron is affected by a specific region in the retina, called the neuron’s receptive field. CNNs consist of three main layers: the input layer, the feature learning layer, and the classifier layer. Each of these has several sub-layers. Through a convolution procedure, every region of neurons (receptive field) from layer
was connected to one neuron in layer
, which resulted in extracting abstractions (informative features from data) from layer
and transferring them to the next layer. CNNs use activation functions, therefore, they can solve non-linear classification tasks and have enabled advancements in computer vision systems, including image classification [
22], image segmentation [
37,
38], and object detection [
39]. In the current study, CNNs were applied for the classification of EEG frequency images into responder and non-responder groups. As shown in
Figure 3a, 2D images with a size of 32 × 32 pixels were harnessed as EEG frequency inputs into the CNN model. We employed ReLU as the activation function in the convolutional layers. The softmax was applied in the last activation layer for final classification and prediction.
2.3.2. EEG Class Prediction Based on Functional Connectivity
In this section, EEG data was transformed into functional connectivity representing the relationship between different cortical regions over time. Graph theory was applied to each EEG frequency domain in order to convert it into a graph-based representation of the frequency domain, shown in
Figure 3b. We employed squared coherence (
to compute the degree of correlation between every two EEG channels
and
in the frequency domain, as shown in Equation (2). Here, the three frequency bands of theta, alpha and beta were used.
where
Pxy represents the cross-spectral density of the signal
x and
y. Parameters
Pxx and
Pyy show the power spectral density. The generated EEG functional connectivity graphs were used as inputs to train a deep neural network. Hereafter, we employed MLP as the classifier.
Figure 4 shows an example of EEG functional connectivity graphs generated from one EEG sample (one second recording) for eight patients at two stages (before and after treatment). The results demonstrate that both the responder and non-responder groups have increased functional connectivity after receiving treatment, with the responder showing a greater increment in their EFC in the cortical areas measured by F1, Fz, F2, FC1, FCz, FC2, C1, Cz, and C2 electrodes. Functional connectivity in non-responder patients increased more significantly after treatment and was spread across all electrodes.
3. Results of Tinnitus Outcome Prediction Using EEG Data
3.1. Prediction Treatment Outcomes Based on EEG Frequency Domain
As explained in
Section 2.1, we created 6642 EEG samples, each with a recording length of 1 s with 256 data points. These samples were transformed into frequency images (explained in
Section 2.3.1) and employed as inputs to the CNN model. Before training the CNN model, we visualised an example of frequency images of 5 s of EEG data related to two patients from the non-responder group (in
Figure 5) and the responder group (in
Figure 6).
For the prediction of tinnitus treatment outcomes, only the baseline EEG data from patients were used to train the CNN model. The class label information for the baseline EEG data was defined according to the patients’ TFI scores after treatment. The trained CNN model was then tested using patients’ baseline EEG data, which were excluded from the training to predict whether the patients are likely to be classified as non-responders or responders. For the training and testing of the CNN, the EEG samples were split into training and test datasets in a ratio of 8:2, meaning that there were 5314 training samples and 1328 test samples. As reported in
Table 2, the prediction accuracy of the non-responder group is 99.07%, while the responder group was predicted with 98.86% accuracy.
Figure 7 shows the accuracy and the loss curves during CNN training. In the experiments, we used different epoch numbers (from 20–100) and learning rates (0.01, 0.001, 0.0001, and 0.00001) with the aim of finding the optimal parameters that generate the highest accuracy. After 80 epochs of training with learning rate of 0.00001, the overall accuracy of the model was reached to 98.94%.
To assure that the CNN model was robust and stable, we also applied
k-fold cross validation for training and testing the CNN model. The results of different numbers of fold validation are reported in
Table 3, which illustrates the model robustness.
3.2. Prediction of Treatment Outcomes Based on EEG Functional Connectivity
In our experiment, the raw EEG signals were transformed into functional connectivity graphs with respect to the coherence measured between every pair of EEG channels as defined in Equation (1). The functional connectivity is a dynamic graph in which the nodes are the EEG channels, and the arcs demonstrate the squared coherence between the channels. The graph was updated with respect to the changes in EEG signal over time, thereby, generating time series information that can be used for the training of deep learning neural networks based on MLP.
Figure 8 visualises the functional connectivity graphs of 5 s EEG data related to two patients from the responder and non-responder classes. Here, we set the coherence visualisation threshold for the graph arcs to 0.75. This means that if the squared coherence between two EEG channels was greater than 0.75, then there was a connection between the two nodes in the functional connectivity graph.
Figure 8 shows functional connectivity graphs generated from 80 s of EEG data from non-responsive and responsive patients before and after treatment. This can be used to better understand the functional changes as a result of treatment. For instance, the functional connectivity increased to a wider area of the brain, including Fpz, AFz, Fz, FCz, AF3, and AF4) after treatment in the responder group. This increment, however, was not seen in the non-responder group.
We applied MLP as the model to predict classes (treatment outcomes) based on the EEG functional connectivity graphs, each with a size of 64*64 cells (representing pairwise correlations between 64 EEG channels). We applied 8:2 training and testing split for the classification reported in
Table 4. The total accuracy was 99.41% (the non-responder group accuracy was 99.28%, and the responder group accuracy was 99.50%).
4. Feature Selection for Identification of Tinnitus Predictive EEG Variables
The experiments for the prediction of tinnitus treatment were conducted using all 64 EEG channels. In order to reduce the dimensionality of the data and the computational model complexity, those EEG channels that were not significant for tinnitus diagnosis were detected and eliminated. We applied a variety feature selection method to determine the most important EEG variables affected by tinnitus treatment. Here, we developed a new approach, called greatest change channel selection (GCCS), for measuring the importance of the EEG channels in classifying the EEG samples to pre- and post-treatment states. This feature-selection method focused on finding the channels with the greatest changes caused by treatment. Equation (3) calculates
, which is the average degree of treatment effect on each EEG channel denoted by
.
where
M is the number of patients,
i is the
i-th channel, and
j is the
j-th patient.
is the pre-treatment amplitude of the patient
j in EEG channel
i.
is the post-treatment amplitude of the same patient
j in the same EEG channel
i. Here the range of I is [1–64] and j is [1–8].
For the GCCS feature selection method, we calculated the changes in the amplitude of each EEG channel before and after treatment. After calculating the average degree of treatment effect for all 64 channels in all patients, we identified the top 30 EEG channels that demonstrated the greatest average of across all patients. We also tested five other feature selection methods, including F_Regression (FR), random forest (RF), ExtraTrees (ET), and RFE. The feature selection methods are called model-based ranking. We first calculated the average amplitude on each EEG channel (across patients) before and after treatment and obtained two (before and after treatment) 64 × 8 matrices, in which 64 represents the number of EEG channels, and 8 represents the number of patients. Then, another average was taken across 8 patients on each channel, thus generating two 64 × 1 matrices, one represents the EEG status before the treatment (called before-matrix) and one represents the EEG status after the treatment (called after-matrix). We entered these two 64 × 1 matrices into the feature-selection methods as hereafter described.
For the FR method, the model returned the F-statistic and p-values as the criteria for ranking the importance of the channels. The larger the p-values, the higher the importance of the corresponding channel. This was calculated by measuring the correlation between the before-matrix data and the after-matrix data based on r-regression. Then, the cross correlation was converted to an F-score and then to a p-value.
The RF method ranked the importance of channels for each tree according to the impurity, which was calculated based on variance.
The ET method ranked the importance of the channels according to the impurity-based feature importance. The higher the rank, the more important the channel. The importance of a channel was computed as the (normalized) total changes in the value of the channels.
The RFE method determined the importance of each channel through the feature importance attribute returned by the model. Then, the least-important channels were removed from the current set of channels. This step was repeated recursively on the channel set until the required number of channels was finally reached.
Table 5 shows the top 30 EEG channels selected by the aforementioned feature selection methods as well as our proposed GCCS method.
Table 6 demonstrates that the GCCS method resulted in the highest prediction accuracy (99.47%), followed by F_Regression (99.39%) and RFE (99.09%).
5. Conclusions
AI diagnostic prediction is an extremely promising new field for the treatment of tinnitus. Due to the richness and variety of treatment methods for tinnitus and the variation in tinnitus treatment outcomes amongst individuals, the use of AI diagnosis and prediction may lead to a better treatment plan. This shortens the treatment time and allows patients to receive more targeted and personalized treatment. The goal of this study was to build artificial intelligence models to predict the outcome of tinnitus treatment. This research applied different neural networks (including CNN and MLP) to learn from patients’ EEG data for predicting their treatment outcome. The neural networks modelled EEG frequency features and functional connectivity features, and they resulted in up to 99% accuracy of for prediction of patients who were responders or non-responders to treatment. To the best of our knowledge, no other method for predicting the treatment outcome of tinnitus by analysing frequency features and functional connectivity has been published. Moreover, there is a lack of EEG studies investigating tinnitus treatment. Compared with previous works using time-domain features as the object of analysis [
40,
41,
42,
43], we use frequency-domain features and functional-connectivity features to identify signal changes or patterns with higher feasibility, rather than just observing a single time domain information. As time-domain features depend on the nature of the signal, using frequency-domain features and functional-connectivity features provides stronger discriminative power. Compared with works that only use the frequency-domain information of a single frequency band for analysis [
43], our work merges three frequency bands in order to obtain a low-dimensional representation, which enriches the features and reduces the computational cost of the model. Furthermore, we extracted correlations (functional connectivity features) between the signals of different sensors and found that these functional connectivity features can be linked to treatment outcomes in tinnitus patients.
So far, the highest accuracy of our predictions comes from the frequency-domain model, at 99.52%. At the same time, the functional connectivity-based model also performed very well, with an accuracy of 99.41%. In the process of model training, we used 70% of the data for training and the remaining 30% for network performance testing. We also conducted 5-fold cross-validation on the model, and the results were all above 90% accuracy. This suggest that our model is sufficiently robust. This is strong evidence that EEG signal analysis by AI models can reliably predict the outcome of tinnitus treatment. In addition, we applied five conventional feature-selection techniques (FR, RF, ET, and RFE) to identify the top predictive EEG variables that lead to increasing the prediction accuracy. We also proposed a new feature-selection approach, called GCCS, which resulted in the best prediction accuracy when compared to other methods. The GCCS method identified FC3, P8, P4, T8, and CP5 as the main predictive EEG variables.
This research also enabled visualisation of the topography of the human brain for frequency-domain features of the EEG, and visualisation of the functional connectivity of the human brain. The frequency-domain visualization (
Figure 5 and
Figure 6) allowed for investigation of the EEG amplitude changes in theta, alpha, and beta bands as a result of tinnitus treatment. According to the functional connectivity visualization, both the responder and non-responder groups have increased functional connectivity after receiving treatment, with the responder showing a greater increment in their connectivity in the cortical areas measured by F1, Fz, F2, FC1, FCz, FC2, C1, Cz, and C2 electrodes. Functional connectivity in non-responsive patients increased more significantly after treatment and was spread across all electrodes.
The research proved that within EEG data, both frequency and functional connectivity contain significant information showing brain changes as a response to treatment. Each of the frequency images and the functional connectivity graphs was generated using a 1 s EEG signal (256 time points), and they were used as inputs to deep learning and achieved a greater accuracy of prediction compared to using the whole time points of raw EEG signals.
For the future, a real-time prognostic digital health system is planned to be developed based on a small number of EEG variables (selected through feature selection) for the potential design of a wearable system for patients as home. To this end, a more robust AI model needs to be trained using more EEG data from tinnitus patients. This will allow us to identify a generalised group of EEG channels associated with tinnitus and the effect of treatment.
Author Contributions
M.D. contributed to the conceptualization, methodological design, algorithm development, analyzing the results, and writing the manuscript. X.L. contributed to the model and algorithm implementation, conducting the experiments, data modelling, visualization and writing the manuscript. Z.D. contributed to the methodological design, data pre-processing, reviewing the results, and writing the manuscript. Y.S. contributed to TFI analysis and reviewing the manuscript. G.S. contributed to the methodology, data collection and pre-processing, reviewing the results and writing the manuscript. P.S. contributed to data collection, data pre-processing and reviewing the manuscript. G.Y.W. contributed to reviewing the results and manuscript. A.S. contributed to reviewing the results and manuscript. W.Q.Y. contributed to reviewing the results and manuscript. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the University of Auckland Human Participants Ethics Committee (UAHPEC) (protocol code 019549 and date of approval 22 September 2017).
Informed Consent Statement
All subjects gave written informed consent in accordance with the Declaration of Helsinki.
Data Availability Statement
Data collection was approved by the University of Auckland Human Participants Ethics Committee.
Acknowledgments
The authors would like to acknowledgement the Faculty of Design and Creative Technology of Auckland University of Technology for the DCT fund and the Te Titoki Mataora Medtech research translator.
Conflicts of Interest
Grant D. Searchfield has a commercial interest in Tinnitus Tunes and TrueSilence which are tinnitus therapy companies.
References
- Watts, E.J.; Fackrell, K.; Smith, S.; Sheldrake, J.; Haider, H.; Hoare, D.J. Why Is tinnitus a problem? A qualitative analysis of problems reported by tinnitus patients. Trends Hear. 2018, 22, 2331216518812250. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Han, B.I.; Lee, H.W.; Kim, T.Y.; Lim, J.S.; Shin, K.S. Tinnitus: Characteristics, causes, mechanisms, and treatments. J. Clin. Neurol. 2009, 5, 11–19. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Hornstein, S.; Forman-Hoffman, V.; Nazander, A.; Ranta, K.; Hilbert, K. Predicting therapy outcome in a digital mental health intervention for depression and anxiety: A machine learning approach. Digital Health 2021, 7, 20552076211060659. [Google Scholar] [CrossRef] [PubMed]
- Fennell, P.G.; Zuo, Z.; Lerman, K. Predicting and explaining behavioral data with structured feature space decomposition. EPJ Data Sci. 2019, 8, 23. [Google Scholar] [CrossRef] [Green Version]
- Paxton, A.; Griffiths, T.L. Finding the traces of behavioral and cognitive processes in big data and naturally occurring datasets. Behav. Res. Methods 2017, 49, 1630–1638. [Google Scholar]
- Delgadillo, J.; Duhne, P.G.S. Targeted prescription of cognitive–behavioral therapy versus person-centered counseling for depression using a machine learning approach. J. Consult. Clin. Psychol. 2020, 88, 14–24. [Google Scholar] [CrossRef] [Green Version]
- Alzhrani, W.; Doborjeh, M.; Doborjeh, Z.; Kasabov, N. Emotion recognition and understanding using EEG data in a brain-inspired spiking neural network architecture. In Proceedings of the 2021 International Joint Conference on Neural Networks, Shenzhen, China, 18–22 July 2021; IEEE: New York, NY, USA, 2021; pp. 1–9. [Google Scholar]
- Doborjeh, M.; Doborjeh, Z.; Merkin, A.; Bahrami, H.; Sumich, A.; Krishnamurthi, R.; Medvedev, O.N.; Crook-Rumseyed, M.; Morgangh, C.; Kirkgh, I.; et al. Personalised predictive modelling with brain-inspired spiking neural networks of longitudinal MRI neuroimaging data and the case study of dementia. Neural Netw. 2021, 144, 522–539. [Google Scholar] [CrossRef]
- Yin, W.; Li, L.; Wu, F. Deep learning for brain disorder diagnosis based on fMRI images. Neurocomputing 2022, 469, 332–345. [Google Scholar] [CrossRef]
- Shoushtarian, M.; Alizadehsani, R.; Khosravi, A.; Acevedo, N.; McKay, C.M.; Nahavandi, S.; Fallon, J.B. Objective measurement of tinnitus using functional near-infrared spectroscopy and machine learning. PLoS ONE 2020, 15, e0241695. [Google Scholar] [CrossRef]
- Hirano, R.; Emura, T.; Nakata, O.; Nakashima, T.; Asai, M.; Kagitani-Shimono, K.; Kishima, H.; Hirata, M. Fully-automated spike detection and dipole analysis of epileptic MEG using deep learning. IEEE Trans. Med. Imaging 2022, 41, 2879–2890. [Google Scholar] [CrossRef]
- Hergès, H.O.; Vithal, R.; El-Merhi, A.; Naredi, S.; Staron, M.; Block, L. Machine learning analysis of heart rate variability to detect delayed cerebral ischemia in subarachnoid hemorrhage. Acta Neurol. Scand. 2022, 145, 151–159. [Google Scholar] [CrossRef]
- Sun, Z.-R.; Cai, Y.-X.; Wang, S.-J.; Wang, C.-D.; Zheng, Y.-Q.; Chen, Y.-H.; Chen, Y.-C. Multi-view intact space learning for tinnitus classification in resting state EEG. Neural Process. Lett. 2019, 49, 611–624. [Google Scholar] [CrossRef]
- Mohagheghian, F.; Makkiabadi, B.; Jalilvand, H.; Khajehpoor, H.; Samadzadehaghdam, N.; Eqlimi, E.; Deevband, M.R. Computer-aided tinnitus detection based on brain network analysis of EEG functional connectivity. J. Biomed. Phys. Eng. 2019, 9, 687–698. [Google Scholar] [CrossRef]
- Pereira, T.; Ferreira, F.L.; Cardoso, S.; Silva, D.; de Mendonça, M.G.A.; Madeira, S.C. Neuropsychological predictors of conversion from mild cognitive impairment to Alzheimer’s disease: A feature selection ensemble combining stability and predictability. BMC Med. Inform. Decis. Mak. 2018, 18, 137. [Google Scholar] [CrossRef] [Green Version]
- Liu, J.; Shi, M. A hybrid feature selection and ensemble approach to identify depressed users in online social media. Front. Psychol. 2021, 12, 802821. [Google Scholar] [CrossRef]
- Washington, P.; Paskov, K.M.; Kalantarian, H.; Stockham, N.; Voss, C.; Kline, A.; Patnaik, R.; Chrisman, B.; Varma, M.; Tariq, Q.; et al. Feature selection and dimension reduction of social autism data. In Proceedings of the Pacific Symposium ON Biocomputing 2020, Kohala Coast, HI, USA, 3–7 January 2020; pp. 707–718. [Google Scholar]
- Bamakana, S.M.H.; Gholami, P. A novel feature selection method based on an integrated data envelopment analysis and entropy model. Procedia Comput. Sci. 2014, 31, 632–638. [Google Scholar] [CrossRef] [Green Version]
- Bashivan, P.; Rish, I.; Yeasin, M.; Codella, N. Learning representations from EEG with deep recurrent-convolutional neural networks. arXiv 2015, arXiv:1511.06448. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 84–90. [Google Scholar] [CrossRef] [Green Version]
- Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
- Rabcan, J.; Levashenko, V.; Zaitseva, E.; Kvassay, M. Review of methods for EEG signal classification and development of new fuzzy classification-based approach. IEEE Access 2020, 8, 189720–189734. [Google Scholar] [CrossRef]
- Thangarajoo, R.G.; Reaz, M.B.I.; Srivastava, G.; Haque, F.; Ali, S.H.M.; Bakar, A.A.A.; Bhuiyan, M.A.S. Machine learning-based epileptic seizure detection methods using wavelet and EMD-based decomposition techniques: A review. Sensors 2021, 21, 8485. [Google Scholar] [CrossRef]
- Durai, M.; Doborjeh, Z.; Sanders, P.J.; Vajsakovic, D.; Wendt, A.; Searchfield, G.D. Behavioral outcomes and neural network modeling of a novel, putative, recategorization sound therapy. Brain Sci. 2021, 11, 554. [Google Scholar] [CrossRef]
- Meikle, M.B.; Henry, J.A.; Griest, S.E.; Stewart, B.J.; Abrams, H.B.; McArdle, R.; Myers, P.J.; Newman, C.W.; Sandridge, S.; Turk, D.C.; et al. The tinnitus functional index: Development of a new clinical measure for chronic, intrusive tinnitus. Ear Hear. 2012, 33, 153–176. [Google Scholar] [CrossRef]
- Chandra, N.; Chang, K.; Lee, A.; Shekhawat, G.S.; Searchfield, G.D. Psychometric validity, reliability, and responsiveness of the tinnitus functional index. J. Am. Acad. Audiol. 2018, 29, 609–625. [Google Scholar] [CrossRef]
- Searchfield, G.D.; Kobayashi, K.; Hodgson, S.-A.; Hodgson, C.; Tevoitdale, H.; Irving, S. Spatial masking: Development and testing of a new tinnitus assistive technology. Assist. Technol. 2016, 28, 115–125. [Google Scholar] [CrossRef]
- Spiegel, D.P.; Linford, T.; Thompson, B.; Petoe, M.A.; Kobayashi, K.; Stinear, C.M.; Searchfield, G.D. Multisensory attention training for treatment of tinnitus. Sci. Rep. 2015, 5, 10802. [Google Scholar]
- Sanders, P.; Doborjeh, Z.; Doborjeh, M.; Kasabov, N.; Searchfield, G. Prediction of acoustic residual inhibition of tinnitus using a brain-inspired spiking neural network model. Brain Sci. 2021, 11, 52. [Google Scholar] [CrossRef]
- Lovibond, P.F.; Lovibond, S.H. The structure of negative emotional states: Comparison of the depression anxiety stress scales (DASS) with the beck depression and anxiety inventories. Behav. Res. Ther. 1995, 33, 335–343. [Google Scholar] [CrossRef] [PubMed]
- Watson, D.; Clark, L.A.; Tellegen, A. Development and validation of brief measures of positive and negative affect: The PANAS scales. J. Personal. Soc. Psychol. 1988, 54, 1063. [Google Scholar] [CrossRef] [PubMed]
- Bertet, S.; Baskind, A.; Londero, A.; Bonfils, L.; Viaud-Delmon, I.; Warusfel, O. Design and evaluation of tinnitus synthesis methods: From spectral to spatial matching. Am. J. Otolaryngol. 2013, 34, 121–132. [Google Scholar] [CrossRef]
- Murugappan, M.; Murugappan, S. Human emotion recognition through short time Electroencephalogram (EEG) signals using Fast Fourier Transform (FFT). In Proceedings of the 2013 IEEE 9th International Colloquium on Signal Processing and its Applications, Kuala Lumpur, Malaysia, 8–10 March 2013; IEEE: New York, NY, USA, 2013; pp. 289–294. [Google Scholar]
- Alfeld, P. A trivariate clough—Tocher scheme for tetrahedral data. Comput. Aided Geom. Des. 1984, 1, 169–181. [Google Scholar] [CrossRef]
- Strypsteen, T.; Bertrand, A. End-to-end learnable EEG channel selection for deep neural networks with Gumbel-softmax. J. Neural Eng. 2021, 18, 0460a9. [Google Scholar] [CrossRef] [PubMed]
- Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef]
- Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Wachinger, C.; Reuter, M.; Klein, T. DeepNAT: Deep convolutional neural network for segmenting neuroanatomy. NeuroImage 2018, 170, 434–445. [Google Scholar] [CrossRef] [PubMed]
- He, K.; Hang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: New York, NY, USA, 2016; pp. 770–778. [Google Scholar]
- Widadi, R.; Zulherman, D.; Ari, S.R.F. Time domain features for eeg signal classification of four class motor imagery using artificial neural network. In Proceedings of the 1st International Conference on Electronics, Biomedical Engineering, and Health Informatics, Surabaya, Indonesia, 8–9 October 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 605–612. [Google Scholar]
- Diykh, M.; Li, Y.; Wen, P. EEG sleep stages classification based on time domain features and structural graph similarity. IEEE Trans. Neural Syst. Rehabil. Eng. 2016, 24, 1159–1168. [Google Scholar] [CrossRef] [PubMed]
- Yazıcı, M.; Ulutaş, M. Classification of EEG signals using time domain features. In Proceedings of the 2015 23rd Signal Processing and Communications Applications Conference (SIU), Malatya, Turkey, 16–19 May 2015; IEEE: New York, NY, USA, 2015; pp. 2358–2361. [Google Scholar]
- Wang, X.-W.; Nie, D.; Lu, B.-L. EEG-based emotion recognition using frequency domain features and support vector machines. In Proceedings of the International Conference on Neural Information Processing, Shanghai, China, 13–17 November 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 734–743. [Google Scholar]
Figure 1.
The protocol of the study: (a) seventeen chronic tinnitus patients undertook tinnitus treatment, and their multimodal data were collected. EEG data were collected from each patient at baseline (pre-treatment) and at three follow-up times. (b) The EEG features were transformed into frequency and functional connectivity domains and used as input data to the neural network algorithms. (c) Procedure for data modelling and predicting the treatment outcomes (responder and non-responder).
Figure 1.
The protocol of the study: (a) seventeen chronic tinnitus patients undertook tinnitus treatment, and their multimodal data were collected. EEG data were collected from each patient at baseline (pre-treatment) and at three follow-up times. (b) The EEG features were transformed into frequency and functional connectivity domains and used as input data to the neural network algorithms. (c) Procedure for data modelling and predicting the treatment outcomes (responder and non-responder).
Figure 2.
An example of one time-frame image of three frequency bands from one patient at baseline (a) and after treatment (b). These colour maps are the power topographies of each frequency band. The colours from blue to yellow represent powers from weak to strong.
Figure 2.
An example of one time-frame image of three frequency bands from one patient at baseline (a) and after treatment (b). These colour maps are the power topographies of each frequency band. The colours from blue to yellow represent powers from weak to strong.
Figure 3.
The proposed deep learning-based architecture for modelling of EEG data and the prediction of treatment outcomes using (a) dynamic frequency images via CNN, and (b) functional connectivity via MLP.
Figure 3.
The proposed deep learning-based architecture for modelling of EEG data and the prediction of treatment outcomes using (a) dynamic frequency images via CNN, and (b) functional connectivity via MLP.
Figure 4.
EEG functional connectivity maps were generated for four patients using EEG data from baseline (pre-treatment) and post-treatment. (
a) Maps from 2 patients from the responder group. (
b) Maps from 2 patients from the non- responder group. Nodes represent the EEG channels, while the lines represent the correlation between every two channels calculated using Equation (1).
Figure 5 and
Figure 6 demonstrate the frequency images (theta, alpha, and beta) from two randomly selected patients belonging to non-responder and responder groups, respectively, that were generated from 80 s of EEG. These frequency images provide an opportunity to further investigate the brain areas affected by the tinnitus therapy. For example,
Figure 5 shows that for a patient from the non-responder group, the power of alpha frequency was increasing over time across most of the brain regions before treatment, and a similar pattern was seen after treatment. On the other hand,
Figure 6 demonstrates that theta and alpha were shown to be more prominent after treatment in the responder patient. These patterns of changes were used as inputs for the training of deep neural networks to distinguish who is likely to respond to treatment.
Figure 4.
EEG functional connectivity maps were generated for four patients using EEG data from baseline (pre-treatment) and post-treatment. (
a) Maps from 2 patients from the responder group. (
b) Maps from 2 patients from the non- responder group. Nodes represent the EEG channels, while the lines represent the correlation between every two channels calculated using Equation (1).
Figure 5 and
Figure 6 demonstrate the frequency images (theta, alpha, and beta) from two randomly selected patients belonging to non-responder and responder groups, respectively, that were generated from 80 s of EEG. These frequency images provide an opportunity to further investigate the brain areas affected by the tinnitus therapy. For example,
Figure 5 shows that for a patient from the non-responder group, the power of alpha frequency was increasing over time across most of the brain regions before treatment, and a similar pattern was seen after treatment. On the other hand,
Figure 6 demonstrates that theta and alpha were shown to be more prominent after treatment in the responder patient. These patterns of changes were used as inputs for the training of deep neural networks to distinguish who is likely to respond to treatment.
Figure 5.
Visualization of the EEG frequency domain features from a randomly selected non-responsive patient before (a) and after (b) treatment.
Figure 5.
Visualization of the EEG frequency domain features from a randomly selected non-responsive patient before (a) and after (b) treatment.
Figure 6.
The visualization EEG frequency domain features from a randomly selected responsive patient before (a) and after (b) treatment.
Figure 6.
The visualization EEG frequency domain features from a randomly selected responsive patient before (a) and after (b) treatment.
Figure 7.
Accuracy and loss curves during the training.
Figure 7.
Accuracy and loss curves during the training.
Figure 8.
The functional connectivity graphs generated from 80 s of EEG data (averaged across patients) in the non-responder and responder groups, pre- and post-treatment.
Figure 8.
The functional connectivity graphs generated from 80 s of EEG data (averaged across patients) in the non-responder and responder groups, pre- and post-treatment.
Table 1.
The data features including cognitive, psychological, and TFI scores.
Table 1.
The data features including cognitive, psychological, and TFI scores.
Test | General | TSNS | DASS | PANAS |
---|
Features | - -
Intrusiveness - -
Sense of Control - -
Concentration - -
Sleep quality - -
Auditory - -
Relaxation - -
Quality of Life - -
Emotion
| - -
Overall (1 = Not a problem, 2 = Small, 3 = Moderate, 4 = Big, 5 = Very big) - -
Strong - -
Uncomfortable - -
Annoying - -
Ignore Unpleasant
| - -
Depression - -
Anxiety - -
Stress
| - -
Positive Affect - -
Negative Affect
|
Total score | 8 | 6 | 3 | 2 |
Table 2.
Confusion table for the prediction of the non-responder class and the responder class using the CNN model which is trained and tested by pre-treatment EEG frequency images to predict the outcome after the treatment (training and testing sample ratio is 8:2).
Table 2.
Confusion table for the prediction of the non-responder class and the responder class using the CNN model which is trained and tested by pre-treatment EEG frequency images to predict the outcome after the treatment (training and testing sample ratio is 8:2).
Prediction Label\True Label | Non-Responder | Responder | Accuracy | Specificity | Sensitivity |
---|
Non-responder | 533 | 9 | 99% | 98.9% | 98.34% |
Responder | 5 | 782 |
Table 3.
The accuracy of the prediction of non-responder and responder groups using CNN via k-fold cross validations.
Table 3.
The accuracy of the prediction of non-responder and responder groups using CNN via k-fold cross validations.
Folds | 5-Fold | 6-Fold | 7-Fold | 8-Fold |
---|
Accuracy | 99.37% | 99.52% | 98.13% | 98.79% |
Table 4.
The predictive accuracies using MLP classifies based on EEG functional connectivity features.
Table 4.
The predictive accuracies using MLP classifies based on EEG functional connectivity features.
Prediction Label\True Label | Non-Responder | Responder | Accuracy | Specificity | Sensitivity |
---|
Non-responder | 138 | 1 | 99.28% | 99.5% | 99.5% |
Responder | 1 | 202 | 99.50% |
Total | 139 | 203 | 99.41% |
Table 5.
Top 30 most-important EEG channels as selected by different feature selection methods, including F_Regression (FR), random forest (RF), ExtraTrees (ET), RFE, and our proposed GCCS method.
Table 5.
Top 30 most-important EEG channels as selected by different feature selection methods, including F_Regression (FR), random forest (RF), ExtraTrees (ET), RFE, and our proposed GCCS method.
Rank\Method | FR | ET | RFE | RF | GCCS |
---|
1 | F6 | F6 | Fp1 | Fp1 | FC3 |
2 | P2 | P2 | AF7 | F6 | P8 |
3 | FC5 | FT8 | F3 | TP7 | P4 |
4 | CP3 | T8 | F5 | AF8 | T8 |
5 | Fp2 | Fp2 | F7 | P2 | CP5 |
6 | Fp1 | FC5 | FT7 | CP3 | F3 |
7 | AF8 | Fp1 | FC5 | CPz | CPz |
8 | P8 | T7 | C5 | F3 | F6 |
9 | TP7 | CPz | T7 | FC1 | Fz |
10 | Oz | F2 | TP7 | Fpz | FT7 |
11 | CP6 | CP3 | CP5 | FC2 | Fpz |
12 | O2 | F1 | CP3 | P5 | O2 |
13 | O1 | O2 | CP1 | F4 | Poz |
14 | CPz | Pz | P7 | P8 | TP7 |
15 | CP1 | Fpz | P9 | Oz | FT8 |
16 | F4 | C2 | Iz | FT8 | FC4 |
17 | FC3 | P6 | Oz | PO4 | F4 |
18 | P5 | P5 | Pz | F1 | AF8 |
19 | F3 | C5 | CPz | FC5 | Oz |
20 | PO8 | CP6 | Fpz | AF3 | FC5 |
21 | C2 | F3 | AF8 | Fp2 | AF4 |
22 | TP8 | Fz | F2 | CP4 | C4 |
23 | P7 | AFz | F6 | AF4 | F1 |
24 | Fpz | AF4 | F8 | T7 | AF7 |
25 | C5 | FC3 | C6 | Fz | AF3 |
26 | PO7 | Iz | CP6 | P1 | T7 |
27 | FC6 | AF8 | P2 | C2 | C1 |
28 | PO3 | PO4 | P8 | Cz | CP6 |
29 | C6 | CP5 | P10 | F5 | P5 |
30 | F8 | FT7 | O2 | C3 | PO8 |
Table 6.
The prediction accuracy of CNN using smaller sets of EEG channels (top 10, top 20, and top 30) selected via our proposed GCCS method and compared with five other methods including RFE, F_Regression (FR), random forest (RF), ExtraTrees (ET), and RFE.
Table 6.
The prediction accuracy of CNN using smaller sets of EEG channels (top 10, top 20, and top 30) selected via our proposed GCCS method and compared with five other methods including RFE, F_Regression (FR), random forest (RF), ExtraTrees (ET), and RFE.
Feature Selection Technique\the Number of Channels | Top10 | Top20 | Top30 | Model Configuration |
---|
FR | 90.86 | 97.96 | 99.39 | Center = True Force_finite = True |
RF | 89.46 | 97.29 | 97.74 | max_features: Auto |
ET | 94.80 | 98.11 | 98.04 | n_estimators = 100 |
RFE | 93.52 | 96.99 | 99.09 | criterion: squared_error |
GCCS | 95.71 | 97.59 | 99.47 | Not Applicable |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).