Next Article in Journal
Ampullary Large-Cell Neuroendocrine Carcinoma, a Diagnostic Challenge of a Rare Aggressive Neoplasm: A Case Report and Literature Review
Next Article in Special Issue
Current Status of 68Ga-Pentixafor in Solid Tumours
Previous Article in Journal
Chlamydia trachomatis as a Current Health Problem: Challenges and Opportunities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism

Department of IS, College of Computer Science and Engineering, Taibah University, Madinah Al Munawara 43353, Saudi Arabia
Diagnostics 2022, 12(8), 1796; https://doi.org/10.3390/diagnostics12081796
Submission received: 5 June 2022 / Revised: 18 July 2022 / Accepted: 21 July 2022 / Published: 25 July 2022

Abstract

:
Parkinson’s disease (PD) is the second most common neurodegenerative disorder after Alzheimer’s disease. It has a slow progressing neurodegenerative disorder rate. PD patients have multiple motor and non-motor symptoms, including vocal impairment, which is one of the main symptoms. The identification of PD based on vocal disorders is at the forefront of research. In this paper, an experimental study is performed on an open source Kaggle PD speech dataset and novel comparative techniques were employed to identify PD. We proposed an unsupervised autoencoder feature selection technique, and passed the compressed features to supervised machine-learning (ML) algorithms. We also investigated the state-of-the-art deep learning 1D convolutional neural network (CNN-1D) for PD classification. In this study, the proposed algorithms are support vector machine, logistic regression, random forest, naïve Bayes, and CNN-1D. The classifier performance is evaluated in terms of accuracy score, precision, recall, and F1 score measure. The proposed 1D-CNN model shows the highest result of 0.927%, and logistic regression shows 0.922% on the benchmark dataset in terms of F1 measure. The major contribution of the proposed approach is that unsupervised neural network feature selection has not previously been investigated in Parkinson’s detection. Clinicians can use these techniques to analyze the symptoms presented by patients and, based on the results of the above algorithms, can diagnose the disease at an early stage, which will allow for improved future treatment and care.

1. Introduction

Parkinson’s disease (PD) is a slow progressing neurodegenerative disorder, causing impaired motor function with slow movements, tremors, and gait and balance disturbances. Various non-motor symptoms are common, and include disturbed autonomic function with orthostatic hypotension, constipation and urinary disturbances, sleep disorders, and neuropsychiatric symptoms. The onset of PD is insidious, with peak age of onset being 55–65 years. The disease has drastic effects on the lives of millions of people all over the world [1]. The progression rate of PD is slow; however, as it progresses, the affected person loses control over their movement, resulting in serious issues. The main cause is the loss of dopaminergic neurons in the substantia nigra and the decrease in the level of dopamine in the striatum. PD clinical diagnosis is possible from the following four cardinal motor symptoms: a tremor at rest, bradykinesia, postural instability, and rigidity. The diagnosis of PD based on symptoms becomes possible when almost 60% of dopaminergic neurons are already dead [2].
PD symptoms can vary from person to person due to its variety of symptoms, as exemplified by the multiple types of possible PD tremors, such as limb rigidity, tremors in the hands, and balance and gait problems. PD symptoms are generally classified into two categories: motor (movement-related) and non-motor (unrelated to movement). Patients experiencing non-motor symptoms face more severe conditions compared with those with motor symptoms. PD is ranked as the 14th leading cause of death in the United States according to the report of Central Disease Control and Prevention (CDC) [3], and is the second most common neurological disease after Alzheimer’s disease [4]. There is a dire need for a mechanism to detect and diagnose PD at an early stage, as the timely detection of PD facilitates rapid treatment and removes or lessens PD symptoms. Early-stage PD detection is a key element to slowing down its progression, and increasing the success rate of patient recovery.
However, researchers are still attempting to identify the most appropriate treatment for Parkinson’s disease [5]. Multiple diagnostic tests and various symptoms are used in combination with PD detection. The biomarkers are investigated for early detection to slow down the disease process. Different approaches were developed based on voice data for PD detection [6,7,8], force tracking data [9], and gait patterns. The current methods for evaluating Parkinson’s disease mostly rely on human efforts and expertise [10]. In the PD diagnosis literature, there is an immense body of work from researchers using speech measurement for general voice disorders and in PD [11,12,13]. In these studies, speech tests are recorded using a microphone, and then the recorded signals are analyzed using software algorithms to detect the properties of these signals.
As the median age of the population is increasing, the number of elderly people is rising. This scenario is contributing to the number of people affected by neurological disorders, and there is a growing need for fast and effective measures to monitor them. It is reported that nearly 90% of PD sufferers experience impairment of the vocal cords, which causes dysphonia; i.e., having an abnormal voice. Articulatory deficits can also manifest along with vocal impairment in the more severe stage [14]. Voice disorder detection at an early stage can play an important role in predicting the progression of the disease. The traditional sound measurement standards are shimmer, jitter, noise to harmonics ratio, pitch, and pressure level of absolute sound. Other measurements were also proposed that are inspired by non-linear dynamical systems theory. The tools recurrence density entropy (RPDE) and detrended fluctuation analysis are used for the detection of voice disorders [15].
The convolutional neural network (CNN) can extract features automatically and brought revolution, especially in the field of computer vision. The common applications of CNN in computer vision tasks are [16] image classification [17], object detection [18], image segmentation [19], medical domain [20], agriculture domain [21], scene understanding [22], and text classification.
This study aims to use the unsupervised autoencoder feature selection technique, and then pass the optimal features to supervised ML models. We also use a 1D convolutional neural network model to verify the deep learning solution for PD detection. The proposed approach, convolutional neural network-1D (CNN-1D), has better results when compared with the autoencoder feature selection technique.

1.1. Motivation

PD is a slow progressive neurodegenerative disease that affects almost 2% of people over the age of 65, and is the second most commonly found neurodegenerative disease after Alzheimer’s disease. PD affects almost 6.3 million people all over the world. PD patients lose their dopamine-producing neurons, which causes motor and non-motor symptoms. PD is characterized by motor and non-motor symptoms, and sufferers can exhibit the following; movement slowness, sidedness, tremors, vocal disorders, and jerkiness. There is still no effective cure and no means of slowing the progression of this fatal disease, and early detection is key to treatment.
The hypothesis of this research study was that PD could be detected in its early stages in a person by feeling the changes in the characteristics of speech as they talk. Such changes could be used to distinguish and classify people with PD disease and without PD disease.

1.2. Research Problem and Objective

Due to a rise in PD sufferers, the early detection of PD using an automated approach is becoming increasingly vital to improve the level of care for PD sufferers. The voice-based dataset contains important features for analysing PD. These type of data undergo many processes such as scaling, normalization, aggregation, and sampling to obtain an accurate prediction in the health care system. Previous researchers conducted a large number of studies to address and propose a solution for PD detection. The deep-learning-based solution brought about revolutions, and offers new opportunities for researchers to implement novel techniques to readily detect PD at an early stage. This study compares the performance of different deep-learning and machine-learning techniques. The aim of this paper is to address the following research questions:
RQ1: Are the autoencoder feature selection techniques providing better results compared to traditional ML algorithms [23]?
RQ2: Is the proposed dimensionality reduction technique providing better results than the recursive feature selection technique [24]?
RQ3: Is the proposed unsupervised autoencoder feature selection providing better results than the 1D-convolution neural network that has automatic feature selection?
The major contributions of this study are as follows:
  • The proposed CNN-1D is the first approach used for the classification of PD, and proves that the deep-learning model returns outstanding results in a large amount of data without feature selection;
  • We conducted a comparison between a statistical recursive feature elimination technique (RFE) and a deep-learning-based unsupervised autoencoder to select the optimal subset of features;
  • Our proposed approach performs analysis between different dimensionality reduction techniques and provides outstanding results.
The remaining portion of the paper is organized as follows. Section 2 reviews the literature, detailing previous studies of this domain. In Section 3, a brief description of the dataset, proposed methodology, and the proposed method can be found. Section 4 describes the results of this work and, finally, Section 5 describes the conclusion of the study and future direction.

2. Literature Review

PD is a fatal neurodegenerative disease, which is still impossible to diagnose in its early stages, where it could easily be treated, and the degeneration process could be slowed down. PD affects the quality of sufferers’ lives all over the world [25]. PD symptoms vary from individual to individual, as it has a variety of symptoms. Some patients experience tremors mainly at rest, such as tremors in their hands, irregular sequence of foot movement, and issues with balancing their body. Two types of symptoms are common in PD: motor and non-motor. Patients with non-motor symptoms are more affected than patients with motor symptoms. Non-motor symptoms include sleep disorders, depression, cognitive impairment, and loss of sense of smell. It is reported in the clinical literature that one of the most common symptoms of PD is vocal cord impairment, resulting in vocal loudness, frequency abnormality, and instability. Vocal impairments and voice breaks are the key factors that provide information about PD patients. Speech processing techniques are used to find the anomalies to detect PD. In 2015, the Movement Disorder Society published their research on a clinical diagnostic criterion for PD [26]. This criterion proved helpful in excluding the possibility of any other diseases that may have a similar set of symptoms, and requires the patient to have a minimum of two clinical symptoms from the given set of symptoms for a positive diagnosis of PD. The neurologist assessment for PD is a subjective assessment based on patient interview and some physical tests. The aforementioned criterion does not have a standard format of test sequence and questionnaire. In 1980, the Unified PD Rating Scale (UPDRS) was developed, which was later modified in 2008 [27]. Research studies extensively use UPDRS-based symptoms for categorizing PD [28,29].
Feature selection is a technique to identify the features best suited for the model from a large set of possible features. Feature selection reduces the computational requirements and enhances the generalization of the ML method. Three feature selection methods, (i) sequential forward feature selection (SFS), (ii) sequential backward feature selection (SBS) [30], and (iii) minimum redundancy maximum relevance (mRMR) [31], are used in this research, and their performance is compared. The maximum relevance feature selection technique selects features based on maximum mutual information by adding additional features in the subset. However, the selected subset may contain redundant data. The minimum redundancy maximum relevancy (mRMR) optimizes the relevance and reduces the redundancy of the selected features. In the mRMR feature selection technique, two features are compared based on mutual information, and the one that affects the result the most (maximum relevance) is selected, yet they have a different minimum redundancy. The mRMR has been used extensively in the classification and detection of early PD from voice recordings [32], classification of pedestrian information, and identifying physical activities [33]. The mRMR could also be suitable for assessing gait features of PD patients.
The past decade witnessed a phenomenal growth in the literature in the field application of ML in medical diagnosis, such as the diagnosis of PD, using different features such as vocal impairment. The following table lists some of the studies where ML was used to diagnose PD. Table 1 provides a summary review of the literature.

3. Methodology

The proposed methodology is based on the following main steps: data pre-processing, dimensionality reduction, classifiers training on training data, and classifiers evaluation on test data. To the best of our knowledge, the research scope of the proposed study makes it the first study to use a non-linear dimensionality reduction technique autoencoder in Parkinson’s disease detection. The autoencoder neatly compressed the features and sends striking performance results to the ML classifiers. We explored the proposed technique with reference to the linear dimensionality reduction techniques, and conclude that the proposed technique provides outstanding results in term of accuracy and F1 measure.
The overall proposed methodology can be illustrated as:

3.1. Data Set

In this research, a vocal-based dataset was used that contains healthy and affected individuals’ vocal recording instances. The dataset for this study was taken from California University, Irvine ML dataset library. This dataset has records for 81 female and 107 male PD patients (i.e., 188 patients altogether). The healthy group has 64 samples (23 male and 41 female), with an age range of 41 to 82 years old. The final version of the dataset contains 754 attributes with 756 records [44,45]. Table 2 provides dataset description.
We also analyzed the class distribution of the dataset, which was skewed and means our dataset was imbalanced. The classification accuracy tends towards the majority class. To address the issue of the imbalanced dataset, we used an oversampling technique, in which class distribution 0 and 1 are equal. Oversampling increases the distribution in the minority class and makes an equal distribution of both classes. Figure 1 shows the class distribution of the target value without oversampling.
The dataset has a split ratio of 30:70, where 30% is the test data while 70% is the training. Training data is used for models training, while test data is used for model evaluation. We used anaconda distribution environment (Jupyter) for the implementation of the proposed models and pre-processing.

3.2. Proposed Models

In this study, two different techniques for Parkinson’s disease detection at an early stage were used. We used comparative analysis of the unsupervised neural network feature selection autoencoder that compressed the input features, and then passed them to supervised ML algorithms for classification, as ML algorithms are not suitable for automatic feature engineering, while deep-learning algorithms are capable of automatic feature selection. We also used Conv1D for the early detection of PD.

3.2.1. Autoencoder

An autoencoder is an unsupervised dimensionality reduction neural network technique, where the neural network is powered to learn the representations. An autoencoder has two main parts: the encoder and the decoder. The encoder encodes the input features into a lower dimensional space, called latent vector, and the decoder decodes the latent vector into its original form. The loss function is used to minimize the difference between input features and output features, (see Figure 2).
In this research, a sparse autoencoder was used with a single hidden layer representation, Table 3 lists autoencoder parameters. The input features were encoded under this single hidden layer, and effort was made to minimize the error and generate a flattened representation with optimal score [39]. The sparse autoencoder is an autoencoder model with the addition of a sparse penalty term and adds more constraints for the extraction of the latent vector and enhanced feature learning. In the network, the activation function sigmoid was used. In order to reduce the activation functions value in the hidden layer neuron, a sparse penalty was embedded [40]. Due to this sparsity, some undesired activation in the hidden layer is possible. The activation function can be represented as:
a = s i g W x + b
where:
W = weight matrix.
b = deviation vector.
The neuron activation function in the hidden layer can be represented as follows:
p j = 1 n i = 1 n a j x i
(KL) divergence method is used as a penalty term. KL divergence can be represented as follows:
k l ( p | | p j ) = pln p p j + 1 p l n 1 p 1 r h o j
The deviation causes a gradual increase in the value of Kullback–Leibler (KL), and it becomes zero as it deviates from p. The loss value of this network is represented by C(w, b). The following equation shows the loss function with the addition of the sparse penalty term:
C sparse   =   C w ,   b + β j = 1 S 2 kl ( p | | p j )
where β is the weight of the sparse penalty term and S2 is the neuron count in the internal layer [38].

3.2.2. Logistic Regression Model

Logistic regression is a ML algorithm that can only solve the classification tasks using the probability value 0.5 to predict the target variable. It uses the complex cost function, known as sigmoid function or logistic function. Logistic regression analysis has become a powerful predictor in health care research over the last few decades [47]. It is mostly used where the occurrence of a binary outcome is predicted for one or more variables [48]. The useful property of logistic regression is that its output always lies between 0 and 1. The basis of a logistic regression is the sigmoid function [49] that is represented as:
f x = e x 1 + e x = 1 1 + e z
where:
e = Euler number,
x =Value of independent variable X.

3.2.3. Support Vector Machine (SVM)

This is a supervised ML technique that is widely employed for classification, pattern recognition, and regression analysis. In SVM, the training data are treated as data points in space, where different classes are divided by a hyper-plane that is far from the nearest data point. New input features are treated as training points, and classified according to their category as they fall into the hyper-plane. In cases where the data points are not linearly separable, kernel tricks with different possible kernel functions, such as polynomial or radial basis functions, are used. These kernels are helpful to map the high dimensionality feature space, and determine a possible high dimensional hyper-plane. The advantage of the support vector machine is that it is effective for high dimensional spaces in which data points are divided into subsets of training points by creating a decision boundary. SVM is versatile because of various possible kernel functions. SVM has the demerit that it does not have probability estimates for classification problems, and regularization terms are not necessary to avoid over-fitting. The SVM model diagram shows the margin between predictor and target variables.

3.2.4. Naïve Bayes Model (NBM)

This model is a supervised ML approach that can solve classification tasks. The naïve Bayes model works based on conditional probability, by using learning probability of certain features belonging to a particular group. The naïve Bayes model is also known as “Naïve” as it assumes that the occurrence of a particular feature is independent of another feature occurrence. The naïve Bayes theorem finds the probability of event occurrences that have already occurred. The naïve Bayes model maps a series of input features X = [x1, x2, ..., xn] to a series of probability value of Y = [y1, y2, y3, ..., yn]. For a particular set of observations X = (x1, x2, ..., xn), we determined those in which Y belongs to yi, and made the classification of the particular object. We must find the highest value of yi.
The mathematical representation of a naïve Bayes theory is given below:
P A | B = P A     P B | A P B

3.2.5. Random Forest Model (RFM)

This model is an ensemble technique that can solve classification, as well as regression, tasks. RFM uses a modified tree-learning algorithm during the learning process, selecting a random subset of the features. The technique only determines a random subset of variables to find the best split at each node. The input vector is fed to each tree in the RFM for the classification task, and each tree votes for a class. The RFM selects the class that has the highest number of votes. The RFM can handle large input datasets compared to other models. The random forest model overcomes the problem of over-fitting by averaging or combining the results of different decision trees. RFM has the same hypermeters as in the decision tree model or bagging models.

3.2.6. Convolutional Neural Network-1D

Convolutional neural network comes under the umbrella of deep learning. Deep learning is a subfield of ML, and the algorithms of deep learning have the automatic feature selection and good performance over a large dataset. The deep-learning model performs well in speech, language translation, and object detection. Conv2D is used for 2D data such as images and video. CNN-1D is used for 1D and sequential type datasets. In the proposed study, two layers of Conv1D with batch normalization are used. Each layer also has the dropout function to overcome the over-fitting. Layers 1 and 2 also have a 1D max-pooling layer that contains only high feature values. The third is a fully connected layer and the fourth is the output layer. Sigmoid activation function was used because the target attribute is a binary classification task. The number of iterations is 20, with a learning rate of 0.0001. The CNN-1D architecture is given below in Figure 3.

4. Results and Discussion

4.1. Evaluation Metrics

To verify the performance of the proposed algorithms, we used the following classification metrics:
Accuracy: It is the simple ratio between the numbers of correctly classified samples out of total samples. The formula of accuracy score is:
Accuracy : TN + TP TN + FP + FN + TP × 100 %
Precision: It is the fraction of the correctly classified samples from the total classified samples. The precision formula is:
Precision :   TP TP + FP × 100 %
Recall: It is the fraction of the correctly classified samples from the total classified samples. The formula of recall is:
Recall :   TP TP + FN × 100 %
F1 Score: It is the harmonic mean of precision and recall. It can be expressed by the following formula:
F 1 Score = 2 × precision × recall precision + recall

4.2. Discussion

RQ1: Are the autoencoder feature selection techniques providing better results compared to traditional ML algorithms [23]?
To verify the proposed technique, we made comparisons with traditional ML classifiers with dimensionality reduction (autoencoder) and without dimensionality reduction techniques. The experimental study proves that logistic regression and support vector machine models with the dimensionality reduction technique provide better results in terms of accuracy score, (see Table 4).
RQ2: Does the proposed dimensionality reduction technique provide better results than the recursive feature selection technique [24]?
The basic aim of dimensionality reduction and feature selection is to reduce the number of features from the original set of features. The feature selection is the subset of the original dataset, while dimensionality reduction transforms the features into lower dimensions. Dimensionality reduction and feature selection have various techniques. In this study, we evaluate the comparisons of the autoencoder with the study [50] that used the recursive feature elimination technique. We found that the proposed model with dimensionality reduction provides better results compared to the feature selection technique RFE, see Table 5.
RQ3: Does the proposed unsupervised autoencoder feature selection provide better results than the 1D-convolution neural network that has automatic feature selection?
In this paper, we also investigate the power of the deep-learning model compared with dimensionality reduction ML models. In our experimental study, we conclude that the 1D-CNN model shows good performance compared to manual dimensionality reduction of supervised ML models (see Table 6).
We examined the comparisons between linear dimensionality reduction techniques (PCA and LDA) with non-linear dimensionality reduction techniques (autoencoder). We found that the autoencoder provides the optimal features to the ML model, and obtains higher results in terms of accuracy and F1 measure. The overall comparison of dimensionality reduction is given below in Table 7.
The results of the autoencoder with supervised ML are analyzed in an ROC curve. The ROC curve analysis shows (Figure 4) that the random forest model has the highest AUC value, 0.929%.
The results are analyzed in terms of a confusion matrix (Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9). The confusion matrix provides information about the correctly classified and misclassified number of instances.
Caution should be applied in concluding that our approach is perfect for PD detection at early stage, as this study was conducted using only a single dataset. We also did not analyze the various data balancing approaches to avoid misclassification of the results. The classification accuracy could be further improved in future studies by using different hyper-parameter techniques.

5. Conclusions

Initial stage of PD detection is a very important step in gaining a better understanding of the causes of the disease, thus, ensuring the correct measures are implemented and therapeutic interventions are initiated. This study is an effort to outline a broad review of the diagnosis system of PD, by applying various robust ML techniques. The different ML dimensionality reduction techniques are evaluated to achieve optimal features, and comparisons are made to obtain the best performance on the benchmark dataset. In this study, we performed analysis of PD in four different ways: (a) evaluation of PD detection without any feature selection technique; (b) analysis of four different machine-learning-based methods on three different dimensional reduction techniques; (c) we also performed analysis of recursive feature elimination technique with dimensionality reduction techniques; and (d) in last section of our experimental approach, we also conducted our evaluation on 1D-CNN with traditional machine-learning models. We conclude that the 1D-CNN model shows better results compared to the manual feature selection technique; the 1D-CNN model shows the highest accuracy of 0.85%, and logistic regression with autoencoder dimensionality reduction shows the highest accuracy of 0.75%. Based on this experimental evaluation, we can say that deep-learning-based 1D-CNN shows better performance as compared to manual-feature-selection-techniques-based machine-learning models. The reason behind this is that the deep-learning model performs better with a large number of features, and has the capability of an automatic feature selection. Our technique has some limitations: we conducted analysis on only one dataset, and we do not consider the dataset balancing methods. In future studies, we are interested in exploring the ensemble and boosting techniques with different feature selection techniques on more than one PD dataset. The outcome of this work can be viewed as a promising first step towards the application of cutting-edge research for the early detection of PD.

Funding

This research received no external funding.

Conflicts of Interest

The author declare no conflict of interest.

References

  1. Fahn, S. Description of Parkinson’s disease as a clinical syndrome. Ann. N. Y. Acad. Sci. 2003, 991, 1–14. [Google Scholar] [CrossRef] [PubMed]
  2. Cummings, J.L.; Henchcliffe, C.; Schaier, S.; Simuni, T.; Waxman, A.; Kemp, P. The role of dopaminergic imaging in patients with symptoms of dopaminergic system neurodegeneration. Brain 2011, 134, 3146–3166. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Lang, A.E.; Lozano, A.M. Parkinson’s disease First of two parts. N. Engl. J. Med. 1998, 339, 1044–1053. [Google Scholar] [CrossRef] [PubMed]
  4. De Rijk, M.D.; Launer, L.J.; Berger, K.; Breteler, M.M.; Dartigues, J.F.; Baldereschi, M.; Hofman, A. Prevalence of Parkinson’s disease in Europe: A collaborative study of population-based cohorts. Neurologic Diseases in the Elderly Research Group. Neurology 2020, 54, S21–S23. [Google Scholar]
  5. Singh, N.; Pillay, V.; Choonara, Y.E. Advances in the treatment of Parkinson′s disease. Prog. Neurobiol. 2007, 81, 29–44. [Google Scholar] [CrossRef] [PubMed]
  6. Gunduz, H. Deep Learning-Based Parkinson’s Disease Classification Using Vocal Feature Sets. IEEE Access 2019, 7, 115540–115551. [Google Scholar] [CrossRef]
  7. Lahmiri, S.; Shmuel, A. Detection of Parkinson’s disease based on voice patterns ranking and optimized support vector machine. Biomed. Signal Process. Control 2019, 49, 427–433. [Google Scholar] [CrossRef]
  8. Braga, D.; Madureira, A.M.; Coelho, L.; Ajith, R. Automatic detection of Parkinson’s disease based on acoustic analysis of speech. Eng. Appl. Artif. Intell. 2018, 77, 148–158. [Google Scholar] [CrossRef]
  9. Bilgin, S. The impact of feature extraction for the classification of amyotrophic lateral sclerosis among neurodegenerative diseases and healthy subjects. Biomed. Signal Process. Control 2017, 31, 288–294. [Google Scholar] [CrossRef]
  10. Zhao, A.; Qi, L.; Li, J.; Dong, J.; Yu, H. A hybrid spatio-temporal model for detection and severity rating of Parkinson’s disease from gait data. Neurocomputing 2018, 315, 1–8. [Google Scholar] [CrossRef] [Green Version]
  11. Cho, C.-W.; Chao, W.-H.; Lin, S.-H.; Chen, Y.-Y. A vision-based analysis system for gait recognition in patients with Parkinson’s disease. Expert Syst. Appl. 2009, 36, 7033–7039. [Google Scholar] [CrossRef]
  12. Betarbet, R.; Sherer, T.B.; Greenamyre, J.T. Animal models of Parkinson’s disease. Bioessays 2002, 24, 308–318. [Google Scholar] [CrossRef] [PubMed]
  13. Boyanov, B.; Hadjitodorov, S. Acoustic analysis of pathological voices. IEEE Eng. Med. Biol. Mag. 1997, 16, 74–82. [Google Scholar] [CrossRef]
  14. Rahn, D.A.; Chou, M.; Jiang, J.J.; Zhang, Y. Phonatory Impairment in Parkinson’s Disease: Evidence from Nonlinear Dynamic Analysis and Perturbation Analysis. J. Voice 2007, 21, 64–71. [Google Scholar] [CrossRef] [PubMed]
  15. Prashanth, R.; Roy, S.D. Early detection of Parkinson’s disease through patient questionnaire and predictive model-ling. Int. J. Med. Informat. 2018, 199, 75–87. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Kumari, D.; Chokkalingam, S.P. Deep convolutional neural networks (CNN) for medical image analysis. Int. J. Eng. Adv. Technol. (IJEAT) 2019. [Google Scholar] [CrossRef]
  17. Song, Y.; Yu, Z.; Zhou, T.; Teoh, J.Y.C.; Lei, B.; Choi, K.S.; Qin, J. CNN in CT Image Segmentation: Beyond Loss Function for Exploiting Ground Truth Images. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 325–328. [Google Scholar] [CrossRef]
  18. Francis, M.; Deisy, C. Disease Detection and Classification in Agricultural Plants Using Convolutional Neural Networks—A Visual Understanding. In Proceedings of the 2019 6th International Conference on Signal Processing and Integrated Networks (SPIN), Noida, India, 7–8 March 2019; pp. 1063–1068. [Google Scholar] [CrossRef]
  19. Sarvamangala, D.R.; Kulkarni, R.V. Convolutional neural networks in medical image understanding: A survey. Evol. Intell. 2021, 15, 1–22. [Google Scholar] [CrossRef]
  20. Alaskar, H. Convolutional neural network application in biomedical signals. J. Comput. Sci. Inform. Tech. 2018, 6, 45–59. [Google Scholar] [CrossRef]
  21. Arashpour, M.; Ngo, T.; Li, H. Scene understanding in construction and buildings using image processing methods: A comprehensive review and a case study. J. Build. Eng. 2021, 33, 101672. [Google Scholar] [CrossRef]
  22. Jang, B.; Kim, M.; Harerimana, G.; Kang, S.U.; Kim, J.W. Bi-LSTM model to increase accuracy in text classification: Combining Word2vec CNN and at-tention mechanism. Appl. Sci. 2020, 10, 5841. [Google Scholar] [CrossRef]
  23. Wang, W.; Lee, J.; Harrou, F.; Sun, Y. Early Detection of Parkinson’s disease Using Deep Learning and ML. IEEE Access 2016, 4, 147635–147646. [Google Scholar]
  24. Nissar, I.; Rizvi, D.R.; Masood, S.; Mir, A.N. Voice-Based Detection of Parkinson’s disease through Ensemble ML Approach: A Performance Study, 23 August. EAI Endorsed Trans. Pervasive Health Technol. 2019, 5, e2. [Google Scholar]
  25. Postuma, R.B.; Berg, D.; Stern, M.; Poewe, W.; Olanow, C.W.; Oertel, W.; Deuschl, G. MDS clinical diagnostic criteria for Parkinson’s disease. Mov. Disord. 2015, 30, 1591–1601. [Google Scholar] [CrossRef]
  26. Movement Disorder Society. The Unified Parkinson’s Disease Rating Scale (UParkinson’s diseaseRS): Status and rec-ommendations. Mov. Disord. Off. J. Mov. Disord. Soc. 2003, 18, 738–750. [Google Scholar] [CrossRef] [PubMed]
  27. Eskofier, B.M.; Lee, S.I.; Daneault, J.F.; Golabchi, F.N.; Ferreira-Carvalho, G.; Vergara-Diaz, G.; Bonato, P. Recent ML advancements in sensor-based mobility analysis: Deep learning for Parkinson’s disease assessment. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 655–658. [Google Scholar]
  28. Arora, S.; Venkataraman, V.; Donohue, S.; Biglan, K.M.; Dorsey, E.R.; Little, M.A. High accuracy discrimination of Parkinson’s disease participants from healthy controls using smartphones. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 3641–3644. [Google Scholar]
  29. Aha, D.W.; Bankert, R.L. Comparative Evaluation of Sequential Feature Selection Algorithms. In Learning from Data; Springer: Cham, Switzerland, 1996; pp. 199–206. [Google Scholar]
  30. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevanceand min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  31. Chen, H.L.; Wang, G.; Ma, C.; Cai, Z.N.; Liu, W.B.; Wang, S.J. An efficient hybrid kernel extreme learning machine approach for early diagnosis of Parkinson’s disease. Nero Comput. 2016, 184, 131–144. [Google Scholar]
  32. Zaki, M.H.; Sayed, T. Using automated walking gait analysis for the identification of pedestrian attributes. Transp. Res. C Emerg. Technol. 2014, 48, 16–36. [Google Scholar] [CrossRef]
  33. Jatoba, L.C.; Grossmann, U.; Kunze, C.; Ottenbacher, J.; Stork, W. Context-aware mobile health monitoring: Evaluation of different pattern recognition methods for classification of physical activity. In Proceedings of the 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Vancouver, BC, Canada, 20–25 August 2008; pp. 5250–5253. [Google Scholar]
  34. Tsanas, A.; Little, M.A.; McSharry, P.E.; Spielman, J.; Ramig, L.O. Novel speech signal processing algorithms for high-accuracy classification of Parkinson′s disease. IEEE Trans. Biomed. Eng. 2012, 59, 1264–1271. [Google Scholar] [CrossRef] [Green Version]
  35. Rouzbahani, H.K.; Daliri, M.R. Diagnosis of Parkinson’s disease in human using voice signals. Basic Clin. Neurosci. 2011, 2, 12. [Google Scholar]
  36. Parisi, L.; Chandran, N.; Manaog, M. Feature-driven ML to improve early diagnosis of Parkinson’s disease. Expert Syst. Appl. 2018, 110, 182–190. [Google Scholar] [CrossRef]
  37. Caliskan, A.; Badem, H.; Basturk, A.; Yuksel, M. Diagnosis of the parkinson disease by using deep neural network classifier. IU-J. Electr. Electron. Eng. 2017, 17, 3311–3318. [Google Scholar]
  38. Wroge, T.J.; Özkanca, Y.; Demiroglu, C.; Si, D.; Atkins, D.C.; Ghomi, R.H. Parkinson’s disease diagnosis using machine learning and voice. In Proceedings of the 2018 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), IEEE, Philadelphia, PA, USA, 1 December 2018. [Google Scholar]
  39. Rathore, V.S.; Worring, M.; Mishra, D.K.; Joshi, A.; Maheshwari, S. Parkinson Disease Prediction Using ML Algorithm. In Emerging Trends in Expert Applications and Security; Springer: Singapore, 2019; pp. 357–363. [Google Scholar]
  40. Yasar, A.; Saritas, I.; Sahman, M.A.; Cinar, A.C. Classification of Parkinson Disease Data with Artificial Neural Networks. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2019; Volume 675. [Google Scholar]
  41. Oh, S.L.; Hagiwara, Y.; Raghavendra, U.; Yuvaraj, R.; Arunkumar, N.; Murugappan, M.; Acharya, U.R. A deep learning approach for Parkinson’s disease diagnosis from EEG signals. Neural. Comput. Appl. 2020, 32, 10927–10933. [Google Scholar] [CrossRef]
  42. Silveira-Moriyama, L.; Petrie, A.; Williams, D.R.; Evans, A.; Katzenschlager, R.; Barbosa, E.R.; Lees, A.J. The use of a color coded probability scale to interpret smell tests in sus-pected parkinsonism. Mov. Dis. 2009, 24, 1144–1153. [Google Scholar] [CrossRef] [PubMed]
  43. Chang, D.; Alban-Hidalgo, M.; Hsu, K. Diagnosing Parkinson’s Disease from Gait. In Stanford; NCPI: Bethesda, MD, USA, 2015. [Google Scholar]
  44. Ng, A. Sparse autoencoder. CS294A Lect. Notes 2011, 72, 1–19. [Google Scholar]
  45. Makhzani, A.; Frey, B. K-sparse autoencoders. arXiv 2013, arXiv:1312.5663. [Google Scholar]
  46. Ahed, F. Image Compression Using Autoencoders in Keras. Available online: https://blog.paperspace.com/autoencoder-image-compression-keras/ (accessed on 11 January 2022).
  47. Zhang, C.; Cheng, X.; Liu, J.; He, J.; Liu, G. Deep sparse autoencoder for feature extraction and diagnosis of locomotive adhesion status. J. Control Sci. Eng. 2018, 2018, 8676387. [Google Scholar] [CrossRef]
  48. Oommen, T.; Baise, L.G.; Vogel, R.M. Sampling Bias and Class Imbalance in Maximum-Likelihood Logistic Regression. Math. Geosci. 2011, 43, 99–120. [Google Scholar] [CrossRef]
  49. Cramer, J.S. The Origins of Logistic Regression (December 2002). Tinbergen Institute Working Paper No. 2002-119/4. Available online: https://ssrn.com/abstract=360300 (accessed on 11 January 2022).
  50. Ghosh, J.; Li, Y.; Mitra, R. On the Use of Cauchy Prior Distributions for Bayesian Logistic Regression. Bayesian Anal. 2018, 13, 359–383. [Google Scholar] [CrossRef]
Figure 1. Flow chart of PD detection.
Figure 1. Flow chart of PD detection.
Diagnostics 12 01796 g001
Figure 2. Structural diagram of autoencoder model, adopted from [46].
Figure 2. Structural diagram of autoencoder model, adopted from [46].
Diagnostics 12 01796 g002
Figure 3. 1D-CNN architecture.
Figure 3. 1D-CNN architecture.
Diagnostics 12 01796 g003
Figure 4. Proposed model with autoencoder ROC curve.
Figure 4. Proposed model with autoencoder ROC curve.
Diagnostics 12 01796 g004
Figure 5. Logistic regression.
Figure 5. Logistic regression.
Diagnostics 12 01796 g005
Figure 6. Support vector machine.
Figure 6. Support vector machine.
Diagnostics 12 01796 g006
Figure 7. Random forest.
Figure 7. Random forest.
Diagnostics 12 01796 g007
Figure 8. Naïve Bayes.
Figure 8. Naïve Bayes.
Diagnostics 12 01796 g008
Figure 9. 1D-CNN.
Figure 9. 1D-CNN.
Diagnostics 12 01796 g009
Table 1. Review of the literature.
Table 1. Review of the literature.
StudyTechniques UsedRemarks
Tsanas and Athanasios, et al. [34]Relief and local learning-based feature selection (LLBFS), minimum redundancy maximum relevance (mRMR), and least absolute shrinkage and selection operatorFeatures such as HNR, shimmer, and expiation of vocal fold produces 98.6% precision rate,
while LLBFS produces the feature set with the lowest classification error.
Rouzbahani et al. [35]Fisher’s discriminate ratio, correlation rates, and t-test.Highest accuracy rate of 94% observed in kNN.
Parisi et al. [36]Multi-layer perceptron (MLP) with custom cost function and Lagrangian support vector machine (LSVM) are used for classification.The proposed algorithm achieves 100% of accuracy rate.
Abdullah et al. [37]DNN classifier with softmax layer and stacked-auto-encoder (SAE). Several different datasets were used to test the proposed model.
DNN classifier is found to be the most suitable classifier for the early diagnosis of the disease.
Timothy et al. [38]DNN classifier.
Minimum redundancy maximum relevance and MFCC used to extract the features.
The study reports 85% accuracy for DNN model.
Mathur et al. [39]kNN, Adaboost.The study reports 91.28% classification accuracy for kNN and Adaboost
Yasar et al. [40]Artificial neural networks (ANN).The study reports that the proposed model achieves 94.93% accuracy in identifying diseased individuals.
ShuLih et al. [41]Thirteen-layer CNN architecture was used on a dataset of EEG signals (20 normal subjects and 20 PD sufferers).88.25% accuracy reported for the proposed CNN architecture.
Laura et al. [42]Logistic regression model on smell identification and Sniffin’ Sticks test.Model shows 82.8% accuracy for smell identification and 85.3% accuracy for Sniffin’ Sticks test.
Daryl et al. [43] SVM and random forest algorithm.SVM shows AUC of 92.3% and an accuracy of 85.3%. A random forest achieves 76.3% AUC with 75.6% accuracy.
Table 2. Dataset description.
Table 2. Dataset description.
DetailSource Information
Dataset propertyUCI ML Repository
Dataset namePD
Dataset attributes754
Dataset records756
Target variable(0: control, 1: PD). Binary class problem
TaskBinary classification
Table 3. Sparse autoencoder parameters.
Table 3. Sparse autoencoder parameters.
ParametersValues
No. of epochs200
Weight decay20−5
Optimization techniqueLbfgs
Sparse penalty weight3
Sparsity0.1
Table 4. Classifiers with dimensionality reduction and without dimensionality reduction.
Table 4. Classifiers with dimensionality reduction and without dimensionality reduction.
ClassifiersClassifiers with AutoencoderML without Dimensionality Reduction Technique
Logistic regression0.8750.865
Support vector machine0.8630.842
Random forest0.8280.814
Naïve Bayes0.8360.711
Table 5. Classifier with dimensionality reduction and feature selection technique.
Table 5. Classifier with dimensionality reduction and feature selection technique.
ClassifiersClassifiers with AutoencoderClassifiers with RFE
Logistic regression0.8750.840
Support vector machine0.8630.823
Random forest0.8280.822
Naïve Bayes0.8360.743
Table 6. Overall result analysis of proposed classifier.
Table 6. Overall result analysis of proposed classifier.
ClassifiersAccuracy ScorePrecisionRecallF1 Score
Logistic regression0.8750.9180.9260.922
Support vector machine0.8630.8670.9710.916
Random forest0.8280.8230.9890.898
Naïve Bayes0.8360.8930.9010.897
1D-CNN 0.8850.9070.9480.927
Table 7. Classifiers accuracy and F1 score analysis with various dimensionality reduction techniques.
Table 7. Classifiers accuracy and F1 score analysis with various dimensionality reduction techniques.
ML ModelAccuracyF1 Measure
PCALDAAutoencoderPCALDAAutoencoder
Logistic regression0.7370.6180.8750.8290.7070.922
Support vector machine0.8420.6250.8630.9100.7160.916
Random forest0.8160.6180.8280.8850.7040.898
Naïve Bayes0.7700.6250.8360.8560.7140.897
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mian, T.S. An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism. Diagnostics 2022, 12, 1796. https://doi.org/10.3390/diagnostics12081796

AMA Style

Mian TS. An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism. Diagnostics. 2022; 12(8):1796. https://doi.org/10.3390/diagnostics12081796

Chicago/Turabian Style

Mian, Tariq Saeed. 2022. "An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism" Diagnostics 12, no. 8: 1796. https://doi.org/10.3390/diagnostics12081796

APA Style

Mian, T. S. (2022). An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism. Diagnostics, 12(8), 1796. https://doi.org/10.3390/diagnostics12081796

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop