Download as pdf or txt
Download as pdf or txt
You are on page 1of 4

2020 IEEE Region 10 Symposium (TENSYMP), 5-7 June 2020, Dhaka, Bangladesh

Classification of Parkinson’s Disease by Analyzing


Multiple Vocal Features Sets
Kazi Amit Hasan Md. Al Mehedi Hasan
Department of Computer Science & Engineering Department of Computer Science & Engineering
Rajshahi University of Engineering & Technology Rajshahi University of Engineering & Technology
Rajshahi, Bangladesh Rajshahi, Bangladesh
Email: kaziamithasan89@gmail.com Email: mehedi ru@yahoo.com

Abstract—Parkinson’s disease (PD) is a growing and chronic these researches, numerous features were obtained by using
neurodegenerative disease with a great amount of motor and non- some speech signal processing methods and extracted features
motor symptoms. In the initial stages, most of the PD patients were fed into different machine learning algorithms.
face difficulties in regular movements. Vocal disorders are one of
the common symptoms of them. Vocal disorder centric diagnosis The analysis and investigation of PD are difficult in some
systems are one of the leading areas in recent PD detection cases because of their overlapping symptoms. Only 75% of
studies. In this paper, the dataset was taken from the UCI the clinical investigation of PD is validated to be idiopathic
Machine Learning repository and a feature extraction technique Parkinson’s Disease at autopsy due to the symptoms overlap
was applied. The Analysis of Variance (ANOVA) is used for with other diseases [6]. An enormous amount of research
extracting the features as the dataset was full of features and the
topmost 50 features are selected according to ANOVA F-score. activities are performed in the PD classification. Classification
Multiple machine learning classification methods were applied methods can escalate the accuracy and efficiency of the re-
and compared with other related existing works. Experimental ferred diagnosis system as well as can make the diagnosis time
results show that the highest accuracy score of 0.91 was achieved more effective. In recent studies, TQWT features produced
with the Random Forest Classifier method by feeding the selected better results than other vocal features. MFCC and TQWT
features. ANOVA as a feature extraction technique successfully
extracted the significant features that differentiate PD patients combined features also improved the classification accuracy
from healthy individuals and also improve the classification [4].
accuracy. In this paper, different classifiers are used and applied to
Index Terms—Parkinson’s Disease (PD), Feature Extraction, recognize Parkinson’s disease. A feature selection technique,
Analysis of Variance (ANOVA), Classification Analysis of Variance (ANOVA) is used in this research and the
topmost 50 features are selected according to ANOVA F-value.
I. I NTRODUCTION The extracted features were fed into different classification
In recent years, health informatics systems perform a vital algorithms and final prediction is generated from that. The
role in identifying and monitoring different diseases. Parkin- overall workflow is shown in Fig.2 in Methodology section.
son’s disease (PD) is a rapidly growing neurodegenerative Various classification methods are implemented for evaluating
disease with a great amount of motor and non-motor signs the accuracy of the classifiers. As per the classification accu-
[1]. The early death of dopamine generative neurons in the racy, random forest classifier yields 91% correct classification
substantia nigral region results in Parkinson’s disease (PD) [2]. rate.
As PD progresses, the amount of dopamine produced in the
II. DATASET
brain diminishes gradually and the affected person becomes
unable to control his/her actions normally. People ranging A. Dataset Description
from 55-75 years are more at risk of affecting by PD. The The dataset used in this research was collected from the UCI
increasing risk expands the necessity for accurate PD diagnosis Machine Learning Repository [7]. This specific dataset was
and monitoring. In the US, there are about one million people also used in research [4], [8]. The dataset contains the data of
who got affected by Parkinson’s disease. Medications and 188 Parkinson disease affected patients, where 107 of them are
surgeries are possible treatments to cure their symptoms. Still, men and 81 of them are female. The age span of Parkinson’s
no effective solution and therapy for Parkinson’s disease is Disease patients is varied between 33 and 87 years. Along
discovered [3]. with this, the control group contains 64 healthy individuals
PD detection methods focus on observing and estimating where 23 of them are men and 41 of them are female. The
the severity and austerity of the indications using numerous age varies between 41 to 82 years of the control healthy
types of devices. One of the most popular signs of PD is the individuals. This dataset was collected from the Department
vocal difficulties, faced by 90% of PD victims in their earlier of Neurology in Cerrahpasa Faculty of Medicine, Istanbul
stages of the disease [4]. This vocal problem-based systems University. The frequency response remained fixed to 44.1 kHz
are in the uppermost positions on PD detection studies [5]. In of the microphone while gathering the data. After the doctor’s

978-1-7281-7366-5/20/$31.00 ©2020 IEEE


Data

Data Preprocessing

Feature Selection

Classification

Fig. 1: Correlation among the features before extracting them Results

from the dataset.


Fig. 2: Graphical representation of overall proposed PD clas-
sification system.
examination the repeated renewal of the vowel /a/ letter was
collected from each individual with three replicates. The
dataset contains various vocal features of different feature sets
1) Calculating Sum of Squared Error (SSE)
like Mel-Frequency Cepstral Coefficients (MFCCs), Tunable
2) Calculating Degree of Freedom (DF)
Q-Factor Wavelet Transform (TQWT), vocal fold along with
3) Calculating Mean Squares (MS)
other baseline features [4].
4) Calculating F-value.
B. Exploratory Data Analysis (EDA) In this research, the prime 50 features were chosen and sorted
There is a total number of 130 male and 122 female indi- according to the best ANOVA F-value shown in Fig.3. The
viduals data that are used in the dataset. This also represents scikit-learn [11] tool was used to determine the F-value as well
that the dataset is balanced from the gender perspective. As as the best 50 feature names. When the numerical features are
this dataset has many features, the correlation heatmap will grouped by the target vector, the F-value tests that the means
describe it more clearly. The correlation heat map in Fig.1 for each group are significantly different from each other [12].
clearly indicates the highly correlated features in the dataset. In Fig.3, the best 50 features are sorted in terms of their
This correlation problem is handled by ignoring the extremely F-values. With the ANOVA feature selection method it is
correlated features and selecting the best features by feature confirmed that most of the features are MFCCs and TQWT
selection method described in the Methodology section. features which also obtained in the study [4]. In our re-
search, the highest ANOVA F-value 142.51 was obtained by
III. M ETHODOLOGY
mean MFCC 2nd coef feature among all the features. TQWT
A. Feature Extraction features also have higher ANOVA F-values than the rest of the
The dataset has 754 attributes. The correlation among the features.
features is shown in Fig.1. As this dataset is full of features, The correlation among the features is also handled using
an univariate feature selection method is used in our research ANOVA. The correlation heatmap in Fig.4 is clearly better
to find out the most significant features out of these. than Fig.1.
1) Analysis of Variance (ANOVA): The dataset used in this
research is full of features. For this, it is a tactful decision B. Classification
to use a feature selection algorithm to find out the optimal
features. Feature selection methods also enhance prediction In this study, the entire dataset was divided into 20%
quality. In this study, as a feature selection technique the training and 80% testing sets. The best scores are shown in
analysis of variance (ANOVA) was used and it is also used Table I in Experiment and Results section.
in other different studies [9], [10]. ANOVA is a simple and 1) Random Forest: Random Forest classification is also
strong statistical method that examines the means of a couple used in previous PD classification studies [13], [14]. It is an
or more groups. It also determines how much the groups extended version of the bagging algorithm, fast and robust
are significantly different from each other. Several steps are to over-fitting that takes its origin from decision tree’s nature
performed in ANOVA [10]. [15].
(a) AUC acheived with Decision (b) AUC acheived with Random
Tree classification method Forest classification method

Fig. 5: AUC curves of Decision Tree and Random Forest


classification method.

It is also used in the study [19]. Conditional probability is


used in naive Bayes classifier to get the results [20].
6) Support Vector Machine (SVM): SVM which is also
known as decision boundary function by finding a hyperplane.
SVM is also used in the study of PD classification [21]. In
this research, Linear and Radial Basis Function (RBF) kernels
are applied as classification methods.
Fig. 3: Top 50 features with their ANOVA F-value. IV. E XPERIMENT AND R ESULTS
A. Evaluation Metrics
In this study, five statistical measures accuracy, precision,
recall, f-1 score and AUC curve were used as evaluation
metrics to estimate the performance and efficiency of the
classifiers [22].
TP + TN
Accuracy = (1)
TP + FP + FN + TN
TP
(a) Correlation heat-map of first 10 (b) Correlation heat-map of second P recision = (2)
features after feature extraction. 10 features after feature extraction. TP + FP
TP
Fig. 4: Correlation among the features after feature extraction Recall = (3)
in the dataset. TP + FN
2 ∗ (Recall ∗ P recision)
F 1 Score = (4)
(Recall + P recision)
2) k-Nearest Neighbour: kNN is a simple, powerful su-
where, TP, TN, FP, FN means True Positives, True Negative,
pervised learning algorithm used in previous PD classification
False Positives and False Negatives values respectively [23].
studies [16]. Different values of k in kNN classifier are applied
AUC curves were also used in this research to examine how
and compared with one another. Then the impacts are analyzed
well the probabilities from the positive classes are separated
and pointed out the best value. In our research, the best
from the negative classes. The AUC curve is the degree of
accuracy score was achieved with k=5.
True Positive and False Positive rate which maps the sensitivity
3) Decision Tree: Decision tree based classification is also against 1 - specificity that signifies the overall effectiveness of
used in the studies [17]. It builds classification or regression the model [17].
models in the type of tree structure. In this study, the best
AUC score is achieved with the decision tree. B. Results
4) Logistic Regression (LR): Logistic Regression based The comparison between this research work and other re-
classification is used in different PD classification studies [18]. search work [4] is shown in the Table I. Different classification
LR also used in many classification problems because of being methods random forest, k-nearest neighbour, decision tree,
binary in nature. gradient boosting, logistic regression, naive bayes and support
5) Naive Bayes: Naive Bayes is a supervised learning vector machine are used in this research. The best accuracy
classifier that works on the Bayesian Probabilistic theorem. score is 0.91 achieved with Random Forest classification
TABLE I: Comparison of accuracy, precision, recall, f-1 score [5] B. E. Sakar, M. E. Isenkul, C. O. Sakar, A. Sertbas, F. Gurgen, S. Delil,
and AUC score. H. Apaydin, and O. Kursun, “Collection and analysis of a parkinson
speech dataset with multiple types of sound recordings,” IEEE Journal
Accu- Preci- F-1 AUC of Biomedical and Health Informatics, vol. 17, no. 4, pp. 828–834, 2013.
Methods Recall
racy sion Score score [6] M. Ene, “Neural network-based approach to discriminate healthy people
Naive from those with parkinson’s disease,” Annals of the University of
Bayes 0.83 - - 0.83 - Craiova-Mathematics and Computer Science Series, vol. 35, pp. 112–
[4] 116, 2008.
Logistic [7] D. Dua and C. Graff, “UCI machine learning repository,” 2017.
Regression 0.85 - - 0.84 - [Online]. Available: http://archive.ics.uci.edu/ml
[4] [8] H. Gunduz, “Deep learning-based parkinson’s disease classification
Other using vocal feature sets,” IEEE Access, vol. 7, pp. 115 540–115 551,
kNN [4] 0.85 - - 0.82 -
Papers 2019.
Random [9] H. Ding, P.-M. Feng, W. Chen, and H. Lin, “Identification of bacte-
Forest 0.85 - - 0.84 - riophage virion proteins by the anova feature selection and analysis,”
[4] Molecular BioSystems, vol. 10, no. 8, pp. 2229–2235, 2014.
SVM [10] Z. A. Bakar, D. I. Ispawi, N. F. Ibrahim, and N. M. Tahir, “Classification
(Linear) 0.83 - - 0.82 - of parkinson’s disease based on multilayer perceptrons (mlps) neural net-
[4] work and anova as a feature extraction,” in 2012 IEEE 8th International
SVM Colloquium on Signal Processing and its Applications. IEEE, 2012,
(RBF) 0.86 - - 0.84 - pp. 63–67.
[4] [11] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion,
Random O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vander-
0.91 0.93 0.96 0.94 0.83
Forest plas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duch-
kNN 0.85 0.88 0.90 0.89 0.70 esnay, “Scikit-learn: Machine learning in Python,” Journal of Machine
Decision Learning Research, vol. 12, pp. 2825–2830, 2011.
0.86 0.94 0.87 0.91 0.86
Tree [12] C. Albon, “Anova f-value for feature selection,” Dec 2017.
This Logistic [Online]. Available: https://chrisalbon.com/machine learning/feature
0.88 0.88 0.97 0.93 0.74
Paper Regression selection/anova f-value for feature selection/
Naive [13] S. Arora, V. Venkataraman, S. Donohue, K. M. Biglan, E. R. Dorsey,
0.84 0.90 0.88 0.89 0.79
Bayes and M. A. Little, “High accuracy discrimination of parkinson’s disease
SVM participants from healthy controls using smartphones,” in 2014 IEEE
0.84 0.88 0.92 0.91 0.73
(Linear) International Conference on Acoustics, Speech and Signal Processing
SVM (ICASSP). IEEE, 2014, pp. 3641–3644.
0.86 0.86 0.96 0.91 0.70
(RBF) [14] S. Joshi, D. Shenoy, V. S. GG, P. Rrashmi, K. Venugopal, and L. Pat-
naik, “Classification of alzheimer’s disease and parkinson’s disease by
using machine learning and neural network methods,” in 2010 Second
method. The highest AUC score is 0.86 achieved with the International Conference on Machine Learning and Computing. IEEE,
2010, pp. 218–222.
Decision Tree classification method shown in Fig.5. [15] L. Breiman, “Random forests,” Machine learning, vol. 45, no. 1, pp.
5–32, 2001.
V. C ONCLUSION [16] S. Bind, A. K. Tiwari, A. K. Sahani, P. Koulibaly, F. Nobili, M. Pagani,
In this research, multiple classification algorithms are imple- O. Sabri, T. Borght, K. Laere, and K. Tatsch, “A survey of machine learn-
ing based approaches for parkinson disease prediction,” International
mented in the processed dataset and high accuracy is achieved Journal of Computer Science and Information Technologies, vol. 6,
with our proposed method than other existing works in this no. 2, pp. 1648–1655, 2015.
domain. As a feature selection technique, Analysis of Variance [17] R. Das, “A comparison of multiple classification methods for diagnosis
of parkinson disease,” Expert Systems with Applications, vol. 37, no. 2,
(ANOVA) is applied. The top 50 features selected by ANOVA pp. 1568–1572, 2010.
are fed into classification methods. Most of the features are [18] M. Tinazzi, C. Del Vesco, E. Fincati, S. Ottaviani, N. Smania,
TQWT, MFCC, vocal fold, bandwidth, formant feature etc. G. Moretto, A. Fiaschi, D. Martino, and G. Defazio, “Pain and motor
complications in parkinson’s disease,” Journal of Neurology, Neuro-
Seven popular machine learning algorithms are used in this surgery & Psychiatry, vol. 77, no. 7, pp. 822–825, 2006.
research and the best accuracy score of 0.91 achieved with [19] S. A. Mostafa, A. Mustapha, S. H. Khaleefah, M. S. Ahmad, and M. A.
random forest classification method. Though these scores are Mohammed, “Evaluating the performance of three classification methods
in diagnosis of parkinson’s disease,” in International Conference on Soft
higher than other existing works, it still could be improved by Computing and Data Mining. Springer, 2018, pp. 43–52.
using other enhanced techniques. [20] J. D. Rennie, L. Shih, J. Teevan, and D. R. Karger, “Tackling the poor
assumptions of naive bayes text classifiers,” in Proceedings of the 20th
R EFERENCES international conference on machine learning (ICML-03), 2003, pp.
[1] J. Jankovic, “Parkinson’s disease: clinical features and diagnosis,” Jour- 616–623.
nal of neurology, neurosurgery & psychiatry, vol. 79, no. 4, pp. 368–376, [21] I. Bhattacharya and M. P. S. Bhatia, “Svm classification to distinguish
2008. parkinson disease patients,” 01 2010, p. 14.
[2] W. Poewe, K. Seppi, C. M. Tanner, G. M. Halliday, P. Brundin, [22] D. M. Powers, “Evaluation: from precision, recall and f-measure to roc,
J. Volkmann, A.-E. Schrag, and A. E. Lang, “Parkinson disease,” Nature informedness, markedness and correlation,” 2011.
reviews Disease primers, vol. 3, no. 1, pp. 1–21, 2017. [23] M. Sokolova and G. Lapalme, “A systematic analysis of performance
[3] R. G. Ramani and G. Sivagami, “Parkinson disease classification using measures for classification tasks,” Information processing & manage-
data mining algorithms,” International journal of computer applications, ment, vol. 45, no. 4, pp. 427–437, 2009.
vol. 32, no. 9, pp. 17–22, 2011.
[4] C. O. Sakar, G. Serbes, A. Gunduz, H. C. Tunc, H. Nizam, B. E. Sakar,
M. Tutuncu, T. Aydin, M. E. Isenkul, and H. Apaydin, “A comparative
analysis of speech signal processing algorithms for parkinson’s disease
classification and the use of the tunable q-factor wavelet transform,”
Applied Soft Computing, vol. 74, pp. 255–263, 2019.

You might also like