- Open Access
- Total Downloads : 16
- Authors : Choudhari Priyanka, T Jayakumari
- Paper ID : IJERTCONV3IS27079
- Volume & Issue : NCRTS – 2015 (Volume 3 – Issue 27)
- Published (First Online): 30-07-2018
- ISSN (Online) : 2278-0181
- Publisher Name : IJERT
- License: This work is licensed under a Creative Commons Attribution 4.0 International License
Classification of Parkinson’s Disease by Comparing Multi Layer Perceptron Neural Network and Least Square Support Vector Machine
Choudhari Priyanka1, Smt. T Jayakumari2 1M.Tech, Dept of Computer Science and Engineering, BTL Institute of Technology, Bangalore, India
2Assistant Professor, Dept of Computer Science and Engineering, BTL Institute of Technology, Bangalore, India
Abstract: Parkinson disease (PD) is the second most common neurodegenerative disorder. Here, the performance of Multi- Layer Perceptron Neural Network (MLPNN) and Least Square Support Vector Machines (LS SVM) regression methods are used to classify the effective detection of Parkinsons disease (PD). Classification technique is used to predict whether the particular patient is Parkinsons disease affected or not. In many cases, not all the tests contribute towards effective diagnosis of a disease. This study comprises to classify the presence of Parkinsons disease with reduced number of attributes. Originally, 23 attributes are included in the classification. After applying correlation filter, 12 attributes are took away and 11 attributes are used. Results indicate that the MLPNN provides the best performance among the other and accuracy for MLPNN yields 60% when compared with LS SVM which yields 41.05%.
Keywords: Parkinsons disease, Multi Layer Perceptron Neural Network, Least Square Support Vector Machine
Data Mining is defined as extracting the information from the large amount of data. In other words, data mining is mining the knowledge from data. There is huge amount of data available in Information Industry. This data is of no use until it is converted into some useful information. So analyzing this huge amount of data and extracting useful information from it is necessary.
Classification is a data mining (machine learning) technique used to predict group membership for data instances. Classification technique is used to predict whether the particular patient is Parkinsons disease affected or not. Classification is the process of finding a model that describes the data classes. The purpose of the model is to predict the class of objects whose class label is unknown. This derived model is based on analysis of set of training data. Classification predicts the class of objects whose class label is unknown. Its objective is to find a derived model that describes and distinguishes data classes. The Derived Model is based on the analysis of set of
training data that is data object whose class label is well known.
(Tsanas et al., 2010a) . In this research paper linear regression least squares (LS), iteratively re-weighted least squares (IRLS), least absolute shrinkage and selection operator (LASSO), and a nonlinear regression method (CART) are applied to the prediction of Parkinsons disease progression. It is reported that CART method achieved the smallest prediction error, and tracked the linearly interpolated UPDRS more accurately. AIC and BIC methods are applied on the features for optimal subset selection, and reported low testing errors with the six features (FS-1) of dysphonia. In a recent paper (Tsanas et al., 2010b) ; another feature set (logtransform of FS-2) using the LASSO linear regression algorithm was introduced for telemonitoring of PD progression.
CORINNA CORTES and VLADIMIR VAPNIK .In this
research paper the support-vector network combines 3 ideas: the solution technique from optimal hyperplanes (that allows for an expansion of the solution vector on support vectors),the idea of convolution of dot-product (that extends the solution surfaces from linear to non- linear), and the notion of soft margins (to allow for errors on the training set) The algorithm has been tested and compared to the performance of other classical algorithms. Despite the simplicity of the design in its decision surface the new algorithm exhibits a very fine performance in the comparison study. Other characteristics like capacity control and ease of changing the implemented decision surface render the support-vector network an extremely powerful and universal learning machine.
M.A. Razi, K. Athappilly . In this research a three-way comparison of prediction accuracy is performed involving nonlinear regression, NNs and CART models. NNs and CART models, produced better prediction accuracy than non-linear regression model. Brown et al. (1993) shows
that NNs do better than CART models on multimodal classification problems where data sets are large with few attributes. However, the authors also point out that CART outperforms NNs models where data sets are smaller with large numbers of irrelevant attributes. NNs and CART models are shown to outperform linear discriminate analysis on problems where the data is non-linear (Curram& Mingers, 1994).
Parkinsons disease is a chronic neurodegenerative disorder of unknown etiology, which usually affects people over the age of fifty (Tsanas, Little, McSharry, & Ramig, 2010b) .It is reported that it affects over 1 million people in North America, apart from thousands of undetected cases (Lang & Lozano, 1998) . As the worldwide population is growing older, number of PD affected patients are even expected to increase. Although medication is available to reduce the symptoms, there is no complete treatment for PD (Das, 2010) . Therefore, early diagnosis is critical and important to help patients to improve and maintain their quality of life (Tsanas, Little, McSharry, &Ramig, 2010a). However, PD may be difficult to diagnose accurately, especially at the early stages of the illness, due to symptom overlap with other disease (Ene, 2008) .
PD symptoms include (Sakar & Kursun,2009 ; Skodda & Sclegel, 2008 ; Tsanaset al.. 2010a).
tremor of the limbs
slowness of movement
difficulty with walking
balance and coordination
difficulty in eating and swallowing
cognition and mood disturbances
The data is collected at the patients home, transmitted over the internet, and processed appropriately in the clinic. The data is collected using the Intel At-Home Testing Device (AHTD), which is a telemonitoring system designed to facilitate remotely located, Internet-enabled measurement of a variety of PD-related motor impairment symptoms. Each patient specified a day and time of the week during which they have to complete the protocol, prompted with an automatic alarm reminder on the device. The collected data is encrypted and transmitted to a dedicated server automatically when the USB stick is inserted in a computer with internet connection. Further details of the AHTD apparatus and trial protocol can be found in the work of Goetz et al .
Voice measurement has shown a great progress in the advancement of Parkinson Disease detection. About 90%
of people with Parkinsons disease present some kind of vocal deterioration. And hence, in this paper dataset which mainly focus on the speech signals is chosen. This dataset is taken from UCI machine learning database . The features of dataset are given in Table I.
The dataset is composed of a range of biomedical voice measurements, there are 195 instances comprising 48 normal and 147 PD cases in the dataset. The purpose of the data is to discriminate PD unaffected patients from those with PD affected. Thus, the dataset is divided into two classes accoring to its "status" column which is set to 0 for PD unaffected patients and 1 for those with PD affected patients. It is a two-decision classification problem.
A correlation filter is applied and out of 23 attributes 12 are removed. The correlation coefficient, which are less than 0.95 are not considered for classification accuracy, thus that attribute is removed. Likewise, a total of 11 attributes are kept after the correlation filter is has been applied. Table II indicates which features are kept. First 10 attributes are used as inputs to the classifiers.
MLPNN (Multi Layer Perceptron Neural Networks)
In this study the MLPNN regression method is used. For this the number of input nodes is determined by the number of attributes decided i.e. 10 neurons; the number of hidden nodes is determined through trial and error; and [tansig, purelin, trainlm] transfer functions are used for input,hidden,output layers. The number of output nodes is represented as a class label that is 0 for PD unaffected or 1 for PD affected. Each neuron at one layer receives a weighted sum from all neurons in the previous layer and provides an input to all neurons for the later layer. For MLPNN regression method, Levenberg Marquardt (LM) backpropagation learning algorithm has been used in the feed forward single hidden layer neural network.
The MLPNN is a common neural network for both regression and process modeling and has often been used for diagnosis and tracking of disease (Kumar, 2005 ; Ubeyli, 2009) . It consists of an input layer, hidden layer and output layer. Each neuron of a layer is connected to all the neurons of the following layer. The input vectors are fed to the input layer and get multiplied by interconnection weights, known as weight factor, as they are passed from the input layer to the hidden layer. Hidden layer then calculates the transfer function and sends its signal to the output layer. The output layer calculates the transfer function to produce the neural network output. It is also necessary to assign the transfer functions for each neuron layer. Backpropagation algorithm employed is a supervised learning method for MLPNN introduced by (Rumelhart, Hinton, and Williams (1986)) , for which LevenbergMarquardt training function to update the weight and bias values are used. After training the network for a set of data, the network performance is evaluated with a new set of test data. The MLPNN outputs are then compare with the outputs obtained from the other regression methods. The detailed information about MLPNN architecture may be found in the literature (Kumar, 2005;Razi & Athappilly, 2005;Rumelhart et al., 1986) .
LS-SVM(Least Square-Support Vector Machine)
Support vector machine uses training samples to build a model that will classify information, usually visualized as a scatter plot with a wide space between categories. When new information is fed into the machine, it is plotted on the graph. The data are then classified based on which category the information falls closest to on the graph. This method works only when there are two options to choose from.
Least squares support vector machines (LS-SVM) are least squares versions of support vector machines (SVM), which are a set of related supervised learning methods that analyze data and recognize patterns, and which are used for classification and regression analysis. In this version one
finds the solution by solving a set of linear equations instead of a convex quadratic programming problem for classical SVMs. Least squares SVM classifiers, were proposed by (Suykens and Vandewalle). LS-SVMs are a class of kernel-based learning methods. LS-SVM is an alternate formulation of SVM introduced by (Suykens and Vandewalle (1999)) . This model uses a set of linear equations instead of quadratic programming problem typical of classical SVM. LS-SVM method considers the regression problem as the following optimization problem:
Subject to constraints
where ek is the prediction error term for data point k, is a regularization constant and denotes an infinite dimensional feature map. By using the Lagrangian method, the LS-SVM model can be expressed as:
where k (k = 1,. . . ,m) is the introduced Lagrange multipliers and b is the bias term, Among the Kernel function known as linear kernel, multilayer perceptron kernel, and polynomial kernel, here the linear kernel is used and is given by,
RESULTS AND DISCUSSION
As proposed, the Multi Layer Perceptron Neural Network and Least Square Support Vector Machine are implemented in MATLAB R2013a. To evaluate this classification UCI Machine Learning Repository Dataset is used. There are 195 instances comprising 48 are PD unaffected and 147 PD affected cases in the dataset. To classify the data between PD affected and unaffected people. Thus, the dataset is divided into two classes according to its "status" column which is set to 0 for PD unaffected cases and 1 for those with PD affected.
To find the performance metrics such as sensitivity, specificity and accuracy, a distinguished confusion matrix is obtained based on the classification results from these regression methods. Confusion matrix is a matrix representation of the classification results as shown in Table III
Accuracy is the percentage of predictions that are correct. The precision is the measure of accuracy provided that a specific class has been predicted.
Sensitivity is the percentage of positive labeled instances that were predicted as positive.
These performance criterions for the classifiers in the disease detection are evaluated as follows from the confusion matrix.
Accuracy = (TP+TN) / (TP+FP+TN+FN)
Sensitivity = TP / (TP+FN) Specificity = TN / (FP+TN)
For MLPNN regression method, Levenberg Marquardt (LM) backpropagation learning algorithm has been used in the feed forward single hidden layer neural network. 10 neurons are used for the input layer and tansig, purelin, trainlm transfer functions are used for input,hidden,output layer. 2 neurons for output layer is represented as a class label that is 0 for PD unaffected or 1 for PD affected. The outcome of MLPNN algorithm are tabulated in TABLE
IV. MLPNN is developed using MATLAB Neural Network Toolbox generated the results as shown in TABLE V. There are 195 instances, out of which 48 patients are normal and 147 patients are PD Effected. MLPNN shows 60% accuracy.
For LS-SVM, the Kernel Function, Linear is used. Linear kernel. In this case, svmtrain finds optimal separating plane in the space. The outcome of LS SVM algorithm are tabulated in TABLE V. The implementation of LS SVM in MATLAB generated the results as shown in TABLE VI. Accuracy for LS SVM shows 41.05%.
Early detection of any kind of disease is an essential factor. This helps in treating the patient well ahead. In this paper the performance of LS-SVM and MLPNN regression methods are used to classify the effective detection of Parkinsons disease (PD). The results indicates that MLPNN regression method yields 60% accuracy when compared to LS SVM regression method which yields 41.05%. Thus MLPNN gives best result comparison with LS SVM regression method.
UCI Machine Learning Repository- Center for Machine Learning and Intelligent System, http://archive.ics.uci.edu.
Tsanas, Little, McSharry, &Ramig, 2010b). Enhanced classical dysphonia measure and sparse regression for telemonitoring of Parkinsons disease progression. International conference on Accoustics , speech and signal processing, 594-597.
Lang A.E., & Lozano, A.M. (1998) Parkinsons Disease- First of two parts. New England journal Medicine, 339, 1044-1053.
Das, R. (2010). A comparison of multiple classification methods for diagnosis of Parkinson disease. Expert Systems with Applications, 37(2), 15681572.
(Tsanas, Little, McSharry, &Ramig, 2010a). Accurte telemonitoring of parkinsons disease progression by non invasive speech tests. IEE Transactions on Biomedical Engineering, 57, 884-893.
Ene, M. (2008). Neural network-based approach to discriminate healthy people from those with Parkinsons disease. Annals of the University of Craiova, Mathematics and Theoretical Computer Science, 35, 112116.
Sakar, C, O., & Kursun, O, (2009). Telemonitoring of parkinsons disease using measurements of dysphonia. Journel of medical system., 34(4), 591-599.
Skodda, S., & Schlegel, U. (2008). Speech rate and rhythm in parkinsond disease. Movemen Disorders, 23,985-992.
Tsanaset al.. 2010a)
Goetz, C.G. et al. Testing objective measures of motor impairment in early Parkinsons disease: Feasibility study of an at-home testing device.
Kumar, U. A. (2005) Comparison of neural network and regression analysis: A new insight. Expert systems with applications, 29, 424- 430.
Ubeyli, E. D. (2009) Combined neural network for diagnosis of erythematosquamous disease. Expert systems with Applications(36), 5107-5112.
Rumelhart, D. E., Hinton, G. E., and Williams R. J. (1986). Learning international representation by error propagation . parallel and
distributed processing: Explorations in the microstructure of cognition: Foundations (vol 1). Cambridge, MA: MIT Press.
Razi, M. A., & Athappilly, K (2005). A comparative predictive analysis of neural networks(NNs), non linear regression and classification and regression tree (CART) models. Expert systems with Applications, 29, 65-74.
Suykens,J. A. K., and Vandewalle (1999). Least squares support vector machines classifiers. Neural processing Letters, 9(3), 293- 300.
Cortes, C., & Vapnik, V. (1995). Support vector networks. Machine Learning, 20(3), 273297.