Literature DB >> 29075547

Can a Smartphone Diagnose Parkinson Disease? A Deep Neural Network Method and Telediagnosis System Implementation.

Y N Zhang1,2.   

Abstract

Parkinson's disease (PD) is primarily diagnosed by clinical examinations, such as walking test, handwriting test, and MRI diagnostic. In this paper, we propose a machine learning based PD telediagnosis method for smartphone. Classification of PD using speech records is a challenging task owing to the fact that the classification accuracy is still lower than doctor-level. Here we demonstrate automatic classification of PD using time frequency features, stacked autoencoders (SAE), and K nearest neighbor (KNN) classifier. KNN classifier can produce promising classification results from useful representations which were learned by SAE. Empirical results show that the proposed method achieves better performance with all tested cases across classification tasks, demonstrating machine learning capable of classifying PD with a level of competence comparable to doctor. It concludes that a smartphone can therefore potentially provide low-cost PD diagnostic care. This paper also gives an implementation on browser/server system and reports the running time cost. Both advantages and disadvantages of the proposed telediagnosis system are discussed.

Entities:  

Year:  2017        PMID: 29075547      PMCID: PMC5624169          DOI: 10.1155/2017/6209703

Source DB:  PubMed          Journal:  Parkinsons Dis        ISSN: 2042-0080


1. Introduction

Parkinson's disease (PD) is a disorder of brain nervous system which can cause partial or full loss in movement, behavior, and mental processing, especially speech function [1]. Generally, PD can be observed in elderly people and causes disorders in speech [2]. At present, about 1% of the worldwide population over the age of fifty is suffering from PD [3]. Until now, many effective methods and medicines [4-6] are invented for relieving the symptoms of PD. Therefore, an early-diagnosis in time and available treatment can improve the prognosis of PD [7]. However, many patients diagnosed with PD are later found which resulted in delays in patient condition [8]. Moreover, many patients with Parkinson's disease in the community still remain undiagnosed and more patients get worse because of poor medical conditions in low income areas [9]. Even though many clinical examinations and diagnostic to PD have been proposed [10-14], it is important that we should exert more effort in automated diagnosis and telediagnosis [15] in real world. In the research of Esteva et al. [16], they deduced that billions of smartphones have the potential to provide medical care for skin cancer diagnosis. Inspired by this novel idea, a smartphone also has the potential to diagnose PD. Many clinical reports reveal that the dysphonic indicator is an important reference index in diagnosis [17]. Therefore, our method is based on the study of vocal impairment symptoms (dysphonia) (90% of people with PD have such symptom) [18]. The purpose of this research is to design a machine learning based telediagnosis PD system for patients by using a smartphone. We found that deep neural network [19] with K nearest neighbor (KNN) [20] method can achieve better performance on available speech datasets than other comparative methods. Not only is the proposed telediagnosis system compared with other researchers' methods, but also the running time cost is tested in browser/server system. To conclude, the contribution of this paper includes (1) achieving high performance and (2) proposing a feasible implementation with empirical test. This paper is organized as follows: in Section 2, background and related works are presented. In Section 3, the proposed method and telediagnosis system are described. Section 4 shows the experimental results. Conclusions are drawn in Section 5.

2. Background and Related Works

2.1. Parkinson's Disease and Speech Disorders

Many researchers have exerted much effort on PD in their researches. In 2006, Rao et al. [23] discussed diagnosis and treatment for PD. The author insisted that psychosis is usually drug induced; further it can be managed initially by reducing antiparkinsonian medications. Jankovic and Aguilar [24] reviewed approaches to the treatment of PD and the authors think that the new treatments are not necessarily better than the established conventional therapy and that the treatment options must be individualized and tailored to the needs of each individual patient. In 2010, Varanese et al. [25] showed treatment of advanced PD, and the research paper concluded that supportive care, including physical and rehabilitative interventions, speech therapy, occupational therapy, and nursing care, has a key role in the late stage of disease. Yitayeh and Teshome [26] reviewed the effectiveness of physiotherapy treatment on balance dysfunction and postural instability in persons with Parkinson's disease and the author also presented meta-analysis results, in 2016. So far, it has been reported that of the 89% of PD patients with voice and speech disorders [27, 28], the reduced speech ability to communicate is considered to be one of the most important aspects of PD by many patients [29]. The common perceptual features of reduced loudness (hypophonia), reduced pitch variation (monotone), breathy and hoarse voice quality, and imprecise articulation [30], together with lessened facial expression (masked faces), contribute to limitations in communication in the majority of people with PD [31]. Figure 1 shows the comparison of waveforms between people with PD and healthy people. Intuitively, it can be observed that the waveform of healthy people is smooth and continuous, but the waveform of people with PD contains unexpected vibrations. The reason for this phenomenon is that the people with PD lose the ability of precise muscle control [32-34]. The vibrations can be detected by time frequency analysis; moreover machine learning based audio analysis methods are appropriate for diagnosing PD.
Figure 1

Waveform of voice records from Istanbul University [21, 22] (x-axis: time duration, y-axis: amplitude of the signal).

2.2. Machine Learning for Parkinson's Disease Diagnostic Care

Many researchers had proposed effective methods based on machine learning in automated diagnosis research. In 2014, Shahbakhi et al. [43] proposed a method using genetic algorithm and support vector machine for analysis of speech for diagnosis of PD. Little et al. [44] used support vector machine (SVM) with Gaussian radical basis kernel to diagnose PD. Shahbaba and Neal [35] showed a nonlinear model for the PD classification which is based on Dirichlet mixtures. Sakar and Kursun [45] applied mutual information measure to combine with SVM. Those methods achieve high classification accuracy but telediagnosis PD needs a better method with higher classification performance. Recently, deep neural networks have shown potentials in speech recognition tasks; the classification and recognition accuracy is superior to conventional machine learning method. We proposed a method using deep neural network (stacked autoencoders, SAE) to reduce dimensions and K nearest neighbor classifier to diagnose PD. We also implemented a telediagnosis system based on the proposed method. Results of empirical test on smartphone are presented in Section 4.

3. Methodology

3.1. Structure of PD Telediagnosis Method and System Structure

The proposed structure of the PD telediagnosis method is as shown in Figure 2. A patient provides personal information and speech records by following instructions of smartphone. The personal information includes gender, age, and a brief health history. Patient is also asked to read a given text; then the speech records of the patient are parsed to be time frequency based features which are extracted from the voice samples. After the processing of SAE and KNN, patient can receive the diagnosis result.
Figure 2

Acquisition and diagnosis of PD telediagnosis method.

Figure 3 shows the workflow of the proposed method in the view of machine learning, an appropriate set of time frequency features, SAE, and classifier dictated diagnostic accuracy. Therefore, the most important work is how to build a high accuracy diagnostic method.
Figure 3

Workflow of the proposed method.

Figure 4 briefly illustrates the architecture of the proposed method on B/S (browser/server) structure and details are shown in Section 4.4. The server should be installed on an operation system. In next step, an appropriate version of web service software should be deployed on this server. Usually a smartphone embedded Internet browser software (such as Google Chrome App). Therefore the smartphone can send/receive data (text and voice records) to server by Internet browser. The connection between smartphone and server can be 2/3/4G mobile network or WIFI. The server receives and processes audio files as Figures 1 and 2 present. Result of PD diagnosis will be displayed on patient's smartphone. This B/S structure is not limited to smartphone, but it is suitable for other electronic devices too, such as iPad, notebook, and even a smart watch. Moreover, the automated voice service system (such as banks' telephone services) should fulfill the same task as B/S structure theoretically.
Figure 4

B/S (browser/server) structure.

3.2. Speech Features

Dysphonia [17] is a typical speech problem of people with PD. Dysphonia is a human vocal problem which includes five major clinical features: loudness, decrease, breathiness, roughness, and exaggerated vocal tremor in voice. All those indications can be detected by analyzing time frequency in speech records. A set of 26 features is listed with the considering of the previous works held on this field of study [21, 46]. In Table 1, 6 types of parameters are listed; they are frequency parameters, pulse parameters, amplitude parameters, voicing parameters, pitch parameters, and harmonicity parameters. 26 features are also presented in Table 1.
Table 1

Time frequency features.

Parameter typeFeatures
FrequencyJitter (local)
ParametersJitter (rap)
(Number of features: 5)Jitter (local, absolute)
Jitter (ppq5)Jitter (ddp)

HarmonicityAutocorrelation
ParametersNoise-to-harmonic
(Number of features: 3)Harmonic-to-noise

PulseNumber of pulses
ParametersMean period
(Number of features: 4)Number of periods
Standard dev. of period

AmplitudeShimmer (local)
ParametersShimmer (apq3)
(Number of features: 6)Shimmer (local, dB)
Shimmer (apq5)
Shimmer (dda)
Shimmer (apq11)

PitchMedian pitch
ParametersMean pitch
(Number of features: 5)Minimum pitch
Maximum pitch
Standard deviations

VoicingFraction of locally
Parametersunvoiced frames
(Number of features: 4)Number of voice breaks
Degree of voice breaks

Types: 6Total features: 26

3.3. Stacked Autoencoders

Autoencoder [47] has been wildly used in unsupervised feature learning and speech recognition tasks. It can be built as a special three-layer neural network: the input layer, the hidden layer, and the reconstruction layer (as shown in Figure 5).
Figure 5

Autoencoder.

An autoencoder has two parts: (1) The encoder receives an input x0 ∈ R to the hidden layer (latent representation feature) x1 ∈ R via a mapping fencoder:where sencoder is the activation function, whose input is called the activation of the hidden layer, and {W1, b1} is the parameter set with a weight matrix W1 ∈ R and a bias value vector b1 ∈ R. (2) The decoder maps the hidden representation x1 back to reconstruction x2 ∈ R via mapping function fdecoder: s decoder is the activation function of the decoder with parameters {W2, b2}, W2 ∈ R, b2 ∈ R. The input of sdecoder is defined as the activation of the reconstruction layer. Parameters are learned through backpropagation by minimizing the loss L(x0, x2): In (3), L(x0, x2) consists of the reconstruction error L and the L2 regularization of W1 and W2. By minimizing the reconstruction error L, the hidden feature should be able to reconstruct the original input x0 as much as possible. The stacked autoencoders (SAE) [48] are multiple layers of autoencoders. They are a deep learning approach for dimensionality reduction and feature learning. As Figure 6 shows, there are n autoencoders which are trained one by one. The input vectors are fed to the first autoencoder. After finishing training the first autoencoder, the output hidden representation is propagated to the second auto layer. A typical activation function is sigmoid function which is used for activation functions sencoder and sdecoder. After this pretraining stage, the whole SAE is fine-tuned [48] based on a predefined objective. The last hidden layer of the SAE encoder can further cooperate with other applications, such as SVM for classification task.
Figure 6

Stacked autoencoders.

3.4. KNN Classifier

The KNN [49] classifier is quite simple: given a speech record of undiagnosed patient s, the system finds the K nearest neighbors to give diagnosis result. Formally, the decision rule can be written as Above, KNN(S) indicates the set of K nearest neighbors of speech records S. δ(S, c) is the classification for undiagnosed patient s with respect to class c, and For undiagnosed patient, the patient could be given diagnosis result. K of KNN is 1 in this paper.

4. Experimental Results

4.1. Speech Datasets, Evaluation Criteria, and Classifiers

For comparison, we chose two research papers [21, 50] and the PD speech datasets they had used. We use Matlab as programming tool and all classification algorithms are determined by grid search method. In the first paper, Sakar et al. [21] collected a speech dataset of diagnosis of PD and donated this speech dataset to UCI machine learning group for other researchers. This speech dataset contains a training set file, a testing set file, and a ZIP package of WMA files. Betul Erdogdu Sakar et al. designed a novel speech test; PD patients were asked to say only the sustained vowels “a” and “o” three times. Their work of speech dataset collection was finished at Istanbul University and we call the dataset “Istanbul Dataset” in the following experiments. In the second paper, Ma et al. [50] proposed a kernel extreme learning machine with subtractive clustering features weighting approach. Their method compared with 15 researches' methods. Total 16 methods are compared in their research paper, and the method of Ma et al. gained top position. The dataset of their research was created by Little et al. [44] of the University of Oxford, in collaboration with the National Centre for Voice and Speech, Denver, Colorado, who recorded the speech signals. We call the dataset “Oxford Dataset” in the following experiments. Accuracy, sensitivity, and other performance indexes are compared among those classifiers and these performance indexes are defined as follows:where TP, FP, TN, and FN are the true and false positive and true and false negative classifications of a classifier and they are defined as shown in Table 2.
Table 2

Confusion matrix.

Prediction as people with PDPrediction as healthy people
Actual people with PDTrue positive (TP)False negative (FN)
Actual healthy peopleFalse positive (FP)True negative (TN)
Sensitivity: Specificity: F-score: MCC: Involved classifiers are kernel extreme learning machine (KELM), Linear Support Vector Machine (LSVM), Multilayer perceptrons Support Vector Machine (MSVM), Radial basis function Support Vector Machine (RSVM), Classification and Regression Tree (CART), KNN, Linear Discriminant Analysis (LDA), and Naive Bayesian (NB) method.

4.2. Experiment Results: Oxford Dataset and Algorithm Parameters

The Oxford Dataset [22, 46] consisted of voice measurements from 31 people and 23 of them were diagnosed with PD. There are total 195 samples comparing 147 PD and 48 healthy samples in the dataset. There are no missing values in the Oxford Dataset and each feature is real type value. The whole 22 features are contained in Table 1. The SAE of the proposed method has two hidden layer: layer 1 and layer 2. The size of layer 1 is set as 10, 9, or 8 neurons and the size of layer 2 is set as 8, 7, or 6 neurons. The batch size of SAE is 20 in training and fine tuning step. All comparative classifiers are optimized by grid search method. We use 50-50% training-testing method as Polat [40] and Daliri [39] did. This experiment method tests comparative approaches with much less training data than 10-fold cross validation. Each classifier was tested 10 times and the results are presented in tables.

4.2.1. Classifiers without Deep Neural Network on Oxford Dataset

Table 3 summarized the detailed results of classification accuracy after 10 runs. From this table, it can be found that the classification performance of KNN is apparently differential. We can see that the KNN outperforms that with other 7 classifiers with a max, mean, and min accuracy of 90.53%, 82.76%, and 76.84%. All comparative classifiers give low classification results because the input samples are 22-dimensional data.
Table 3

Results of comparative classifiers without SAE on Oxford Dataset.

ClassifierClassification accuracy (%)MaxMeanMin
KELMACC83.2371.3268.49
LSVMACC81.0564.2144.21
MSVMACC84.2161.9843.16
RSVMACC85.2674.3469.47
CARTACC89.4773.9558.95
KNNACC 90.53 82.76 76.84
LDAACC87.3769.6153.68
NBACC75.7969.7461.05

4.2.2. Classifiers with Deep Neural Network on Oxford Dataset

Tables 4 and 5 presented the comparison result of the classification accuracy and other performance indexes. Table 4 lists details of accuracy. For SAE, the influence of two layers in dimension reduction has been investigated. In this study, two hidden layers are tested as a subgrid search. As seen from Table 4, 1NN classifier gives more correct diagnosis results than other methods.
Table 4

Results of comparative classifiers with SAE on Oxford Dataset.

Classification accuracy (%)Classifiers
KELMLSVMMSVMRSVMCARTKNNLDANB
SAE 10-8Max93.4599.9999.9099.9099.90100.0099.6999.19
Mean77.7495.8994.4796.3296.04 97.81 95.1794.29
Min66.6793.6689.2393.4393.8193.2392.8593.73

SAE 10-7Max96.43100.0098.46100.00100.00100.0099.6099.58
Mean80.1896.2095.7696.5896.82 97.63 96.1695.22
Min63.6994.1393.4693.5294.5394.5293.0193.61

SAE 10-6Max89.29100.0099.8199.99100.00100.0099.7399.15
Mean71.7396.2694.7496.5196.43 96.97 94.5995.18
Min45.8393.7489.1393.5194.6493.2291.3793.62

SAE 9-8Max90.48100.0097.40100.0099.99100.0099.7099.71
Mean78.2795.8795.0496.8796.87 97.89 96.0895.12
Min56.5592.3689.0493.9994.2492.7092.7193.54

SAE 9-7Max93.45100.0098.17100.00100.00100.0099.7498.74
Mean80.7796.2393.88 97.75 96.8297.1397.0195.82
Min65.4894.2289.0493.5994.4993.1493.2893.93

SAE 9-6Max98.81100.0099.71100.00100.00100.0099.6699.36
Mean84.5996.2295.6396.4696.01 98.01 94.0395.51
Min70.8393.3190.5894.1694.3093.2693.5593.94

SAE 8-8Max93.45100.0099.5299.99100.00100.0099.5899.51
Mean76.6196.0295.3995.6496.22 96.98 95.4594.32
Min55.9594.1992.2193.3794.3692.7593.7592.25

SAE 8-7Max98.81100.0099.62100.00100.00100.0099.7699.79
Mean84.2396.7195.4896.8296.2396.6395.69 97.07
Min70.8394.4092.2193.9294.5794.1092.9994.14

SAE 8-6Max91.07100.0099.4299.9999.99100.0099.7899.37
Mean78.2296.0194.5296.2195.91 96.92 95.5795.98
Min57.7493.3790.8793.5992.5393.4293.4492.49
Table 5

Performance indexes of comparative classifiers with SAE on Oxford Dataset (average of 10 runs).

Performance indexesClassifiers
KELMLSVMMSVMRSVMCARTKNNLDANB
SAE 10-8Specificity0.20830.08890.08160.81480.7302 0.9565 0.48720.0286
Sensitivity0.04230.70.21740.38240.0625 0.9306 0.14290.2
F-score0.06450.55560.1980.52530.0784 0.9571 0.19050.102
MCC−0.7719−0.2635−0.70950.1897−0.245 0.8425 −0.3992−0.8053

SAE 10-7Specificity0.34550.27940.34210.32560.7273 0.9710 0.66670.8571
Sensitivity0.7750.74070.71930.76920.5616 0.9231 0.55810.7778
F-score0.57940.41670.66670.66120.6833 0.9231 0.70070.6512
MCC0.13040.02040.06530.10580.2438 0.8941 0.1320.565

SAE 10-6Specificity0.57140.33870.72730.2581 0.8621 0.74320.05560.8806
Sensitivity0.56720.90910.75810.93750.39390.90480.0779 0.9286
F-score0.64960.57690.79660.81630.54170.64410.1200 0.8387
MCC0.12660.27150.46980.2760.25360.5489−0.7927 0.7696

SAE 9-8Specificity0.44620.20000.02740.12070.4615 0.6066 0.25000.6316
Sensitivity0.8333 0.9091 0.31820.32430.18290.82350.20690.7368
F-score0.54950.72990.14000.24000.2885 0.6512 0.32430.6437
MCC0.2710.1567−0.7202−0.5726−0.2897 0.4142 −0.34710.3612

SAE 9-7Specificity0.25970.1250.04550.36670.4000 0.9492 0.47920.2917
Sensitivity0.05560.43660.21920.28570.62220.50000.2766 0.7447
F-score0.02630.50410.29090.2410 0.7517 0.63160.30590.6034
MCC−0.5503−0.3827−0.6362−0.33540.0102 0.5251 −0.24930.0408

SAE 9-6Specificity0.89290.84620.06060.78570.3404 0.9500 0.08790.9241
Sensitivity0.76120.25580.04840.02560.7500 0.8182 0.25000.5000
F-score0.84300.35480.06250.03850.6261 0.8824 0.02270.5333
MCC0.60210.1269−0.8850−0.27000.0992 0.7586 −0.41560.4477

SAE 8-8Specificity0.1522 1.0000 0.92860.26830.84090.98890.61110.3542
Sensitivity0.14290.76470.53730.61110.54900.6000 0.9870 0.234
F-score0.14740.86670.68570.56410.65120.6667 0.9500 0.2472
MCC−0.705 0.8529 0.4336−0.12640.40310.65480.7056−0.4146

SAE 8-7Specificity0.72220.75000.35290.75000.1600 0.9808 0.21880.4118
Sensitivity 0.7805 0.12050.07690.0110.31430.74420.12900.4103
F-score0.72730.20830.12630.02150.3894 0.8421 0.09410.5333
MCC0.4980−0.1252−0.5701−0.3344−0.4651 0.7579 −0.6174−0.1374

SAE 8-6Specificity0.84620.13330.3750.1250.24140.0308 0.9231 0.557
Sensitivity0.34150.60000.50570.74190.33330.1667 0.7857 0.6875
F-score0.50000.68090.64710.41820.40000.1020 0.8544 0.3548
MCC0.1387−0.2028−0.0663−0.1667−0.3928−0.8271 0.6974 0.1831
Comparing Tables 3 and 4, other 7 classifiers gave more correct classification result with the using of SAE (hidden layer 1 contains 9 neurons; hidden layer 2 contains 7 neutrons). KELM has obtained max classification accuracy as 98.81%; it is 16% higher than 82.23% in Table 3. LSVM gives 96.71% classification accuracy after dimensional reduction by SAE 8-7. In SAE 10-7 row, MSVM increased 11% in terms of classification accuracy. RSVM gives 97.75% classification accuracy in the row of SAE 9-7. Similarly, CART produced average classification accuracy as 96% when applying SAE. LDA and NB also have got better performance than without using SAE; the classification accuracy of those 7 classifiers are improved by 10–15% with the using of SAE.

4.3. Experiment Results: Istanbul Dataset and Parameters Settings

The Istanbul University built a PD database. This database consists of training and test files. The training data belongs to 20 people with PD (6 females, 14 males) and 20 healthy individuals (10 females, 10 males) who participate in the PD database project at the Department of Neurology in Cerrahpasa Faculty of Medicine, Istanbul University. From all subjects, multiple types of sound recordings (26 voice samples including sustained vowels, numbers, words, and short sentences) are taken. A group of 26 linear and time frequency based features are extracted from each voice sample. Those features are presented in Section 3.2, as Table 2 shows. 28 PD patients are asked to say only the sustained vowels “a” and “o” three times, respectively, which makes a total of 168 recordings. The same 26 features are extracted from voice samples of this dataset. The researchers of the Istanbul University declared that [22] the PD dataset can be used as an independent test set to validate the results obtained on training set. The training file contains 1040 recordings and the testing file contains 168 files. Therefore, we used the training file and testing file to compare the proposed method with other methods. The SAE of the proposed method has two hidden layer: layer 1 and layer 2. Layer 1 is set as 10, 9, or 8 neurons and layer 2 is set as 8, 7, or 6 neurons. The batch size of SAE is 20 in training and fine tuning step.

4.3.1. Classifiers without Deep Neural Network on Istanbul Dataset

Table 6 summarized the results of classification accuracy. According to the declaration of Istanbul University, fixed training set and testing set are given. Only one run will obtain final result. From this table, it can be found that the classification performance of RSVM is apparently differential. We can see that the RSVM outperforms that with other 7 classifiers with a classification accuracy of 76.41%. All comparative classifiers give low classification results because the input samples are 26-dimensional data. The Naive Bayesian method gives worst result among all classifiers.
Table 6

Results of comparative classifiers without SAE on Istanbul Dataset.

ClassifierClassification accuracy (%)Value
KELMACC57.83
L-SVMACC39.88
M-SVMACC75.60
R-SVMACC 76.41
CARTACC50.00
KNNACC55.95
LDAACC57.28
NBACC30.95

4.3.2. Classifiers with Deep Neural Network on Istanbul Dataset

Tables 7 and 8 presented the comparison result of the classification accuracy and other 4 performance indexes. Table 7 lists details of classification accuracy. To SAE, two hidden layers are set as Oxford Dataset experiment. As seen from Table 6, KNN classifier gives more correct diagnosis results than other methods. The max, mean, and min classification accuracy of KNN classifier are not affected by network structure of SAE.
Table 7

Results of comparative classifiers with SAE on Istanbul Dataset.

Classification accuracy (%)Classifiers
KELMLSVMMSVMRSVMCARTKNNLDANB
SAE 10-8Max93.45100.0099.90100.00100.00100.00100.00100.00
Mean77.7467.8388.0379.9091.5893.4583.9381.97
Min66.6757.0877.2074.2471.7489.2375.9580.31

SAE 10-7Max96.43100.0098.46100.00100.00100.00100.00100.00
Mean80.1867.7388.2480.4891.2593.9983.8882.59
Min63.6956.3476.8074.4772.3189.5976.3779.80

SAE 10-6Max89.29100.0099.81100.00100.00100.00100.0099.90
Mean71.7367.3688.1880.6591.7393.7284.7082.09
Min45.8356.7977.0173.7272.4590.0676.6079.70

SAE 9-8Max90.48100.0097.40100.00100.00100.0099.81100.00
Mean78.2767.9088.6380.5791.5593.9884.3982.29
Min56.5556.8777.1073.9272.2389.8276.3280.27

SAE 9-7Max93.45100.0098.17100.00100.00100.0099.9099.71
Mean80.7767.8588.6779.7691.1994.1784.3581.88
Min65.4856.4576.9874.0472.0790.2176.1179.73

SAE 9-6Max98.81100.0099.71100.00100.00100.00100.00100.00
Mean84.5967.6788.5080.3691.1793.7884.6982.02
Min70.8357.0177.6274.0471.9490.1375.8979.68

SAE 8-8Max93.45100.0099.52100.00100.00100.0099.81100.00
Mean76.6167.9688.4780.0891.1494.1084.2381.89
Min55.9557.2476.7274.2172.2989.3175.7680.40

SAE 8-7Max98.81100.0099.62100.00100.00100.00100.00100.00
Mean84.2368.2787.9280.3491.5694.3584.5582.08
Min70.8357.1277.1773.9572.6189.8776.2679.74

SAE 8-6Max91.07100.0099.42100.00100.00100.00100.00100.00
Mean78.2267.4987.8579.8991.0193.8184.2481.75
Min57.7456.6276.6973.8271.8890.0575.8080.05
Table 8

Performance indexes of comparative classifiers without SAE on Istanbul Dataset (average values of 10 runs).

Performance indexesClassifiers
KELMLSVMMSVMRSVMCARTKNNLDANB
SAE 10-8Specificity0.44740.74450.66670.84290.36840.80000.89040.8807
Sensitivity0.86920.35480.91490.57140.97990.94120.80000.6949
F-score0.85610.28570.92470.48480.95110.96000.84920.7257
MCC0.32970.08640.55770.3690.46620.63910.68450.5884

SAE 10-7Specificity0.95510.92590.80460.52630.92310.95370.8860.8193
Sensitivity0.62030.43680.9630.83890.90520.90000.72220.8235
F-score0.74240.58020.88640.88340.93330.90760.73580.8235
MCC0.61790.41220.77370.28790.80210.85690.61440.6428

SAE 10-6Specificity0.79580.25450.97140.83970.95930.92750.88750.8293
Sensitivity0.26920.87610.73020.33330.8000.93940.80680.7778
F-score0.22580.78260.82140.19510.83720.94420.84520.6931
MCC0.05730.16450.74730.11790.78290.86510.69430.5703

SAE 9-8Specificity0.98480.68420.91070.09090.95950.93750.97780.9067
Sensitivity0.02780.67120.86610.91100.55000.93180.78860.1111
F-score0.05130.64470.90650.88960.59460.93710.87780.1176
MCC0.03910.35300.74980.00220.54710.86890.68840.0187

SAE 9-7Specificity0.88740.53260.90630.85840.94780.92630.7830.4706
Sensitivity0.05880.84210.87500.65450.76470.95890.93550.8543
F-score0.05710.69950.92250.67290.77610.93330.81120.8927
MCC−0.05240.38780.6880.52070.72050.88070.69390.2558

SAE 9-6Specificity0.27780.70500.97640.72650.98330.94690.82571
Sensitivity0.91330.51720.58540.98040.87040.90910.88140.6517
F-score0.91330.35290.70590.75190.92610.90090.80000.7891
MCC0.19110.17820.6570.65020.82520.85210.68320.6841

SAE 8-8Specificity0.79560.69640.89090.85000.650.98450.57140.7538
Sensitivity0.61290.64290.33330.78130.94590.79490.90980.8544
F-score0.48720.57140.09090.85470.94920.86110.89960.8502
MCC0.35300.32440.09380.55720.58360.82820.49770.6100

SAE 8-7Specificity0.78300.84550.72730.9470.92920.91300.67920.9766
Sensitivity0.93550.22220.91110.25000.87270.95080.92170.3000
F-score0.81120.27030.92130.34620.86490.95870.89080.4364
MCC0.69390.07940.61810.27530.79830.85270.63070.4131

SAE 8-6Specificity0.98130.44680.98940.81650.98260.90480.83970.7300
Sensitivity0.42620.76030.72970.76270.73580.98410.83330.9412
F-score0.58430.76990.83720.72580.82980.91850.42550.8050
MCC0.52590.20340.76080.56770.77730.86960.42680.6612
Comparing Tables 6 and 5, the other 7 classifiers gave more correct classification result with the using of SAE. And LSVM, RSVM, and CART show much more stability in classification accuracy. But KELM still did not obtain 100.00% classification accuracy, Table 6. LSVM gives 92.00% classification accuracy after dimensional reduction by SAE 9-7. In SAE 10-7, SAE10-6, and SAE 9-6 row, LSVM did not give perfect performance. MSVM, LDA, and NB increased classification accuracy but still lack stability. In conclusion, every classifier has got better performance than without using SAE.

4.4. Implementation on B/S System

As Figure 7 shows, our smartphone is running Android system and it was installed from Google Chrome web browser. The server installed Windows operation system and Internet Information Services. We also installed Matlab 2016a and Visual Studio 2013 on this server. Programming techniques contain Matlab, C#, and HTML5.
Figure 7

Implementation of the proposed system.

We used Chrome to connect server via 4G mobile network in the street; we send 28 test speech records (WMA files) of Istanbul Dataset; a transmission of 12.1 MB data took less than 5 seconds to our laboratory. The server received speech records and ran the empirical experiments as Sections 4.2 and 4.3 present. No more than 2 minutes, results of all comparative classifiers were displayed on our smartphone. We also recorded our speech (as healthy samples) and send it to server for testing, but it should be noticed that a quite environment is necessary. It is hard to achieve satisfied classification accuracy in a noisy circumstance.

4.5. Discussions and Future Work

Table 9 summarized the comparative results achieved from related researches. It can be seen that the proposed method achieves better results than other methods and our method shows a relatively fewer training samples. The reduced training dataset is meaningful when applying the proposed method in reality. The performance of SAE and KNN is robust to number of hidden neutrons of deep neural network.
Table 9

Comparative experiment results.

Related researchesTest methodClassification accuracy (%)Training set
Shahbaba and Neal [35]5-fold CV87.7080%
Psorakis et al. [36]10-fold CV89.4790%
Guo et al. [37]10-fold CV93.1090%
Ozcift and Gulten [38]10-fold CV87.1090%
Daliri [39]50-50%training-testing91.20 50%
Polat [40]50-50%training-testing97.93 50%
Chen et al. [41]10-fold CV96.0790%
Zuo et al. [42]10-fold CV97.4790%
Rao et al. [23]10-fold CV99.4990%
This study50-50%training-testing94.00–98.00 50%
We focus on classification accuracy, sensitivity, and other performance indexes to evaluate machine learning based PD diagnosis method; then we choose a B/S structure to test the proposed architecture. Advantages contain time saving, being convenient, and low cost. But we did not achieve 100% correct classification accuracy on smartphone. We deduced that there are still potential problems: (1) a good enough microphone should be taken into consideration. (2) Speech denoising may be fulfilled by multiple microphones. (3) An evolutionary system can overcome big testing data in real world. The future investigation will pay much attention on an evolutionary telediagnosis PD system and solve above issues. Moreover, we also noticed that two public datasets cannot guarantee that a machine learning based telediagnosis system can be trained very well and it is hard to satisfy hospitals, clinics, and other medical institutions. And the stability and robustness of a telediagnosis system still need to be built urgently.

5. Conclusion

For building a convenient and feasible telediagnosis PD service via smartphone, we proposed a machine learning based method on browser/server architecture. In this paper, the proposed method contains stacked autoencoders and KNN classifier which is used to process speech records. The proposed method can remap time frequency features in low dimensional space. Results show that the proposed method with KNN classifier can give doctor-level classification results on public PD speech records. An experimental system is also built for testing; it is projected that telediagnosis of PD on a smartphone will be in the future.
  35 in total

1.  Defective cortical drive to muscle in Parkinson's disease and its improvement with levodopa.

Authors:  Stephan Salenius; Sari Avikainen; Seppo Kaakkola; Riitta Hari; Peter Brown
Journal:  Brain       Date:  2002-03       Impact factor: 13.501

2.  Parkinson's disease: a reduced ability to shift to a new grouping if not prompted or guided.

Authors:  A R Cools; J H van den Bercken; M W Horstink; K P van Spaendonck; H J Berger
Journal:  Mov Disord       Date:  1990       Impact factor: 10.338

Review 3.  Diagnosis and treatment of Parkinson disease: molecules to medicine.

Authors:  Joseph M Savitt; Valina L Dawson; Ted M Dawson
Journal:  J Clin Invest       Date:  2006-07       Impact factor: 14.808

4.  Speech and swallowing symptoms associated with Parkinson's disease and multiple sclerosis: a survey.

Authors:  L Hartelius; P Svensson
Journal:  Folia Phoniatr Logop       Date:  1994       Impact factor: 0.849

5.  Telediagnosis of Parkinson's disease using measurements of dysphonia.

Authors:  C Okan Sakar; Olcay Kursun
Journal:  J Med Syst       Date:  2009-03-14       Impact factor: 4.460

6.  Suitability of dysphonia measurements for telemonitoring of Parkinson's disease.

Authors:  Max A Little; Patrick E McSharry; Eric J Hunter; Jennifer Spielman; Lorraine O Ramig
Journal:  IEEE Trans Biomed Eng       Date:  2009-04       Impact factor: 4.538

7.  Defective motor control of coughing in Parkinson's disease.

Authors:  G A Fontana; T Pantaleo; F Lavorini; F Benvenuti; S Gangemi
Journal:  Am J Respir Crit Care Med       Date:  1998-08       Impact factor: 21.405

Review 8.  Accuracy of clinical diagnosis of Parkinson disease: A systematic review and meta-analysis.

Authors:  Giovanni Rizzo; Massimiliano Copetti; Simona Arcuti; Davide Martino; Andrea Fontana; Giancarlo Logroscino
Journal:  Neurology       Date:  2016-01-13       Impact factor: 9.910

9.  An efficient diagnosis system for Parkinson's disease using kernel-based extreme learning machine with subtractive clustering features weighting approach.

Authors:  Chao Ma; Jihong Ouyang; Hui-Ling Chen; Xue-Hua Zhao
Journal:  Comput Math Methods Med       Date:  2014-11-18       Impact factor: 2.238

Review 10.  Clinical diagnostic criteria for dementia associated with Parkinson's disease.

Authors:  Murat Emre; Dag Aarsland; Richard Brown; David J Burn; Charles Duyckaerts; Yoshikino Mizuno; Gerald Anthony Broe; Jeffrey Cummings; Dennis W Dickson; Serge Gauthier; Jennifer Goldman; Christopher Goetz; Amos Korczyn; Andrew Lees; Richard Levy; Irene Litvan; Ian McKeith; Warren Olanow; Werner Poewe; Niall Quinn; Christina Sampaio; Eduardo Tolosa; Bruno Dubois
Journal:  Mov Disord       Date:  2007-09-15       Impact factor: 10.338

View more
  9 in total

1.  Automated Detection of Parkinson's Disease Based on Multiple Types of Sustained Phonations Using Linear Discriminant Analysis and Genetically Optimized Neural Network.

Authors:  Liaqat Ali; Ce Zhu; Zhonghao Zhang; Yipeng Liu
Journal:  IEEE J Transl Eng Health Med       Date:  2019-10-07       Impact factor: 3.316

2.  Prevalence and Diagnosis of Neurological Disorders Using Different Deep Learning Techniques: A Meta-Analysis.

Authors:  Ritu Gautam; Manik Sharma
Journal:  J Med Syst       Date:  2020-01-04       Impact factor: 4.460

Review 3.  Internet of Things Technologies and Machine Learning Methods for Parkinson's Disease Diagnosis, Monitoring and Management: A Systematic Review.

Authors:  Konstantina-Maria Giannakopoulou; Ioanna Roussaki; Konstantinos Demestichas
Journal:  Sensors (Basel)       Date:  2022-02-24       Impact factor: 3.576

Review 4.  Digital Technology in Movement Disorders: Updates, Applications, and Challenges.

Authors:  Jamie L Adams; Karlo J Lizarraga; Emma M Waddell; Taylor L Myers; Stella Jensen-Roberts; Joseph S Modica; Ruth B Schneider
Journal:  Curr Neurol Neurosci Rep       Date:  2021-03-03       Impact factor: 6.030

5.  Prediction of Freezing of Gait in Parkinson's Disease Using a Random Forest Model Based on an Orthogonal Experimental Design: A Pilot Study.

Authors:  Zhonelue Chen; Gen Li; Chao Gao; Yuyan Tan; Jun Liu; Jin Zhao; Yun Ling; Xiaoliu Yu; Kang Ren; Shengdi Chen
Journal:  Front Hum Neurosci       Date:  2021-03-22       Impact factor: 3.169

6.  Insight into an unsupervised two-step sparse transfer learning algorithm for speech diagnosis of Parkinson's disease.

Authors:  Yongming Li; Xinyue Zhang; Pin Wang; Xiaoheng Zhang; Yuchuan Liu
Journal:  Neural Comput Appl       Date:  2021-02-09       Impact factor: 5.606

7.  A mobile-assisted voice condition analysis system for Parkinson's disease: assessment of usability conditions.

Authors:  Javier Carrón; Yolanda Campos-Roca; Mario Madruga; Carlos J Pérez
Journal:  Biomed Eng Online       Date:  2021-11-21       Impact factor: 2.819

8.  Resting-state electroencephalography based deep-learning for the detection of Parkinson's disease.

Authors:  Mohamed Shaban; Amy W Amara
Journal:  PLoS One       Date:  2022-02-24       Impact factor: 3.240

9.  Assessment of the Status of Patients with Parkinson's Disease Using Neural Networks and Mobile Phone Sensors.

Authors:  Yulia Shichkina; Elizaveta Stanevich; Yulia Irishina
Journal:  Diagnostics (Basel)       Date:  2020-04-12
  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.