Literature DB >> 29740058

Model-based and Model-free Machine Learning Techniques for Diagnostic Prediction and Classification of Clinical Outcomes in Parkinson's Disease.

Chao Gao1,2, Hanbo Sun1,3, Tuo Wang1,3, Ming Tang1,2, Nicolaas I Bohnen4,5,6, Martijn L T M Müller4,5,6, Talia Herman7, Nir Giladi7,8, Alexandr Kalinin1,6,9, Cathie Spino2,6, William Dauer5,6, Jeffrey M Hausdorff7,10,11, Ivo D Dinov12,13,14,15.   

Abstract

In this study, we apply a multidisciplinary approach to investigate falls in PD patients using clinical, demographic and neuroimaging data from two independent initiatives (University of Michigan and Tel Aviv Sourasky Medical Center). Using machine learning techniques, we construct predictive models to discriminate fallers and non-fallers. Through controlled feature selection, we identified the most salient predictors of patient falls including gait speed, Hoehn and Yahr stage, postural instability and gait difficulty-related measurements. The model-based and model-free analytical methods we employed included logistic regression, random forests, support vector machines, and XGboost. The reliability of the forecasts was assessed by internal statistical (5-fold) cross validation as well as by external out-of-bag validation. Four specific challenges were addressed in the study: Challenge 1, develop a protocol for harmonizing and aggregating complex, multisource, and multi-site Parkinson's disease data; Challenge 2, identify salient predictive features associated with specific clinical traits, e.g., patient falls; Challenge 3, forecast patient falls and evaluate the classification performance; and Challenge 4, predict tremor dominance (TD) vs. posture instability and gait difficulty (PIGD). Our findings suggest that, compared to other approaches, model-free machine learning based techniques provide a more reliable clinical outcome forecasting of falls in Parkinson's patients, for example, with a classification accuracy of about 70-80%.

Entities:  

Mesh:

Year:  2018        PMID: 29740058      PMCID: PMC5940671          DOI: 10.1038/s41598-018-24783-4

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

PD clinical characteristics, current state-of-the-art techniques, societal impact

Parkinson’s disease (PD) is a common neurodegenerative disorder that affects over 10 million people worldwide. PD affects about 1% of people over 60 years of age and the prevalence increases with age. People with PD experience a range of motor and non-motor symptoms that include tremor, rigidity, bradykinesia, postural instability, gait disturbances such as freezing of gait (FoG), autonomic disturbances, affective disorders, sleep disturbances, and cognitive deficits[1]. These symptoms markedly impact and curtail health related quality of life[2]. Freezing of gait and associated falls represent one of the most serious consequences of PD[3]. Falls are much more common in patients with PD than in age-matched controls and falls often lead to reduced functional independence, increased morbidity, and higher mortality[4]. The ability to better identify future fallers from non-fallers could inform more effective treatment and personalized medicine planning. The hallmark pathology of PD is loss of dopamine in the striatum secondary to progressive degeneration of dopaminergic cells in the substantia nigra pars compacta, accompanied by the formation of Lewy bodies[5]. A variable combination of tremor, rigidity, and bradykinesia symptoms may present along with postural instability and gait difficulty (PIGD) features. Because of primary involvement of the basal ganglia in PD, it has often been asserted that these motor features are mainly attributable to nigrostriatal dopaminergic loss. A common dopamine replacement therapy to ameliorate PD motor symptom is levodopa (L-DOPA). A recent study from Vu et al.[6] showed that L-DOPA potency was lowest for PIGD features compared to other cardinal motor features. In the Sydney Multicenter Study of PD, patients have been followed for about two decades. Results of this study indicate that dopamine non-responsive problems dominate 15 years after initial assessments and include frequent falls, which occurs in 81% of the patients[7]. Similar findings were recently reported by López et al. after following de novo PD patients for 10 years[8]. These authors reported good responses to dopaminergic treatment in the first year with a progressive decline, becoming more manifest especially after 3 years. Significant PIGD motor disabilities arose at 10 years in 71% of patients that were mainly caused by non-dopamine-responsive features such as freezing of gait (FoG)[8]. The L-DOPA resistance of PIGD motor features has been proposed to include non-dopaminergic structures in widespread brain regions[9]. As axial motor impairments, in particular falls, do not respond well to dopaminergic medications there is a need to identify early predictors of falls. Such predictors may provide potential clues about underlying mechanism of falls that may more effectively inform future treatment interventions. The main goal of this study was to identify clinical and MR imaging predictors of falls from two independent archives containing clinical and imaging data of PD patients.

Machine Learning methods for prediction, classification, forecasting and data-mining

Both model-based and model-free techniques may be employed for prediction of specific clinical outcomes or diagnostic phenotypes. The application of model-based approaches heavily depends on the a priori statistical statements, such as specification of relationship between variables (e.g. independence) and the model-specific assumptions regarding the process probability distributions (e.g., the outcome variable may be required to be binomial). Examples of model-based methods include generalized linear models. Logistic regression is one of the most commonly used model-based tools, which is applicable when the outcome variables are measured on a binary scale (e.g., success/failure) and follow Bernoulli distribution[10]. Hence, the classification process can be carried out based on the estimated probabilities. Investigators have to carefully examine and confirm the model assumptions and choose appropriate link functions. Since the statistical assumptions do not always hold in real life problems, especially for big incongruent data, the model-based methods may not be applicable or may generate biased results. In contrast, model-free methods adapt to the intrinsic data characteristics without the use of any a priori models and with fewer assumptions. Given complicated information, model-free techniques are able to construct non-parametric representations, which may also be referred as (non-parametric) models, using machine learning algorithms or ensembles of multiple base learners without simplification of the problem. In the present study, several model-free methods are utilized, e.g., Random Forest[11], AdaBoost[12], XGBoost[13], Support Vector Machines[14], Neural Network[15], and SuperLearner[16]. These algorithms benefit from constant learning, or retraining, as they do not guarantee optimized classification/regression results. However, when trained, maintained and reinforced properly and effectively, model-free machine learning methods have great potential in solving real-world problems (prediction and data-mining). The morphometric biomarkers that were identified and reported here may be useful for clinical decision support and assist with diagnosis and monitoring of Parkinson’s disease. There are prior reports of using model-free machine-learning techniques to diagnose Parkinson’s disease. For instance, Abos et al. explored connection-wise patterns of functional connectivity to discriminate PD patients according to their cognitive status[17]. They reported an accuracy of 80.0% for classifying a validation sample independent of the training dataset. Dinesh and colleagues employed (boosted) decision trees to forecast PD. Their approach was based on analyzing variations in voice patterns of PD patients and unaffected subjects and reported average prediction accuracy of 91–95%[18]. Peng et al. used machine learning method for detection of morphometric biomarkers in Parkinson’s disease[19]. Their multi-kernel support vector machine classifier performed well with average accuracy = 86%, specificity = 88%, and sensitivity = 88%. Another group of researchers developed a novel feature selection technique to predict PD based on multi-modal neuroimaging data and using support vector classification[20]. Their cross-validation results of predicting three types of patients, normal controls, subjects without evidence of dopaminergic denervation (SWEDDs), and PD patients reported classification accuracy about 89–90%. Bernad-Elazari et al. applied a machine learning approach to distinguish between subjects with and without PD. Their objective characterization of daily living transitions in patients with PD used a single body-fixed sensor, successfully distinguishing mild patients from healthy older adults with an accuracy of 86%[21]. Previously identified biomarkers, as well as the salient features determined in our study, may be useful for improving the diagnosis, prognosticating the course, and tracking the progression of the disease over time.

Study Goals

This study aims to address four complementary challenges. To address the need for effective data management and reliable data accumulation, Challenge 1 involves designing a protocol for harmonizing and aggregating complex, multisource, and multi-site Parkinson’s disease data. We applied machine learning techniques and controlled variable selection, e.g., knockoff filtering[22], to address Challenge 2, identify salient predictive features associated with specific clinical traits, e.g., patient falls. Challenge 3 involves forecasting patient falls using alternative techniques based on the selected features and evaluating the classification performance using internal (statistical) and external (prospective data) validation. Finally, Challenge 4, addresses the need to forecast other clinically relevant traits like Parkinson’s phenotypes, e.g., tremor dominance (TD) vs. posture instability and gait difficulty (PIGD)[23].

Predictive Analytic Strategy

The datasets used in this study were collected independently at two sites – the University of Michigan Udall Center of Excellence in Parkinson’s Disease Research (Michigan data) and the Sourasky Medical Center, Israel (Tel-Aviv data). Both the datasets include high dimensional data consisting of several hundred demographic and clinical features for about a couple of hundred PD patients. This research is focused primarily on the prediction of patients’ falls, although alternative clinical outcomes and diagnostic phenotypes can be explored using the same approach. As not all of the features in the clinical record are strongly associated with each specific response, our goal is to identify some important critical features, build the simplest statistical models, and demonstrate reproducible computational classifies that produce higher prediction accuracy while avoiding overfitting. Figure 1 shows a high-level schematic of the study-design, including the complementary training and testing strategies.
Figure 1

Predictive Analytics Strategy: (Top) Identify critical features and build predictive models independently on the Michigan and the Tel-Aviv datasets, respectively. (Bottom) Harmonize and merge the two data archives and perform the same analytics on the aggregate data. The bottom-right branch of the diagram illustrates the process of training the models on one of the datasets and (externally) validating their accuracy on the other complementary dataset.

Predictive Analytics Strategy: (Top) Identify critical features and build predictive models independently on the Michigan and the Tel-Aviv datasets, respectively. (Bottom) Harmonize and merge the two data archives and perform the same analytics on the aggregate data. The bottom-right branch of the diagram illustrates the process of training the models on one of the datasets and (externally) validating their accuracy on the other complementary dataset. In general, model-free statistical learning methods (e.g. Random Forest, Support Vector Machines) make fewer assumptions and often outperform model-based statistical techniques like logistic regression, which is often considered a baseline method, on large and complex biomedical data[24-27]. To quantify the forecasting results, we used established evaluation metrics such as overall accuracy, sensitivity, specificity, positive and negative predictive power, and log odds ratio. For clinical datasets with a large number of features, it is difficult to avoid the multi-collinearity problem, which causes problems with maximum likelihood estimation of model-based techniques[28]. As the machine learning techniques have minimal statistical assumptions, they may provide more flexible and reliable predictions. This manuscript is organized as follows: The methods section describes the study design, the characteristics of the data and meta-data, the preprocessing, harmonization, aggregation and analysis methods, as well as the evaluation strategies. The results section reports the findings for each of the study designs shown in Fig. 1. Finally, the discussion section explains the findings, identifies potential drawbacks and suggests prospective translational studies.

Methods

All methods and analyses reported in the manuscript were carried out in accordance with relevant institutional, state and government guidelines and regulations. The experimental protocols were approved by the institutional review boards of the University of Michigan (HUM00022832) and Tel Aviv Sourasky Medical Center (0595–09TLV). Informed consent was obtained from all participating volunteers prior to enrollment in the study and data collection.

Data sources and management

Below we describe the two main sources of data (University of Michigan and Tel Aviv Sourasky Medical Center) and discuss the data management, wrangling, preprocessing, imputation, harmonization, aggregation, and analytics.

Michigan data

The University of Michigan archive included data collected as part of a NIH-funded clinical and neuroimaging study of PD. Additional information about inclusion/exclusion criteria and data dictionary are provided in Supplementary Materials Section I.1.a. Briefly, the raw dataset compiled at Michigan contains study subjects’ demographics, PET, behavioral and sensory assessments, Mattis Dementia Rating Scale, sleep questionnaires, genetics, number of falls, clinical measures and MR neuroimaging (207 variables in total). Among the 225 study subjects, there were 148 patients with Parkinson’s disease and 77 healthy participants.

Tel-Aviv data

The Tel-Aviv archive includes demographic, clinical, gait, balance and imaging data. The dataset was originally gathered to study the role of white matter changes in PD and putative relationships to motor phenotypes[29,30]. The study included 110 patients with idiopathic PD recruited by referrals from specialists at the outpatient movement disorders unit, and from other affiliated clinics. Additional information about inclusion/exclusion criteria and data dictionary are provided in Supplementary Materials Section I.1.b.

Michigan + TelAviv Data Aggregation

The preprocessed Tel-Aviv and Michigan datasets are harmonized and merged using 133 shared variables, which include Subject ID, PD subtype (TD vs. PIGD), Tremor score, PIGD score, gender, age, weight, height, BMI, Geriatric Depression Scale (short form), the Timed up and go test, specific items from Part I, II and III of the Movement Disorder Society (MDS)-sponsored version of the UPDRS, Hoehn and Yahr scale, Montreal Cognitive Assessment (MoCA), and 56 derived neuroimaging features. Notably, the UPDRS Part III sub items from the two datasets were both measured under the “OFF” medication cycle, i.e., approximately 12 hours of antiparkinsonian medication withdrawal prior to the assessments. The aggregated dataset consists of 251 subjects and 133 variables.

Model-based and Model-free machine learning methods

The Supplementary Materials Section I.2 (Predictive Analytics) includes the mathematical descriptions of the model-based (e.g., Logistic Regression) and model-free (e.g., Random Forest, Adaptive and gradient boosting, Support Vector Machines, Neural networks, SuperLearner) techniques used for prediction and classification. The Knockoff filtering and random-forest feature selection methods are detailed in Supplementary Materials Section I.3 (Feature Selection).

Statistical validation strategies and evaluation metrics

Classification

To validate the prediction performance for binary classes, we usually construct a 2 × 2 contingency table (confusion matrix) as illustrated on Table 1:
Table 1

The confusion matrix provides a mechanism to assess the accuracy of binary diagnostic classification.

Reference
FallNon-fall
PredictionFallTPFP
Non-fallFNTN
The confusion matrix provides a mechanism to assess the accuracy of binary diagnostic classification. True Positive(TP): Number of observations that correctly classified as “Fall” group. True Negative(TN): Number of observations that correctly classified as “Non-Fall” group. False Positive(FP): Number of observations that incorrectly classified as “Fall” group. False Negative(FN): Number of observations that incorrectly classified as “Non-Fall” group. Accuracy(ACC): ACC = (TF + TN)/Total number of observations. Sensitivity (SENS) & specificity (SPEC): Sensitivity measures the proportion of “Falls” that are correctly classified while specificity measures the proportion of “Non-fall” that are correctly identified: Positive Predictive Value (PPV) & Negative Predictive Value (NPV): Positive Predicted Value measures the proportion of true “Fall” observations among predicted “Fall” observations. Similarly, Negative Predicted Value measures the proportion of true “Non-fall” observations among predicted “Non-fall” observations: ROC Curve & Area Under the Curve (AUC): The Receiver Operating Characteristic (ROC) curve explicates the relation between true positive rate (i.e., sensitivity) and false positive rate (i.e. 100%-specificity) for various cut-offs of a continuous diagnostic test[31]. The performance of the test may be summarized by the aggregate area under the ROC curve (AUC); and higher AUC indicates better performance. In this study, 5-fold cross validation is applied, the AUC is calculated for each repeated iteration, and the average AUC is reported as an overall quantitative estimate of classification performance, which can be used to compare alternative classifiers[32].

Statistical tests

A number of critical features from Michigan/Tel-Aviv/Combined datasets were identified during feature selection. As observed in density plots, data of clinical measurements were not normally distributed within sub-patient groups, hence two-sample t-test cannot be used. When comparing two independent samples (fall and non-fall patient group), non-parametric tests are implemented as they have the advantage of making no assumption about data distribution.

Mann-Whitney-Wilcoxon (MWW) test

Frequently treated as the non-parametric equivalent of the two-sample t-test, the MWW test is used to determine whether two independent samples from populations having the same distributions with the same median without assuming normal distributions[33]. The calculation is based on the order of the observation in samples. In this study, we used R-based wilcox.test() to carry out two-sided hypothesis testing procedure: H0: The distributions of two samples do not differ by a location shift. H1: The distribution of one population is shifted to the left or right of the other. MWW test statistic: , where is the rank sum statistic of one group and is the number of observations in the other group whose ranks were not summed. The statistic is reported and labeled as [34].

Kolmogorov–Smirnov (KS) test

Named after Andrey Kolmogorov and Nikolai Smirnov, it is one of the most useful and general non-parametric method that determines whether two independent samples differ significantly in both location and shape of the one-dimensional probability distributions. KS test[35] quantifies the distance between the empirical distribution functions of two sample: H0: The samples are drawn from the same distribution. H1: The samples are not drawn from the same distribution. The empirical distribution function: , where n is the number of observations. Then, the KS test statistic is:where and are the empirical distribution functions of the first and second sample.

Results

Overall Summaries

Table 2 shows the basic summary statistics for the three datasets and Fig. 2 illustrates correlation heatmaps of some core data features. There are some differences between the paired correlations between features and across data archives. For instance, gait-speed is strongly negatively correlated with tremor score, PIGD score, BMI, Hoehn and Yahr scale (H&Y), and GDS-SF (Geriatric Depression Scale - short form), whereas PIGD (MDS_PIGD) is strongly-positively correlated with TUG (Timed Up and Go test), GDS-SF, BMI, and Hoehn and Yahr scale. We also found that gait speed is negatively correlated with postural stability (pos_stab). The presence of more severe postural instability and gait difficulties is not robustly correlated with the non-motor experiences of daily living in the patient. The non-motor experiences of daily living reflect impairments of cognition, mood, sleep and autonomic functions. Although axial impairments are generally associated with cognitive impairments in PD, the lack of significant associations with overall non-motor experiences of daily living may be due to the heterogeneous (cognitive and non-cognitive) nature of this MDS UPDRS subscale.
Table 2

A summary table, with selected feature pair correlations, separately for each of the three datasets used in the study.

CohortOriginal Size(n)Effective Size(m)#Features*
Michigan225(48)148**(45)179
Tel-Aviv105(41)103(41)165
Aggregated330(89)251(86)129

The values in parentheses represent the numbers of patients that had falls.

*Number of features after preprocessing.

**77 healthy controls were excluded.

Figure 2

Pair correlations of some features, separately for each of the three datasets used in the study. (A) Michigan data boxplots illustrating significant differences in MDS_TREM (p = 0.5465), MDS_PIGD (p < 0.001), H and Y scale (p < 0.001), gaitSpeed_Off (p < 0.001) between PD patients with and without a history of falls, based on MWW test. (No = 0, Yes = 1). (B) Tel-Aviv data boxplots illustrating significant differences in Tremor_score (p = 0.01094), PIGD_score (p < 0.001), H and Y scale (p < 0.001) and FOG_Q (p < 0.001) between PD patients with and without a history of falls, based on MWW test. (No = 0, Yes = 1).

A summary table, with selected feature pair correlations, separately for each of the three datasets used in the study. The values in parentheses represent the numbers of patients that had falls. *Number of features after preprocessing. **77 healthy controls were excluded. Pair correlations of some features, separately for each of the three datasets used in the study. (A) Michigan data boxplots illustrating significant differences in MDS_TREM (p = 0.5465), MDS_PIGD (p < 0.001), H and Y scale (p < 0.001), gaitSpeed_Off (p < 0.001) between PD patients with and without a history of falls, based on MWW test. (No = 0, Yes = 1). (B) Tel-Aviv data boxplots illustrating significant differences in Tremor_score (p = 0.01094), PIGD_score (p < 0.001), H and Y scale (p < 0.001) and FOG_Q (p < 0.001) between PD patients with and without a history of falls, based on MWW test. (No = 0, Yes = 1).

EDA Plots for Michigan and Tel-Aviv Data

Figure 3 demonstrates exploratory data analytics (EDA) including univariate and multivariate distributions contrasting the Michigan and Tel-Aviv populations, also see Supplementary Figures S.3 and S.4.
Figure 3

Exploratory data analytics illustrating some of the relations between falling and several clinical measures for the Michigan dataset (A) and the Tel-Aviv dataset (B), separately.

Exploratory data analytics illustrating some of the relations between falling and several clinical measures for the Michigan dataset (A) and the Tel-Aviv dataset (B), separately.

Missing Data Plots

Figure 4 illustrates the missing data patterns for both, the Michigan and the Tel-Aviv datasets. This lower dimensional projection suggests that the two cohorts are quite entangled, which may present a challenge in classification of falls/no-fall.
Figure 4

Missing patterns of Michigan (top) and Tel-Aviv (bottom) datasets. Approximately 30% of the Michigan study subjects have complete information, e.g., many cases have unrecorded genetic biomarkers. Data completeness is higher in Tel-Aviv data, missingness only occurred in about 19% of the participants.

Missing patterns of Michigan (top) and Tel-Aviv (bottom) datasets. Approximately 30% of the Michigan study subjects have complete information, e.g., many cases have unrecorded genetic biomarkers. Data completeness is higher in Tel-Aviv data, missingness only occurred in about 19% of the participants.

Challenge 1. Harmonizing and aggregating complex multi-source and multisite Parkinson’s disease data

Data Aggregation: Since the data were acquired in independent studies at two separate institutions, not all the features collected were homologous. Even common features contained in both archives had some with substantially different distributions, according to Kolmogorov–Smirnov test, Fig. 5.
Figure 5

Results of KS tests on 126 features comparing the distributions in Michigan and Tel-Aviv data. The red horizontal line represents the cutoff of −log(α), where α (desired FDR) = 0.01.

Results of KS tests on 126 features comparing the distributions in Michigan and Tel-Aviv data. The red horizontal line represents the cutoff of −log(α), where α (desired FDR) = 0.01. Figure 5 shows the Kolmogorov–Smirnov tests carried out on all the numeric features (126 in total) that were common in both, Michigan and Tel-Aviv, datasets. Some extremely small -values were slightly transformed, i.e., replaced by the minimum of the other non-zero -values, to ensure that the logarithmic y-axis scale is correctly plotted. False Discovery Rate (FDR) was used to control the false-positive rate at the level of 0.01. Thus, among the set of rejected null hypotheses, the expected proportion of false discoveries is limited to 1%. Assuming the tests are independent, the FDR control is achieved by calculating -values (Benjamini/Hochberg FDR adjusted -value[36] for each test and rejecting those with -value <0.01. The red line in Fig. 5 represents the −log(0.01) cutoff value. Table 3 shows the level of similarity between Michigan and Tel-Aviv datasets in two different types of variables (clinical/demographic and neuroimaging).
Table 3

Some of the clinical/demographic variables and many of the neuroimaging features exhibit significantly different distributions between the two datasets.

Feature CategoryFeatures of significant incongruence
Clinical/Demographic24 out of 70 (34%)
Neuroimaging54 out of 56 (96%)
Some of the clinical/demographic variables and many of the neuroimaging features exhibit significantly different distributions between the two datasets. Figure 6 includes examples of feature distributions in these two datasets showing some similarity and some differences.
Figure 6

Similarities and differences between feature distributions in the Michigan and Tel-Aviv datasets.

Similarities and differences between feature distributions in the Michigan and Tel-Aviv datasets. As the study subjects in both Michigan and Tel-Aviv datasets represent Parkinson’s disease patients, an aggregate dataset was generated to increase the number of training and testing cases and examine the performance of the predictive analytics on the complete data. We used normalization (centering and scaling) of the data elements prior to their aggregation. Figure 7 shows batch effects on the aggregate dataset using two alternative standardization techniques – normalize two data sets separately prior to aggregation vs. aggregate and normalize the combined data. To illustrate the similarities and differences between the pair of standardization techniques we show 2D projections of the data in each paradigm (top and bottom) using both multidimensional scaling (MDS)[37] (left) and t-distributed Stochastic Neighbor Embedding (tSNE)[32,38] (right).
Figure 7

Visualization of batch effects of the aggregated data using different data aggregation strategies (normalize the two data sets separately vs. normalize the combined data) using two alternative dimensionality reduction methods - MDS (left) and tSNE (right).

Visualization of batch effects of the aggregated data using different data aggregation strategies (normalize the two data sets separately vs. normalize the combined data) using two alternative dimensionality reduction methods - MDS (left) and tSNE (right). Batch effects do not represent underlying biological variability. Rather, they reflect technical sources of data variation due to handling of the samples. To untangle batch technical variation from intrinsic biomedical process variability we need to carefully select the data harmonization, normalization and aggregation strategies to avoid unintended bias. In this case, we chose to normalize each of the two datasets separately prior to their aggregation into the combined Michigan+TelAviv dataset.

Challenge 2: Identification of salient predictors associated with patients’ falls

In this part, we aim to identify for the strongest predictors for patients’ falls for each of the three datasets, Michigan, Tel-Aviv, and the aggregated Michigan+TelAviv. We carry out feature selection using two different methods: random forest (RF)[11,39] and Knockoff filtering (KO)[40]. For each dataset, both feature selection techniques identify the top 20 selected variables. MWW test and KS test are used to compare the distributions of these features between patient subgroups (Falls vs. No-falls). We aim to identify commonly selected features by both techniques that also show significant differences on the MWW and KS tests.

Michigan dataset

We consider common variables selected by both LASSO[41] and Knockoff (FDR = 0.35) as the “potentially falls-associated features”. In addition, candidate features that are significantly different on both MWW and KS tests across two cohorts (“fall” and “non-fall”) are considered “falls-associated features”. Regularized (LASSO) linear modeling rejects all genetic features, the only set of multi-level categorical features in Michigan dataset. This fact facilitates our implementation of Knockoff filtering, which is not directly applicable for multi-level categorical variables. Excluding all genetic variables, we apply Random Forest (RF) and Knockoff (KO) variable selections on all other numeric or binary features. The feature selection results are shown on Table 4 with a corresponding variable importance plots on Fig. 8. The common features selected by both methods, RF and KO, are annotated (*). The Supplementary Materials include the technical details of the two alternative feature selection strategies. RF feature section is based on fitting a number of decision trees where each node represents a single feature condition split the dataset into two branches according to an impurity measure (e.g., Gini impurity, information gain, entropy). The feature ranking reported in Table 4 reflect the frequencies that each of these top variables decreases the weighted impurity measure in multiple decision trees. KO feature selection relies on pairing each feature with a decoy variable, which resembles its characteristics but carries no signal, and optimizes an objective function that jointly estimates model coefficients and variable selection, by minimizing a the sum of the model fidelity and a regularization penalty components. The discrepancy between a real feature () and its decoy (knockoff) counterpart () is measured by a statistic like , which effectively measures how much more important is relative to . The strength of the importance of relative to is measured by the statistic magnitude, . There is a strong evidence of the importance of the commonly selected features (*) by RF and KO, see Table 4 and Fig. 8.
Table 4

Feature selection for the Michigan data using RF (left) and KO (right). Six common features (*) are selected by both methods: MDS_PIGD, gaitSpeed_Off, MOT_EDL, NON_MOTOR_EDL, walk, pos_stab.

Random ForestsKnockoff
FeaturesFrequencyFeaturesFrequency
MDS_PIGD*0.888hx_smoke0.764
gaitSpeed_Off*0.860high_bp0.751
R_middle_temporal_gyrus0.662walk*0.718
R_inferior_temporal_gyrus0.618MDS_PIGD*0.672
Caudate_DA0.554SLEEP_APNEA0.602
Striatum_DA0.534head_inj0.598
MOT_EDL*0.516SLEEP_RBD0.552
time_upgo0.494out_bed0.515
L_middle_temporal_gyrus0.436gaitSpeed_Off*0.502
NON_MOTOR_EDL*0.418HY0.477
UPSIT400.410NON_MOTOR_EDL*0.440
Putamen_DA0.408hal_psy0.415
R_middle_orbitofrontal_gyrus0.364Chair0.415
walk*0.354pos_stab*0.407
R_fusiform_gyrus0.336Caudate_DA0.403
BMI0.324MOT_EDL*0.398
L_inferior_temporal_gyrus0.322gait0.374
MDRS_PERSEV0.320gender0.361
L_insular_cortex0.318turn0.361
pos_stab*0.318depression0.324
Figure 8

Results of feature selection for the Michigan dataset using random forest (top) and knockoff filtering (bottom). The barplots present the exact number of times the top listed features are selected.

Feature selection for the Michigan data using RF (left) and KO (right). Six common features (*) are selected by both methods: MDS_PIGD, gaitSpeed_Off, MOT_EDL, NON_MOTOR_EDL, walk, pos_stab. Results of feature selection for the Michigan dataset using random forest (top) and knockoff filtering (bottom). The barplots present the exact number of times the top listed features are selected. Table 5 shows the results comparing the distributions between fallers and no-fallers in the Michigan data, using the top six common features identified by RF and KO controlled feature selection.
Table 5

MWW test and KS tests of group differences performed on the commonly selected features.

Selected FeaturesMann-Whitney-Wilcoxon TestKolmogorov-Smirnov Tests
Wp-valueDp-value
MDS_PIGD1011.53.933e-080.429341.935e-05
gaitSpeed_Off34125.082e-060.376910.0002733
MOT_EDL12538.713e-060.415753.974e-05
NON_MOTOR_EDL1486.50.00051820.276810.01647
walk11951.643e-070.418553.432e-05
pos_stab12531.255e-060.374110.0003118
MWW test and KS tests of group differences performed on the commonly selected features. Figure 9 depicts the density plots of the top six selected clinical features that have significantly different distributions between falls and no-fall subpopulations in the Michigan dataset.
Figure 9

Density plots showing the top six clinical features with significantly different distributions between falls and no-fall cohorts within the Michigan study.

Density plots showing the top six clinical features with significantly different distributions between falls and no-fall cohorts within the Michigan study.

Tel-Aviv data

Table 6 illustrates the top features selected by RF and KO methods solely on the Tel-Aviv dataset. Again, commonly selected features by both strategies are labeled (*). Figure 10 presents the Tel-Aviv RF and KO feature selection results. Table 7 contains the MWW and KS test results comparing the distributions of fallers and no-fallers. Figure 11 shows the density plots of the top 10 selected clinical features separately for falls and no-fall groups.
Table 6

13 features (*) are selected by both methods (RF and KO): gaitSpeed_Off, ABC, BMI, PIGD_score, cerebellum, X2.11, partII_sum, Attention, DGI, Tremor_score, FOG_Q, R_fusiform_gyrus, H_and_Y_OFF.

Random ForestsKnockoff
FeaturesFrequencyFeaturesFrequency
gaitSpeed_Off*0.924gender0.917
ABC*0.874X2.11*0.753
BMI*0.824ABC*0.488
PIGD_score*0.644gaitSpeed_Off*0.452
TUG_OFF0.614partII_sum*0.425
cerebellum*0.596H_and_Y_OFF*0.421
X2.110.568cerebellum*0.386
partII_sum*0.522PIGD_score*0.359
brainstem0.406FOG_Q*0.351
L_inferior_occipital_gyrus0.402X1.80.351
L_supramargiNAl_gyrus0.402BMI*0.347
Attention*0.392X3.10gait_off0.339
DGI*0.378DGI*0.296
L_hippocampus0.344Attention*0.296
L_fusiform_gyrus0.342R_fusiform_gyrus*0.238
Tremor_score*0.336X2.130.226
FOG_Q*0.328X3.17d0.211
R_fusiform_gyrus*0.328X4.30.187
R_parahippocampal_gyrus0.318Tremor_score*0.176
H_and_Y_OFF*0.308X3.130.172
Figure 10

Results of feature selection for the Tel-Aviv dataset using random forest (top) and knockoff (bottom) methods. The bar plots present the exact number of times the top features are selected.

Table 7

MWW test and KS test are performed on selected features in the Tel-Aviv data. Cerebellum, Tremor_score and R_fusiform_gyrus are excluded because their p-values > 0.05, for the KS test.

Selected FeaturesMann-Whitney-Wilcoxon TestKolmogorov-Smirnov Tests
Wp-valueDp-value
gaitSpeed_Off19573.861e-060.442170.0001288
ABC19771.927e-060.483081.988e-05
BMI8410.0038080.382770.001447
PIGD_score6271.132e-050.473253.162e-05
cerebellum*16920.0046110.253740.06936
X2.114903.008e-080.481512.143e-05
partII_sum669.55.007e-050.376480.001831
Attention17100.0031330.296620.026
DGI18624.841e-050.334780.007917
Tremor_score*1648.50.010940.272620.05103
FOG_Q8020.00010010.35090.004586
R_fusiform_gyrus*16650.0080220.254520.06705
H_and_Y_OFF752.50.00025070.341860.006249
Figure 11

Density plots showing that the top 10 selected clinical features have significantly different distributions between falls and no-fall patient groups.

13 features (*) are selected by both methods (RF and KO): gaitSpeed_Off, ABC, BMI, PIGD_score, cerebellum, X2.11, partII_sum, Attention, DGI, Tremor_score, FOG_Q, R_fusiform_gyrus, H_and_Y_OFF. Results of feature selection for the Tel-Aviv dataset using random forest (top) and knockoff (bottom) methods. The bar plots present the exact number of times the top features are selected. MWW test and KS test are performed on selected features in the Tel-Aviv data. Cerebellum, Tremor_score and R_fusiform_gyrus are excluded because their p-values > 0.05, for the KS test. Density plots showing that the top 10 selected clinical features have significantly different distributions between falls and no-fall patient groups.

Aggregated (Michigan+TelAviv)

Similar results, corresponding to the separate Michigan and Tel-Aviv results shown above, are included below for the aggregate Michigan+TelAviv dataset, Tables 8 and 9, Figs 12 and 13.
Table 8

Top seven features (*) are selected by both methods (RF and KO): gaitSpeed_Off, PIGD_score, partII_sum, BMI, X2.11, H_and_Y_OFF, X3.10gait_off.

Random ForestsKnockoff
FeaturesFrequencyFeaturesFrequency
gaitSpeed_Off*0.992X2.11*0.822
PIGD_score*0.992PIGD_score*0.784
partII_sum*0.878Gender0.742
TUG_OFF0.856X3.10gait_off*0.621
BMI*0.806H_and_Y_OFF*0.579
X2.11*0.788partII_sum*0.566
R_middle_temporal_gyrus0.632gaitSpeed_Off*0.544
H_and_Y_OFF*0.586X2.120.394
R_inferior_temporal_gyrus0.558X1.80.355
R_middle_orbitofrontal_gyrus0.406BMI*0.346
partI_sum0.404X2.80.333
L_middle_temporal_gyrus0.392MoCA0.256
L_gyrus_rectus0.384X2.90.246
X3.10gait_off*0.376X3.17d0.240
L_middle_occipital_gyrus0.354X1.90.211
R_fusiform_gyrus0.354X3.12pull_test_off0.202
L_lateral_orbitofrontal_gyrus0.352X1.100.195
L_middle_orbitofrontal_gyrus0.326X2.130.192
R_angular_gyrus0.290L_middle_frontal_gyrus0.157
L_superior_occipital_gyrus0.282X2.100.154
Table 9

MWW test and KS test results for top selected features. Weight is excluded as its p-value > 0.05 in the MWW test.

Selected FeaturesMann-Whitney-Wilcoxon TestKolmogorov-Smirnov Tests
Wp-valueDp-value
gaitSpeed_Off104428.745e-100.371393.373e-07
PIGD_score32491.172e-120.444124.128e-10
partII_sum37629.742e-100.369563.933e-07
BMI5283.50.00090810.280830.0002681
X2.1132589.779e-140.405141.741e-08
H_and_Y_OFF39181.102e-090.342923.363e-06
X3.10gait_off41893.258e-090.358141.006e-06
Figure 12

Results of feature selection for the aggregated dataset. The bar plot presents the exact number of times that the top features are selected by random forests (top) and knockoff (bottom).

Figure 13

Density plots showing 7 selected clinical features with significantly different distributions between falls and no-fall groups.

Top seven features (*) are selected by both methods (RF and KO): gaitSpeed_Off, PIGD_score, partII_sum, BMI, X2.11, H_and_Y_OFF, X3.10gait_off. MWW test and KS test results for top selected features. Weight is excluded as its p-value > 0.05 in the MWW test. Results of feature selection for the aggregated dataset. The bar plot presents the exact number of times that the top features are selected by random forests (top) and knockoff (bottom). Density plots showing 7 selected clinical features with significantly different distributions between falls and no-fall groups.

Challenge 3. Classification of patients’ falls

Below, we report the prediction results for the model-based logistic regression, used as a reference method, and machine learning classification using the normalized datasets. The results are reported separately for the Michigan only, Tel-Aviv only, and the aggregate Michigan+TelAviv datasets.

Michigan data

Table 10 shows the binary classification of fall/no-fall (5-fold CV) using all features. The columns represent seven complementary performance estimating measures: accuracy (acc), sensitivity (sens), specificity (spec), positive and negative predictive values (ppv and npv), and area under the receiver operating curve (auc).
Table 10

Performance of model-based and model-free methods (using all features).

Methodaccsensspecppvnpvlorauc
Logistic Regression0.4390.4000.4560.2430.635−0.5810.630
Random Forests0.7640.3560.9420.7270.7702.1880.727
AdaBoost0.7030.3330.8640.5170.7481.1560.695
XGBoost0.7300.3330.9030.6000.7561.5370.710
SVM0.7430.2000.9810.8180.7372.5360.750
Neural Network0.6550.4440.7480.4350.7550.863
Super Learner0.7230.2890.9130.5910.7461.445
Performance of model-based and model-free methods (using all features). Table 11 shows the binary classification of fall/no-fall (5-fold CV) using only the top 6 selected features (MDS_PIGD, gaitSpeed_Off, MOT_EDL, NON_MOTOR_EDL, walk, pos_stab).
Table 11

Performance of model-based and model-free methods (using top 6 features).

Methodaccsensspecppvnpvlorauc
Logistic Regression0.7360.2890.9320.6500.7501.7180.781
Random Forests0.7770.4440.9220.7140.7922.2510.697
AdaBoost0.7500.4440.8830.6250.7841.8030.693
XGBoost0.7770.4670.9130.7000.7972.2130.657
SVM0.7570.4670.8830.6360.7911.8920.742
Neural Network0.6690.4000.7860.4500.7500.898
Super Learner0.7840.4670.9220.7240.7982.341
Performance of model-based and model-free methods (using top 6 features).

Tel Aviv data

Table 12 illustrates the results of the binary classification of fall/no-fall (5-fold CV) using all features.
Table 12

Performance of model-based and model-free methods (using all features).

Methodaccsensspecppvnpvlorauc
Logistic Regression0.5050.3900.5810.3810.590−0.1210.603
Random Forests0.6890.5370.7900.6290.7211.4730.702
AdaBoost0.7180.6100.7900.6580.7541.7730.719
XGBoost0.6700.6100.7100.5810.7331.3400.711
SVM0.7570.5120.9190.8080.7402.4820.767
Neural Network0.6800.6590.6940.5870.7541.474
Super Learner0.6700.5120.7740.6000.7061.281
Performance of model-based and model-free methods (using all features). Table 13 shows the binary classification of fall/no-fall (5-fold CV) using top 10 selected features (gaitSpeed_Off, ABC, BMI, PIGD_score, X2.11, partII_sum, Attention, DGI, FOG_Q, H_and_Y_OFF).
Table 13

Performance of model-based and model-free methods (using top 10 selected features).

Methodaccsensspecppvnpvlorauc
Logistic Regression0.7280.5370.8550.7100.7361.9200.774
Random Forests0.7960.6830.8710.7780.8062.6770.821
AdaBoost0.6890.6100.7420.6100.7421.5020.793
XGBoost0.6990.7070.6940.6040.7821.6990.787
SVM0.7090.5610.8060.6570.7351.6720.822
Neural Network0.6990.6100.7580.6250.7461.588
Super Learner0.7380.6830.7740.6670.7871.999
Performance of model-based and model-free methods (using top 10 selected features). Improving Classification Sensitivity: We attempted to further improve the classification sensitivity, which is important in this clinical setting. As Random Forest outperforms the other methods, we focused our performance tuning on RF classification. By optimizing the RF parameters, using grant weights, setting cut off points for two classes and the number of features used for each decision tree branch split, we obtained a classification model with higher sensitivity and LOR. Although, there is more room to further improvement of the sensitivity, it is also important to keep specificity within a reasonable range. Table 14 shows the best RF results on the Tel-Aviv data. Note that improving the classifier sensitivity trades off with (compromising) it’s sensitivity.
Table 14

Fine-tuned RF classification results on the Tel-Aviv dataset.

Methodaccsensspecppvnpvlor
Random Forests0.7670.8050.7420.6730.8522.473
Fine-tuned RF classification results on the Tel-Aviv dataset. Fall prediction with a subset of important features: We applied a logit model for a low dimensional case-study. Our results show 74% prediction accuracy using four variables, Table 15. Prior work by Paul, et al.[42] reported accuracy about 80% using three variables, including “fall in the previous year” as an additional predictor, which may be very strongly associated with the clinical outcome of interest—whether a patient is expected to fall or not.
Table 15

Logit model prediction of falls in the Tel-Aviv case, using only four features.

Selected FeaturesAccSensSpecppvnpvlor
PIGD_score, FOG_Q, H&Y(OFF), gaitSpeed(Off)0.7380.4390.9350.8180.7162.429
Logit model prediction of falls in the Tel-Aviv case, using only four features. Table 16 and Fig. 14 show the areas under the ROC curve of the Random Forest classification using several different study-designs. The results suggest that four features provide sufficient predictive power to forecast fall sin PD patients (area under the ROC curve is approximately 0.8).
Table 16

Performance of the RF falls/no-fall classifier under different conditions.

All FeaturesTD/PIGD + OthersRemove UPDRSImage FeaturesSelected FeaturesFour Features
AUC0.6690.6710.6400.5590.7790.796
Figure 14

ROC plot for random Forest, lines in different colors represents the results under 6 different training conditions: (1) All features; (2) TD/PIGD classification and other clinical/demographic information; (3) Remove all UPDRS items; (4) Neuroimaging features only; (5) 10 selected features and (6) 4 vital features (PIGD Score, H_and_Y_Off, FOG_Q, gaitSpeed_Off). Corresponding Area Under the ROC Curve (AUC) are listed in Table 16.

Performance of the RF falls/no-fall classifier under different conditions. ROC plot for random Forest, lines in different colors represents the results under 6 different training conditions: (1) All features; (2) TD/PIGD classification and other clinical/demographic information; (3) Remove all UPDRS items; (4) Neuroimaging features only; (5) 10 selected features and (6) 4 vital features (PIGD Score, H_and_Y_Off, FOG_Q, gaitSpeed_Off). Corresponding Area Under the ROC Curve (AUC) are listed in Table 16. Truncated classification of multiple-falls vs. no-falls (5-fold CV): A natural consideration is that some patients with prior falls might be attributed to unrelated accidents. Therefore, we tried to accurately identify patients with multiple falls. Further, for patients who had a history of falls, including one or more falls, the observations who had presence of fall by accident could mask the key demographic/clinical predictors, associated with falls. Table 17 shows the proportion of participants with two or more falls vs. no falls and Table 18 shows the classification results using all features.
Table 17

Distribution of patients without fall history compared to patients with two or more falls.

no-fallstwo or more falls
Number of cases (%)62 (69%)28 (31%)
Table 18

Performance of model-based and model-free methods (using all features) for Tel-Aviv dataset to predict no fall or at least two falls, contrast to results in Table 10 (fall/no-fall), using the same features.

Methodaccsensspecppvnpvlorauc
Random Forests0.7670.4640.9030.6840.7892.0900.821
AdaBoost0.7890.5360.9030.7140.8122.3770.836
XGBoost0.7110.3930.8550.5500.7571.3380.848
SVM0.7330.6430.7740.5630.8281.8200.839
Neural Network0.7330.6790.7580.5590.8391.889
Super Learner0.7440.3930.9030.6470.7671.798
Distribution of patients without fall history compared to patients with two or more falls. Performance of model-based and model-free methods (using all features) for Tel-Aviv dataset to predict no fall or at least two falls, contrast to results in Table 10 (fall/no-fall), using the same features. Finally, Table 19 shows the classification using only the commonly selected features.
Table 19

Performance of model-based and model-free methods (using selected features) for Tel-Aviv dataset to predict no fall or at least two falls, contrast to results in Table 11 (falls/no-fall).

Methodaccsensspecppvnpvlorauc
Random Forests0.8110.7140.8550.6900.8692.6890.880
AdaBoost0.8220.7500.8550.7000.8832.8720.886
XGBoost0.8110.6430.8870.7200.8462.6490.885
SVM0.8330.7140.8870.7410.8732.9780.881
Neural Network0.7220.6070.7740.5480.8141.667
Super Learner0.8000.6430.8710.6920.8442.497
Performance of model-based and model-free methods (using selected features) for Tel-Aviv dataset to predict no fall or at least two falls, contrast to results in Table 11 (falls/no-fall). The best results were obtained using adaptive boosting (Adaboost)[12] and SVM with Gaussian kernel[43].

Aggregate Michigan + TelAviv Data

Table 20 shows the binary falls/non-fall classification of the mixed/aggregated data using all features (5-fold CV).
Table 20

Performance of model-based and model-free methods (using all features) on aggregated data.

Methodaccsensspecppvnpvlorauc
Logistic Regression0.5940.4880.6480.4200.7090.5660.639
Random Forests0.7370.4070.9090.7000.7461.9260.772
AdaBoost0.7170.4070.8790.6360.7401.6050.753
XGBoost0.6890.4190.8300.5630.7331.2590.734
SVM0.6290.5580.6670.4660.7430.9270.768
Neural Network0.6410.4880.7210.4770.7300.904
Super Learner0.7290.4300.8850.6610.7491.758
Performance of model-based and model-free methods (using all features) on aggregated data. Table 21 illustrates the results of the mixed/aggregated data (5-fold CV) classification using only the seven commonly selected features: gaitSpeed_Off, PIGD_score, partII_sum, BMI, X2.11, H_and_Y_OFF, X3.10gait_off.
Table 21

Performance of model-based and model-free methods (using selected features) on aggregated data.

Methodaccsensspecppvnpvlorauc
Logistic Regression0.7730.4300.9520.8220.7622.6960.817
Random Forests0.7050.4530.8360.5910.7461.4450.774
AdaBoost0.7170.5580.8000.5930.7761.6200.765
XGBoost0.7450.5470.8480.6530.7821.9090.781
SVM0.7770.5120.9150.7590.7822.4250.785
Neural Network0.6610.5120.7390.5060.7441.089
Super Learner0.7290.4530.8730.6500.7541.739
Performance of model-based and model-free methods (using selected features) on aggregated data. Train on Michigan and Test on Tel-Aviv Data: Table 22 shows the falls/no-fall classification (training on Michigan and testing on Tel-Aviv data) results using the selected features.
Table 22

Performance of model-based and model-free methods. Train on Michigan and test on Tel-Aviv data.

Methodaccsensspecppvnpvlorauc
Logistic Regression0.7180.3900.9350.8000.6992.2280.832
Random Forests0.7380.5370.8710.7330.7402.0560.796
AdaBoost0.6990.4630.8550.6790.7071.6260.791
XGBoost0.7090.4630.8710.7040.7111.7630.758
SVM0.6890.2680.9680.8460.6672.3980.827
Neural Network0.6310.5850.6610.5330.7071.014
Super Learner0.7570.5620.8870.7670.7532.31
Performance of model-based and model-free methods. Train on Michigan and test on Tel-Aviv data.

Train on Tel-Aviv and Test on Michigan Data

Table 23 shows the opposite falls/no-fall classification (training on Tel-Aviv and testing on Michigan data) results using only the commonly selected features.
Table 23

Performance of model-based and model-free methods. Train on Tel Aviv and test on Michigan.

Methodaccsensspecppvnpvlorauc
Logistic Regression0.7770.4890.9030.6880.8022.1860.794
Random Forests0.7090.6670.7280.5170.8331.6780.755
AdaBoost0.6890.6440.7090.4920.8201.4840.780
XGBoost0.7300.6000.7860.5510.8181.7090.748
SVM0.7970.4440.9510.8000.7972.7520.805
Neural Network0.6220.6440.6120.4200.7971.049
Super Learner0.7700.6440.8250.6170.8422.15
Performance of model-based and model-free methods. Train on Tel Aviv and test on Michigan.

Challenge 4. Morbidity phenotype (TD/PIGD) Classification

Next, ignoring the UPDRS subitems, we performed predictive analytics of tremor dominant (TD) vs. posture instability and gait disorder (PIGD) classification using only the demographic and clinical information (neuroimaging features were excluded).

Michigan Data

Table 24 shows that compared to prediction of falls using all features, the overall accuracy for both logistic regression and AdaBoost TD/PIGD classification is improved, compare to Table 10.
Table 24

Performance of prediction for TD/PIGD class label on Michigan dataset.

Methodaccsensspecppvnpvlor
Logistic Regression0.6150.3110.7480.3500.7130.291
Random Forests0.7430.3560.9130.6400.7641.751
AdaBoost0.7430.4220.8830.6130.7781.712
Performance of prediction for TD/PIGD class label on Michigan dataset.

Tel-Aviv Data

Table 25 demonstrated improved sensitivity of TD/PIGD classification, as compared to prediction of falls using all features (Table 12). This indicates TD/PIGD classification may also be an important predictor of patients’ falls.
Table 25

Performance of prediction for TD/PIGD class label on Tel-Aviv dataset.

Methodaccsensspecppvnpvlor
Logistic Regression0.7380.7070.7580.6590.7972.024
Random Forests0.7380.6100.8230.6940.7611.980
AdaBoost0.7280.6340.7900.6670.7661.877
Performance of prediction for TD/PIGD class label on Tel-Aviv dataset.

Aggregated Michigan+TelAviv Data

Table 26 shows a slightly higher sensitivity for random forest and AdaBoost TD/PIGD classification, compared to falls prediction using all features (Table 20). Yet, compared to within archive training with internal CV assessment, the performance of both classifiers on the aggregated dataset is less impressive, which may be explained by the heterogeneity of the sets discussed in Challenge 1.
Table 26

Performance of the prediction of TD/PIGD label on the aggregated dataset.

Methodaccsensspecppvnpvlor
Logistic Regression0.7130.2790.9390.7060.7141.792
Random Forests0.6890.4300.8240.5610.7351.264
AdaBoost0.7130.4770.8360.6030.7541.538
Performance of the prediction of TD/PIGD label on the aggregated dataset.

Discussion and Conclusions

Regarding Challenge 1 (data compilation), we carefully examined, harmonized and aggregated the two independently acquired PD datasets. The merged dataset was used to retrain the algorithms and validate their classification accuracy using internal statistical cross validation. The substantial biomedical variability in the data may explain the fact that the predictive accuracy of the falls/no-fall classification results were lower in the merged aggregated data compared to training and testing the forecasting methods on each dataset separately. Challenge 2 (feature selection for prediction of falls) showed that three variables appear to be consistently chosen in the feature selection process across Michigan, Tel-Aviv and aggregated datasets – the MDS-UPDRS PIGD subscore (MDS_PIGD), gait speed in the off state, and sum score for MDS-Part II: Motor Aspects of Experiences of Daily Living (M-EDL). This is consistent with expectations as PIGD has been previously related to fall risk in PD. In the third Challenge (prediction of falls), we found some differences between the classification results obtained by training of the three different datasets. For instance, training on the Michigan data, the highest overall classification accuracy was about 78%, with a lower sensitivity, ~47%. Whereas, training on the Tel-Aviv data, the accuracy and sensitivity rates reached 80% and 68%, respectively. For the Tel-Aviv data, the prediction model can be tuned to yield a sensitivity of 81% and accuracy of 77%. Furthermore, training on the Tel-Aviv data yields better results when the classification outcome corresponds to discriminating PD patients with multiple falls from those without falls. When training on the aggregated dataset, the falls/no-fall classification accuracy is about 70% with sensitivity around 55%. The most realistic, yet difficult, case involves external out-of-bag validation, training on one of the datasets and testing on the other. For instance, training an RF classifier on the Tel-Aviv dataset and tested it out of-bag on the Michigan dataset yields accuracy of 71% and sensitivity of 67%. The results of the last Challenge (TD/PIGD) suggest that tremor dominant (TD) vs. postural instability and gait difficulty (PIGD) classification is reliable. For example, training and statistically validating on the Tel-Aviv data yields accuracy of 74%, sensitivity of 61% and specificity of 82%. The classification performance of different machine learning methods varies with respect to the testing and training datasets. Overall, the random forests classifier works best on most combinations of training/testing datasets and feature selection strategies. The boosting method also showed high predictive classification accuracy on Tel-Aviv data. When the number of features is small, logistic regression may provide a viable model for predicting patient falls and it has always the benefit of easy intuitive interpretation within the scope of the problem. The reported variable importance results may be useful for selecting features that may be important biomarkers helping clinicians quantify the risk of falls in PD patients. This study may have some potential pitfalls and limitations. For instance, the sample sizes are relatively small, Michigan (N1 = 148) and Tel-Aviv (N2 = 103). There was significant heterogeneity of the feature distributions between the Michigan and Tel-Aviv datasets. It is not clear if there were underlying biological, clinical, physiological, or technological reasons for the observed variation. This is a common challenge in all Big data analytic studies relying on multisource heterogeneous data. Features that were completely incongruent between the two data archives were removed from the subsequent analyses and were not included in the aggregated dataset. Finally, the classifiers trained on one of the datasets (Tel-Aviv) performed better when tested either via internal statistical cross-validation or via external out-of-bag valuation (using the Michigan test data). Our study of falls primarily focused on the binary indicator of falls. The frequency of falls, or the severity of falls, were not examined due to lack of sufficient information in either data archive. However, both frequency and severity of falls require further examination.

Clinical impact

The study findings indicate that clinical markers of PIGD motor features were more robust predictors of falls than striatal dopamine bindings as measured by DTBZ VMAT2 brain PET imaging. Along the same line, typical clinical predictors of nigrostriatal dopaminergic losses, such as distal bradykinesias did not significantly predict falls in the analyses. These findings underscore the notion that falls are more related to extra-striatal and non-dopaminergic mechanisms than striatal dopamine level per se. The presented results suggest a need for new approaches for determining fall risk and motor phenotypes among patients with PD. If the conclusions are replicated on a larger scale and reproduced in prospective studies, then the methods described here can contribute to the diagnosis and prognosis, and perhaps to personalized or individualized treatment approaches.

Synergies with previous studies

We have previously shown that PD fallers did not differ in nigrostriatal dopaminergic nerve terminal integrity but had lower cholinergic brain activity compared to the PD no-fallers[44,45]. We have also shown in prior analyses that freezing of gait is most prominent with extra-striatal non-dopaminergic changes, in particular the combined presence of cholinergic denervation and β-amyloid plaque deposition[46]. Some of the clinical predictors of falls in this study, such as slow gait speed or PIGD motor feature severity have been found to associate with cortical cholinergic and β-amyloid plaque deposition, respectively[47,48] and were independent from the degree of nigrostriatal nerve terminal losses. Another interesting observation in our analyses is that brain MRI morphometry measures did not appear to be robust predictors of fall status. It should be noted that mobility functions are subserved by a widespread network of interconnected brain and extra-cranial structures (e.g., spinal cord, nerves). Therefore, it is unlikely that individual brain structures may be highly salient predictive features. In this study, infratentorial brain structures, such as the cerebellum and brainstem, performed relatively better than supratentorial brain regions. Another factor is that the etiology of falls is multi-factorial (cognitive impairment, freezing of gait, sarcopenia, postural instability) and thereby involving multiple neural and neuromuscular structures and connections. It is plausible, however, that more precise clinical sub-typing of specific fall mechanisms, may identify more vulnerable brain regions or networks of regions. There are enormous opportunities for expanding this work to include additional classifiers, explore alternative features, validate on new cohorts and translate into clinical practice. For example, utilizing novel computational models and genomic biomarkers (e.g., noncoding RNA) may improve the automated PD diagnosis. For example, publicly available archives including long noncoding RNAs[49,50], micro RNAs[51,52], or other sequence, expression, or functional data may provide additional power to reduce classification error and enhance the forecasting reproducibility. Extreme Gradient Boosting Machine or other powerful classifiers may be able to improve the diagnostic prediction by capitalizing on RNA functional similarity, disease semantic similarity, and other RNA-disease associations[53]. Knowledge-based machine learning is an alternative strategy for disease classification[54]. Combinatorial genomic signature sets[55] and molecular signaling networks[56,57] may also be useful to predict, prognosticate, or forecast motor and cognitive decline in PD. In addition, combining these approaches with metrics extracted from long-term (e.g., 24/7) monitoring of movement also holds promise for enhancing this line of work[21,58]. The present transdisciplinary work illustrates some of the advantages of open-science principles, collaborative research, and independent validation of findings. We have compiled and are sharing the entire data preprocessing pipeline, visualization tools, and analytic protocol. This promotes community-wide validation, improvements, and collaborative transdisciplinary research into other complex healthcare and biomedical challenges. The R-based Predictive Analytics source-code is released under permissive LGPL license on our GitHub repository (https://github.com/SOCR). Supplementary Materials (Appendix)
  38 in total

1.  Prospective assessment of falls in Parkinson's disease.

Authors:  B R Bloem; Y A Grimbergen; M Cramer; M Willemsen; A H Zwinderman
Journal:  J Neurol       Date:  2001-11       Impact factor: 4.849

2.  Extra-nigral pathological conditions are common in Parkinson's disease with freezing of gait: an in vivo positron emission tomography study.

Authors:  Nicolaas I Bohnen; Kirk A Frey; Stephanie Studenski; Vikas Kotagal; Robert A Koeppe; Gregory M Constantine; Peter J H Scott; Roger L Albin; Martijn L T M Müller
Journal:  Mov Disord       Date:  2014-06-07       Impact factor: 10.338

3.  RandomForest4Life: a Random Forest for predicting ALS disease progression.

Authors:  Torsten Hothorn; Hans H Jung
Journal:  Amyotroph Lateral Scler Frontotemporal Degener       Date:  2014-09       Impact factor: 4.092

4.  A multilevel-ROI-features-based machine learning method for detection of morphometric biomarkers in Parkinson's disease.

Authors:  Bo Peng; Suhong Wang; Zhiyong Zhou; Yan Liu; Baotong Tong; Tao Zhang; Yakang Dai
Journal:  Neurosci Lett       Date:  2017-04-21       Impact factor: 3.046

5.  How to identify tremor dominant and postural instability/gait difficulty groups with the movement disorder society unified Parkinson's disease rating scale: comparison with the unified Parkinson's disease rating scale.

Authors:  Glenn T Stebbins; Christopher G Goetz; David J Burn; Joseph Jankovic; Tien K Khoo; Barbara C Tilley
Journal:  Mov Disord       Date:  2013-02-13       Impact factor: 10.338

6.  A prospective study of falls in relation to freezing of gait and response fluctuations in Parkinson's disease.

Authors:  Yasuyuki Okuma; Ana Lígia Silva de Lima; Jiro Fukae; Bastiaan R Bloem; Anke H Snijders
Journal:  Parkinsonism Relat Disord       Date:  2017-10-19       Impact factor: 4.891

7.  Heterogeneity of cholinergic denervation in Parkinson's disease without dementia.

Authors:  Nicolaas I Bohnen; Martijn L T M Müller; Vikas Kotagal; Robert A Koeppe; Michael R Kilbourn; Sid Gilman; Roger L Albin; Kirk A Frey
Journal:  J Cereb Blood Flow Metab       Date:  2012-05-09       Impact factor: 6.200

8.  Objective characterization of daily living transitions in patients with Parkinson's disease using a single body-fixed sensor.

Authors:  Hagar Bernad-Elazari; Talia Herman; Anat Mirelman; Eran Gazit; Nir Giladi; Jeffrey M Hausdorff
Journal:  J Neurol       Date:  2016-05-23       Impact factor: 4.849

Review 9.  Deep learning for computational biology.

Authors:  Christof Angermueller; Tanel Pärnamaa; Leopold Parts; Oliver Stegle
Journal:  Mol Syst Biol       Date:  2016-07-29       Impact factor: 11.429

10.  EGBMMDA: Extreme Gradient Boosting Machine for MiRNA-Disease Association prediction.

Authors:  Xing Chen; Li Huang; Di Xie; Qi Zhao
Journal:  Cell Death Dis       Date:  2018-01-05       Impact factor: 8.469

View more
  23 in total

1.  XGBoost based machine learning approach to predict the risk of fall in older adults using gait outcomes.

Authors:  Byungjoo Noh; Changhong Youm; Eunkyoung Goh; Myeounggon Lee; Hwayoung Park; Hyojeong Jeon; Oh Yoen Kim
Journal:  Sci Rep       Date:  2021-06-09       Impact factor: 4.379

2.  Longitudinal clustering analysis and prediction of Parkinson's disease progression using radiomics and hybrid machine learning.

Authors:  Mohammad R Salmanpour; Mojtaba Shamsaei; Ghasem Hajianfar; Hamid Soltanian-Zadeh; Arman Rahmim
Journal:  Quant Imaging Med Surg       Date:  2022-02

3.  Prediction of Acute Kidney Injury after Liver Transplantation: Machine Learning Approaches vs. Logistic Regression Model.

Authors:  Hyung-Chul Lee; Soo Bin Yoon; Seong-Mi Yang; Won Ho Kim; Ho-Geol Ryu; Chul-Woo Jung; Kyung-Suk Suh; Kook Hyun Lee
Journal:  J Clin Med       Date:  2018-11-08       Impact factor: 4.241

Review 4.  The Role of Movement Analysis in Diagnosing and Monitoring Neurodegenerative Conditions: Insights from Gait and Postural Control.

Authors:  Christopher Buckley; Lisa Alcock; Ríona McArdle; Rana Zia Ur Rehman; Silvia Del Din; Claudia Mazzà; Alison J Yarnall; Lynn Rochester
Journal:  Brain Sci       Date:  2019-02-06

5.  Prediction of future gastric cancer risk using a machine learning algorithm and comprehensive medical check-up data: A case-control study.

Authors:  Junichi Taninaga; Yu Nishiyama; Kazutoshi Fujibayashi; Toshiaki Gunji; Noriko Sasabe; Kimiko Iijima; Toshio Naito
Journal:  Sci Rep       Date:  2019-08-27       Impact factor: 4.379

6.  Applications of machine learning techniques to predict filariasis using socio-economic factors.

Authors:  Phani Krishna Kondeti; Kumar Ravi; Srinivasa Rao Mutheneni; Madhusudhan Rao Kadiri; Sriram Kumaraswamy; Ravi Vadlamani; Suryanaryana Murty Upadhyayula
Journal:  Epidemiol Infect       Date:  2019-09-02       Impact factor: 2.451

7.  Decision Support Systems in Temporomandibular Joint Osteoarthritis: A review of Data Science and Artificial Intelligence Applications.

Authors:  Jonas Bianchi; Antonio Ruellas; Juan Carlos Prieto; Tengfei Li; Reza Soroushmehr; Kayvan Najarian; Jonathan Gryak; Romain Deleat-Besson; Celia Le; Marilia Yatabe; Marcela Gurgel; Najla Al Turkestani; Beatriz Paniagua; Lucia Cevidanes
Journal:  Semin Orthod       Date:  2021-05-19       Impact factor: 1.340

8.  Derivation and Validation of Machine Learning Approaches to Predict Acute Kidney Injury after Cardiac Surgery.

Authors:  Hyung-Chul Lee; Hyun-Kyu Yoon; Karam Nam; Youn Joung Cho; Tae Kyong Kim; Won Ho Kim; Jae-Hyon Bahk
Journal:  J Clin Med       Date:  2018-10-03       Impact factor: 4.241

9.  Selecting Clinically Relevant Gait Characteristics for Classification of Early Parkinson's Disease: A Comprehensive Machine Learning Approach.

Authors:  Rana Zia Ur Rehman; Silvia Del Din; Yu Guan; Alison J Yarnall; Jian Qing Shi; Lynn Rochester
Journal:  Sci Rep       Date:  2019-11-21       Impact factor: 4.996

10.  A comparative study of machine learning algorithms for predicting acute kidney injury after liver cancer resection.

Authors:  Lei Lei; Ying Wang; Qiong Xue; Jianhua Tong; Cheng-Mao Zhou; Jian-Jun Yang
Journal:  PeerJ       Date:  2020-02-25       Impact factor: 2.984

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.