Literature DB >> 31218287

Machine Learning Versus Logistic Regression Methods for 2-Year Mortality Prognostication in a Small, Heterogeneous Glioma Database.

Sandip S Panesar1, Rhett N D'Souza2, Fang-Cheng Yeh2,3, Juan C Fernandez-Miranda1.   

Abstract

BACKGROUND: Machine learning (ML) is the application of specialized algorithms to datasets for trend delineation, categorization, or prediction. ML techniques have been traditionally applied to large, highly dimensional databases. Gliomas are a heterogeneous group of primary brain tumors, traditionally graded using histopathologic features. Recently, the World Health Organization proposed a novel grading system for gliomas incorporating molecular characteristics. We aimed to study whether ML could achieve accurate prognostication of 2-year mortality in a small, highly dimensional database of patients with glioma.
METHODS: We applied 3 ML techniques (artificial neural networks [ANNs], decision trees [DTs], and support vector machines [SVMs]) and classical logistic regression (LR) to a dataset consisting of 76 patients with glioma of all grades. We compared the effect of applying the algorithms to the raw database versus a database where only statistically significant features were included into the algorithmic inputs (feature selection).
RESULTS: Raw input consisted of 21 variables and achieved performance of accuracy/area (C.I.) under the curve of 70.7%/0.70 (49.9-88.5) for ANN, 68%/0.72 (53.4-90.4) for SVM, 66.7%/0.64 (43.6-85.0) for LR, and 65%/0.70 (51.6-89.5) for DT. Feature selected input consisted of 14 variables and achieved performance of 73.4%/0.75 (62.9-87.9) for ANN, 73.3%/0.74 (62.1-87.4) for SVM, 69.3%/0.73 (60.0-85.8) for LR, and 65.2%/0.63 (49.1-76.9) for DT.
CONCLUSIONS: We demonstrate that these techniques can also be applied to small, highly dimensional datasets. Our ML techniques achieved reasonable performance compared with similar studies in the literature. Although local databases may be small versus larger cancer repositories, we demonstrate that ML techniques can still be applied to their analysis; however, traditional statistical methods are of similar benefit.

Entities:  

Keywords:  ANN, Artificial neural network; AUC, Area under the curve; CI, Confidence interval; DT, Decision tree; Diagnosis; Gliomas; LR, Logistic regression; Logistic regression; ML, Machine learning; Machine learning; NLR, Negative likelihood ratio; NPV, Negative predictive value; Neuro-oncology; PLR, Positive likelihood ratio; PPV, Positive predictive value; Prognostication; SVM, Support vector machine; WHO, World Health Organization

Year:  2019        PMID: 31218287      PMCID: PMC6581022          DOI: 10.1016/j.wnsx.2019.100012

Source DB:  PubMed          Journal:  World Neurosurg X        ISSN: 2590-1397


Introduction

Gliomas are a heterogeneous class of tumors comprising approximately 30% of all brain malignancies. Previously, the World Health Organization (WHO) grading system stratified them by histologic origin (i.e., astrocytoma, oligodendroglioma, mixed oligoastrocytoma, ependymoma), with additional grading (I–IV) according to pathologic features of aggression. In 2016, the WHO presented a novel classification system with incorporation of molecular biomarkers including isocitrate dehydrogenase (IDH1/IDH2) mutations, O6-methylguanine-DNA methyltransferase (MGMT) methylation, p53 and phosphate and tensin homolog (PTEN) deletion,4, 5 epidermal growth factor receptor (EGFR) amplification, 1p/19q deletions,7, 8 9p(16q) deletions, and Ki67 index. The phenotypic expression of these markers by a glioma carries unique prognostic and therapeutic implications.6, 7, 11, 12, 13, 14 Moreover, the prognostic implications of the relationship between a tumor possessing more than 1 molecular marker and a patients' baseline clinical and demographic status is not fully understood.15, 16 Existing prognostic systems separate patients into low-grade (i.e., WHO grades I and II) or high-grade (i.e., WHO grades III and IV) groups, and incorporate additional clinical features such as performance status, age, and tumor size13, 17, 18, 19, 20, 21 into their stratifications. Although some newer studies have incorporated limited molecular classification features, it is clear that older prognostic indices are likely to become obsolete in the molecular medicine era. Machine learning (ML) is a subset of computer science, whereby a computer algorithm learns from prior experience. Using specified training data with known input and output values, the ML algorithm is able to devise a set of rules which can be used as predictors for novel data with similar input characteristics to the training data. Previously, a human investigator would have to approach data collection and analysis using a set of a priori assumptions to prevent the burden of collecting data irrelevant to their hypothesis. The risk of this approach is that potentially meaningful trends caused by disregarded variables go unnoticed. ML lends itself naturally to trend delineation in large, unprocessed datasets. It may also be used for clinical prediction using known inputs and desired outputs (e.g., mortality). Moreover, when implemented in a local database, ML-derived prognosticators may take into account unique features of the local population and treatment infrastructure, making them potentially more useful than evidence from noncontiguous populations. Local databases may however be considerably smaller than large-scale cancer repositories, limiting their academic study, but potentially providing the local clinician with meaningful clinical information. Bearing these factors in mind, we aimed to apply a selection of ML algorithms to a database of 76 glioma cases to devise a 2-year mortality predictor. The complex histologic and molecular pathologic features of gliomas, combined with a series of clinical prognosticators, such as performance status, age, and treatment techniques, make them an ideal multidimensional application for ML techniques. Additionally, because of our database characteristics, we aimed to compare the performance of ML algorithms using an unprocessed dataset with a dataset where only statistically significant variables had been preselected.

ML Methods

Logistic Regression

Logistic regression (LR) (Figure 1A) is a traditional statistical method used for binary classification and has been adopted as a basic ML model. It differs from linear regression (Figure 1B) because it uses a sinusoidal curve, delineating a boundary between 2 categories. Similar to linear regression, the logarithmic function is derived from weighted transformation of the categorical data points. The regression function therefore categorizes novel inputs into 1 of 2 categories based on what side of the line its coordinates fall on.
Figure 1

Graphical representation of traditional statistical approaches to regression, with logistic (A) and linear regression (B) on the top row. The bottom row demonstrates machine learning approaches graphically, with support vector machine (C), artificial neural network (D), and decision tree (E) approaches.

Graphical representation of traditional statistical approaches to regression, with logistic (A) and linear regression (B) on the top row. The bottom row demonstrates machine learning approaches graphically, with support vector machine (C), artificial neural network (D), and decision tree (E) approaches.

Support Vector Machines

Support vector machines (SVMs) (Figure 1C) are based on the LR method and assign training examples to 1 of 2 categories, with a bisecting hyperplane separating the data points. Unlike LR, however, the optimal hyperplane bisects the points representing the largest separation between the 2 categories, and its shape may not be defined by a simple function. The algorithm is tasked with finding the data points (support vectors) defining the hyperplane and derivative line coefficients. The function can then categorize novel input values into groups falling on either side of the hyperplane, similar to LR.

Artificial Neural Networks

Artificial neural networks (ANNs) (Figure 1D) are so called because they are modelled after the layer-like histologic stratification of neurons. The input and output values represent the most superficial-but-opposing layers of the network, whereas the inner hidden layers consist of successive transformations of the input values. The algorithm therefore learns from the training set by progressive transformation of initial inputs. Values of these transformed inputs are then used by the model to predict output values.

Decision Trees

Decision tree (DT) (Figure 1E) algorithms split data into binary categories using progressive iterations. ML algorithms aim to find optimal features at which to perform data splitting, creating a branching tree–shaped diagram. Each node represents a point at which the data are split, and the leaves at the end of the tree are the output variables. Because the method involves binary classification, categorical data are preferred, whereas noncategorical data are preferably discretized prior to input.

Methods

Study Population

Our study population consisted of 76 patients (40 women and 36 men) with WHO grade I–IV gliomas, presenting to the neurosurgical oncology service at the University of Pittsburgh Medical Center from 2009 to 2017. At the end of the 2-year follow-up period, 52 patients were alive, whereas 24 had died. The mean age for the whole population at diagnosis was 47.3 ± 16.8 years. Interventions included total or subtotal resection (as stated by the operating surgeon), stereotactic biopsy, gamma knife therapy, or no intervention. Other information collected included radiologic maximum tumor diameter (centimeters); tumor location (lobe); pre- and postoperative Eastern Co-operative Oncology Group (ECOG) Performance Status score (0–5); whether the patient underwent subsequent chemotherapy, radiotherapy, or vaccine therapy; or had more than 1 surgical intervention. Surgical histopathology data included the presence of EGFR amplification, PTEN deletion, p53 mutation, 1p deletion, 19q deletion, 9p(p16) deletion, IDH1/IDH2 mutations, MGMT methylation, and Ki67 proliferation index.

Study Design

Because of the relatively small number of subjects in our database (N = 76), and the high dimensionality of the data, with 21 variables, we adopted 2 approaches to ML for this population (Figure 2). The first was to apply the algorithms to the raw dataset, for which input variables had not been preselected. The second was to apply χ2 (for categorical variables) and independent samples t tests (for continuous variables) to the dataset, as outlined by Oermann et al. to discern features with influence upon mortality (“feature selection”). As this involved a number of independent statistical tests, Bonferroni correction was subsequently applied. Fourteen variables were therefore identified for which there was a significant difference between subjects who survived 2 years and those who did not. Nonsignificant variables were excluded from the input (Table 1).
Figure 2

Diagram of study design demonstrating both approaches to machine learning (ML). Left demonstrates the raw approach, and right demonstrates the initial statistical testing of variable significance prior to ML input (feature selection). Outputs (performance) were analyzed independently and then compared. ANN, artificial neural network; DT, decision tree; LR, logistic regression; SVM, support vector machine.

Table 1

Demographic and Variable Features of the Population Categorized by 2-Year Survival

VariableTotal (N = 76)Dead at 2 Years (n = 24)Alive at 2 Years (n = 52)StatisticP Value
Age (years)47.29 ± 16.7860.48 ± 14.0343.10 ± 15.854.81<0.05
Sex
 Male3714233.840.25
 Female391029
Average diameter (cm)3.41 ± 1.613.403.42–0.060.95
Initial intervention10.69<0.05
 Total resection29524
 Subtotal resection381820
 Biopsy only606
 Gamma knife211
 None101
ECOG Performance Status
 Preoperative score1.70 ± 0.671.921.601.980.05
 Postoperative score1.55 ± 0.851.921.382.44<0.05
Adjunctive treatment0.060.97
 Chemotherapy511833
 Radiotherapy481830
 Vaccine312
Number of surgeries0.220.64
 1511734
 >125718
Lobe10.160.12
 Frontal28523
 Temporal221111
 Parietal211
 Occipital202
 Brainstem202
 Other303
 Multiple17710
WHO grade16.73<0.05
 1606
 224222
 3826
 4382018
Molecular features (number unknown)23.71<0.05
 EGFR amplification2112 (1)9 (2)
 PTEN deletion3016 (1)14 (2)
 p53 mutation296 (1)23 (2)
 1p deletion152 (1)13 (2)
 19q deletion195 (1)14 (2)
 9p(p16) deletion3414 (1)20 (2)
 IDH1 mutation243 (1)21 (2)
 IDH2 mutation30 (1)3 (2)
 MGMT methylation3510 (1)25 (2)
 Ki67 index18.80 ± 16.7327.9014.603.74<0.05

Values are mean ± SD, number of patients, or as otherwise indicated.

WHO, World Health Organization; MGMT, O6-methylguanine-DNA methyltransferase; PTEN, phosphate and tensin homolog.

Statistic is either χ2 (categorical variables) or T statistic (continuous variables).

Because multiple independent statistical tests were performed, P values have been adjusted via application of Bonferroni correction.

Diagram of study design demonstrating both approaches to machine learning (ML). Left demonstrates the raw approach, and right demonstrates the initial statistical testing of variable significance prior to ML input (feature selection). Outputs (performance) were analyzed independently and then compared. ANN, artificial neural network; DT, decision tree; LR, logistic regression; SVM, support vector machine. Demographic and Variable Features of the Population Categorized by 2-Year Survival Values are mean ± SD, number of patients, or as otherwise indicated. WHO, World Health Organization; MGMT, O6-methylguanine-DNA methyltransferase; PTEN, phosphate and tensin homolog. Statistic is either χ2 (categorical variables) or T statistic (continuous variables). Because multiple independent statistical tests were performed, P values have been adjusted via application of Bonferroni correction.

Data Collection, Information Encoding, and Dataset Splitting

The raw data were collected using Microsoft Excel (Microsoft Corp., Redmond, Washington, USA). The data were parsed using Python 2.7 programming language (Python Software Foundation, Beaverton, OR, USA), using a custom written code. We used binary notation for ordinal variables (i.e., yes = 1, no = 0). Categorical and continuous variables were scaled (e.g., for the Ki67 index and age at diagnosis) to values between 0 and 1. Scaling was done using normalization (unit length vectors) and minimum-maximum scaling techniques, implemented in scikit-learn's preprocessing libraries. The continuous variables were age, maximum tumor diameter, and Ki67 index. Categorical variables were total resection, ECOG Performance Status, lobe/area of brain affected, and WHO grade. There were 3 subjects whose surgical pathology results were unavailable. Instead of discarding these from analysis, we assigned a value of 0.5 for each variable (e.g., IDH1/IDH2, PTEN). This was done to reflect the common situation where clinical data is partly missing from records. All the features were then normalized using a normal vector. The dataset was partitioned using a 70/30 training/testing split, meaning that 53 subjects were used for training and 23 were used for testing for each cycle of each algorithm.

ML Algorithms

All ML and LR models were imported from the scikit-learn library. All models were run 15 times for each model, in an attempt to reduce the problem of overfitting due to small database size. Each cycle consisted of a training and testing stage, where the dataset was repetitively partitioned. Per cycle, the same subject was not used for both training and testing. The subjects used (and their characteristics) for training and testing varied between cycles and algorithms. The number of dead and alive participants in the training and testing sets did vary between cycles however. Metrics presented are averaged figures from the 15 testing cycles for each ML method.

ANN Method

Our ANN method used a single layer of neurons between the input and output layers. The intermediate layer contained 100 neurons, each with a mini-batch size of 5. The network was trained using 1000 epochs, using an Adam optimizer, with a default 0.001 learning rate. Briefly, the Adam optimizer is an algorithm for first-order gradient-based optimization, which is an extension to stochastic gradient descent.

DT Method

The criteria used to split each node was determined by the Gini index, a standard measure of information gain in DT applications. This represents a more intuitive approach than randomly selecting criteria at which to split data. The minimum number of samples for each leaf was 1, whereas the minimum number of samples to split a node was 2.

SVM Method

Our SVM model used a radial basis function Kernel, with a C-penalty parameter of 100 and a gamma value of 0.1.

LR

LR was the benchmark, traditional statistical method we used for comparison with the performance of the ML algorithms. Nevertheless, it was also implemented using the same platform (scikit-learn) as the ML algorithms. The penalization parameter used was l2 norm. The C parameter was 150.0, and the optimization algorithm used was coordinate descent.

Data Processing

The averaged output values from the 15 cycles were then tabulated into standardized 4 × 4 confusion matrices. The sensitivity, specificity, positive likelihood ratio (PLR), negative likelihood ratio (NLR), positive predictive value (PPV), negative predictive value (NPV), and overall accuracy were calculated. All probabilities were calculated to 95% certainty. Receiver operating curves and the area under the receiver operating curves were additionally calculated and tabulated using the roc_curve model imported from the scikit-learn toolbox. To optimize comparison between accuracy (percentages) and area under the curve (AUC) (ratio), we multiplied AUC results by 100.

Results

Comparison of Diagnostic Performance

For raw data, the ANN method performed best in terms of sensitivity (81.54%), followed by the SVM (79.31%), LR (76.75%), and DT (73.65%) methods. Using a feature-selected dataset, sensitivity decreased for DT (68.93%), ANN (78.39%), and LR (74.26%), but increased slightly for SVM (80.54%). Using a feature-selected dataset, the specificity of all algorithms increased for all methods, with ANN performance showing the biggest increase (+11.62%) and DT showing the smallest (+7.56%). Using a feature-selected versus a raw dataset, all methods demonstrated a performance increase in terms of PPV (SVM = +7.69%; ANN = +7.08%; LR = +7.03%; DT = +5.87%), whereas all (DT = −6.21%; ANN = −3.79%; LR = −3.37%) but SVM (+0.54%) demonstrated a decrease in NPV performance. Likewise, ANN (+0.42), SVM (+0.36), LR (+0.28), and DT (+0.14) demonstrated an increase in PLR performance using a feature-selected dataset. In terms of NLR, all methods (SVM = −0.10; LR = −0.04; ANN = −0.02) aside from DT (+0.55) demonstrated a decrease in NLR prediction. All methods demonstrated an increase in accuracy using the feature-selected dataset (SVM = +5.38%; ANN = +2.71%; LR = +2.62%; DT = +0.17%). Finally, feature-selection increased overall performance, as represented by the AUC for all methods (LR = +8.58%; ANN = +6.21%; SVM = +2.83%) aside from DT, which demonstrated a decrease in the AUC (−7.54%) (Table 2; Figure 3).
Table 2

Performance for All Machine Learning Categories

ANN: Raw Data
SVM: Raw Data
DT: Raw Data
LR: Raw Data
AliveDeadAliveDeadAliveDeadAliveDead
Predicted alive19.136.20Predicted alive18.676.67Predicted alive17.336.40Predicted alive18.066.53
Predicted dead4.336.26Predicted dead4.875.80Predicted dead6.206.07Predicted dead5.475.93

Smaller tables are 2 × 2 confusion matrices containing the averaged output variables of the 15 cycles of machine learning for each algorithm. Underneath each confusion matrix is the performance of each test, calculated from the matrix and given to 95% CI. The upper 2 rows of the tables are for the raw datasets, and the lower 2 rows of tables are for the feature selected datasets.

ANN, artificial neural network; SVM, support vector machine; DT, decision tree; LR, logistic regression; CI, confidence interval; PLR, positive likelihood ratio; NLR, negative likelihood ratio; PPV, positive predictive value; NPV, negative predictive value.

Figure 3

A 4 × 3 array of figures demonstrating the algorithm performance using both raw data (far left column) and feature-selected (middle column) approaches. Change in performance between raw and feature-selected data is demonstrated in the far-right column. The first row shows sensitivity versus specificity performance; the second row shows positive predictive value versus negative predictive value performance; the third row shows positive likelihood ratio versus negative likelihood ratio performance; and the fourth row shows accuracy versus area under the curve performance. *Area under the curve metrics have been scaled to 100 to correlate with accuracy. ANN, artificial neural network; AUC, area under the curve; DT, decision tree; LR, logistic regression; NLR, negative likelihood ratio; NPV, negative predictive value; PLR, positive likelihood ratio; PPV, positive predictive value; SVM, support vector machine.

Performance for All Machine Learning Categories Smaller tables are 2 × 2 confusion matrices containing the averaged output variables of the 15 cycles of machine learning for each algorithm. Underneath each confusion matrix is the performance of each test, calculated from the matrix and given to 95% CI. The upper 2 rows of the tables are for the raw datasets, and the lower 2 rows of tables are for the feature selected datasets. ANN, artificial neural network; SVM, support vector machine; DT, decision tree; LR, logistic regression; CI, confidence interval; PLR, positive likelihood ratio; NLR, negative likelihood ratio; PPV, positive predictive value; NPV, negative predictive value. A 4 × 3 array of figures demonstrating the algorithm performance using both raw data (far left column) and feature-selected (middle column) approaches. Change in performance between raw and feature-selected data is demonstrated in the far-right column. The first row shows sensitivity versus specificity performance; the second row shows positive predictive value versus negative predictive value performance; the third row shows positive likelihood ratio versus negative likelihood ratio performance; and the fourth row shows accuracy versus area under the curve performance. *Area under the curve metrics have been scaled to 100 to correlate with accuracy. ANN, artificial neural network; AUC, area under the curve; DT, decision tree; LR, logistic regression; NLR, negative likelihood ratio; NPV, negative predictive value; PLR, positive likelihood ratio; PPV, positive predictive value; SVM, support vector machine.

Receiver Operating Curves and Confidence Intervals

When comparing the receiver operating curves performance to that of y = x, with an area of 0.5 (50), the SVM (AUC = 71.88) demonstrated the best performance, followed by DT (AUC = 70.54), ANN (AUC = 69.19), and LR (AUC = 64.29). Although these were higher than 0.5, the 95% confidence intervals (CIs) for both the ANN (49.86–88.52) and LR (43.63–84.95) both included 50, indicating non-significance. Even though the SVM (53.40–90.36) and DT (51.62–89.46) algorithms had CI values more than 50, these were only marginally greater than 50. The feature-selected datasets provided a performance increase for all but the DT algorithms, which demonstrated a decrease in AUC value. The performance benefit was indicated by higher AUC values, with ANN (AUC = 75.40) performing best, followed by SVM (AUC = 74.71), LR (AUC = 72.87), and DT (AUC = 63.00). Using feature-selected data also yielded overall narrower 95% CIs, with all methods aside from DT demonstrating at least a 10-unit increase of lower CI boundary above 50, indicating significance over random guessing and use of raw data. Nevertheless, for both feature-selected and raw data, none of the ML methods demonstrated significant performance improvement versus LR, nor over one another (Table 3; Figure 4).
Table 3

Receiver Operating Curve Characteristics for Uncensored and Censored Approaches

Uncensored
Feature Selected
AlgorithmAUCSE95% CI (Lower)95% CI (Upper)AlgorithmAUCSE95% CI (Lower)95% CI (Upper)
ANN69.199.8649.8688.52ANN75.406.4062.9087.90
SVM71.889.4353.4090.36SVM74.716.5062.1087.40
DT70.549.6551.6289.46DT63.007.1049.1076.90
LR64.2910.5443.6384.95LR72.876.6060.0085.80

Machine learning versus LR methods for 2-year mortality prognostication in a small, heterogeneous glioma database.

AUC, area under the curve; CI, confidence interval; ANN, artificial neural network; SVM, support vector machine; DT, decision tree; LR, logistic regression.

Figure 4

Receiver operating curves for raw (left) and feature-selected (right) data. Lines are color coordinated using the figure legend in the bottom right corner of the graph. Perforated diagonal line is y = x, with an area under the curve of 0.5 (indicative of random guessing). The performance increase is discerned by the increased distance between all curves and that of y = x. The lower right column chart demonstrates the scaled 95% confidence intervals of the area under the curves calculated for each machine learning method. Artificial neural network and logistic regression methods were not statistically different from 0.5. Support vector machine and decision tree methods were better than random guessing; however, their statistical significance was weak. Performance of algorithms using raw datasets can be compared with the chart for the feature-selected dataset in the lower right corner. These are the 95% confidence intervals for all machine learning methods, and can be concluded to be not only further away from 0.5 but also narrower, indicating greater significance. ANN, artificial neural network; AUC, area under the curve; C.I., confidence interval; DT, decision tree; LR, logistic regression; ROC, receiver operating curve; SVM, support vector machine.

Receiver Operating Curve Characteristics for Uncensored and Censored Approaches Machine learning versus LR methods for 2-year mortality prognostication in a small, heterogeneous glioma database. AUC, area under the curve; CI, confidence interval; ANN, artificial neural network; SVM, support vector machine; DT, decision tree; LR, logistic regression. Receiver operating curves for raw (left) and feature-selected (right) data. Lines are color coordinated using the figure legend in the bottom right corner of the graph. Perforated diagonal line is y = x, with an area under the curve of 0.5 (indicative of random guessing). The performance increase is discerned by the increased distance between all curves and that of y = x. The lower right column chart demonstrates the scaled 95% confidence intervals of the area under the curves calculated for each machine learning method. Artificial neural network and logistic regression methods were not statistically different from 0.5. Support vector machine and decision tree methods were better than random guessing; however, their statistical significance was weak. Performance of algorithms using raw datasets can be compared with the chart for the feature-selected dataset in the lower right corner. These are the 95% confidence intervals for all machine learning methods, and can be concluded to be not only further away from 0.5 but also narrower, indicating greater significance. ANN, artificial neural network; AUC, area under the curve; C.I., confidence interval; DT, decision tree; LR, logistic regression; ROC, receiver operating curve; SVM, support vector machine.

Discussion

We have successfully demonstrated the application of 3 ML techniques and a ML-implemented LR technique to a database of 76 patients with glioma of all stages, molecular phenotypes, and heterogeneous clinical characteristics. Relative to older, published prognostic studies, which do not incorporate molecular features, our study involves considerably fewer subjects. We accomplished our goal of applying ML techniques to this database with a relatively low subject number/variable ratio; furthermore, we demonstrate that ML can be applied with a reasonable level of confidence to make prognostic inferences from this data.

Comparison with Similar Studies

In the neuro-oncology literature, much focus of ML application has been directed toward discernment of magnetic resonance imaging characteristics of central nervous system tumors (subsequently discussed). Only 1 non–imaging-focused study has used ML for glioma outcome prediction, whereas the study by Oermann et al. used a similar methodology for cerebral metastasis prognostication. The study by Malhotra et al. applied a novel data mining algorithm to extract relevant features pertaining to treatment and molecular patterns in a database of 300 newly diagnosed glioblastoma multiforme cases. The ML component of their study involved the extraction of relevant treatment and pathologic features, which were then classified and subjected to classical statistical methods for prognostication. This is in effect the opposite approach to our method of using feature-selected data, as the authors used data mining to extract relevant features which were subsequently subjected to statistical testing, whereas we conducted statistical tests of significance for feature-selection prior to ML implementation. They achieved maximal C values of 0.85 using LR and 0.84 using Cox multivariate regression. The study by Oermann et al., although pertaining to cerebral metastases rather than gliomas, used a similar methodology to our feature-selected approach to prognosticate 1-year survival in a total of 196 patients. In this study, the pooled voting results of 5 independent ANNs (AUC = 84%) significantly outperformed traditional LR methods (AUC = 75%). Further, they found that ML techniques were more accurate at predicting 1-year survival than 2 traditional prognostic indices. Because our study used data from gliomas of all stages, we did not compare our results to existing prognostic indices, which specifically differentiate patients into low- and high-grade categories. Using a feature-selected approach, our best performing algorithm (which was coincidentally also an ANN) achieved approximately 10-unit lower AUC metric than their ANN approach. We suspect that this is for 2 reasons. First, their training set consisted of 98 patients, which was over twice the size of our training set of 40, offering more examples to learn from. Second, their method only used 6 input variables, compared to 21 for our raw approach and 14 for our feature-selected approach. It is therefore likely that increased proportionality of subjects to variables in their dataset also enhanced predictive performance of the ML algorithm by providing a less-noisy dataset. From this, it is apparent that smaller datasets may require feature selection prior to ML application if predictive performance is to be maximized. We cannot conclude that for small, highly dimensional datasets, ML approaches including ANN, DT, or SVM offer any significant performance advantage over traditional LR methods. Nevertheless, we achieved reasonably good predictive metrics using feature selected data with all ML approaches.

Future Directions

ML algorithms have been intuitively applied to data-rich magnetic resonance imaging sequences in an effort to quantitatively discern characteristic imaging features of gliomas.32, 33, 34, 35, 36 These methods have yielded the ability to discern occult imaging features not detectable by humans and which indiate the presence of MGMT methylation, IDH1 mutation,37, 38, 39 and 1p/19q co-deletion. This approach may potentially allow for the noninvasive identification40, 41 and even prognostication41, 42 of gliomas using imaging characteristics alone. It is not unreasonable to suggest that the next generation of prognostic indices will be derived from a combination of clinical database mining techniques, such as our present study combined with novel techniques of image-based ML. This will represent a substantial step forward because previous prognostic systems relied on invasive methods for definitive diagnosis, prognostication, and treatment stratification. It may also permit clinicians to prognose the clinical course of low-grade tumors noninvasively and with greater accuracy by using information from local databases to guide clinical decision making, rather than relying upon data from non-contiguous populations, which may be subject to confounding (and potentially clinically-significant) genetic and environmental effects. Depending on the integrity and scale of the localized database, predictions can be made with reasonable accuracy, as we have demonstrated in the present study.

Limitations

Although we achieved acceptable predictive performance using feature-selected data, our study has highlighted potential difficulties of ML application to smaller, highly dimensional clinical databases. Feature selection of relevant data may optimize ML algorithms in studies using smaller subject sets; however, censoring of particular variables may result in weaker trends going unnoticed. We also anticipate that the predictive accuracy and AUC would improve by increasing the number of subjects included in the training set. Despite our study using a training set less than half the size of that of Oermann et al., we achieved only slightly weaker prognostic performance. Nevertheless, the relative success of our algorithms could also be attributed to the potential effect of including both low- and high-grade tumors, which have significantly different prognostic profiles and which may exert a skewing effect on the data. An alternative to the manual statistical testing of variables for feature selection is principal component analysis, which is an entirely ML-based method of reducing dataset dimensionality, therefore reducing scope for human error or bias. Another important concern with small datasets is overfitting of the data, which is when the models, having had few data to train with, cannot appropriately anticipate novel data with fundamentally different data parameters, or when outlier values in the data exert a substantial effect which is not realistic and which impose a penalty on the models' overall accuracy. We attempted to minimize this problem by running 15 cycles of each algorithm using the same split proportions and with different subjects used for training and testing in each cycle. Feature selection reduced the variability (and dimensionality) of data without reducing database size and improved algorithmic performance, which is another method of reducing the effect of data overfitting. Other methods to reduce the issue of overfitting include early stoppage of training (i.e., before accuracy decreases), ensembling (using multiple models in parallel), and dropouts for neural networks. When selecting variables to include for our data collection, we attempted to extract as much information as possible for each subject. Nevertheless, some of the features we included are considered unconventional from a traditional neurosurgical and oncologic standpoint (e.g., use of intraoperatively determined total or subtotal resection extent [which was not a significant effector of survivability and was therefore excluded during the process of feature selection anyway]). Nevertheless, the purpose of this paper was to demonstrate the potential of ML in the neurosurgical domain and provide a blueprint for neurosurgeons wishing to implement the methodology on their own datasets.

Conclusions

As clinical approaches to gliomas are beginning to adapt to the molecular-medicine era, the small size of a local database does not provide a barrier to the implementation of ML techniques for prognostication purposes. Although our study was purely academic, it demonstrates the potential for ML to provide meaningful insight into the diagnosis and treatment of these heterogeneous tumors at a local level.
  39 in total

1.  What is principal component analysis?

Authors:  Markus Ringnér
Journal:  Nat Biotechnol       Date:  2008-03       Impact factor: 54.908

2.  Radiotherapy and temozolomide for newly diagnosed glioblastoma: recursive partitioning analysis of the EORTC 26981/22981-NCIC CE3 phase III randomized trial.

Authors:  René-Olivier Mirimanoff; Thierry Gorlia; Warren Mason; Martin J Van den Bent; Rolf-Dieter Kortmann; Barbara Fisher; Michele Reni; Alba A Brandes; Jüergen Curschmann; Salvador Villa; Gregory Cairncross; Anouk Allgeier; Denis Lacombe; Roger Stupp
Journal:  J Clin Oncol       Date:  2006-06-01       Impact factor: 44.544

3.  PTEN mutation, EGFR amplification, and outcome in patients with anaplastic astrocytoma and glioblastoma multiforme.

Authors:  J S Smith; I Tachibana; S M Passe; B K Huntley; T J Borell; N Iturria; J R O'Fallon; P L Schaefer; B W Scheithauer; C D James; J C Buckner; R B Jenkins
Journal:  J Natl Cancer Inst       Date:  2001-08-15       Impact factor: 13.506

4.  Long-term efficacy of early versus delayed radiotherapy for low-grade astrocytoma and oligodendroglioma in adults: the EORTC 22845 randomised trial.

Authors:  M J van den Bent; D Afra; O de Witte; M Ben Hassel; S Schraub; K Hoang-Xuan; P-O Malmström; L Collette; M Piérart; R Mirimanoff; A B M F Karim
Journal:  Lancet       Date:  2005 Sep 17-23       Impact factor: 79.321

5.  Clinical and molecular characteristics of malignant transformation of low-grade glioma in children.

Authors:  Alberto Broniscer; Suzanne J Baker; Alina N West; Melissa M Fraser; Erika Proko; Mehmet Kocak; James Dalton; Gerard P Zambetti; David W Ellison; Larry E Kun; Amar Gajjar; Richard J Gilbertson; Christine E Fuller
Journal:  J Clin Oncol       Date:  2007-02-20       Impact factor: 44.544

6.  Molecular subclasses of high-grade glioma predict prognosis, delineate a pattern of disease progression, and resemble stages in neurogenesis.

Authors:  Heidi S Phillips; Samir Kharbanda; Ruihuan Chen; William F Forrest; Robert H Soriano; Thomas D Wu; Anjan Misra; Janice M Nigro; Howard Colman; Liliana Soroceanu; P Mickey Williams; Zora Modrusan; Burt G Feuerstein; Ken Aldape
Journal:  Cancer Cell       Date:  2006-03       Impact factor: 31.743

7.  Clinical prognostic factors in patients with malignant glioma treated with combined modality approach.

Authors:  Branislav Jeremic; Biljana Milicic; Danica Grujicic; Aleksandar Dagovic; Jasna Aleksandrovic; Nebojsa Nikolic
Journal:  Am J Clin Oncol       Date:  2004-04       Impact factor: 2.339

8.  Temozolomide for low-grade gliomas: predictive impact of 1p/19q loss on response and outcome.

Authors:  G Kaloshi; A Benouaich-Amiel; F Diakite; S Taillibert; J Lejeune; F Laigle-Donadey; M-A Renard; W Iraqi; A Idbaih; S Paris; L Capelle; H Duffau; P Cornu; J-M Simon; K Mokhtari; M Polivka; A Omuro; A Carpentier; M Sanson; J-Y Delattre; K Hoang-Xuan
Journal:  Neurology       Date:  2007-05-22       Impact factor: 9.910

9.  IDH1 and IDH2 mutations in gliomas.

Authors:  Hai Yan; D Williams Parsons; Genglin Jin; Roger McLendon; B Ahmed Rasheed; Weishi Yuan; Ivan Kos; Ines Batinic-Haberle; Siân Jones; Gregory J Riggins; Henry Friedman; Allan Friedman; David Reardon; James Herndon; Kenneth W Kinzler; Victor E Velculescu; Bert Vogelstein; Darell D Bigner
Journal:  N Engl J Med       Date:  2009-02-19       Impact factor: 176.079

10.  p53 and Pten control neural and glioma stem/progenitor cell renewal and differentiation.

Authors:  Hongwu Zheng; Haoqiang Ying; Haiyan Yan; Alec C Kimmelman; David J Hiller; An-Jou Chen; Samuel R Perry; Giovanni Tonon; Gerald C Chu; Zhihu Ding; Jayne M Stommel; Katherine L Dunn; Ruprecht Wiedemeyer; Mingjian J You; Cameron Brennan; Y Alan Wang; Keith L Ligon; Wing H Wong; Lynda Chin; Ronald A DePinho
Journal:  Nature       Date:  2008-10-23       Impact factor: 49.962

View more
  15 in total

1.  A Brief History of Machine Learning in Neurosurgery.

Authors:  Andrew T Schilling; Pavan P Shah; James Feghali; Adrian E Jimenez; Tej D Azad
Journal:  Acta Neurochir Suppl       Date:  2022

Review 2.  Artificial intelligence-based clinical decision support in pediatrics.

Authors:  Sriram Ramgopal; L Nelson Sanchez-Pinto; Christopher M Horvat; Michael S Carroll; Yuan Luo; Todd A Florin
Journal:  Pediatr Res       Date:  2022-07-29       Impact factor: 3.953

3.  Machine Learning Algorithms for understanding the determinants of under-five Mortality.

Authors:  Rakesh Kumar Saroj; Pawan Kumar Yadav; Rajneesh Singh; Obvious N Chilyabanyama
Journal:  BioData Min       Date:  2022-09-24       Impact factor: 4.079

4.  Artificially-reconstructed brain images with stroke lesions from non-imaging data: modeling in categorized patients based on lesion occurrence and sparsity.

Authors:  Stephanie Sutoko; Hirokazu Atsumori; Akiko Obata; Ayako Nishimura; Tsukasa Funane; Masashi Kiguchi; Akihiko Kandori; Koji Shimonaga; Seiji Hama; Toshio Tsuji
Journal:  Sci Rep       Date:  2022-06-16       Impact factor: 4.996

5.  Artificial Neural Network and Cox Regression Models for Predicting Mortality after Hip Fracture Surgery: A Population-Based Comparison.

Authors:  Cheng-Yen Chen; Yu-Fu Chen; Hong-Yaw Chen; Chen-Tsung Hung; Hon-Yi Shi
Journal:  Medicina (Kaunas)       Date:  2020-05-19       Impact factor: 2.430

6.  Automatic evaluation of contours in radiotherapy planning utilising conformity indices and machine learning.

Authors:  Samsara Terparia; Romaana Mir; Yat Tsang; Catharine H Clark; Rushil Patel
Journal:  Phys Imaging Radiat Oncol       Date:  2020-12-01

7.  Machine Learning Models for Predicting Neonatal Mortality: A Systematic Review.

Authors:  Cheyenne Mangold; Sarah Zoretic; Keerthi Thallapureddy; Axel Moreira; Kevin Chorath; Alvaro Moreira
Journal:  Neonatology       Date:  2021-07-14       Impact factor: 4.035

8.  Predictive Modeling of Outcomes After Traumatic and Nontraumatic Spinal Cord Injury Using Machine Learning: Review of Current Progress and Future Directions.

Authors:  Omar Khan; Jetan H Badhiwala; Jamie R F Wilson; Fan Jiang; Allan R Martin; Michael G Fehlings
Journal:  Neurospine       Date:  2019-12-31

9.  Comparison of deep learning with regression analysis in creating predictive models for SARS-CoV-2 outcomes.

Authors:  Ahmed Abdulaal; Aatish Patel; Esmita Charani; Sarah Denny; Saleh A Alqahtani; Gary W Davies; Nabeela Mughal; Luke S P Moore
Journal:  BMC Med Inform Decis Mak       Date:  2020-11-19       Impact factor: 2.796

10.  Efficacy of deep learning methods for predicting under-five mortality in 34 low-income and middle-income countries.

Authors:  Adeyinka Emmanuel Adegbosin; Bela Stantic; Jing Sun
Journal:  BMJ Open       Date:  2020-08-16       Impact factor: 2.692

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.