Literature DB >> 32790207

Deep learning-based survival analysis for brain metastasis patients with the national cancer database.

Noah Bice1, Neil Kirby1, Tyler Bahr1, Karl Rasmussen1, Daniel Saenz1, Timothy Wagner1, Niko Papanikolaou1, Mohamad Fakhreddine1.   

Abstract

PURPOSE: Prognostic indices such as the Brain Metastasis Graded Prognostic Assessment have been used in clinical settings to aid physicians and patients in determining an appropriate treatment regimen. These indices are derivative of traditional survival analysis techniques such as Cox proportional hazards (CPH) and recursive partitioning analysis (RPA). Previous studies have shown that by evaluating CPH risk with a nonlinear deep neural network, DeepSurv, patient survival can be modeled more accurately. In this work, we apply DeepSurv to a test case: breast cancer patients with brain metastases who have received stereotactic radiosurgery.
METHODS: Survival times, censorship status, and 27 covariates including age, staging information, and hormone receptor status were provided for 1673 patients by the NCDB. Monte Carlo cross-validation with 50 samples of 1400 patients was used to train and validate the DeepSurv, CPH, and RPA models independently. DeepSurv was implemented with L2 regularization, batch normalization, dropout, Nesterov momentum, and learning rate decay. RPA was implemented as a random survival forest (RSF). Concordance indices of test sets of 140 patients were used for each sample to assess the generalizable predictive capacity of each model.
RESULTS: Following hyperparameter tuning, DeepSurv was trained at 32 min per sample on a 1.33 GHz quad-core CPU. Test set concordance indices of 0.7488 ± 0.0049, 0.6251 ± 0.0047, and 0.7368 ± 0.0047, were found for DeepSurv, CPH, and RSF, respectively. A Tukey HSD test demonstrates a statistically significant difference between the mean concordance indices of the three models.
CONCLUSION: Our results suggest that deep learning-based survival prediction can outperform traditional models, specifically in a case where an accurate prognosis is highly clinically relevant. We recommend that where appropriate data are available, deep learning-based prognostic indicators should be used to supplement classical statistics.
© 2020 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals LLC on behalf of American Association of Physicists in Medicine.

Entities:  

Keywords:  brain metastasis; deep learning; machine learning; survival

Mesh:

Year:  2020        PMID: 32790207      PMCID: PMC7285927          DOI: 10.1002/acm2.12995

Source DB:  PubMed          Journal:  J Appl Clin Med Phys        ISSN: 1526-9914            Impact factor:   2.243


Introduction

Clinical motivation

The median survival time for patients with brain metastases is on the order of months; however, some groups of patients can significantly outlive the median survival. Physicians have many treatment options to choose from, where the potential for disease‐free recovery is strongly connected to the treatment intensity. The brain met graded prognostic assessment (GPA) is one clinical tool that allows physicians to predict the longevity of patients with brain metastasis and thus select an appropriate treatment based on expected patient lifetime. For example, patients expected to live longer than 6 months are more likely to benefit from the short‐term memory protection offered by pin‐point radiosurgery treatment. On the other hand, patients with more limited life expectancy may be just as well served with a simpler whole brain radiotherapy, as they may not live long enough to experience the longer‐term cognitive effects of radiotherapy. Brain met GPA uses multivariate Cox regression (MCR) and recursive partitioning analysis (RPA) to determine factors that significantly contribute to survival predictions. In the specific case of breast cancer, the factors determined to be most significant include Karnofsky performance status, number of metastases, and hormone receptor characterization. The weights according to MCR are used to compute an index, or scale from 0.0 to 4.0, that maximizes separation between survival curves between groups. New patients are placed in a group according to a few features and given a highly nonspecific survival estimate as in Fig. 1. ,
Fig. 1

GPA is an online free‐to‐use tool utilized by some oncologists for patient prognosis (www.brainmetgpa.com). This validity of this tool has been established by Sperduto et al. in several journal publications with timely updates (reference: PMID: 22203767). The GPA tool uses five covariates determined to be significant by MCR to make survival predictions for patients with brain metastases and breast primary site. In this image, a breast cancer patient with age> 60, KPS in the range 90–100, tumor subtype luminal B triple‐positive, etc. is given a score of 2.5, indicating an expected survival of 11.1–47 months.

GPA is an online free‐to‐use tool utilized by some oncologists for patient prognosis (www.brainmetgpa.com). This validity of this tool has been established by Sperduto et al. in several journal publications with timely updates (reference: PMID: 22203767). The GPA tool uses five covariates determined to be significant by MCR to make survival predictions for patients with brain metastases and breast primary site. In this image, a breast cancer patient with age> 60, KPS in the range 90–100, tumor subtype luminal B triple‐positive, etc. is given a score of 2.5, indicating an expected survival of 11.1–47 months. In this study, we focus on predicting survival probabilities for patients with brain metastases and breast primary site using a deep neural network. We expect that a representation learning approach to prognostic assessment will produce more accurate survival estimates than MCR and RPA for our dataset. Similar work in machine learning for patient prognosis has been done by Alcorn et al. Their work focuses on the application of random survival forests specifically to the problem of prognosis for patients with bone metastases.

Cox proportional hazards

Proportional hazards models are regarded as the gold standard for survival analysis. Cox models aim to describe a patient‐specific hazard function (event rate), given a quantitative description of their attributes (covariates, features). According to the proportional hazards assumption, the event rate for patient having covariates at time is modeled with the hypothesis function. Regression with survival data is limited by censoring, or “loss to follow‐up.” There is no meaningful way to ascribe an event time to patients who discontinue communication with record keepers. Therefore, the parameters of the Cox model must be learned with a nonparametric objective function. Parameters that best predict the order of survival times for N patients having covariates { and survival times are obtained by maximizing the Cox partial likelihood: where indicates that patient i was not lost to follow‐up.

DeepSurv

Deep learning has been shown to be an effective tool for modeling nonlinear functions. There have been many breakthroughs in image classification, natural language processing, and other fields due to new methods and increased availability of deep learning platforms. In 2016, Katzman et al. released DeepSurv, a Cox deep neural network for public use. DeepSurv assumes the same structure as the Cox proportional hazards (CPH) model but uses a state‐of‐the‐art neural network to evaluate risk ( in CPH). The deep architecture of the model allows it to create higher‐order representations of features that might be more useful for survival predictions than the features alone. DeepSurv has been used to successfully characterize risk in several datasets including the Worcester Heart Attack Study (WHAS) and Molecular Taxonomy of Breast Cancer International Consortium (METABRIC) datasets. DeepSurv realizes its potential fully when the true risk function is highly nonlinear. DeepSurv’s concordance index improved from CPH by about 15% on toy data with risk generated from a 2D Gaussian.

Random survival forests

For comparison of DeepSurv with a leading nonlinear machine learning method that is not based on deep learning, a random survival forest was implemented in R using the RandomForestSRC package. Random Survival Forests utilize an ensemble of trees generated from bootstrap samples of the training dataset. Trees are “grown” for each data subset by identifying the covariate from a random set of candidate covariates that maximizes the survival difference for the data subset. The surviving fraction of the split data is calculated for each terminal node in the decision tree. For prediction on a test data point, the data are passed through all trees and the ensemble cumulative hazard is used to make survival predictions.

Methods

NCDB dataset

Survival predictions of three machine learning models were compared using 1673 patients’ data from the National Cancer Database. Patients included in the dataset all had brain metastases and a breast primary site. Twenty‐seven features including age, staging information, and hormone receptor status were used to characterize risk. An event indicator and time was provided for each patient. The NCDB dataset can be visualized with dimensionality reduction. t‐Distributed Stochastic Neighbor Embedding is a method of dimensionality reduction that compromises large distances in data space in favor of preserving small distances. A three‐dimensional embedding of the 27 features was created using t‐SNE in Python (Fig. 2). Clustering by event occurrence suggests that there are learnable representations of the dataset that describe patient survival.
Fig. 2

t‐SNE visualization of the NCDB dataset, labeled by event occurrence. Patients with recorded death times have a yellow marker, while those which are lost to follow‐up are labeled purple. Because directions in the embedding space do not correspond to known physical parameters, axis labels in t‐SNE visualization are arbitrary.

t‐SNE visualization of the NCDB dataset, labeled by event occurrence. Patients with recorded death times have a yellow marker, while those which are lost to follow‐up are labeled purple. Because directions in the embedding space do not correspond to known physical parameters, axis labels in t‐SNE visualization are arbitrary. Monte Carlo cross‐validation with 50 samples (84% training, 16% test) was used to assess model performance. The three models considered were DeepSurv, CPH, and a random survival forest (RSF). The RSF serves as a benchmark nondeep state‐of‐the‐art survival analysis method, based on recursive partitioning analysis. For more on RSFs, see. ,

Survival model evaluation

The generalization error for a model with parameters can be described by its concordance index on a test data set, , The concordance index considers all admissible pairs of patients and computes the fraction of patients that are correctly ordered by the model according to their true survival times. A pair of patients is considered admissible if patient A has survival time and experiences an event, . Concordance indices range from 0.0 (100% discordant) to 1.0 (100% concordant), with suggesting that predictions were made randomly.

Model implementation

Cox regression was implemented in Python with stochastic gradient ascent. Parameters were updated according to their derivatives with respect to the Cox partial log‐likelihood: where α is the learning rate. Updates were halted when the validation concordance appeared to converge to a maximum (60 iterations). A Wald test with significance level α = 0.01 was used to identify the parameters which are likely to be truly nonzero in the Cox framework. Features and significance levels are displayed in Table 1.
Table 1

The significance of various covariates according to the proportional hazards assumption is listed in the table.

Variable P Variable P Variable P
Age0.02Sex0.64Race0.51
Charleon/Deyo0.19 Grade 0.01 Tumor Size0.87
Regional LNs Positive0.94 AJCC Clinical N <0.005 AJCC Pathologic N0.32
Bone Mets at DX0.59 Brain Mets at DX <0.005 Lung Mets at DX0.85
ER Assay0.45PR Assay0.93 HER2 Summary <0.005
Multigene Signature Method0.62Multigene Signature Results0.91 Treatment Days After DX 0.01
Rad Days After DX0.38Radiation Type0.97 Volume Irradiated <0.005
Regional Dose0.99Chemotherapy Type0.16Hormone Therapy Type0.26
Immunotherapy Type <0.005 Mets at DX0.42Systemic Surgery Sequence0.04

Some factors that are obviously significant in survival predictions have significant P‐values. The many covariates with insignificant P‐values will not greatly contribute to risk calculation in the CPH framework. Bold indicates statistical significance.

The significance of various covariates according to the proportional hazards assumption is listed in the table. Some factors that are obviously significant in survival predictions have significant P‐values. The many covariates with insignificant P‐values will not greatly contribute to risk calculation in the CPH framework. Bold indicates statistical significance. The deep neural network DeepSurv was implemented in Python with L2 regularization, batch normalization, dropout, Nesterov momentum, and learning rate decay. A six‐dimensional box in hyperparameter space was uniformly sampled 100 times and DeepSurv’s performance was evaluated with a validation dataset. The hyperparameters that yielded the highest validation accuracy (Fig. 3) were chosen for deployment. DeepSurv was then trained for 7000 epochs per sample at about 32 min per sample on a 1.33 GHz quad‐core CPU.
Fig. 3

Left: Hyperparameters that yielded the highest validation accuracy. Right: A plot of learning rates versus validation accuracy demonstrates the impact of one hyperparameter choice on generalization error. A learning rate of was chosen for deployment.

Left: Hyperparameters that yielded the highest validation accuracy. Right: A plot of learning rates versus validation accuracy demonstrates the impact of one hyperparameter choice on generalization error. A learning rate of was chosen for deployment.

Results

The three models were independently trained and validated 50 times on the randomly split dataset. Test set concordance indices of , , and , were found for each model. DeepSurv and the RSF significantly outperform the CPH model. Concordance indices are displayed in Fig. 4.
Fig. 4

The test set concordance indices for three models are shown in this plot. The 95% confidence intervals of the median are given by notches in the boxes. DeepSurv appears to yield the smallest generalization error.

The test set concordance indices for three models are shown in this plot. The 95% confidence intervals of the median are given by notches in the boxes. DeepSurv appears to yield the smallest generalization error. A Tukey Honestly Significant Difference test was used to evaluate the difference of mean concordances for each model (Table 2). With a significance level of 0.05, one can reject the null hypothesis that the mean concordances for DeepSurv and the RSF equal the mean concordance for the proportional hazards model. There is not enough evidence to suggest that the mean concordance of DeepSurv is different from the mean concordance of the random forest Fig. 5.
Table 2

A Tukey HSD post‐hoc test supports the hypothesis that DeepSurv and the RSF have greater mean test concordances than CPH.

Tukey HSD – Multiple comparison of means
Group 1Group 2Difference of meansLower boundUpper boundReject null
DeepSurvCPH0.12370.10740.1399True
RSFCPH0.11170.09570.1277True
DeepSurvRSF0.0120‐0.00420.0282False
Fig. 5

Survival probability curves for three patients predicted by each model according to a Kaplan–Meier estimate of the baseline survival. The true times of death for these patients are indicated by the vertical lines.

A Tukey HSD post‐hoc test supports the hypothesis that DeepSurv and the RSF have greater mean test concordances than CPH. Survival probability curves for three patients predicted by each model according to a Kaplan–Meier estimate of the baseline survival. The true times of death for these patients are indicated by the vertical lines. To highlight the clinical utility of working survival models, we have included predicted survival curves for three test patients in one cross‐validation sample. The CPH and DeepSurv predictions are calculated by exponentiating a Kaplan–Meier estimate of the baseline survival function (from the training set) with the predicted hazard. The survival forest instead creates a unique curve based on all the terminal nodes to which the patient belongs. With the assumption of a baseline survival function, the expected value for each patient in the validation dataset can be evaluated with . This technique can be used to better understand the biases of the working models. We consider the distribution of errors for each model (Fig. 6). All three techniques tend to overestimate the true survival time.
Fig. 6

The distribution of errors predicted with for the validation dataset suggests all three models overestimate patient survival times.

The distribution of errors predicted with for the validation dataset suggests all three models overestimate patient survival times.

Discussion/Conclusion

In this study we demonstrate a highly clinically relevant scenario where deep learning‐based survival estimates outperform the gold‐standard survival analysis technique — MCR. With an appropriate baseline hazard estimate, survival predictions generated by DeepSurv might prove beneficial to patients and physicians in determining an appropriate treatment option. One failure of deep learning‐based survival analysis is the challenge of interpretation. Neural networks utilize complex interactions between features, which improves classification performance at the cost of interpretability. This is one weakness of deep learning compared to conventional methods. The interpretability of the inner workings of models in artificial intelligence is active area of research beyond the scope of this work. , We therefore recommend the supplement of traditional techniques with deep learning, rather than outright replacement. If future work allows intuitive interpretability of the “black box” of neural networks, they may well replace current models such as CPH. Advanced survival analysis techniques suffer from inaccessibility. Brain met GPA is widely popular because it is online and easy to use. Any future deep learning‐based approaches to patient prognosis should be accessible to physicians in the form of a webpage or easy‐to‐use software. One weakness of the deep risk framework is the lack of a time‐dependent hazard estimation; DeepSurv acts as an extension of the classic CPH model. Luck et al. have recently shown that by directly modeling the survival function, as opposed to risk, they can obtain concordance indices on par with those generated by DeepSurv. A final implementation of this work for clinical use might benefit from an effort to include time dependence. Currently, there does not appear to be any significant benefit to using DeepSurv over the Random Forest. However, deep learning is a very rapidly growing field. DeepSurv, despite utilizing several state‐of‐the‐art training techniques (dropout, batch normalization, L2 regularization), is architecturally quite simple. It does not take advantage of expected patterns in survival data in the way that convolutional networks handle images, and recurrent networks handle language. Our group expects that deep learning‐based methods will continue to improve in the near future.

CONFLICT OF INTEREST

The authors declare no conflict of interest.
  5 in total

1.  A new prognostic index and comparison to three other indices for patients with brain metastases: an analysis of 1,960 patients in the RTOG database.

Authors:  Paul W Sperduto; Brian Berkey; Laurie E Gaspar; Minesh Mehta; Walter Curran
Journal:  Int J Radiat Oncol Biol Phys       Date:  2007-10-10       Impact factor: 7.038

Review 2.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

3.  Summary report on the graded prognostic assessment: an accurate and facile diagnosis-specific tool to estimate survival for patients with brain metastases.

Authors:  Paul W Sperduto; Norbert Kased; David Roberge; Zhiyuan Xu; Ryan Shanley; Xianghua Luo; Penny K Sneed; Samuel T Chao; Robert J Weil; John Suh; Amit Bhatt; Ashley W Jensen; Paul D Brown; Helen A Shih; John Kirkpatrick; Laurie E Gaspar; John B Fiveash; Veronica Chiang; Jonathan P S Knisely; Christina Maria Sperduto; Nancy Lin; Minesh Mehta
Journal:  J Clin Oncol       Date:  2011-12-27       Impact factor: 44.544

4.  Regression modelling strategies for improved prognostic prediction.

Authors:  F E Harrell; K L Lee; R M Califf; D B Pryor; R A Rosati
Journal:  Stat Med       Date:  1984 Apr-Jun       Impact factor: 2.373

5.  DeepSurv: personalized treatment recommender system using a Cox proportional hazards deep neural network.

Authors:  Jared L Katzman; Uri Shaham; Alexander Cloninger; Jonathan Bates; Tingting Jiang; Yuval Kluger
Journal:  BMC Med Res Methodol       Date:  2018-02-26       Impact factor: 4.615

  5 in total
  5 in total

1.  Applying Artificial Neural Networks to Develop a Decision Support Tool for Tis-4N0M0 Non-Small-Cell Lung Cancer Treated With Stereotactic Body Radiotherapy.

Authors:  Takafumi Nemoto; Atsuya Takeda; Yukinori Matsuo; Noriko Kishi; Takahisa Eriguchi; Etsuo Kunieda; Ryusei Kimura; Naoko Sanuki; Yuichiro Tsurugai; Masamichi Yagi; Yousuke Aoki; Yohei Oku; Yuto Kimura; Changhee Han; Naoyuki Shigematsu
Journal:  JCO Clin Cancer Inform       Date:  2022-06

2.  An automated approach for predicting glioma grade and survival of LGG patients using CNN and radiomics.

Authors:  Chenan Xu; Yuanyuan Peng; Weifang Zhu; Zhongyue Chen; Jianrui Li; Wenhao Tan; Zhiqiang Zhang; Xinjian Chen
Journal:  Front Oncol       Date:  2022-08-12       Impact factor: 5.738

3.  An energy-efficient in-memory computing architecture for survival data analysis based on resistive switching memories.

Authors:  Andrea Baroni; Artem Glukhov; Eduardo Pérez; Christian Wenger; Enrico Calore; Sebastiano Fabio Schifano; Piero Olivo; Daniele Ielmini; Cristian Zambelli
Journal:  Front Neurosci       Date:  2022-08-09       Impact factor: 5.152

4.  Prognostic risk stratification of gliomas using deep learning in digital pathology images.

Authors:  Pranathi Chunduru; Joanna J Phillips; Annette M Molinaro
Journal:  Neurooncol Adv       Date:  2022-07-14

5.  Deep Learning-Based Time-to-Death Prediction Model for COVID-19 Patients Using Clinical Data and Chest Radiographs.

Authors:  Toshimasa Matsumoto; Shannon Leigh Walston; Michael Walston; Daijiro Kabata; Yukio Miki; Masatsugu Shiba; Daiju Ueda
Journal:  J Digit Imaging       Date:  2022-08-08       Impact factor: 4.903

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.