Literature DB >> 30290849

Effect of machine learning methods on predicting NSCLC overall survival time based on Radiomics analysis.

Wenzheng Sun1,2, Mingyan Jiang3, Jun Dang4, Panchun Chang5, Fang-Fang Yin2.   

Abstract

BACKGROUND: To investigate the effect of machine learning methods on predicting the Overall Survival (OS) for non-small cell lung cancer based on radiomics features analysis.
METHODS: A total of 339 radiomic features were extracted from the segmented tumor volumes of pretreatment computed tomography (CT) images. These radiomic features quantify the tumor phenotypic characteristics on the medical images using tumor shape and size, the intensity statistics and the textures. The performance of 5 feature selection methods and 8 machine learning methods were investigated for OS prediction. The predicted performance was evaluated with concordance index between predicted and true OS for the non-small cell lung cancer patients. The survival curves were evaluated by the Kaplan-Meier algorithm and compared by the log-rank tests.
RESULTS: The gradient boosting linear models based on Cox's partial likelihood method using the concordance index feature selection method obtained the best performance (Concordance Index: 0.68, 95% Confidence Interval: 0.62~ 0.74).
CONCLUSIONS: The preliminary results demonstrated that certain machine learning and radiomics analysis method could predict OS of non-small cell lung cancer accuracy.

Entities:  

Keywords:  Machine learning; Non-small cell lung cancer; Overall survival; Radiomics analysis

Mesh:

Year:  2018        PMID: 30290849      PMCID: PMC6173915          DOI: 10.1186/s13014-018-1140-9

Source DB:  PubMed          Journal:  Radiat Oncol        ISSN: 1748-717X            Impact factor:   3.481


Background

Lung cancer is the leading cause of cancer-related deaths worldwide [1]. Lung cancer could be clinically divided into several groups: 1) the non-small cell lung cancer (NSCLC, 83.4%), 2) the small cell lung cancer (SCLC, 13.3%), 3) not otherwise specified lung cancer (NOS, 3.1%), 4) Sarcoma lung cancer (0.2%), and 5) other specified lung cancer (0.1%) [2]. The ability to predict clinical outcomes accurately is crucial for it allows clinicians to judge the most appropriate therapies for patients. Radiomics analysis can extract a large number of imaging features quantitatively, which could offer a cost-effective and non-invasive approach for individual medicine [3-5]. Several studies have shown the predictive and diagnostic ability of radiomics features in different kinds of cancers using various medical imaging modalities, such as PET [6-8], MRI [9] and CT [4, 10, 11]. It is also demonstrated that the radiomic features are associated with the overall survival. Besides, these associations can be used to establish positive predictive models. Machine-learning (ML) can be resumptively defined as the computational methods utilizing data/experience to obtain precise predictions [12]. The ML method can first learn laws from the data and then establish accuracy and efficiency prediction model based on these laws automatically. Moreover, an appropriate model is essential for the success use of radiomics. Hence, it is crucial to compare the performance of different ML models for clinical biomarkers based on radiomics analysis. Besides, appropriate feature selection methods should be applied first for the high-throughput radiomics features who may cause serious overfitting problems. In this study, we investigated the effect of 8 ML and 5 feature selection methods on predicting OS for non-small cell lung cancer based on radiomics analysis. The effectiveness of ML and feature selection methods on the prediction of OS were evaluated utilizing the concordance index (CI) [6, 13–16].

Methods

Data acquisition

The data used in this study was obtained from the ‘NSCLC-Radiomics’ collection [4, 17, 18] in the Cancer Imaging Archive which was an open access resource [19]. All the NSCLC patients in this data set were treated at MAASTRO Clinic, the Netherlands. For each patient, manual region of interest (ROI), CT scans and survival time (including survival status) were available. All the ROIs in this data set were the 3D volume of the gross tumor volume (GTV) delineated by a radiation oncologist.

Prediction process

The flow chart of the prediction process [20, 21] for all the ML methods in this study was outlined in Fig. 1. The performance of each ML and feature selection methods for the 283 NSCLC patients were evaluated using the cross-validation (CV) method (3-CV in this study). For each CV process, the total patients were divided into three folds, in which two folds (training fold) for training the machine learning model and the third (validation fold) for evaluating the model.
Fig. 1

The flow chart of predicted process for each ML method. (I) Dividing total data into three folds using the cross validation method. (II) Training each ML model using the selected radiomics features of the training fold. (III) Validating the prediction performance of each ML model on the validation fold

The flow chart of predicted process for each ML method. (I) Dividing total data into three folds using the cross validation method. (II) Training each ML model using the selected radiomics features of the training fold. (III) Validating the prediction performance of each ML model on the validation fold For each training fold, the training algorithm required both the training inputs (for prediction) and the prediction targets (for validation) data. The training inputs referred to the selected radiomics features, while the prediction targets referred to the OS of the patients. The radiomics features were first extracted from the images and then selected (dimension reduction) using the filter based feature selection methods to reduce the risk of overfitting. Finally, the selected features would be used to optimize and train all the ML models. In this study, the Bayesian optimization method was applied to determine the optimal parameters [22]. For each validation fold, the corresponding selected radiomics features were first extracted from the images and then transferred into the trained model. Finally, the prediction OS would be used to evaluate the goodness of each model.

Image pre-processing and Radiomics features extraction

Prior to extracting the radiomics features, we fixed the bin number (32 bins) of all the pre-treatment CT scans to discretize the image intensities. It should be noted that the original voxels for the images were used in this study. Then, the radiomics features were automatically extracted from the GTV region of the CT images by our in-house developed radiomics image analysis software and the Wavelet toolbox based on the Matlab R2017a (The Mathworks, Natick, MA). Total 43 unique quantitative features in 4 categories (Fig. 2) were extracted:
Fig. 2

Radiomics features used in this study. The definitions of radiomics features could be found in the IBSI document [26]. (I) Intensity features (1–4): 3.4.19, 3.4.18, 3.3.4 and 3.3.3 sections; (II) Fine texture features (5–26): 3.6.20, 3.6.23, 3.6.22, 3.6.21, 3.6.12, 3.6.19, 3.6.7, 3.6.5, 3.6.11, 3.6.4, 3.6.14, 3.6.16, 3.6.24, 3.6.25, 3.6.17, 3.6.15, 3.6.18, 3.6.1, 3.6.8, 3.6.10, 3.6.9 and 3.6.3 sections; (III) Coarse texture features (27–37): 3.7.1, 3.7.2, 3.7.9, 3.7.11, 3.7.13 and 3.7.3–3.7.8 sections; (IV) Morphological feature: 3.1.5, 3.1.6, 3.1.8, 3.1.7, 3.1.3 and 3.1.1 sections

Radiomics features used in this study. The definitions of radiomics features could be found in the IBSI document [26]. (I) Intensity features (1–4): 3.4.19, 3.4.18, 3.3.4 and 3.3.3 sections; (II) Fine texture features (5–26): 3.6.20, 3.6.23, 3.6.22, 3.6.21, 3.6.12, 3.6.19, 3.6.7, 3.6.5, 3.6.11, 3.6.4, 3.6.14, 3.6.16, 3.6.24, 3.6.25, 3.6.17, 3.6.15, 3.6.18, 3.6.1, 3.6.8, 3.6.10, 3.6.9 and 3.6.3 sections; (III) Coarse texture features (27–37): 3.7.1, 3.7.2, 3.7.9, 3.7.11, 3.7.13 and 3.7.3–3.7.8 sections; (IV) Morphological feature: 3.1.5, 3.1.6, 3.1.8, 3.1.7, 3.1.3 and 3.1.1 sections 1) Intensity features: to describe the shape characteristics of the CT volume’s gray-level intensity histogram, i.e., a probability density function (PDF) of gray-level distribution. 2) Fine texture features: to describe the high-resolution heterogeneity in the ROI. These features were derived from the ROI’s Gray-Level Co-Occurrence Matrix (GLCOM), a joint PDF that measures the frequency of co-occurring adjacent voxel pairs having the same grayscale intensity at a given direction [23]. 3) Coarse texture features: to describe the low-resolution heterogeneity in the ROI. These features were calculated from the ROI’s Gray-Level Run Length Matrix (GLRLM), a joint PDF that measures the size of a set of consecutive voxels with the same grayscale intensity at a given direction [24]. 4) Morphological features: to describe the morphological characteristics of the ROI [25]. Here, the first category and the following two (second and third) categories required the intensity histogram and textural image processing steps, respectively. Both the above two image processing steps and the 43 radiomics features used in this study matched benchmarks of the Image Biomarker Standardization Initiative (IBSI) [26]. Moreover, these radiomics features were also extracted from different wavelet decompositions of the original CT image by a three levels wavelet transformation [27, 28]. However, the morphological features weren’t extracted from the images with the wavelet decompositions for the wavelet transformation didn’t have effect on these features. Hence in total, 339 features were extracted for each patient in this study.

Features selection and machine learning methods

Pearson’s (PCC) [29], Kendall’s (KCC), [30] Spearman’s linear correlation coefficient (SCC) [31], Mutual information (MI) [32] and CI [15] were used as the filter based feature selection methods to reduce the dimensions of radiomics features in this study. In order to make sure the reliability of the selected features, we repeated each feature selection process 100 times using the bootstrap samples of each training fold and recorded the selected feature subset each time. Then, we selected the most frequently selected radiomics features as the final features which were used to train the ML models [6]. In this study, the first four feature selected methods (PCC, KCC, SCC and MI) were implemented using the Matlab R2017a and the following one method (CI) was implemented using the R software 3.5.1. All the feature selection methods would be performed on each training fold. The effect of 8 ML methods were investigated in this study, including: Cox proportional hazards model (Cox) [33], gradient boosting linear models based on Cox’s partial likelihood (GB-Cox) [34], gradient boosting linear models based on CI’s partial likelihood (GB-Cindex) [34], Cox model by likelihood based boosting (CoxBooxt) [35], bagging survival tree (BST) [36], random forests for survival model (RFS) [37], survival regression model (SR) [38] and support vector regression for censored data model (SVCR) [39, 40]. All the machine learning methods were implemented on each training fold using the R software 3.5.1. The specifics of the packages for each feature selection and ML method were showed in the Table 1. Besides, the descriptions of each feature selection and ML method could be found in the Additional file 1: Supplementary A and B, respectively.
Table 1

The specifics of the packages for each feature selection and machine learning method

MethodsSoftwarePackagesWebsite Links
PCCSML toolboxcorr https://ww2.mathworks.cn/help/stats/corr.html
KCC
SCC
MIMIToolboxmi https://github.com/Craigacp/MIToolbox
CIHiscrcorr.cens https://github.com/harrelfe/Hmisc
Coxsurvivalcoxph https://github.com/therneau/survival
GB-Coxmboostmboost https://github.com/boost-R/mboost
GB-Cindexmboostmboost https://github.com/boost-R/mboost
CoxBoostCoxBoostCoxBoost https://github.com/binderh/CoxBoost
BSTipredbagging https://github.com/cran/ipred
RFSrandomForestSRCrfsrc https://github.com/kogalur/randomForestSRC
SRsurvivalsurvreg https://github.com/therneau/survival
SVCRsurvivalsvmsurvivalsvm https://git-hub.com/imbs-hl/survivalsvm

SML statistics and machine learning

The specifics of the packages for each feature selection and machine learning method SML statistics and machine learning

Parameters tuning

For each ML method, the parameters were selected from the combination of parameters that produced the best performance using the three-fold CV on each training fold. Similar procedures were implemented in Brungard et al. [41] and Heung B et al [42]. The range of parameters used in this study was showed in Table 1. The GB-Cox, GB-Cindex, SVCR and SR methods just required one parameter to tune while the Cox method did not require parameterization. The complex models, such as the BTS and RFS, were time consuming for tuning parameters. The parameters from all of these models, such as the average terminal node size of forest and the number of trees for the RFS model, the minimum number of observations that must exist in a node (Minsplit) and the number of trees for BST, made up a large range of parameter permutation and combination choices. It should be noted that the feature number selected by the feature selection methods were also used as a tuning parameter (range [3, 29]) for all the ML methods.

Evaluation methods

CI with confidence interval (CFI) based on bootstrapping technique (the number of bootstrap samples was 2000 in this study) was used to assess the performance of difference ML methods on the merged validation fold (merged all the three validation folds). The percentage of CFI was 95% in this study. A nonparametric analytical approach method proposed by Kang L et al. [43] and the z-score test method were used to compare the significance between pairs of machine learning algorithms for each validation fold. Besides, the survival curves were evaluated by the Kaplan-Meier algorithm and compared by the log-rank tests [44] for each validation fold.

Results

Figure 3 depicted the performance of ML (in rows) and feature selection methods (in columns) on the merged validation fold. Besides, the maximum CI with confidence interval for each ML method on the merged validation fold was showed in Table 2. The GB-Cox method using the CI feature selection method obtained the best performance (CI: 0.682, 95% CFI: [0.620, 0.744]). However, the CoxBoost method using CI feature selection method also obtained a favorable performance (CI: 0.674, 95% CFI: [0.615, 0.731]). We found only the above mentioned two prediction method’s CIs were close. Hence, we just calculated the p-value using the z-test between the above two methods. The p-value of CI between these two methods was 0.5, indicating that the difference of prediction performance between these two methods wasn’t significant. The values selected for the hyper-parameters mentioned in Table 3, as well as the number of selected features on each validation fold could be found in the Additional file 1: Supplementary C.
Fig. 3

The performance of feature selection and machine learning methods on the merged validation fold

Table 2

Maximum CI with confidence interval for each machine learning method on the merged validation fold

MethodsFSMaximum CICFI of Maximum CI
GB-CoxCI0.682[0.620, 0.744]
CoxBoostCI0.674[0.615, 0.731]
CoxMI0.646[0.578, 0.714]
GB-CindexSCC0.357[0.290, 0.423]
RFSPCC0.627[0.558, 0.695]
SRMI0.380[0.310, 0.452]
BSTSCC0.385[0.318, 0.450]
SVCRKCC0.405[0.341, 0.470]

FS feature selection method

Table 3

The range of parameter tuning

MethodsParametersRange of Parameters
Cox
GB-CoxNumber of boosting steps[1, 500]
GB-CindexNumber of boosting steps[1, 500]
CoxboostNumber of boosting steps[1, 500]
BSTMinsplit[1, 10]
Number of trees[1, 500]
RFSAverage terminal node size of forest[1, 10]
Number of trees[1, 500]
SRAssumed distributionWeibull, Gaussian, Exponential
SVCRParameter of regularization[0.01, 1]
The performance of feature selection and machine learning methods on the merged validation fold Maximum CI with confidence interval for each machine learning method on the merged validation fold FS feature selection method The range of parameter tuning Patients on each validation fold were divided into two groups (low- and high- risk group) based on the predicted risk of each radiomics model at the cut-off value. The cut-off value utilized for stratification was the median of each training fold which would be applied to the corresponding validation fold unchanged. Then, the Kaplan-Meier and log-rank tests methods were used to evaluate and compare the survival curves for each validation fold, respectively. Among all the ML methods, the GB-Cox method with the CI feature selection method obtained the best stratified result on the 3 CV folds (Fig. 4). Besides, the p-value of the CoxBoost method with the PCC feature selection method was also significant for each validation fold. The heatmap of p-values on each validation fold for all the ML methods was showed in the Additional file 1: Supplementary D.
Fig. 4

Examples of the Kaplan-Meier evaluations. All the NSCLC patients on each validation fold were stratified into low- and high- risk groups based on the cut-off values determined by the corresponding training fold. Here, (a), (b) and (c) presented the Kaplan-Meier curve of the three CV validation folds, respectively

Examples of the Kaplan-Meier evaluations. All the NSCLC patients on each validation fold were stratified into low- and high- risk groups based on the cut-off values determined by the corresponding training fold. Here, (a), (b) and (c) presented the Kaplan-Meier curve of the three CV validation folds, respectively

Discussion

Several previous studies have compared the prediction performance of the ML models based on the radiomics analysis. Parmar C et al. [11] identified that three classifiers, included Bayesian, random forest (RF) and nearest neighbor, showed high OS prediction performance for the head and neck squamous cell carcinoma (HNSCC). Parmar C et al. [17] also evaluated the effect of ML models (classifiers) on the OS prediction for NSCLC patients and found that the random forest method with Wilcoxon test feature selection method obtained the highest prediction performance. However, the outcome of interest in these two studies explored by Parmar C et al. was transformed into a dichotomized endpoint. This may lead to the bias of prediction accuracy [13]. Hence, Leger S et al. [13] assessed the prediction performance (OS and loco-regional tumor control) of ML models which could dealt with continuous time-to-event data for HNSCC. His study found that the random forest using maximally selected rank statistics and the model based on boosting trees using CI methods with Spearman feature selection method got the best prediction performance for the loco-regional tumor control. Besides, the survival regression model based on the Weibull distribution, the GB-Cox and the GB-Cindex methods with the random feature selection method achieved the highest prediction performance for the OS. In this study, the effect of 8 ML models and five feature selection methods based on radiomics feature analysis were investigated to predict the time-to-event data (OS) of non-small cell lung cancer. In general, the GB-Cox method obtained the best predictive performance in the systematic evaluation on the merged validation fold. However, the CoxBoost methods with certain feature selection method also showed comparable positive performance compared with the GB-Cox method. Hence, we thought a wide range of ML methods have the potential to be effective radiomics analysis tools. Besides, a significant difference for OS prediction on each validation fold was found between the low- and high- risk groups using the GB-Cox and CoxBoost methods, which showed the clinical potential of ML methods on the OS prediction. As shown in Fig. 3, almost all of the ML methods using the KCC feature selection method didn’t obtain a positive result. This indicated that the feature selection method was also important for the performance of OS prediction. Sometimes, the effect of feature selection methods was even more obvious than the ML models. A large panel of feature selection methods had been used for data mining of high-throughput problems [45, 46]. In general, the feature selection methods would be divided into three categories: the filter based, the wrapper based and the embedded methods. In this study, we only investigated five different filter based methods because this kind of methods were not only less prone to overfitting but also more efficient in computation than other two methods [45, 46]. Moreover, the filter based methods were more independent than the wrapper and embedded methods, which could increase the fairness of ML methods comparison. Some previous studies [4, 5] have shown the potential clinical utility of the prognostic models based on radiomics analysis. This study could be a crucial supplementary reference for the use of prognostic models based on radiomics analysis because we compared a large number of machine-learning methods for the OS prediction of the NSCLC cancer. Such a comparison would be helpful in the selection of the optimal ML methods for OS prediction based on radiomics analysis.

Conclusion

The preliminary results demonstrated that certain machine learning and radiomics analysis method could predict OS of non-small cell lung cancer accuracy. Supplementary A: Feature selection methods. Supplementary B: Machine learning methods. Supplementary C: The values selected for the hyper-parameters on each validation fold. Supplementary D: P-values of the log-rank test for all the feature selection and ML methods on each validation fold. (PDF 625 kb)
  23 in total

1.  Boosting for high-dimensional time-to-event data with competing risks.

Authors:  Harald Binder; Arthur Allignol; Martin Schumacher; Jan Beyersmann
Journal:  Bioinformatics       Date:  2009-02-25       Impact factor: 6.937

2.  Support vector methods for survival analysis: a comparison between ranking and regression approaches.

Authors:  Vanya Van Belle; Kristiaan Pelckmans; Sabine Van Huffel; Johan A K Suykens
Journal:  Artif Intell Med       Date:  2011-08-06       Impact factor: 5.326

Review 3.  Multivariable prognostic models: issues in developing models, evaluating assumptions and adequacy, and measuring and reducing errors.

Authors:  F E Harrell; K L Lee; D B Mark
Journal:  Stat Med       Date:  1996-02-28       Impact factor: 2.373

4.  Responsible Radiomics Research for Faster Clinical Translation.

Authors:  Martin Vallières; Alex Zwanenburg; Bodgan Badic; Catherine Cheze Le Rest; Dimitris Visvikis; Mathieu Hatt
Journal:  J Nucl Med       Date:  2017-11-24       Impact factor: 10.057

5.  The Cancer Imaging Archive (TCIA): maintaining and operating a public information repository.

Authors:  Kenneth Clark; Bruce Vendt; Kirk Smith; John Freymann; Justin Kirby; Paul Koppel; Stephen Moore; Stanley Phillips; David Maffitt; Michael Pringle; Lawrence Tarbox; Fred Prior
Journal:  J Digit Imaging       Date:  2013-12       Impact factor: 4.056

6.  Classification of human lung carcinomas by mRNA expression profiling reveals distinct adenocarcinoma subclasses.

Authors:  A Bhattacharjee; W G Richards; J Staunton; C Li; S Monti; P Vasa; C Ladd; J Beheshti; R Bueno; M Gillette; M Loda; G Weber; E J Mark; E S Lander; W Wong; B E Johnson; T R Golub; D J Sugarbaker; M Meyerson
Journal:  Proc Natl Acad Sci U S A       Date:  2001-11-13       Impact factor: 11.205

7.  Quantitative Analysis of (18)F-Fluorodeoxyglucose Positron Emission Tomography Identifies Novel Prognostic Imaging Biomarkers in Locally Advanced Pancreatic Cancer Patients Treated With Stereotactic Body Radiation Therapy.

Authors:  Yi Cui; Jie Song; Erqi Pollom; Muthuraman Alagappan; Hiroki Shirato; Daniel T Chang; Albert C Koong; Ruijiang Li
Journal:  Int J Radiat Oncol Biol Phys       Date:  2016-05-07       Impact factor: 7.038

8.  A comparative study of machine learning methods for time-to-event survival data for radiomics risk modelling.

Authors:  Stefan Leger; Alex Zwanenburg; Karoline Pilz; Fabian Lohaus; Annett Linge; Klaus Zöphel; Jörg Kotzerke; Andreas Schreiber; Inge Tinhofer; Volker Budach; Ali Sak; Martin Stuschke; Panagiotis Balermpas; Claus Rödel; Ute Ganswindt; Claus Belka; Steffi Pigorsch; Stephanie E Combs; David Mönnich; Daniel Zips; Mechthild Krause; Michael Baumann; Esther G C Troost; Steffen Löck; Christian Richter
Journal:  Sci Rep       Date:  2017-10-16       Impact factor: 4.379

9.  External validation of a Cox prognostic model: principles and methods.

Authors:  Patrick Royston; Douglas G Altman
Journal:  BMC Med Res Methodol       Date:  2013-03-06       Impact factor: 4.615

10.  Radiomics: Images Are More than Pictures, They Are Data.

Authors:  Robert J Gillies; Paul E Kinahan; Hedvig Hricak
Journal:  Radiology       Date:  2015-11-18       Impact factor: 11.105

View more
  19 in total

1.  Radiomics in nuclear medicine: robustness, reproducibility, standardization, and how to avoid data analysis traps and replication crisis.

Authors:  Alex Zwanenburg
Journal:  Eur J Nucl Med Mol Imaging       Date:  2019-06-25       Impact factor: 9.236

Review 2.  Leveraging RSF and PET images for prognosis of multiple myeloma at diagnosis.

Authors:  Ludivine Morvan; Thomas Carlier; Bastien Jamet; Clément Bailly; Caroline Bodet-Milin; Philippe Moreau; Françoise Kraeber-Bodéré; Diana Mateus
Journal:  Int J Comput Assist Radiol Surg       Date:  2019-06-29       Impact factor: 2.924

3.  Metastatic melanoma treated by immunotherapy: discovering prognostic markers from radiomics analysis of pretreatment CT with feature selection and classification.

Authors:  Gulnur Ungan; Anne-Flore Lavandier; Jacques Rouanet; Constance Hordonneau; Benoit Chauveau; Bruno Pereira; Louis Boyer; Jean-Marc Garcier; Sandrine Mansard; Adrien Bartoli; Benoit Magnin
Journal:  Int J Comput Assist Radiol Surg       Date:  2022-06-02       Impact factor: 3.421

4.  Tumor and peritumor radiomics analysis based on contrast-enhanced CT for predicting early and late recurrence of hepatocellular carcinoma after liver resection.

Authors:  Nu Li; Xiaoting Wan; Hong Zhang; Zitian Zhang; Yan Guo; Duo Hong
Journal:  BMC Cancer       Date:  2022-06-17       Impact factor: 4.638

5.  Intratumoral and peritumoral CT-based radiomics strategy reveals distinct subtypes of non-small-cell lung cancer.

Authors:  Xing Tang; Haolin Huang; Peng Du; Lijuan Wang; Hong Yin; Xiaopan Xu
Journal:  J Cancer Res Clin Oncol       Date:  2022-04-17       Impact factor: 4.322

6.  External validation of radiomics-based predictive models in low-dose CT screening for early lung cancer diagnosis.

Authors:  Noemi Garau; Chiara Paganelli; Paul Summers; Wookjin Choi; Sadegh Alam; Wei Lu; Cristiana Fanciullo; Massimo Bellomi; Guido Baroni; Cristiano Rampinelli
Journal:  Med Phys       Date:  2020-06-23       Impact factor: 4.071

7.  Integrating Machine Learning and Tumor Immune Signature to Predict Oncologic Outcomes in Resected Biliary Tract Cancer.

Authors:  Gu-Wei Ji; Ke Wang; Yong-Xiang Xia; Jin-Song Wang; Xue-Hao Wang; Xiang-Cheng Li
Journal:  Ann Surg Oncol       Date:  2020-11-23       Impact factor: 5.344

Review 8.  Radiomics as a personalized medicine tool in lung cancer: Separating the hope from the hype.

Authors:  Isabella Fornacon-Wood; Corinne Faivre-Finn; James P B O'Connor; Gareth J Price
Journal:  Lung Cancer       Date:  2020-06-02       Impact factor: 5.705

9.  Initial Evaluation of Computer-Assisted Radiologic Assessment for Renal Mass Edge Detection as an Indication of Tumor Roughness to Predict Renal Cancer Subtypes.

Authors:  Rahul Rajendran; Kevan Iffrig; Deepak K Pruthi; Allison Wheeler; Brian Neuman; Dharam Kaushik; Ahmed M Mansour; Karen Panetta; Sos Agaian; Michael A Liss
Journal:  Adv Urol       Date:  2019-04-23

10.  Efficacy of Location-Based Features for Survival Prediction of Patients With Glioblastoma Depending on Resection Status.

Authors:  Madjid Soltani; Armin Bonakdar; Nastaran Shakourifar; Reza Babaie; Kaamran Raahemifar
Journal:  Front Oncol       Date:  2021-07-06       Impact factor: 6.244

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.