Literature DB >> 31200519

Deep Convolutional Neural Network-Based Positron Emission Tomography Analysis Predicts Esophageal Cancer Outcome.

Cheng-Kun Yang1, Joe Chao-Yuan Yeh2, Wei-Hsiang Yu3, Ling-I Chien4, Ko-Han Lin5, Wen-Sheng Huang6, Po-Kuei Hsu7.   

Abstract

In esophageal cancer, few prediction tools can be confidently used in current clinical practice. We developed a deep convolutional neural network (CNN) with 798 positron emission tomography (PET) scans of esophageal squamous cell carcinoma and 309 PET scans of stage I lung cancer. In the first stage, we pretrained a 3D-CNN with all PET scans for a task to classify the scans into esophageal cancer or lung cancer. Overall, 548 of 798 PET scans of esophageal cancer patients were included in the second stage with an aim to classify patients who expired within or survived more than one year after diagnosis. The area under the receiver operating characteristic curve (AUC) was used to evaluate model performance. In the pretrain model, the deep CNN attained an AUC of 0.738 in identifying patients who expired within one year after diagnosis. In the survival analysis, patients who were predicted to be expired but were alive at one year after diagnosis had a 5-year survival rate of 32.6%, which was significantly worse than the 5-year survival rate of the patients who were predicted to survive and were alive at one year after diagnosis (50.5%, p < 0.001). These results suggest that the prediction model could identify tumors with more aggressive behavior. In the multivariable analysis, the prediction result remained an independent prognostic factor (hazard ratio: 2.830; 95% confidence interval: 2.252-3.555, p < 0.001). We conclude that a 3D-CNN can be trained with PET image datasets to predict esophageal cancer outcome with acceptable accuracy.

Entities:  

Keywords:  PET; deep convolutional neural network; esophageal cancer

Year:  2019        PMID: 31200519      PMCID: PMC6616908          DOI: 10.3390/jcm8060844

Source DB:  PubMed          Journal:  J Clin Med        ISSN: 2077-0383            Impact factor:   4.241


1. Introduction

Accurate estimates of survival are of paramount importance for patients and oncologists making personalized and patient-centered decisions in the era of precision medicine. Many clinical prediction tools combining key prognostic factors have been proposed to improve risk stratification and prognostication in esophageal cancer. However, few existing tools can be confidently used in current clinical practice [1]. Besides clinical parameters, medical image analysis is also gaining exponential interest in prognostic research. It has been proposed that radiomics analysis, which is defined as the high-throughput extraction of quantitative image metrics, potentially facilitates characterization of tumor phenotypes and even predicts patient survival [2,3,4]. Several studies have developed radiomics prognostic models incorporating positron emission tomography (PET) image features, which contain both anatomic and metabolic information, in lung cancer, colorectal cancer, and esophageal cancer [3,4,5,6]. Recently, the mainstream of computational image analysis has been gradually replaced by deep learning techniques [7,8,9]. The most impactful feature of the deep learning algorithm is self-learning; once a data set has been provided, the program can automatically discover features and adaptively learn without human indication by a backpropagation algorithm and by changing the internal parameters of each layer [7,8]. Deep learning algorithms, in particular convolutional neural networks (CNNs), have been applied in medical image classification, segmentation, and detection [8,9], e.g., skin lesions classification [10], diabetic retinopathy detection [11], and diagnosis of lymph node metastasis in breast cancer [12]. Pilot studies even demonstrated the potential of CNNs to recognize biological features that are overlooked by human experts and predict someone’s risk for certain diseases [13,14]. However, the application of deep learning in cancer survival prediction is limited in the literature. We hypothesize that a deep learning algorithm facilitates PET image analysis in predicting esophageal cancer outcome. In the present study, we trained a CNN to identify esophageal cancer patients with extremely poor prognosis, i.e., expired within one year after diagnosis, using only pretreatment PET images.

2. Methods

2.1. Data Source

The PET scans of patients with the diagnosis of esophageal squamous cell carcinoma between September 2009 and August 2017 at Taipei Veterans General Hospital were collected. The PET scans of patients with stage I lung cancer diagnosed between January 2012 and November 2017 at Taipei Veterans General Hospital were also collected for use in pretraining the CNN. To focus on critical information, only PET images from the hypopharynx to the stomach, which included all of the esophagus and the peri-esophageal regions were used.

2.2. Convolution Neural Network Model Setup and Training

Input images were classified as positive if the patient expired within a year after diagnosis, while those corresponding to patients that survived more than one year after diagnosis were classified as negative. We built a three-dimensional (3D) CNN based on a residual network, ResNet [15], as illustrated in Figure S1. To achieve better performance and arrive at convergence (that is, optimum performance) faster, we used PET scans from patients with stage I lung cancer, whose esophagi were presumed to be normal, and trained the network as our first stage of training. We used the PET scans to pretrain our model with an aim to classify the scans into esophageal cancer (abnormal esophagus) or lung cancer (normal esophagus). For the second stage, which aimed to classify esophageal cancer outcome, model weights were transferred from the aforementioned pretrained model. We used an 18-layer ResNet with the stochastic gradient descent (SGD) optimizer and anisotropic max-pooling as the baseline model and compared different model structures/hyper-parameters. For each experimental setting, one of the components was modified and compared with the baseline model. The experimental conditions included: (1) with or without pretraining; (2) choice of optimizer (adaptive momentum estimation (Adam) vs. SGD); and (3) number of layers (18- vs. 34-layer residual network). The final model contained 34 layers with a bottleneck design containing about 33 million learnable parameters. The detailed data preprocessing, model setup, and hyper-parameter settings are shown in supplementary files.

2.3. Statistics

We randomly shuffled the data index before splitting the entire dataset into five subgroups of equal size. We repeated the process of model training using one subgroup as a validation set and the rest as a training set until all subgroups were used. With each subgroup as a validation set, we trained the deep neural network in five independent runs and used the average area under receiver-operating characteristic (ROC) curve (AUC) to evaluate its performance, as shown in Figure S2. Based on the probability threshold of 0.5, patients were predicted to expire within one year after diagnosis (positive) or not (negative). The prediction results were compared to clinical data and patients were classified into correct or incorrect prediction groups. The correct prediction group included patients whose survival status was the same as predicted status; whereas those who were predicted to be expired (or alive) but were alive (or expired) at one year were classified in the incorrect prediction group. Overall survival (OS) was defined as the time from the date of diagnosis until death or last known follow-up. Survival curves were plotted using the Kaplan–Meier method and compared by the log-rank test. Multivariable survival analysis, which included all available clinical factors, was conducted with a Cox regression model. Survival analysis was performed using Statistical Package for the Social Sciences (version 17, SPSS Inc., Chicago, IL, USA) and a two-sided p-value < 0.05 was considered significant.

3. Results

3.1. Patients

A total of 798 PET scans of esophageal squamous cell carcinoma and 309 PET scans of stage I lung cancer were included in the pretraining stage of this study. In the esophageal cancer outcome classification stage, only the pretreatment PET scans were used and patients without complete clinical follow-up data were excluded. The characteristics of final 548 patients are shown in Table 1.
Table 1

Patient demographics.

TotalCorrect PredictionIncorrect Prediction p
n = 548n = 401n = 147
Age, (years, mean ± SD)61.4 ± 12.461.7 ± 12.760.5 ± 11.70.290
Gender (%) 0.320
Male504 (92.0)366 (91.3)138 (93.9)
Female44 (8.0)35 (8.7)9 (6.1)
Clinical T stage (%) 0.109
T165 (11.9)55 (13.7)10 (6.8)
T2118 (21.5)89 (22.2)29 (19.7)
T3325 (59.3)229 (57.1)96 (65.3)
T440 (7.3)28 (7.0)12 (8.2)
Clinical N stage (%) 0.016
N0178 (32.5)141 (35.2)37 (25.2)
N1196 (35.8)128 (31.9)68 (46.3)
N2102 (18.6)76 (19.0)26 (17.7)
N372 (13.1)56 (14.0)16 (10.9)
Clinical M stage (%) 0.970
M0458 (83.6)335 (83.5)123 (83.7)
M190 (16.4)66 (16.5)24 (16.3)
Tumor location (%) 0.349
Upper third161 (29.4)111 (27.7)50 (34.0)
Middle third242 (44.2)182 (45.4)60 (40.8)
Lower third145 (26.5)108 (26.9)37 (25.2)
Tumor size, (cm, mean ± SD)4.4 ± 3.44.3 ± 3.34.5 ± 3.50.730
SUVMAX, (mean ± SD) 12.5 ± 6.012.3 ± 6.212.9 ± 5.40.308
Tumor markers, (mean ± SD)
SCC, ng/ml 2.6 ± 5.32.6 ± 6.02.6 ± 2.80.987
Cyfra 21-1, ng/ml 4.2 ± 5.94.4 ± 7.24.1 ± 5.90.879
Treatments (%) 0.758
Primary resection141 (25.7)101 (25.2)40 (27.2)
Trimodal treatments160 (29.2)115 (28.7)45 (30.6)
Medical treatment208 (38.0)154 (38.4)54 (36.7)
Supportive care39 (7.1)31 (7.7)8 (5.4)

SD: standard deviation; SUVMAX: maximal standard uptake value; SCC: squamous cell carcinoma antigen. Data were available in 470 (size), 489 (SUV), 387 (SCC), and 139 (Cyfra 21-1) patients owning to respective nature of this study.

3.2. CNN Performance

The performance of each model is shown in Table 2, and distribution of AUC for each experiment is shown in Figure S3. The ROC curve comparison between different hyper-parameters and structure combinations is illustrated in Figure 1. In general, there was no difference between models with pretraining and models without pretraining. The model performances were comparable between the SGD optimizer and the Adam optimizer. Finally, we compared the 18-layer residual network and the 34-layer residual network. In the pretrained models, the 34-layer network achieved better results than the 18-layer network without over-fitting (0.738 AUC vs. 0.717 AUC with the Adam optimizer and 0.720 AUC vs. 0.709 AUC with the SGD optimizer). In the non-pretrained models, the 34-layer network had better results than the 18-layer network with the Adam optimizer (0.740 AUC vs. 0.710 AUC) and had similar results with the SGD optimizer (0.722 AUC vs. 0.724 AUC).
Table 2

Performance of CNN models.

ModelsAUC95% CI
PretrainOptimizerLayers
NoAdam180.7100.688–0.733
NoSGD180.7240.714–0.734
NoAdam340.7400.704–0.776
NoSGD340.7220.697–0.747
YesAdam180.7170.693–0.741
YesSGD180.7090.678–0.741
YesAdam340.7380.714–0.761
YesSGD340.7200.694–0.746

CNN: convolutional neural network, AUC: area under curve; CI: confidence interval.

Figure 1

Receiver-operating characteristic curves of different CNN models.

3.3. Clinical Relevance

The comparison between patients with correct and incorrect prediction (according to the pretrained model of 34-layer network with the Adam optimizer) is shown in Table 1. These two groups were similar with regard to age, gender, clinical T stage, clinical M stage, tumor location, tumor size, maximal standard uptake value (SUVmax), value of tumor on PET scan, serum level of tumor markers, and treatment modalities. The only difference is the clinical N stage. The incorrect prediction group had more clinical N1 patients than the correct prediction group (46.3% vs. 31.9%, p = 0.016). To investigate the prognostic impact of prediction results, survival analysis was performed based on prediction results and clinical status, as shown in Figure 2. Patients who were predicted to survive and survived one year after diagnosis had the highest 5-year survival rate of 50.5%. Interestingly, patients who were predicted to be expired but survived at one year after diagnosis had a 5-year survival rate of 32.6%, with a median survival of 23.2 (95% CI: 19.3–27.1) months, which was significantly worse than the outcome of the patients who were predicted to survive and really survived one year after diagnosis (p < 0.001). In the multivariable survival analysis, the independent prognostic factors for overall survival included age, gender, tumor location at upper third, clinical N stage, clinical M stage, and the prediction results by our CNN model, as shown in Table 3 (hazard ratio (HR): 2.830; 95% confidence interval (CI): 2.252–3.555, p < 0.001).
Figure 2

Prognostic significance of 3D-CNN prediction results. A: yellow: patients predicted alive at one year; purple: patients predicted expired at one year. B: green: patients predicted alive at one year and survived more than one year after diagnosis; blue: patients predicted expired within one year but survived more than one year; tan: patients predicted alive at one year but expired within one year; red: patients predicted expired and died within one year, p < 0.001.

Table 3

Multivariable overall survival analysis.

HR95% CI p
Age1.0181.010–1.027<0.001
Gender
Male1--
Female0.4260.252–0.7200.001
Tumor location
Lower third1--
Middle third1.2970.960–1.7520.090
Upper third1.3331.007–1.7630.044
Clinical T stage
T1/21--
T3/41.2090.946–1.5450.130
Clinical N stage
N(−)1--
N(+)1.4171.082–1.8560.011
Clinical M stage
M01--
M12.0411.571–2.651<0.001
Prediction results
Alive at one year1--
Expired within one year2.8302.252–3.555<0.001

HR: hazard ratio; CI: confidence interval.

4. Discussion

Outcome prediction in cancer care is the foundation for individualized treatment planning. Prediction tools integrating cancer stage, age, sex, comorbidities, treatment received, treatment response, and specific surgical pathology have been developed for esophageal cancer prognostication. However, few existing survival prediction tools are ideal for use in current clinical practice. In a review by Gupta et al., the discrimination ability, which might vary from 0.5 (model predictions are similar to chance) to 1.0 (model predictions are perfect), of sixteen clinical tool-based models in the literature was around 0.63 to 0.77, which was far from confident application in clinical use [1]. In addition to patient specific clinicopathological information, medical image analysis is gaining substantial interest in prognostic research. In oncology, the SUV of 18-fluorodeoxyglucose (FDG), which measures FDG activity in the tumor and correlates with viable tumor cell number and metabolism, obtained from FDG-PET scans is widely used for diagnosing, staging, monitoring response to therapy, as well as outcome prediction [16,17,18]. Correlation between higher SUVmax and worse survival has been reported in several studies [19,20,21]. Owing to non-uniformly standardized PET imaging protocols, variable SUVmax thresholds have been reported and thus limit its use [22]. Besides conventional uptake parameters, texture analysis, which provides numerous quantitative and semiquantitative indices, termed “features” [2,3,4], seems to perform better in characterizing tumor phenotypes [23]. This approach as a whole is named “radiomics” [24]. Several studies have successfully developed radiomic prognostic classifiers that can be associated with metastatic recurrence and survival in several types of cancers [3,4,16]. For example, Lambin et al. demonstrated that the prognostic model based on pretreatment CT radiomic features yielded an AUC of 0.69 in predicting 3-year overall survival of esophageal cancer patients after chemoradiotherapy [25]. Although many different PET segmentation techniques have been proposed and numerous PET-based radiomic features have been described, the results are highly dependent on the method used [4,16,26]. For example, in an external validation study of a prognostic model incorporating quantitative PET image features in esophageal cancer, Foley et al. has shown that results of a developed prognostic model combining clinical factors and PET radiomic features was not replicated in another cohort of patients treated with different regimens [26]. Another limitation in radiomics research is the algorithm that extracts imaging features through manual detection and characterization of tumor regions, which is labor intensive and subjective. In contrast, deep learning algorithms, in particular convolutional neural networks, provide workflows that allow automated selection and quantification of the most robust features. Since tumor segmentation or feature calculation is not required, it simplifies the analysis procedure and is more objective than the classical methods [7,8,9,27]. The application of CNNs in clinical medical imaging is on the horizon. The performance of CNNs can be outstanding with an abundance of well-annotated data; for example, classification of skin cancer [10], classification of dermoscopic melanoma recognition [28], detection of diabetic retinopathy with retinal fundus photographs [11], detection of lymph node metastases with whole slide images in breast cancer [12], and anatomical classification of esophagogastroduodenoscopy images [29]. Intriguingly, it has been proposed that CNNs may be able to astutely reveal subtle biological characteristics that are not visible to physicians. As examples, researchers have used CNNs to predict cardiovascular risk factors with retinal fundus photographs [14]. In thoracic oncology, a CNN has been trained to classify mediastinal lymph nodes of non-small cell lung cancers with FDG-PET images [30]. In another study, Ypsilantis et al. used PET imaging data from 107 patients with esophageal cancer to demonstrate that a CNN has the potential to predict chemotherapy response [31]. Applying deep learning networks in predicting overall survival, Hosny et al. trained a CNN from CT data of lung cancer patients treated with radiotherapy or surgery. The prognostic power attained an AUC of 0.70 and they concluded that the CNN is capable to significantly stratify patients into low and high mortality risk groups [27]. In this study, we trained a CNN to predict the survival status at one year after diagnosis in esophageal cancer patients. Our model attained an AUC of 0.738, which was better than that of clinical factor-based prediction models, as shown in Figure S4. The AUC of our model was also better than that of a CT radiomic feature-based model in predicting survival status in esophageal cancer (AUC: 0.69 [25]), and that of a deep learning-based model in predicting mortality risk in lung cancer (AUC: 0.70 [27]). In the multivariable analysis, the prediction result was an independent prognostic factor, indicating that our model could identify esophageal cancer with aggressive behavior. Interestingly, when we compared correct and incorrect prediction groups, the only difference was the percentage of clinical N1 stage. Indeed, the outcome of patients with minimal nodal involvement was highly dependent on either the extent of surgical resection or response to systemic treatments, which could not be revealed by pretreatment PET scan images. To overcome the limitation of a relatively small dataset, we adopted a stepwise workflow, in which the first stage was to classify normal and abnormal esophagi only. Another limitation is the nature of opaqueness of deep learning networks. Feature definition, extraction, and selection are all automated and occur implicitly. Consequently, the imaging characteristics they measure are highly obscure. This ambiguity is in sharp contrast to the expert-based well-defined radiomic features. For example, it has been proposed that morphometric measures, such as core muscle size, obtained from cross-sectional imaging, was correlated with sarcopenia, which is an independent predictor of clinical outcomes in multiple gastrointestinal cancers [32,33]. The CNN classification may be based on tumor or non-tumor features. Moreover, we did not include PET images from different PET scanners. Further studies are required to test the reliability and robustness of our results. Lastly, our proposed model was not compared against other machine learning-based methods. In the future, it is mandatory to apply other methods on our dataset and compare the AUC performance. In conclusion, a 3D-CNN can be trained with PET image datasets to predict outcome in esophageal cancer with acceptable accuracy. The prediction result remained an independent prognostic factor in multivariable survival analysis. Although our current results cannot be readily applied to clinical decision making, we demonstrated the potential of deep learning. With a larger dataset, the CNN can be trained to achieve a better prediction performance.
  32 in total

1.  Metabolic imaging predicts response, survival, and recurrence in adenocarcinomas of the esophagogastric junction.

Authors:  Katja Ott; Wolfgang A Weber; Florian Lordick; Karen Becker; Raymonde Busch; Ken Herrmann; Hinrich Wieder; Ulrich Fink; Markus Schwaiger; Jörg-Rüdiger Siewert
Journal:  J Clin Oncol       Date:  2006-09-11       Impact factor: 44.544

Review 2.  Radiomics: extracting more information from medical images using advanced feature analysis.

Authors:  Philippe Lambin; Emmanuel Rios-Velazquez; Ralph Leijenaar; Sara Carvalho; Ruud G P M van Stiphout; Patrick Granton; Catharina M L Zegers; Robert Gillies; Ronald Boellard; André Dekker; Hugo J W L Aerts
Journal:  Eur J Cancer       Date:  2012-01-16       Impact factor: 9.162

3.  Maximum standardized uptake values on positron emission tomography of esophageal cancer predicts stage, tumor biology, and survival.

Authors:  Robert J Cerfolio; Ayesha S Bryant
Journal:  Ann Thorac Surg       Date:  2006-08       Impact factor: 4.330

Review 4.  Quantifying tumour heterogeneity in 18F-FDG PET/CT imaging by texture analysis.

Authors:  Sugama Chicklore; Vicky Goh; Musib Siddique; Arunabha Roy; Paul K Marsden; Gary J R Cook
Journal:  Eur J Nucl Med Mol Imaging       Date:  2012-10-13       Impact factor: 9.236

5.  Predictive and prognostic value of metabolic tumour volume and total lesion glycolysis in solid tumours.

Authors:  Christophe Van de Wiele; Vibeke Kruse; Peter Smeets; Mike Sathekge; Alex Maes
Journal:  Eur J Nucl Med Mol Imaging       Date:  2012-11-14       Impact factor: 9.236

6.  Decreased core muscle size is associated with worse patient survival following esophagectomy for cancer.

Authors:  K H Sheetz; L Zhao; S A Holcombe; S C Wang; R M Reddy; J Lin; M B Orringer; A C Chang
Journal:  Dis Esophagus       Date:  2013-01-25       Impact factor: 3.429

7.  Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning.

Authors:  Hoo-Chang Shin; Holger R Roth; Mingchen Gao; Le Lu; Ziyue Xu; Isabella Nogues; Jianhua Yao; Daniel Mollura; Ronald M Summers
Journal:  IEEE Trans Med Imaging       Date:  2016-02-11       Impact factor: 10.048

Review 8.  FDG-PET parameters as prognostic factor in esophageal cancer patients: a review.

Authors:  J M T Omloo; M van Heijl; O S Hoekstra; M I van Berge Henegouwen; J J B van Lanschot; G W Sloof
Journal:  Ann Surg Oncol       Date:  2011-05-03       Impact factor: 5.344

9.  Predicting Response to Neoadjuvant Chemotherapy with PET Imaging Using Convolutional Neural Networks.

Authors:  Petros-Pavlos Ypsilantis; Musib Siddique; Hyon-Mok Sohn; Andrew Davies; Gary Cook; Vicky Goh; Giovanni Montana
Journal:  PLoS One       Date:  2015-09-10       Impact factor: 3.240

10.  Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach.

Authors:  Hugo J W L Aerts; Emmanuel Rios Velazquez; Ralph T H Leijenaar; Chintan Parmar; Patrick Grossmann; Sara Carvalho; Sara Cavalho; Johan Bussink; René Monshouwer; Benjamin Haibe-Kains; Derek Rietveld; Frank Hoebers; Michelle M Rietbergen; C René Leemans; Andre Dekker; John Quackenbush; Robert J Gillies; Philippe Lambin
Journal:  Nat Commun       Date:  2014-06-03       Impact factor: 14.919

View more
  5 in total

1.  Predicting aggressive histopathological features in esophageal cancer with positron emission tomography using a deep convolutional neural network.

Authors:  Joe Chao-Yuan Yeh; Wei-Hsiang Yu; Cheng-Kun Yang; Ling-I Chien; Ko-Han Lin; Wen-Sheng Huang; Po-Kuei Hsu
Journal:  Ann Transl Med       Date:  2021-01

2.  Evaluation of the Effect of Comprehensive Nursing Interventions on Plaque Control in Patients with Periodontal Disease in the Context of Artificial Intelligence.

Authors:  Juan Xu; Lingling Wang; Hongxia Sun; Shanshan Liu
Journal:  J Healthc Eng       Date:  2022-03-23       Impact factor: 2.682

Review 3.  Machine Learning and Radiomics Applications in Esophageal Cancers Using Non-Invasive Imaging Methods-A Critical Review of Literature.

Authors:  Chen-Yi Xie; Chun-Lap Pang; Benjamin Chan; Emily Yuen-Yuen Wong; Qi Dou; Varut Vardhanabhuti
Journal:  Cancers (Basel)       Date:  2021-05-19       Impact factor: 6.639

4.  Lung Nodule Sizes Are Encoded When Scaling CT Image for CNN's.

Authors:  Dmitry Cherezov; Rahul Paul; Nikolai Fetisov; Robert J Gillies; Matthew B Schabath; Dmitry B Goldgof; Lawrence O Hall
Journal:  Tomography       Date:  2020-06

5.  Automatic classification of esophageal lesions in endoscopic images using a convolutional neural network.

Authors:  Gaoshuang Liu; Jie Hua; Zhan Wu; Tianfang Meng; Mengxue Sun; Peiyun Huang; Xiaopu He; Weihao Sun; Xueliang Li; Yang Chen
Journal:  Ann Transl Med       Date:  2020-04
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.