| Literature DB >> 34944970 |
Kuo-Chen Wu1,2, Shang-Wen Chen2,3,4,5, Te-Chun Hsieh6,7, Kuo-Yang Yen6,7, Kin-Man Law2,8, Yu-Chieh Kuo2, Ruey-Feng Chang1,2,9, Chia-Hung Kao2,6,10,11.
Abstract
OBJECTIVES: Neoadjuvant chemoradiotherapy (NCRT) followed by surgery is the mainstay of treatment for patients with locally advanced rectal cancer. Based on baseline 18F-fluorodeoxyglucose ([18F]-FDG)-positron emission tomography (PET)/computed tomography (CT), a new artificial intelligence model using metric learning (ML) was introduced to predict responses to NCRT. PATIENTS AND METHODS: This study used the data of 236 patients with newly diagnosed rectal cancer; the data of 202 and 34 patients were for training and validation, respectively. All patients received pretreatment [18F]FDG-PET/CT, NCRT, and surgery. The treatment response was scored by Dworak tumor regression grade (TRG); TRG3 and TRG4 indicated favorable responses. The model employed ML combined with the Uniform Manifold Approximation and Projection for dimensionality reduction. A receiver operating characteristic (ROC) curve analysis was performed to assess the model's predictive performance.Entities:
Keywords: 18F-fluorodeoxyglucose positron emission tomography/computed tomography; metric learning; neoadjuvant chemoradiotherapy; rectal cancer
Year: 2021 PMID: 34944970 PMCID: PMC8699508 DOI: 10.3390/cancers13246350
Source DB: PubMed Journal: Cancers (Basel) ISSN: 2072-6694 Impact factor: 6.639
Figure 1Structure of proposed model for classifying favorable responses to neoadjuvant chemoradiotherapy in patients with rectal cancer. Triplet loss for Resnet was utilized as a loss function for the ML algorithms. Abbreviations: UMAP = Uniform Manifold Approximation and Projection; SVM = support vector machine.
Patient characteristics of training cohort (n = 202).
| Characteristic | Value |
|---|---|
| Age (years) | 31–86 (median, 58) |
| Gender | Male:139, Female:63 |
| Primary lesion location | |
| low rectum | 83 |
| middle rectum | 103 |
| upper rectum or rectosigmoid junction | 16 |
| CEA (ng/dL) | 17.08 ± 37.92(0.48–241.88) |
| Pretreatment clinical staging (AJCC 7th ed.) | |
| T stage | T2:26, T3:148, T4:28 |
| N stage | N0:60, N1:80; N2:62 |
| M stage | M0:198, M1:4 |
| Differentiation | |
| W-D | 5 |
| M-D | 39 |
| P-D | 4 |
| unknown | 154 |
| Concurrent chemotherapy regimen (%) | |
| Capecitabine | 174 |
| Uracil-Tegafur | 21 |
| Intravenous 5-Fluorouracil based regimen | 7 |
| Interval from the end of radiation to surgery | |
| >4 and <8 week | 102 |
| ≥8 and <12 week | 100 |
| Tumor regression grade (%) | |
| Grade 0 | 0 |
| Grade 1 | 31 |
| Grade 2 | 54 |
| Grade 3 | 93 |
| Grade 4 | 24 |
Abbreviations: JCC = American Joint Committee on Cancer; CEA = carcinoembryonic antigen; W-D = well differentiated; M-D = moderately differentiated; P-D = poorly differentiated.
Classification indices for tumor regression grade 3 or 4 responses in all five sets of the training cohort, comparing patients with and without favorable responses.
| Prediction | RG Grade 3 or 4 Response | Indices | |
|---|---|---|---|
| Positive | Negative | ||
| Positive | 115 | 2 | 98.3% |
| Negative | 3 | 82 | 96.5% |
| Indices | 97.5% | 97.6% | 97.5% |
Figure 2Area under ROC curve for training cohort (A) and validation cohort (B).
Figure 3Distribution of visualized features for training model before (A) and after (B) metric learning and dimensionality reduction using Uniform Manifold Approximation and Projection. Blue spots represent tumor regression grade 3 or 4 responses; red spots correspond to responses lower than grade.
Figure 4Classification performance of training cohort by support vector machine for tumor regression grade 3 or 4 responses following dimensionality reduction of visualized features. Blue spots represent tumor regression grade 3 or 4 responses; red spots correspond to responses lower than grade 3.
Figure 5Representative heat map of tumor with regression grade 3 response on [F18]FDG-PET/CT in last layer of proposed model. (A) original PET, (B) heat map of PET, (C) original CT, and (D) heat map of CT. Note: 1. Heat maps were generated using a commercial software (Grad-CAM). 2. The geographical center of the PET and CT images is the same.
Classification indices with metric learning and Uniform Manifold Approximation and Projection (UMAP).
| AUC | Sensitivity | Specificity | Accuracy | |||||
|---|---|---|---|---|---|---|---|---|
| Training cohort ( | 0.960 (0.951–0.993) | 0.983 (0.962–1.000) | 0.965 (0.936–0.993) | 0.975 (0.960–0.991) | ||||
| Validation cohort ( | 0.962 (0.935–0.999) | 0.950 (0.910–0.990) | 1.000 (1.000–1.000) | 0.982 (0.969–0.997) | ||||
|
|
|
|
|
|
| |||
| Training AUC | 0.993 | 0.944 | 1.000 | 0.964 | 0.959 | 0.972 | ||
| Validation AUC | 1.000 | 0.92 | 0.939 | 0.977 | 1.000 | 0.967 | ||
| Training accuracy | 0.976 | 0.976 | 1.000 | 0.950 | 0.975 | 0.975 | ||
| Validation accuracy | 1.000 | 0.971 | 0.971 | 0.971 | 1.000 | 0.982 | ||
| Training sensitivity | 1.000 | 1.000 | 1.000 | 0.957 | 0.957 | 0.983 | ||
| Validation sensitivity | 1.000 | 0.917 | 0.917 | 0.917 | 1.000 | 0.950 | ||
| Training specificity | 0.941 | 0.941 | 1.000 | 0.941 | 1.000 | 0.965 | ||
| Validation specificity | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | ||
Abbreviation: 95% CI = 95% confidence interval.
Classification indices with traditional deep learning (ResNet) without UMAP or metric leaning.
| AUC | Sensitivity | Specificity | Accuracy | |||||
|---|---|---|---|---|---|---|---|---|
| Training cohort (n = 202) | 0.618 (0.576–0.704) | 0.630 (0.449–0.810) | 0.588 (0.349–0.827) | 0.614 (0.591–0.636) | ||||
| Validation cohort (n = 34) | 0.606 (0.511–0.704) | 0.467 (0.238–0.695) | 0.745 (0.645–0.846) | 0.647 (0.611–0.683) | ||||
|
|
|
|
|
|
| |||
| Training AUC | 0.645 | 0.588 | 0.573 | 0.760 | 0.634 | 0.640 | ||
| Validation AUC | 0.542 | 0.614 | 0.716 | 0.458 | 0.708 | 0.608 | ||
| Training accuracy | 0.634 | 0.634 | 0.575 | 0.625 | 0.600 | 0.614 | ||
| Validation accuracy | 0.618 | 0.618 | 0.706 | 0.618 | 0.676 | 0.647 | ||
| Training sensitivity | 0.625 | 0.958 | 0.565 | 0.391 | 0.609 | 0.630 | ||
| Validation sensitivity | 0.333 | 0.583 | 0.583 | 0.083 | 0.750 | 0.467 | ||
| Training specificity | 0.647 | 0.176 | 0.588 | 0.941 | 0.588 | 0.588 | ||
| Validation specificity | 0.773 | 0.636 | 0.773 | 0.909 | 0.636 | 0.745 | ||
Comparison of classification performance using statistical analysis. AUC value of traditional deep learning (ResNet) was 0.618, which was significantly inferior to that of the proposed model (p = 0.002).
| Number | Correlation | ||
|---|---|---|---|
| Training cohort | 202 | 0.215 | 0.002 |
| Test cohort | 34 | 0.308 | 0.076 |