| Literature DB >> 36262277 |
Qingwen Zeng1,2,3, Hong Li4, Yanyan Zhu5, Zongfeng Feng1,2, Xufeng Shu1, Ahao Wu1, Lianghua Luo1, Yi Cao1, Yi Tu6, Jianbo Xiong1, Fuqing Zhou5, Zhengrong Li1,2.
Abstract
Background: This study aims to develop and validate a predictive model combining deep transfer learning, radiomics, and clinical features for lymph node metastasis (LNM) in early gastric cancer (EGC). Materials and methods: This study retrospectively collected 555 patients with EGC, and randomly divided them into two cohorts with a ratio of 7:3 (training cohort, n = 388; internal validation cohort, n = 167). A total of 79 patients with EGC collected from the Second Affiliated Hospital of Soochow University were used as external validation cohort. Pre-trained deep learning networks were used to extract deep transfer learning (DTL) features, and radiomics features were extracted based on hand-crafted features. We employed the Spearman rank correlation test and least absolute shrinkage and selection operator regression for feature selection from the combined features of clinical, radiomics, and DTL features, and then, machine learning classification models including support vector machine, K-nearest neighbor, random decision forests (RF), and XGBoost were trained, and their performance by determining the area under the curve (AUC) were compared.Entities:
Keywords: convolutional neural networks; deep learning; early gastric cancer (EGC); lymph node metastasis; radiomics
Year: 2022 PMID: 36262277 PMCID: PMC9573999 DOI: 10.3389/fmed.2022.986437
Source DB: PubMed Journal: Front Med (Lausanne) ISSN: 2296-858X
FIGURE 1Inclusion and exclusion criteria for patients with EGC for the training and internal validation cohorts. EGC, early gastric cancer; CT, computed tomography; ESD, endoscopic submucosal dissection.
FIGURE 2Workflow of model development. DTL, deep transfer learning; ROI, regions of interest.
FIGURE 3Radiomics + DTL (Resnet152) + clinical features dimension reduction and performance of the model. (A) LASSO coefficient profiles of the features. Different color line shows corresponding coefficient of each feature. (B) Tuning parameter (λ) selection in LASSO model. (C) Selected features weight coefficients. (D) Area under the curve (AUC) of predictive model based on radiomics + DTL (Resnet152) + clinical features in training and validation cohorts. DTL, deep transfer learning; LASSO, least absolute shrinkage and selection operator.
Characteristics of early gastric cancer (EGC) patient included for classification modeling.
| Characteristics | Training cohort ( | Internal validation cohort ( | External validation cohort ( | |
|
| 58.62 ± 10.84 | 57.47 ± 11.74 | 58.09 ± 12.07 | 0.535 |
|
| ||||
| Male | 237 (61.1) | 107 (64.1) | 48 (60.8) | 0.784 |
| Female | 151 (38.9) | 60 (35.9) | 31 (39.2) | |
|
| 22.82 ± 13.28 | 20.88 ± 11.49 | 20.96 ± 12.09 | 0.181 |
|
| 0.064 | |||
| T1a | 168 (43.3) | 70 (41.9) | 23 (29.1) | |
| T1b | 220 (56.7) | 97 (58.1) | 56 (70.9) | |
|
| 0.075 | |||
| Intestinal type | 232 (59.8) | 94 (56.3) | 46 (58.2) | |
| Diffuse type | 67 (17.3) | 34 (20.4) | 23 (29.1) | |
| Mixed type | 89 (22.9) | 39 (23.3) | 10 (12.7) | |
|
|
| |||
| Poor/undifferentiated | 197 (50.8) | 76 (45.5) | 28 (35.4) | |
| Moderate | 138 (35.6) | 70 (41.9) | 43 (54.4) | |
| Well | 53 (13.6) | 21 (12.6) | 8 (10.2) | |
|
| 0.608 | |||
| Negative | 336 (86.6) | 140 (83.8) | 66 (83.5) | |
| Positive | 52 (13.4) | 27 (16.2) | 13 (16.5) | |
|
| 0.300 | |||
| Negative | 136 (35.1) | 68 (40.7) | 54 (68.4) | |
| Positive | 252 (64.9) | 99 (59.3) | 25 (31.6) |
Quantitative variables were in mean ± SD and qualitative variables are in n (%). †According to the eighth edition AJCC Cancer Staging Manual. The bolded P-value showed statistically significant (P-value < 0.05).
Difference of various deep transfer learning models.
| Models | Cohorts | AUC (95% CI) | Accuracy | Sensitivity | Specificity |
| Resnet152 | Training | 0.909 (0.874–0.943) | 0.830 | 0.872 | 0.846 |
| Internal validation | 0.901 (0.847–0.956) | 0.962 | 0.800 | 0.881 | |
| External validation | 0.915 (0.850–0.981) | 0.861 | 0.882 | 0.806 | |
| Resnet101 | Training | 0.923 (0.889–0.958) | 0.835 | 0.885 | 0.904 |
| Internal validation | 0.887 (0.829–0.945) | 0.826 | 0.735 | 0.898 | |
| External validation | 0.899 (0.803–0.996) | 0.848 | 0.882 | 0.839 | |
| Resnet50 | Training | 0.937 (0.909–0.966) | 0.869 | 0.892 | 0.892 |
| Internal validation | 0.882 (0.825–0.940) | 0.850 | 0.735 | 0.898 | |
| External validation | 0.900 (0.821–0.980) | 0.823 | 0.941 | 0.790 | |
| Resnet34 | Training | 0.916 (0.883–0.949) | 0.832 | 0.872 | 0.858 |
| Internal validation | 0.877 (0.817–0.937) | 0.832 | 0.755 | 0.864 | |
| External validation | 0.884 (0.805–0.963) | 0.823 | 0.941 | 0.694 | |
| Resnet18 | Training | 0.939 (0.911–0.967) | 0.838 | 0.919 | 0.879 |
| Internal validation | 0.831 (0.761–0.901) | 0.820 | 0.592 | 0.932 | |
| External validation | 0.862 (0.762–0.963) | 0.810 | 0.824 | 0.806 | |
| Wide_resnet101_2 | Training | 0.921 (0.889–0.954) | 0.835 | 0.865 | 0.892 |
| Internal validation | 0.859 (0.795–0.922) | 0.826 | 0.755 | 0.881 | |
| External validation | 0.846 (0.742–0.951) | 0.747 | 0.882 | 0.661 | |
| Wide_resnet50_2 | Training | 0.937 (0.909–0.964) | 0.840 | 0.851 | 0.896 |
| Internal validation | 0.868 (0.806–0.929) | 0.844 | 0.714 | 0.907 | |
| External validation | 0.888 (0.800–0.976) | 0.861 | 0.706 | 0.903 | |
| Inception v3 | Training | 0.890 (0.852–0.929) | 0.830 | 0.851 | 0.846 |
| Internal validation | 0.897 (0.844–0.950) | 0.826 | 0.837 | 0.839 | |
| External validation | 0.900 (0.825–0.976) | 0.823 | 0.765 | 0.887 |
AUC, area under the receiver operating characteristic curve; 95% CI, 95% confidence interval.
FIGURE 4Area under the curve (AUC) of various groups of feature fusion in the training and internal validation cohorts. (A) DTL features (Resnet152); (B) clinical features; (C) radiomics features; (D) DTL features (Resnet152) + clinical features; (E) clinical + radiomics features; (F) DTL features (Resnet152) + radiomics features. DTL, deep transfer learning.
Performance of various combined models.
| Models | Cohorts | AUC (95% CI) | Accuracy | Sensitivity | Specificity |
| DTL features | Training | 0.697 (0.642–0.751) | 0.660 | 0.669 | 0.667 |
| Internal validation | 0.687 (0.601–0.773) | 0.725 | 0.857 | 0.476 | |
| External validation | 0.600 (0.449–0.750) | 0.785 | 0.710 | 0.607 | |
| Clinical variables | Training | 0.874 (0.838–0.910) | 0.794 | 0.878 | 0.717 |
| Internal validation | 0.807 (0.731–0.884) | 0.796 | 0.735 | 0.805 | |
| External validation | 0.882 (0.806–0.959) | 0.785 | 0.941 | 0.726 | |
| Radiomics features | Training | 0.823 (0.774–0.872) | 0.724 | 0.770 | 0.788 |
| Internal validation | 0.620 (0.532–0.709) | 0.689 | 0.776 | 0.530 | |
| External validation | 0.637 (0.464–0.811) | 0.747 | 0.706 | 0.629 | |
| DTL features + clinical variables | Training | 0.882 (0.845–0.919) | 0.789 | 0.878 | 0.742 |
| Internal validation | 0.878 (0.819–0.937) | 0.814 | 0.857 | 0.797 | |
| External validation | 0.913 (0.842–0.986) | 0.886 | 0.765 | 0.902 | |
| Radiomics features + clinical variables | Training | 0.952 (0.927–0.977) | 0.822 | 0.900 | 0.915 |
| Internal validation | 0.844 (0.780–0.909) | 0.808 | 0.673 | 0.873 | |
| External validation | 0.849 (0.739–0.959) | 0.835 | 0.765 | 0.839 | |
| DTL + radiomics features | Training | 0.707 (0.655–0.761) | 0.660 | 0.851 | 0.471 |
| Internal validation | 0.673 (0.579–0.766) | 0.719 | 0.694 | 0.619 | |
| External validation | 0.581 (0.415–0.746) | 0.759 | 0.706 | 0.541 | |
| DTL + radiomics features + clinical variables | Training | 0.909 (0.874–0.943) | 0.830 | 0.872 | 0.846 |
| Internal validation | 0.901 (0.847–0.956) | 0.962 | 0.800 | 0.881 | |
| External validation | 0.915 (0.850–0.981) | 0.861 | 0.882 | 0.806 |
AUC, area under the receiver operating characteristic curve; 95% CI, 95% confidence interval.
FIGURE 5Performance of different machine learning classifications based on radiomics + deep transfer learning (Resnet152) + clinical features in the training and internal validation cohorts. (A) support vector machine (SVM); (B) K-nearest neighbor (KNN); (C) random decision forests; and (D) XGBoost.