| Literature DB >> 35686096 |
Yixiao Liu1,2,3, Shen Jin4, Qi Shen1,2,3, Lufan Chang5, Shancheng Fang4, Yu Fan1,2,3, Hao Peng4, Wei Yu1,2,3.
Abstract
Background: Although deep learning systems (DLSs) have been developed to diagnose urine cytology, more evidence is required to prove if such systems can predict histopathology results as well.Entities:
Keywords: convolutional neural network; cyto-histo correlation; deep learning; urine cytology; urothelial carcinoma
Year: 2022 PMID: 35686096 PMCID: PMC9170952 DOI: 10.3389/fonc.2022.901586
Source DB: PubMed Journal: Front Oncol ISSN: 2234-943X Impact factor: 5.738
Figure 1The data acquisition process is illustrated here. We collected the cytology images retrospectively and consecutively from Sep 2014 to Jan 2020 and built a series of data sets for training, validation, and tests. For those who underwent surgeries within the next 1 year, surgical results were also followed. The preliminary test set was only used for cancer cell detection, while the internal and extra test sets were used for both cancer cell detection and malignancy prediction.
Baseline characteristics.
| Training and Validation | Preliminary Test | Negative Cytology with Benign Histopathology | Negative Cytology with Malignant Histopathology | |
|---|---|---|---|---|
| Age | ||||
| <60 | 79 | 10 | 26 | 26 |
| ≥60 | 308 | 44 | 36 | 307 |
| Sex | ||||
| Female | 260 | 11 | 23 | 95 |
| Male | 127 | 43 | 43 | 238 |
| Cytology diagnosis1 | ||||
| I | 0 | 0 | 57 | 248 |
| II | 0 | 0 | 5 | 85 |
| III | 260 | 40 | 0 | 0 |
| IV | 127 | 14 | 0 | 0 |
| V | 0 | 0 | 0 | 0 |
1Cytology was diagnosed following Papanicolaou classification.
Surgical follow-up.
| Training and Validation Set1 | Positive Cytology with Malignant Histology | Negative Cytology with Benign Histology | Extra Test Set | |
|---|---|---|---|---|
| N | 188 | 23 | 62 | 333 |
| Sex | ||||
| Female | 63 | 5 | 19 | 95 |
| Male | 125 | 18 | 43 | 238 |
| Age | ||||
| <60 | 26 | 3 | 26 | 74 |
| ≥60 | 162 | 20 | 36 | 259 |
| Surgery2 | ||||
| TUR-Bt or biopsy | 102 | 16 | 48 | 310 |
| nephroureterectomy | 100 | 6 | 13 | 34 |
| Radical cystectomy | 9 | 1 | 1 | 20 |
| Tumor | ||||
| Negative | 0 | 0 | 62 | 0 |
| upper urinary tract | 89 | 6 | 0 | 15 |
| Lower urinary tract | 92 | 17 | 0 | 301 |
| Synchronous U&L | 7 | 0 | 0 | 17 |
| tumor grade | ||||
| Low grade | 24 | 7 | 117 | |
| High grade | 164 | 16 | 213 | |
| NA3 | 0 | 0 | 3 | |
| Tumor stage | ||||
| Muscle non-invasive | 104 | 16 | 277 | |
| Muscle invasive | 82 | 6 | 52 | |
| NA4 | 2 | 1 | 4 |
Cytology was diagnosed following Papanicolaou criteria. Cancer grade was diagnosed using WHO2004. Tumor stage was diagnosed using TNM staging AJCC UICC 8th edition. Synchronous U&L, synchronous tumors in both upper and lower urinary tract. NA, not available.
1Only cases in the training and validation set who underwent surgery were listed here.
2Many cases undertook more than 1 procedure, either at one time or many times.
3Tumor grades were missing due to the following: a case reported as unable to rule out for low grade (n = 1); grade not reported for a case with in situ carcinoma (n = 1); a case reported as Grade 2 using WHO 1999 but not using WHO 2004 (n=1).
4Tumor stages were missing for those undertaken biopsies with no further operation available (n = 2 + 1 + 4).
Figure 2The overall design of the deep learning system. The ResNet 101 Faster-RCNN detected the UC cells while assigning a possibility to each of the cell, and an additional classifier picked the maximal possibility and predicted the histopathological malignant state according to the set threshold.
Figure 3The examples of snapshot images from positive cases and the results by the deep learning system were provided. The malignant cells detected were labeled by the Faster-RCNN, and the possibilities of each detection were also shown.
Figure 4The DLS performance to detect malignant cells is illustrated. (A) As the system tried to find more malignant cells (to achieve higher sensitivity), it made more mistakes classifying benign cells as malignant. The sensitivities at the optimal threshold for three sets were marked. (B) The accuracy at different thresholds. Thresholds were represented by their middle value (for example, x = 52.5 represented the interval of 50–55 points). Higher thresholds tend to have better detection accuracy. The performance at the optimal threshold of 50–55 points is highlighted by the dashed line. At this point, the increasing rates began to slow down.
Figure 5The DLS performance to predict malignancy is illustrated. (A) The receiver operating characteristics curve for the performance to predict malignancy for the internal and extra tests. (B) The distribution of maximal possibilities in the internal test set. Most images from benign histology were scored below 55. (C) The distribution of maximal possibilities in the extra test set. Most images from negative cases were also scored below 55 (highlighted by the dashed line).
Performance on the internal test set.
| Threshold | Sensitivity | Specificity | F1 Score | Kappa Score |
|---|---|---|---|---|
| 40.00 | 0.94 | 0.61 | 0.67 | 0.45 |
| 42.00 | 0.94 | 0.62 | 0.67 | 0.51 |
| 43.00 | 0.90 | 0.68 | 0.70 | 0.54 |
| 44.00 | 0.90 | 0.71 | 0.72 | 0.56 |
| 45.00 | 0.90 | 0.73 | 0.71 | 0.55 |
| 46.00 | 0.87 | 0.74 | 0.72 | 0.57 |
| 47.00 | 0.87 | 0.76 | 0.73 | 0.57 |
| 49.00 | 0.81 | 0.79 | 0.71 | 0.58 |
| 50.00 | 0.81 | 0.80 | 0.69 | 0.56 |
| 51.00 | 0.74 | 0.83 | 0.71 | 0.56 |
| 53.00 | 0.74 | 0.88 | 0.72 | 0.64 |
| 54.00 | 0.71 | 0.91 | 0.75 | 0.68 |
| 55.00 | 0.71 | 0.94 | 0.77 | 0.68 |
| 57.00 | 0.68 | 0.95 | 0.76 | 0.71 |
| 60.00 | 0.65 | 1.00 | 0.78 | 0.71 |
| 65.00 | 0.61 | 1.00 | 0.76 | 0.68 |
| 68.00 | 0.48 | 1.00 | 0.59 | 0.50 |
| 69.00 | 0.42 | 1.00 | 0.56 | 0.46 |
| 70.00 | 0.39 | 1.00 | 0.41 | 0.32 |
| 71.00 | 0.26 | 1.00 | 0.32 | 0.25 |
| 73.00 | 0.19 | 1.00 | 0.32 | 0.25 |
| 75.00 | 0.16 | 1.00 | 0.23 | 0.17 |
| 76.00 | 0.13 | 1.00 | 0.23 | 0.17 |
| 78.00 | 0.10 | 1.00 | 0.18 | 0.13 |
| 83.00 | 0.06 | 1.00 | 0.12 | 0.09 |
| 88.00 | 0.00 | 1.00 | NA | NA |
NA, not available.
Performance on the extra test set.
| Threshold | Sensitivity | Specificity | F1 Score | Kappa Score |
|---|---|---|---|---|
| 40.00 | 1.00 | 0.56 | 0.52 | 0.35 |
| 41.00 | 1.00 | 0.59 | 0.54 | 0.37 |
| 42.00 | 0.98 | 0.62 | 0.57 | 0.42 |
| 43.00 | 0.98 | 0.67 | 0.58 | 0.44 |
| 44.00 | 0.97 | 0.69 | 0.60 | 0.47 |
| 45.00 | 0.97 | 0.72 | 0.62 | 0.50 |
| 46.00 | 0.94 | 0.75 | 0.63 | 0.51 |
| 47.00 | 0.91 | 0.78 | 0.67 | 0.56 |
| 48.00 | 0.91 | 0.81 | 0.68 | 0.58 |
| 49.00 | 0.88 | 0.84 | 0.68 | 0.59 |
| 50.00 | 0.86 | 0.85 | 0.68 | 0.60 |
| 51.00 | 0.83 | 0.86 | 0.68 | 0.60 |
| 52.00 | 0.80 | 0.88 | 0.69 | 0.60 |
| 53.00 | 0.77 | 0.89 | 0.66 | 0.58 |
| 54.00 | 0.72 | 0.90 | 0.67 | 0.59 |
| 55.00 | 0.67 | 0.92 | 0.66 | 0.58 |
| 56.00 | 0.63 | 0.94 | 0.64 | 0.57 |
| 57.00 | 0.59 | 0.94 | 0.61 | 0.54 |
| 58.00 | 0.55 | 0.95 | 0.58 | 0.51 |
| 59.00 | 0.48 | 0.96 | 0.55 | 0.48 |
| 60.00 | 0.45 | 0.96 | 0.53 | 0.45 |
| 61.00 | 0.42 | 0.96 | 0.52 | 0.45 |
| 62.00 | 0.41 | 0.96 | 0.53 | 0.47 |
| 63.00 | 0.39 | 0.98 | 0.50 | 0.44 |
| 64.00 | 0.36 | 0.98 | 0.49 | 0.44 |
| 65.00 | 0.34 | 0.99 | 0.47 | 0.41 |
| 66.00 | 0.31 | 0.99 | 0.37 | 0.32 |
| 67.00 | 0.23 | 0.99 | 0.24 | 0.20 |
| 68.00 | 0.14 | 1.00 | 0.24 | 0.20 |
| 70.00 | 0.13 | 1.00 | 0.19 | 0.16 |
| 71.00 | 0.11 | 1.00 | 0.14 | 0.11 |
| 72.00 | 0.08 | 1.00 | 0.09 | 0.07 |
| 73.00 | 0.05 | 1.00 | 0.06 | 0.04 |
| 74.00 | 0.03 | 1.00 | NA | NA |
| 78.00 | 0.00 | 1.00 | NA | NA |
NA, not available.
Figure 6Different designs in between different DLSs. The current DLS could make both cytopathological diagnosis and histopathological prediction, yet it did not need histopathological data in its training.