| Literature DB >> 35956236 |
Jiangfen Wu1,2,3, Qian Xu4, Yiqing Shen3,5, Weidao Chen3, Kai Xu4, Xian-Rong Qi1.
Abstract
Background: Deep learning (DL) could predict isocitrate dehydrogenase (IDH) mutation status from MRIs. Yet, previous work focused on CNNs with refined tumor segmentation. To bridge the gap, this study aimed to evaluate the feasibility of developing a Transformer-based network to predict the IDH mutation status free of refined tumor segmentation.Entities:
Keywords: IDH mutation status; ResNet; Swin transformer; bounding box; image inputs
Year: 2022 PMID: 35956236 PMCID: PMC9369996 DOI: 10.3390/jcm11154625
Source DB: PubMed Journal: J Clin Med ISSN: 2077-0383 Impact factor: 4.964
Figure 1Flowchart of patient enrollment and network implementation. (A) Flow diagram of AHXZ dataset enrollment. (B) Patient information for model implementation. IDHm = IDH mutant type; IDHw = IDH wild type.
Figure 2Overview of this study design. This study includes five key steps: tumor delineation, image preprocessing and augmentation, image inputs definition, network development, and hybrid model development.
Figure 3Tumor delineation and image inputs definition. (a) IDH-mutant case, female, 59 years old, WHO grade Ⅲ; (b) IDH-wild case, male, 53 years old, WHO grade Ⅳ. The two cases enrolled from AHXZ. Red voxels represent tumor and cyan voxels represent edema.
Description of the seven pre-processing strategies for the input images.
| Image Inputs | Description | |
|---|---|---|
|
| Tumor slice | the whole slices contained the tumor mask |
|
| Tumor mask | the tumor region alone by setting all outside tumor pixels as zero |
|
| Tumor mask + Edema | The joint region that contained both tumor region and the edema region by setting all outside pixels as zero |
|
| 0.8× Tumor Bbox | downscaled the bounding box of tumor mask by 0.8 |
|
| 1.0× Tumor Bbox | bounding box of tumor mask |
|
| 1.2× Tumor Bbox | enlarged the bounding box of tumor mask by 1.2 times |
|
| 1.5× Tumor Bbox | enlarged the bounding box of tumor mask by 1.5 times |
Note: both Tumor mask and Tumor mask + Edema were defined as refined segmentation input image; The 0.8× Tumor Bbox, 1.0× Tumor Bbox, 1.2× Tumor Bbox and 1.5× Tumor Bbox were defined as free of segmentation input image.
Figure 4The entire classification process and Swin Transformer architecture. LN: layer normalization; MLP: multilayer perceptron; W-MSA: window multi-head self-attention; SW-MSA: shifted-window multi-head self-attention.
Patient characteristics.
| TCIA ( | TCIA | AHXZ ( | AHXZ | ||||
|---|---|---|---|---|---|---|---|
| IDHstatus | IDH-mutant | IDH-wild | - | IDH-mutant | IDH-wild | - | 0.053 |
| 112(43.2%) | 147(56.8%) | 81(34.6%) | 153(65.4%) | ||||
| gender | 0.208 | >0.999 | 0.277 | ||||
| female | 58 | 63 | 34 | 64 | |||
| male | 54 | 83 | 47 | 89 | |||
| age | 51.5 ± 15.7 | <0.05 | 52.2 ± 13.1 | <0.05 | 0.678 | ||
| 42.4 ± 13.8 | 58.5 ± 13.3 | 47.22 ± 11.7 | 54.79 ± 13.1 | ||||
| locationfeatures | <0.05 | 0.002 | <0.05 | ||||
| frontal | 53 | 33 | 36 | 31 | |||
| temporal | 18 | 42 | 9 | 16 | |||
| occipital | 1 | 4 | 1 | ||||
| parietal | 13 | 24 | 1 | 10 | |||
| Lociothers # | 6 | 3 | 5 | 13 | |||
| Multiplelobes | 21 | 41 | 30 | 82 | |||
| hemispheredistribution | 0.924 | 0.097 | 0.01 | ||||
| left | 53 | 73 | 44 | 61 | |||
| right | 53 | 67 | 28 | 64 | |||
| bothsides | 6 | 7 | 9 | 23 | |||
| hemisphereothers ## | 5 | ||||||
| WHOgrade | <0.05 | <0.05 | 0.059 | ||||
| 2 | 62 | 6 | 40 | 23 | |||
| 3 | 44 | 23 | 24 | 17 | |||
| 4 | 6 | 118 | 17 | 113 | |||
Note: IDHm = IDH mutant type; IDHw = IDH wild type. # Loci Others including Insula, basal ganglia, thalamus, cerebellum, brainstem; ## hemisphere others including cerebellum and brain stem.
Patient-level diagnostic performance of the models for the IDH mutation status prediction.
| TCIA Internal Test Set | AHXZ External Test Set | |||
|---|---|---|---|---|
| AUC | ACC | AUC | ACC | |
| ResNet | ||||
| Tumor Slice | 0.933 | 88.5% | 0.763 | 66.5% |
| Tumor Mask | 0.936 | 90.4% | 0.818 | 73.0% |
| Tumor + Edema | 0.911 | 84.6% | 0.823 | 76.8% |
| 0.8× Tumor Bbox | 0.897 | 92.3% | 0.818 | 76.0% |
| 1.0× Tumor Bbox | 0.938 | 92.3% | 0.831 | 77.3% |
| 1.2× Tumor Bbox | 0.924 | 90.4% | 0.802 | 77.7% |
| 1.5× Tumor Bbox | 0.915 | 86.5% | 0.783 | 77.3% |
| average | 0.922 | 89.3% | 0.805 | 74.9% |
| Swin Transformer | ||||
| Tumor Slice | 0.955 | 90.4% | 0.804 | 70.8% |
| Tumor Mask | 0.975 | 90.4% | 0.849 | 73.4% |
| Tumor + Edema | 0.946 | 92.3% | 0.862 | 78.5% |
| 0.8× Tumor Bbox | 0.953 | 88.5% | 0.858 | 77.7% |
| 1.0× Tumor Bbox | 0.975 | 96.2% | 0.868 | 80.7% |
| 1.2× Tumor Bbox | 0.965 | 92.3% | 0.827 | 76.4% |
| 1.5× Tumor Bbox | 0.984 | 96.2% | 0.829 | 79.0% |
| average | 0.965 | 92.3% | 0.842 | 76.6% |
Note: AUC = area under the ROC curve; ACC = accuracy.
Figure 5Visualization of model performance. (a) Radar map of the model AUC. When the same input images were used, the transformer models obtained higher AUC than the corresponding ResNet models. (b) Histogram of the model AUC and accuracy (ACC). All the histogram was presented in the order of model AUC from highest to lowest. Using 1.0× Tumor Bbox images as inputs, both the Swin Transformer and ResNet models achieved the best performance on the external test.
Figure 6ROCs of the top performing image-based models and their corresponding hybrid models.