| Literature DB >> 28710497 |
Zeju Li1, Yuanyuan Wang2,3, Jinhua Yu4,5, Yi Guo1,6, Wei Cao7.
Abstract
Deep learning-based radiomics (DLR) was developed to extract deep information from multiple modalities of magnetic resonance (MR) images. The performance of DLR for predicting the mutation status of isocitrate dehydrogenase 1 (IDH1) was validated in a dataset of 151 patients with low-grade glioma. A modified convolutional neural network (CNN) structure with 6 convolutional layers and a fully connected layer with 4096 neurons was used to segment tumors. Instead of calculating image features from segmented images, as typically performed for normal radiomics approaches, image features were obtained by normalizing the information of the last convolutional layers of the CNN. Fisher vector was used to encode the CNN features from image slices of different sizes. High-throughput features with dimensionality greater than 1.6*104 were obtained from the CNN. Paired t-tests and F-scores were used to select CNN features that were able to discriminate IDH1. With the same dataset, the area under the operating characteristic curve (AUC) of the normal radiomics method was 86% for IDH1 estimation, whereas for DLR the AUC was 92%. The AUC of IDH1 estimation was further improved to 95% using DLR based on multiple-modality MR images. DLR could be a powerful way to extract deep information from medical images.Entities:
Mesh:
Substances:
Year: 2017 PMID: 28710497 PMCID: PMC5511238 DOI: 10.1038/s41598-017-05848-2
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Characteristics of patients in all datasets, first cohort and second cohort.
| Parameters | Total Cases | IDH1 Mutation Status | p-value | |
|---|---|---|---|---|
| Mutation | Wild Type | |||
|
| ||||
| Number of samples | 151 | 112 | 39 | |
|
| ||||
| Male | 81(54.6%) | 58(51.8%) | 23(59.0%) | 0.44 |
| Female | 70(46.4%) | 54(48.2%) | 16(41.0%) | |
|
| ||||
| Mean ± standard deviation | 40.7 ± 10.8 | 38.7 ± 10.7 | 43.5 ± 12.1 | 0.03 |
|
| ||||
| Mean ± standard deviation | 68.1 ± 47.4 | 69.0 ± 48.0 | 65.8 ± 46.8 | 0.843 |
|
| ||||
| Astrocytoma | 81(54.6%) | 54(48.2%) | 27(39.2%) | 0.01 |
| Oligodendroglioma | 31(20.5%) | 27(24.1%) | 4(10.2%) | |
| Oligoastrocytoma | 39(25.8%) | 31(27.7%) | 8(20.5%) | |
|
| ||||
| Number of samples | 119 | 89 | 30 | |
|
| ||||
| Male | 67(56.3%) | 48(54.0%) | 19(63.3%) | 0.37 |
| Female | 52(43.7%) | 41(46.1%) | 11(36.7%) | |
|
| ||||
| Mean ± standard deviation | 39.6 ± 10.2 | 37.9 ± 8.9 | 44.2 ± 11.7 | 0.03 |
|
| ||||
| Mean ± standard deviation | 68.3 ± 57.4 | 66.7 ± 55.6 | 73.1 ± 63.0 | 0.146 |
|
| ||||
| Astrocytoma | 69(58.0%) | 47(52.8%) | 22(73.3%) | 0.03 |
| Oligodendroglioma | 24(20.2%) | 21(23.6%) | 3(10.0%) | |
| Oligoastrocytoma | 26(21.8%) | 21(23.6%) | 5(16.7%) | |
|
| ||||
| Before 2015 | 85(71.4%) | 63(70.8%) | 22(73.3%) | 0.79 |
| After 2015 | 34(28.6%) | 26(29.2%) | 8(26.7%) | |
|
| ||||
| Number of samples | 110 | 76 | 34 | |
|
| ||||
| Male | 54(49.1%) | 33(43.4%) | 21(61.7%) | 0.08 |
| Female | 56(50.9%) | 43(56.6%) | 13(38.2%) | |
|
| ||||
| Mean ± standard deviation | 40.3 ± 11.3 | 39.0 ± 10.7 | 43.4 ± 12.7 | 0.06 |
|
| ||||
| Mean ± standard deviation | 68.1 ± 57.9 | 74.3 ± 62.5 | 54.3 ± 43.8 | 0.06 |
|
| ||||
| Astrocytoma | 55(50.0%) | 31(40.8%) | 24(70.6%) | 0.01 |
| Oligodendroglioma | 21(19.1%) | 18(23.7%) | 3(8.8%) | |
| Oligoastrocytoma | 34(30.9%) | 27(35.5%) | 7(20.6%) | |
Figure 1Tumor segmentation results for single-modal images and multi-modal images using different network structures. (a) Comparison between the indexes of different segmentation results of different CNNs. Conv. indicates the number of convolutional layers in the CNN structures, and fc. indicates the number of neurons in the fully connected layers in the CNN structures. (b) Three typical cases with segmentation results for the CNN with 6 convolutional layers and fully connected layers with 4096 neurons.
Figure 2CNN features from the last convolutional layers. (a) An example of specific CNN features. Deep filter responses showed noticeable differences between wild-type and mutant IDH1, and the Fisher vectors could successfully represent the differences.
Prediction results of different cohorts using different methods.
| Dataset | Methods | AUC | ACC | SENS | SPEC | PPV | NPV | MCC |
|---|---|---|---|---|---|---|---|---|
|
| ||||||||
| Second cohort with leave-one-out cross-validation | Radiomics[ | 0.8572 | 0.8000 | 0.8289 |
| 0.8750 | 0.6579 | 0.5483 |
| DLR |
|
|
| 0.7059 |
|
|
| |
|
| ||||||||
| First cohort with leave-one-out cross-validation | DLR with single modality | 0.8045 | 0.8235 | 0.9326 | 0.5000 | 0.8469 | 0.7143 | 0.4927 |
| DLR with multiple modality | 0.9157 | 0.8655 |
| 0.6333 | 0.8842 | 0.7917 | 0.6246 | |
| DLR with multiple modality improved by further feature selection |
|
|
|
|
|
|
| |
|
| ||||||||
| First cohort with divided test set | DLR with multiple modality | 0.9615 | 0.9118 | 0.9231 | 0.8750 | 0.9600 | 0.7778 | 0.7673 |
(a) Comparison of IDH1 prediction results between radiomics and DLR using T2 flair modal MR images of the first cohort by leave-one-out cross-validation SVM. (b) Prediction results of IDH1 using multi-modal MR images of the second cohort by leave-one-out cross-validation SVM and (c) validation on a divided test set. The data set was divided according to the diagnosis time.
Figure 3ROC curves of the prediction results. (a) ROC curves of the radiomics features and DLR of the second cohort with single-modal images. (b) ROC curves of DLR of the first cohort with multiple modal images are shown on the right.
Comparison of IDH1 prediction results using CNN features from different layers.
| Methods | AUC | ACC | SENS | SPEC | PPV | NPV | MCC |
|---|---|---|---|---|---|---|---|
| Conv.1 | 0.6165 | 0.5630 | 0.5393 | 0.6333 | 0.8136 | 0.3167 | 0.1499 |
| Conv.2 | 0.7109 | 0.6387 | 0.6404 | 0.6333 | 0.8382 | 0.3725 | 0.2402 |
| Conv.3 | 0.8858 | 0.8403 | 0.9213 | 0.6000 | 0.8723 | 0.7200 | 0.5557 |
| Conv.4 | 0.8734 | 0.7899 | 0.8876 | 0.5000 | 0.8404 | 0.6000 | 0.4132 |
| Conv.5 | 0.9004 | 0.8571 | 0.9101 |
|
| 0.7241 | 0.6171 |
| Fc.7 | 0.8614 | 0.8319 | 0.9326 | 0.5333 | 0.8557 | 0.7273 | 0.5212 |
| Fc.8 | 0.7524 | 0.7647 | 0.8876 | 0.4000 | 0.8144 | 0.5455 | 0.3217 |
| Conv.6 |
|
|
| 0.6333 | 0.8842 |
|
|
Conv. means the convolutional layers, and fc. represents the fully connected layers. The same post-processing was applied in each situation.
Figure 4Feature maps of CNN features from different convolutional layers. The four most significant filter banks of different layers were selected. As shown in the figures, feature maps of deeper layers represented more detailed characteristics.
Figure 5An overview of our DLR. Our approach included two selection steps. The first step is to recognize the tumor regions in the MR images based on a state-of-the-art CNN structure. In the second step, deep filter responses were extracted from the last convolutional layer through Fisher vector encoding. Then, the prediction results were evaluated by a leave-one-out cross-validation SVM.