| Literature DB >> 35087174 |
Soumick Chatterjee1,2,3, Faraz Ahmed Nizamani4, Andreas Nürnberger5,6,7, Oliver Speck8,7,9,10.
Abstract
A brain tumour is a mass or cluster of abnormal cells in the brain, which has the possibility of becoming life-threatening because of its ability to invade neighbouring tissues and also form metastases. An accurate diagnosis is essential for successful treatment planning, and magnetic resonance imaging is the principal imaging modality for diagnosing brain tumours and their extent. Deep Learning methods in computer vision applications have shown significant improvement in recent years, most of which can be credited to the fact that a sizeable amount of data is available to train models, and the improvements in the model architectures yield better approximations in a supervised setting. Classifying tumours using such deep learning methods has made significant progress with the availability of open datasets with reliable annotations. Typically those methods are either 3D models, which use 3D volumetric MRIs or even 2D models considering each slice separately. However, by treating one spatial dimension separately or by considering the slices as a sequence of images over time, spatiotemporal models can be employed as "spatiospatial" models for this task. These models have the capabilities of learning specific spatial and temporal relationships while reducing computational costs. This paper uses two spatiotemporal models, ResNet (2+1)D and ResNet Mixed Convolution, to classify different types of brain tumours. It was observed that both these models performed superior to the pure 3D convolutional model, ResNet18. Furthermore, it was also observed that pre-training the models on a different, even unrelated dataset before training them for the task of tumour classification improves the performance. Finally, Pre-trained ResNet Mixed Convolution was observed to be the best model in these experiments, achieving a macro F1-score of 0.9345 and a test accuracy of 96.98%, while at the same time being the model with the least computational cost.Entities:
Mesh:
Year: 2022 PMID: 35087174 PMCID: PMC8795458 DOI: 10.1038/s41598-022-05572-6
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1An example MRI of Low-grade glioma (LGG, on the left) and High-grade glioma (HGG, on the right). Source: BraTS 2019.
Figure 2High-grade glioma structure on T1ce, T2 and FLAIR contrast images (from left to right), (red circle) Necrotic core, (blue circle) Perifocal oedema. Source: BraTS 2019.
Figure 3Schematic representations of the network architectures. (a) ResNet (2+1)D, (b) ResNet Mixed Convolution, and (c) ResNet 3D.
Total number of trainable parameters for each model.
| Model | No. of parameters |
|---|---|
| ResNet3D | 33,150,522 |
| ResNet (2+1)D | 31,297,254 |
| ResNet mixed convolution | 11,472,963 |
Figure 4Confusion matrix for 3-fold cross-validation on pre-trained ResNet(2+1)D.
Figure 5Confusion matrix for 3-fold cross-validation on ResNet mixed convolution.
Figure 6Confusion matrix for 3-fold cross-validation on ResNet3D18.
Figure 7Heatmaps showing the class-wise performance of the classifiers, compared using precision, recall, specificity, and F1-score: (a) LGG, (b) HGG, and (c) healthy.
Low-grade glioma model comparison (*denotes the overall winning model).
| Low-grade glioma | |
|---|---|
| Model | Mean F1 score |
| ResNet 3D | 0.8542 |
| Pre-trained ResNet 3D | 0.8143 |
| ResNet(2+1)D | 0.8448 |
| Pre-trained ResNet(2+1)D | 0.8739 |
| ResNet mixed convolution | 0.7734 |
| Pre-trained ResNet mixed convolution | |
High-grade glioma model comparison (*denotes the overall winning model).
| High-grade glioma | |
|---|---|
| Model | Mean F1 score |
| ResNet 3D | 0.8773 |
| Pre-trained ResNet 3D | 0.8634 |
| ResNet(2+1)D | 0.8730 |
| Pre-trained ResNet(2+1)D | 0.8979 |
| ResNet Mixed Convolution | 0.8231 |
| Pre-trained ResNet mixed convolution | |
Healthy brain model comparison (*denotes the overall winning model).
| Healthy brain | |
|---|---|
| Model | Mean F1 score |
| ResNet 3D | 0.9970 |
| Pre-trained ResNet 3D | |
| ResNet(2+1)D | 0.9927 |
| Pre-trained ResNet(2+1)D | 0.9992 |
| ResNet mixed convolution | 0.9855 |
| Pre-trained ResNet mixed convolution | 0.9963 |
Consolidated comparison of the models (*denotes the overall winning model).
| Consolidated scores | ||
|---|---|---|
| Model | Macro F1 score | Weighted F1 score |
| ResNet 3D | 0.9095 | 0.9269 |
| Pre-trained ResNet 3D | 0.8925 | 0.9171 |
| ResNet(2+1)D | 0.9035 | 0.9220 |
| Pre-trained ResNet(2+1)D | 0.9237 | 0.9393 |
| ResNet mixed convolution | 0.8607 | 0.8881 |
| Pre-trained ResNet mixed convolution | ||
Comparisons against other published works *s=specificity cross-validated State of the art.
| Study | Method | Contrast | Dimension | Test Accuracy |
|---|---|---|---|---|
| Shahzadi et al.[ | CNN with LSTM | T2-FLAIR | 3D | 84.00 % |
| Pei et al.[ | Similar to U-Net for Segmentation, Regular CNN for classification | T1, T1ce, T2, T2-FLAIR | 3D | 74.9% |
| Ge et al.[ | Deep CNN | T1, T2 , T2-FLAIR | 2D | 90.87% |
| Yang et al.[ | Pre-trained GoogLeNet | T1ce | 2D | 94.5% |
| Mzoughi et al.[ | Deep CNN | T1ce | 3D | 96.49% |
| Zhuge et al.[ | Deep CNN | T1, T1ce, T2, T2-FLAIR | 3D | 97.1% s*=0.968 |
| Ouerghi et al.[ | Random forest | T1, T2, T2-FLAIR | 2D | 96.5% |
| This paper | Pre-trained ResNet mixed convolution spatiospatial model |
|
|
|