| Literature DB >> 36010200 |
Yuting Xie1, Fulvio Zaccagna1,2, Leonardo Rundo3, Claudia Testa2,4, Raffaele Agati5, Raffaele Lodi1,6, David Neil Manners1, Caterina Tonon1,2.
Abstract
Convolutional neural networks (CNNs) constitute a widely used deep learning approach that has frequently been applied to the problem of brain tumor diagnosis. Such techniques still face some critical challenges in moving towards clinic application. The main objective of this work is to present a comprehensive review of studies using CNN architectures to classify brain tumors using MR images with the aim of identifying useful strategies for and possible impediments in the development of this technology. Relevant articles were identified using a predefined, systematic procedure. For each article, data were extracted regarding training data, target problems, the network architecture, validation methods, and the reported quantitative performance criteria. The clinical relevance of the studies was then evaluated to identify limitations by considering the merits of convolutional neural networks and the remaining challenges that need to be solved to promote the clinical application and development of CNN algorithms. Finally, possible directions for future research are discussed for researchers in the biomedical and machine learning communities. A total of 83 studies were identified and reviewed. They differed in terms of the precise classification problem targeted and the strategies used to construct and train the chosen CNN. Consequently, the reported performance varied widely, with accuracies of 91.63-100% in differentiating meningiomas, gliomas, and pituitary tumors (26 articles) and of 60.0-99.46% in distinguishing low-grade from high-grade gliomas (13 articles). The review provides a survey of the state of the art in CNN-based deep learning methods for brain tumor classification. Many networks demonstrated good performance, and it is not evident that any specific methodological choice greatly outperforms the alternatives, especially given the inconsistencies in the reporting of validation methods, performance metrics, and training data encountered. Few studies have focused on clinical usability.Entities:
Keywords: brain tumor classification; clinical application; clinical effectiveness; computer-aided diagnosis; convolutional neural network; deep learning; magnetic resonance imaging
Year: 2022 PMID: 36010200 PMCID: PMC9406354 DOI: 10.3390/diagnostics12081850
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
The search queries used to interrogate the PubMed and Scopus databases.
| PubMed | (deep learning OR deep model OR artificial intelligence OR artificial neural network OR autoencoder OR generative adversarial network) OR convolutional OR (neural network) OR neural network OR deep model OR convolutional) | AND |
| (brain tumor OR glioma OR brain cancer OR glioblastoma OR astrocytoma OR oligodendroglioma OR ependymoma) | AND | |
| (classification OR grading OR classify) | AND | |
| (MRI OR Magnetic Resonance OR MR images OR radiographic OR radiology) | IN | |
| Title/Abstract |
Figure 1The PRISMA flowchart of this review. n: number of articles.
Figure 2The basic workflow of a typical CNN-based brain tumor classification study with four high-level steps: Step 1. Input Image: 2D or 3D Brain MR samples are fed into the classification model; Step 2. Preprocessing: several preprocessing techniques are used to remove the skull, normalize the images, resize the images, and augment the number of training examples; Step 3. CNN Classification: the preprocessed dataset is propagated into the CNN model and is involved in training, validation, and testing processes; Step 4. Performance Evaluation: evaluation of the classification performance of a CNN algorithm with accuracy, specificity, F1 score, area under the curve, and sensitivity metrics.
An overview of publicly available datasets.
| Dataset Name | Available Sequences | Size | Classes | Unbiased Gini Coefficient | Source |
|---|---|---|---|---|---|
| TCGA-GBM | T1w, ceT1w, T2w, FLAIR | 199 patients | N/D | N/D | [ |
| TCGA-LGG | T1w, ceT1ce, T2w, FLAIR | 299 patients | N/D | N/D | [ |
| Brain tumor dataset from Figshare (Cheng et al., 2017) | ceT1w | 233 patients (82 MEN, 89 Glioma, 62 PT), 3064 images (708 MEN, 1426 Glioma, 930 PT) | Patients (82 MEN, 89 Glioma, 62 PT), images (708 MEN, 1426 Glioma, 930 PT) | 0.116 (patients), 0.234 (images) | [ |
| Kaggle (Navoneel et al., 2019) | No information given | 253 images (98 normal, 155 tumorous) | 98 normal, 155 tumorous | 0.225 | [ |
| REMBRANDT | T1w, T2w, FLAIR, DWI | 112 patients (30 AST-II, 17 AST-II, 14 OLI-II, 7 OLI-III, 44 GBM) | 30 AST-II, 17 AST-II, 14 OLI-II, 7 OLI-III, 44 GBM | 0.402 | [ |
| BraTS | T1w, ceT1w, T2w, FLAIR | 2019: 335 patients (259 HGG, 76 LGG); 2018: 284 patients (209 HGG, 75 LGG); 2017: 285 patients (210 HGG, 75 LGG); 2015: 274 patients (220 HGG, 54 LGG) | 2019: 259 HGG, 76 LGG;2018: 209 HGG, 75 LGG;2017: 210 HGG, 75 LGG; 2015: 220 HGG, 54 LGG | 0.546 (2019); 0.472 (2018); 0.474 (2017); 0.606 (2015) | [ |
| ClinicalTrials.gov (Liu et al., 2017) | T1w, ceT1w, T2w, FLAIR | 113 patients (52 LGG, 61 HGG) | 52 LGG, 61 HGG | 0.080 | [ |
| CPM-RadPath 2019 | T1w, ceT1w, T2w, FLAIR | 329 patients | N/D | N/D | [ |
| IXI dataset | T1w, T2w, DWI | 600 normal images | N/D | N/D | [ |
| RIDER | T1w, T2w, DCE-MRI, ce-FLAIR | 19 GBM patients (70,220 images) | 70,220 images | N/D | [ |
| Harvard Medical School Data | T2w | 42 patients (2 normal, 40 tumor), 540 images (27 normal, 513 tumorous) | Patients (2 normal, 40 tumorous), images (27 normal, 513 tumorous) | 0.905 (patients), 0.900 (images) | [ |
The imaging configurations and main clinical distinctions of T1w, T2w, ceT1w, and FLAIR.
| Sequence | Sequence Characteristics | Main Clinical Distinctions | Example * |
|---|---|---|---|
| T1w | Uses short TR and TE [ |
Lower signal for a higher water content [ Higher signal for fat [ Higher signal for subacute hemorrhage [ |
|
| T2w | Uses long TR and TE [ |
Higher signal for a higher water content, such as in edema, tumor, infarction, inflammation, infection, or subdural collection [ Lower signal for fat [ Lower signal for fibrous tissue [ |
|
| ceT1w | Uses the same TR and TE as T1w; employs contrast agents [ |
Higher signal for areas of breakdown in the blood–brain barrier that indicate induced inflammation [ |
|
| FLAIR | Uses very long TR and TE; the inversion time nulls the signal from fluid [ |
Highest signal for abnormalities [ Highest signal for gray matter [ Lower signal for cerebrospinal fluid [ |
|
* Pictures from [68]. TR, repetition time. TE, echo time.
Figure 3Data augmentation: (a) original image; (b) 18° rotation. When rotating by an arbitrary number of degrees (non-modulo 90), rotation will result in the image being padded in each corner. Then, a crop is taken from the center of the newly rotated image to retain the largest crop possible while maintaining the image’s aspect ratio; (c) left–right flipping; (d) top–bottom flipping; (e) scaling by 1.5 times; (f) cropping by center cropping to the size 150 × 150; (g) random brightness enhancement; (h) random contrast enhancement.
Figure 4Number of articles published from 2015 to 2022.
Figure 5Usage of preprocessing techniques from 2017 to 2022.
Figure 6Usage of state-of-the-art CNN models from 2015 and 2022.
(a) Overview of included studies that focus on CNN-based deep learning methods for brain tumor classification, with the exception of studies focusing on normal vs. tumorous classification. Datasets, MRI sequences, size of the datasets, and preprocessing methods are summarized. (b) Overview of included studies that focus on CNN-based deep learning methods for brain tumor classification, with the exception of study focusing on normal vs. tumorous classification. Classification tasks, classification architecture, validation methods, and performance metrics are summarized.
| (a) | |||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Author and Year | Datasets | MRI | Size of Dataset | Pre-Processing | Data Augmentation | ||||||||||||
| Patients | Images | Cropping | Normalization | Resizing | Skull Stripping | Registration 1 | Other | Translation 2 | Rotation | Scaling 3 | Reflection 4 | Shearing | Cropping | Other | |||
| Özcan et al. [ | Private dataset | T2w/FLAIR | 104 (50 LGG, 54 HGG) | 518 | x | x | Conversion to BMP | x | x | x | x | ||||||
| Hao et al. [ | BraTS 2019 | T1w, ceT1w, T2w | 335 (259 HGG, 76 LGG) | 6700 | x | x | x | ||||||||||
| Tripathi et al. [ | 1. TCGA-GBM, | T2w | 322 (163 HGG, 159 LGG) | 7392 (5088 LGG, 2304 HGG) | x | x | x | x | x | x | |||||||
| Ge et al. [ | BraTS 2017 | T1w, ceT1w, T2w, FLAIR | 285 (210 HGG, 75 LGG) | x | x | ||||||||||||
| Mzoughi et al. [ | BraTS 2018 | ceT1w | 284 (209 HGG, 75 LGG) | x | x | Contrast enhancement | x | ||||||||||
| Yang et al. [ | ClinicalTrials.gov (NCT026226201) | ceT1w | 113 (52 LGG, 61 HGG) | Conversion to BMP | x | x | x | Histogram equalization, adding noise | |||||||||
| Zhuge et al. [ | 1.TCIA-LGG, 2. BraTS 2018 | T1w, T2w, FLAIR, ceT1w | 315 (210 HGG, 105 LGG) | x | x | Clipping, bias field correction | x | x | x | ||||||||
| Decuyper et al. [ | 1. TCGA-LGG, 2. TCGA-GBM, 3. TCGA-1p19qDeletion, 4. BraTS 2019. 5. GUH dataset | T1w, ceT1w, T2w, FLAIR | 738 (164 from TCGA-GBM, 121 from TCGA-LGG, 141 from 1p19qDeletion, 202 from BraTS 2019, 110 from GUH dataset) (398 GBM vs. 340 LGG) | x | x | x | Interpolation | x | x | Elastic transform | |||||||
| He et al. [ | 1.Dataset from TCIA | FLAIR, ceT1w | 214 (106 HGG, 108 LGG) | x | x | x | x | ||||||||||
| 2. BraTS 2017 | FLAIR, ceT1w | 285 (210 HGG, 75 LGG) | x | x | x | x | |||||||||||
| Hamdaoui et al. [ | BraTS 2019 | T1w, ceT1w, T2w, FLAIR | 285 (210 HGG, 75 LGG) | 53,064 (26,532 HGG, 26,532 LGG) | x | x | x | ||||||||||
| Chikhalikar et al. [ | BraTS 2015 | T2w, FLAIR | 274 (220 HGG, 54 LGG) | 521 | Contrast enhancement | ||||||||||||
| Ahmad [ | BraTS 2015 | No info shared | 124 (99 HGG, 25 LGG) | x | |||||||||||||
| Naser et al. [ | TCGA-LGG | T1W, FLAIR, ceT1w | 108 (50 Grade II, 58 Grade III) | x | x | x | Padding | x | x | x | x | x | |||||
| Allah et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | PGGAN | |||||||||
| Swati et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | |||||||||||
| Guan et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | Contrast enhancement | x | x | ||||||||
| Deepak et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | |||||||||||
| Díaz-Pernas et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | Elastic transform | |||||||||||
| Ismael et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | x | x | x | x | Whitening, brightness manipulation | |||||
| Alhassan et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | ||||||||||||
| Bulla et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | |||||||||||
| Ghassemi et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | ||||||||||
| Kakarla et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | Contrast enhancement | ||||||||||
| Noreen et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | ||||||||||||
| Noreen et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | ||||||||||||
| Kumar et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | ||||||||||||
| Badža et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | x | |||||||||
| Alaraimi et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | x | x | x | x | ||||||
| Lo et al. [ | Dataset from TCIA ** | ceT1w | 130 (30 Grade II, 43 Grade III, 57 Grade IV) | x | x | Contrast enhancement | x | x | x | x | x | ||||||
| Kurc et al. [ | Data from TCGA | ceT1w, T2-FLAIR | 32 (16 OLI, 16 AST) | x | x | Bias field correction | x | x | |||||||||
| Pei et al. [ | 1. CPM-RadPath 2019, 2. BraTS 2019 | T1w, ceT1w, T2w, FLAIR | 398 (329 from CPM-RadPath 2019, 69 from BraTS 2019) | x | x | x | Noise | x | x | x | |||||||
| Ahammed et al. [ | Private dataset | T2w | 20 | 557 (130 Grade I, 169 Grade II, Grade III 103, Grade IV 155) | x | Filtering, enhancement | x | x | x | x | |||||||
| Mohammed et al. [ | Radiopaedia | No info shared | 60 (15 of each class) | 1258 (311 EP, 286 normal, 380 MEN, 281 MB) | x | Denoising | x | x | x | x | x | ||||||
| McAvoy et al. [ | Private dataset | ceT1w | 320 (160 GBM, 160 PCNSL) | 3887 (2332 GBM, 1555 PCNSL) | x | x | Random changes to color, noise sampling | x | |||||||||
| Gilanie et al. [ | Private dataset | T1w, T2w, FLAIR | 180 (50 AST-I, 40 AST-II, 40 AST-III, 50 AST-IV) | 30240 (8400 AST-I, 6720 AST-II, 6720 AST-III, 8400 AST-IV) | x | Bias field correction | x | ||||||||||
| Kulkarni et al. [ | Private dataset | T1w, T2w, FLAIR | 200 (100 benign, 100 malignant) | Denoising, contrast enhancement | x | x | x | x | x | ||||||||
| Artzi et al. [ | Private dataset | T1w, FLAIR, DTI | 158 (22 Normal, 63 PA, 57 MB, 16 EP) | 731 (110 Normal, 280 PA, 266 MB, 75 EP) | x | x | x | Background removal, bias field correction | x | x | x | Brightness changes | |||||
| Tariciotti et al. [ | Private dataset | ceT1w | 121 (47 GBM, 37 PCNSL, 37 Metastasis) | 3597 (1481 GBM, 1073 PCNSL, 1043 Metastasis)) | x | x | Conversion to PNG | ||||||||||
| Ait et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | |||||||||||
| Alanazi et al. [ | 1. Dataset from Kaggle | No info shared | 826 Glioma, 822 MEN, 395 no tumor, and 827 PT | x | x | x | Noise removal | ||||||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | Noise removal | ||||||||||
| Ye et al. [ | Private dataset | ceT1w | 73 | x | x | Image transformation | x | Blurring, ghosting, motion, affining, random elastic deformation | |||||||||
| Gaur et al. [ | MRI dataset by Bhuvaji | No info shared | 2296 | x | Gaussian noise adding | ||||||||||||
| Guo et al. [ | CPM-RadPath 2020 | T1w, ceT1w, T2w, FLAIR | 221 (133 GBM, 54 AST, 34 OLI) | x | x | Bias field correction, Gaussian noise adding | x | x | Random | ||||||||
| Aamir et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | Contrast enhancement | x | x | |||||||||
| Rizwan et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 230 (81 MEN, 90 Glioma, 59 PT) | 3061 (707 MEN, 1425 Glioma, 929 PT) | x | x | Noise filtering and smoothing | salt-noise/grayscale di stortion | |||||||||
| Dataset from TCIA | T1w | 513 (204 Grade II, 128 Grade III, 181 Grade IV) | 70 (32 Grade II, 18 Grade III, 20 Grade IV) | x | x | Noise filtering and smoothing | salt-noise/grayscale di stortion | ||||||||||
| Nayak et al. [ | 1.daataset from Kaggle, 2. Figshare (Cheng et al., 2017) | ceT1w | 1. No info shared, 2. 233 (as shown in | 3260 (196 Normal, 3064 (as shown in | x | Gaussian blurring, noise removal | x | x | x | ||||||||
| Chatterjee et al. [ | 1.BraTS2019, 2. IXI Dataset | ceT1w | 1. 332 (259 HGG, 73 LGG), 2. 259 Normal | x | x | x | x | Affine | |||||||||
| Khazaee et al. [ | BraTS2019 | ceT1w, T2w, FLAIR | 335 (259 HGG, 76 LGG) | 26,904 (13,233 HGG, 13,671 LGG) | x | x | |||||||||||
| Isunuri et al. [ | Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | |||||||||||
| Gu et al. [ | 1. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | ||||||||||||
| 2. REMBRANDT | No info shared | 130 | 110,020 | x | |||||||||||||
| Rajini [ | 1. IXI dataset, REMBRANDT, TCGA-GBM, TCGA-LGG | No info shared | 600 normal images from IXI dataset, 130 patients from REMBRANDT, 200 patients from TCGA-GBM, 299 patients from TCGA-LGG | ||||||||||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | ||||||||||||||
| Anaraki et al. [ | 1: IXI dataset, REMBRANDT, TCGA-GBM, TCGA-LGG, private dataset | no info of IXI, ceT1w from REMBRANDT, TCGA-GBM, TCGA-LGG | 600 normal images from IXI dataset, 130 patients from REMBRANDT, 199 patients from TCGA-GBM, 299 patients from TCGA-LGG, 60 patients from private dataset | x | x | x | x | x | x | ||||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | x | x | x | x | ||||||||
| Sajjad et al. [ | 1. Radiopaedia | No info shared | 121 (36 Grade I, 32 Grade II, 25 Grade III, 28 Grade IV) | x | x | Denoising, bias field correction | x | x | x | Gaussian blurring, sharpening, embossing, skewing | |||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | x | Denoising, bias field correction | x | x | x | Gaussian blurring, sharpening, embossing, skewing | |||||||
| Wahlang et al. [ | 1. Radiopaedia | FLAIR | 11 (2 Metastasis, 6 Glioma, 3 MEN) | x | |||||||||||||
| 2. BraTS 2017 | No info shared | 20 | 3100 | Median filtering | |||||||||||||
| Tandel et al. [ | REMBRANDT | T2w | See 1–4 below | See 1–4 below | x | Converted to RGB | x | x | |||||||||
| 130 | 1. 2156 (1041 normal, 1091 tumorous) | ||||||||||||||||
| 47 | 2. 557 (356 AST-II, 201 AST-III) | ||||||||||||||||
| 21 | 3. 219 (128 OLI-II, 91 OLI-III) | ||||||||||||||||
| 112 | 4. 1115 (484 LGG, 631 HGG) | ||||||||||||||||
| Xiao et al. [ | 1. Private dataset | No info shared | 1109 (495 MT, 614 Normal) | x | |||||||||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | x | |||||||||||||
| 3. Brain Tumor Classification (MRI) Dataset from Kaggle | No info shared | 3264 (937 MEN, 926 Glioma, 901 PT, 500 Normal) | x | ||||||||||||||
| Tandel et al. [ | REMBRANDT | T2w | 112 (30 AST-II, 17 AST-II, 14 OLI-II, 7 OLI-III, 44 GBM) | See 1–5 below | x | x | x | ||||||||||
| 1. 2132 (1041 normal, 1091 tumorous) | |||||||||||||||||
| 2. 2156 (1041 normal, 484 LGG, 631 HGG) | |||||||||||||||||
| 3. 2156 (1041 normal, 557 AST, 219 OLI, 339 GBM) | |||||||||||||||||
| 4. 1115 (356 AST-II, 201 AST-III, 128 OLI-II, 91 OLI-III, 339 GBM) | |||||||||||||||||
| Ayadi et al. [ | 1. Radiopaedia | No info shared | 121 (36 Grade I, 32 Grade II, 25 Grade III, 28 Grade IV) | x | x | Gaussian blurring, sharpening | |||||||||||
| 2. Figshare (Cheng et al., 2017) | ceT1w | 233 (as shown in | 3064 (as shown in | ||||||||||||||
| 3. REMBRANDT | FLAIR, T1w, T2w | 130 (47 AST, 21 OLI, 44 GBM, 18 unknown) | See 1–5 below | x | x | Gaussian blurring, sharpening | |||||||||||
| 1. 2132 (1041 normal, 1091 tumorous) | |||||||||||||||||
| Özcan et al. [ | LGG (grade II) vs. HGG (grade IV) | Custom CNN model | 5-fold CV | SEN = 98.0%, SPE = 96.3%, F1 score = 97.0%, AUC = 0.989 | 97.1 | ||||||||||||
| Hao et al. [ | LGG vs. HGG | Transfer learning with AlexNet | No info shared | AUC = 82.89% | |||||||||||||
| Tripathi et al. [ | LGG vs. HGG | Transfer learning with Resnet18 | No info shared | 95.87 | |||||||||||||
| Ge et al. [ | LGG vs. HGG | Custom CNN model | No info shared | SEN = 84.35%, SPE = 93.65% | 90.7 | ||||||||||||
| Mzoughi et al. [ | LGG vs. HGG | Multi-scale 3D CNN | No info shared | 96.49 | |||||||||||||
| Yang et al. [ | LGG vs. HGG | Transfer learning with AlexNet, GoogLeNet | 5-fold CV | AUC = 0.939 | 86.7 | ||||||||||||
| Zhuge et al. [ | LGG vs. HGG | Transfer learning with ResNet50 | 5-fold CV | SEN = 93.5%, SPE = 97.2% | 96.3 | ||||||||||||
| 3D CNN | 5-fold CV | SEN = 94.7%, SPE = 96.8% | 97.1 | ||||||||||||||
| Decuyper et al. [ | LGG vs. GBM | 3D CNN | No info shared | SEN = 90.16%, SPE = 89.80%, AUC = 0.9398 | 90 | ||||||||||||
| He et al. [ | LGG vs. HGG | Custom CNN model | 5-fold CV | TCIA: SEN = 97.14%, SPE = 90.48%, AUC = 0.9349 | 92.86 | ||||||||||||
| BraTS 2017: SEN = 95.24%, SPE = 92%, AUC = 0.952 | 94.39 | ||||||||||||||||
| Hamdaoui et al. [ | LGG vs. HGG | Transfer learning with stacking VGG16, VGG19, MobileNet, InceptionV3, Xception, Inception ResNetV2, DenseNet121 | 10-fold CV | PRE = 98.67%, F1 score = 98.62%, SEN = 98.33% | 98.06 | ||||||||||||
| Chikhalikar et al. [ | LGG vs. HGG | Custom CNN model | No info shared | 99.46 | |||||||||||||
| Ahmad [ | LGG vs. HGG | Custom CNN model | No info shared | 88 | |||||||||||||
| Khazaee et al. [ | LGG vs. HGG | Transfer learning with EfficientNetB0 | CV | PRE = 98.98%, SEN = 98.86%, SPE = 98.79% | 98.87% | ||||||||||||
| Naser et al. [ | LGG (Grade II) vs. LGG (Grade III) | Transfer learning with VGG16 | 5-fold CV | SEN = 97%, SPE = 98% | 95 | ||||||||||||
| Kurc et al. [ | OLI vs. AST | 3D CNN | 5-fold CV | 80 | |||||||||||||
| McAvoy et al. [ | GBM vs. PCNSL | Transfer learning with EfficientNetB4 | No info shared | GBM: AUC = 0.94, PCNSL: AUC = 0.95 | |||||||||||||
| Kulkarni et al. [ | Benign vs. Malignant | Transfer learning with AlexNet | 5-fold CV | PRE = 93.7%, RE = 100%, F1 score = 96.77% | 96.55 | ||||||||||||
| Transfer learning with VGG16 | 5-fold CV | PRE = 55%, RE = 50%, F1 score = 52.38% | 50 | ||||||||||||||
| Transfer learning with ResNet18 | 5-fold CV | PRE = 78.94%, RE = 83.33%, F1 score = 81.07% | 82.5 | ||||||||||||||
| Transfer learning with ResNet50 | 5-fold CV | PRE = 95%, RE = 55.88%, F1 score = 70.36% | 60 | ||||||||||||||
| Transfer learning with GoogLeNet | 5-fold CV | PRE = 75%, RE = 100%, F1 score = 85.71% | 87.5 | ||||||||||||||
| Wahlang et al. [ | HGG vs. LGG | AlexNet | No info shared | 62 | |||||||||||||
| U-Net | No info shared | 60 | |||||||||||||||
| Xiao et al. [ | MT vs. Normal | Transfer learning with ResNet50 | 3-fold, 5-fold, 10-fold CV | AUC = 0.9530 | 98.2 | ||||||||||||
| Alanazi et al. [ | Normal vs. Tumorous | Custom CNN | No info shared | 95.75% | |||||||||||||
| Tandel et al. [ | 1. Normal vs. Tumorous | DL-MajVot (AlexNet, VGG16, ResNet18, GoogleNet, ResNet50) | 5-fold CV | SEN = 96.76%, SPE = 96.43%, AUC = 0.966 | 96.51 | ||||||||||||
| 2. AST-II vs. AST-III | DL-MajVot (AlexNet, VGG16, ResNet18, GoogleNet, ResNet50) | 5-fold CV | SEN = 94.63%, SPE = 99.44%, AUC = 0.9704 | 97.7 | |||||||||||||
| 3. OLI-II vs. OLI-III | DL-MajVot (AlexNet, VGG16, ResNet18, GoogleNet, ResNet50) | 5-fold CV | SEN = 100%, SPE = 100%, AUC = 1 | 100 | |||||||||||||
| 4. LGG vs. HGG | DL-MajVot (AlexNet, VGG16, ResNet18, GoogleNet, ResNet50) | 5-fold CV | SEN = 98.33%, SPE = 98.57%, AUC = 0.9845 | 98.43 | |||||||||||||
| Tandel et al. [ | Normal vs. Tumorous | Transfer learning with AlexNet | Multiple CV (K2, K5, K10) | RE = 100%, PRE = 100%, F1 score = 100% | 100 | ||||||||||||
| Ayadi et al. [ | Normal vs. Tumorous | Custom CNN model | 5-fold CV | 100 | |||||||||||||
| Ye et al. [ | Germinoma vs. Glioma | Transfer learning with ResNet18 | 5-fold CV | AUC = 0.88 | 81% | ||||||||||||
| 3 classes | |||||||||||||||||
| Allah et al. [ | MEN vs. Glioma vs. PT | PGGAN-augmentation VGG19 | No info shared | 98.54 | |||||||||||||
| Swati et al. [ | MEN vs. Glioma vs. PT | Transfer learning with VGG19 | 5-fold CV | SEN = 94.25%, SPE = 94.69%, PRE = 89.52%, F1 score = 91.73% | 94.82 | ||||||||||||
| Guan et al. [ | MEN vs. Glioma vs. PT | EfficientNet | 5-fold CV | 98.04 | |||||||||||||
| Deepak et al. [ | MEN vs. Glioma vs. PT | Transfer learning with GoogleNet | 5-fold CV | 98 | |||||||||||||
| Díaz-Pernas et al. [ | MEN vs. Glioma vs. PT | Multiscale CNN | 5-fold CV | 97.3 | |||||||||||||
| Ismael et al. [ | MEN vs. Glioma vs. PT | Residual networks | 5-fold CV | PRE = 99.0%, RE = 99.0%, F1 score = 99.0% | 99 | ||||||||||||
| Alhassan et al. [ | MEN vs. Glioma vs. PT | Custom CNN model | k-fold CV | PRE = 99.6%, RE = 98.6%, F1 score = 99.0% | 98.6 | ||||||||||||
| Bulla et al. [ | MEN vs. Glioma vs. PT | Transfer learning with InceptionV3 CNN model | holdout validation, 10-fold CV, stratified 10-fold CV, group 10-fold CV | Under group 10-fold CV: PRE = 97.57%, RE = 99.47%, F1 score = 98.40%, AUC = 0.995 | 99.82 | ||||||||||||
| Ghassemi et al. [ | MEN vs. Glioma vs. PT | CNN-GAN | 5-fold CV | PRE = 95.29%, SEN = 94.91%, SPE = 97.69%, F1 score = 95.10% | 95.6 | ||||||||||||
| Kakarla et al. [ | MEN vs. Glioma vs. PT | Custom CNN model | 5-fold CV | PRE = 97.41%, RE = 97.42% | 97.42 | ||||||||||||
| Noreen et al. [ | MEN vs. Glioma vs. PT | Transfer learning with Inception-v3 | K-fold CV | 93.31 | |||||||||||||
| Transfer learning with Inception model | K-fold CV | 91.63 | |||||||||||||||
| Noreen et al. [ | MEN vs. Glioma vs. PT | Transfer learning with Inception-v3 | No info shared | 99.34 | |||||||||||||
| Transfer learning with DensNet201 | No info shared | 99.51 | |||||||||||||||
| Kumar et al. [ | MEN vs. Glioma vs. PT | Transfer learning with ResNet50 | 5-fold CV | PRE = 97.20%, RE = 97.20%, F1 score = 97.20% | |||||||||||||
| Badža et al. [ | MEN vs. Glioma vs. PT | Custom CNN model | 10-fold CV | PRE = 95.79%, RE = 96.51%, F1 score = 96.11% | 96.56 | ||||||||||||
| Ait et al. [ | MEN vs. Glioma vs. PT | Custom CNN | No info shared | PRE = 98.3%, SEN = 98.6%, F1 score = 98.6% | 98.70% | ||||||||||||
| Alanazi et al. [ | MEN vs. Glioma vs. PT | Custom CNN | No info shared | 96.90% | |||||||||||||
| Gaur et al. [ | MEN vs. Glioma vs. PT | Custom CNN | k-fold CV | 94.64% | |||||||||||||
| Aamir et al. [ | MEN vs. Glioma vs. PT | Custom CNN | 5-fold CV | 98.95% | |||||||||||||
| Rizwan et al. [ | MEN vs. Glioma vs. PT | Custom CNN | No info shared | 99.8% | |||||||||||||
| Isunuri et al. [ | MEN vs. Glioma vs. PT | Custom CNN | 5-fold CV | PRE = 97.33%, SEN = 97.19%, F1 score = 97.26% | 97.52% | ||||||||||||
| Alaraimi et al. [ | MEN vs. Glioma vs. PT | Transfer learning with AlexNet | No info shared | AUC = 0.976 | 94.4 | ||||||||||||
| Transfer learning with VGG16 | No info shared | AUC = 0.981 | 100 | ||||||||||||||
| Transfer learning with GoogLeNet | No info shared | AUC = 0.986 | 98.5 | ||||||||||||||
| Lo et al. [ | Grade II vs. Grade III vs. Grade IV | Transfer learning with AlexNet | 10-fold CV | 97.9 | |||||||||||||
| Pei et al. [ | GBM vs. AST vs. OLI | 3D CNN | No info shared | 74.9 | |||||||||||||
| Gu et al. [ | 1. MEN vs. Glioma vs. PT | Custom CNN model | 5-fold CV | SEN = 94.64%, PRE = 94.61%, F1 score = 94.70% | 96.39 | ||||||||||||
| 2. GBM vs. AST vs. OLI | Custom CNN model | 5-fold CV | SEN = 93.66%, PRE = 95.12%, F1 score = 94.05% | 97.37 | |||||||||||||
| Rajini [ | MEN vs. Glioma vs. PT | Custom CNN model | 5-fold CV | 98.16 | |||||||||||||
| Anaraki et al. [ | MEN vs. Glioma vs. PT | Custom CNN model | 5-fold CV | 94.2 | |||||||||||||
| Sajjad et al. [ | MEN vs. Glioma vs. PT | Transfer learning with VGG19 | No info shared | SEN = 88.41%, SPE = 96.12% | 94.58 | ||||||||||||
| Wahlang et al. [ | Metastasis vs. Glioma vs. MEN | Lenet | No info shared | 48 | |||||||||||||
| AlexNet | No info shared | 75 | |||||||||||||||
| Xiao et al. [ | MEN vs. Glioma vs. PT | Transfer learning with ResNet50 | 3-fold, 5-fold, 10-fold CV | 98.02 | |||||||||||||
| Tandel et al. [ | Normal vs. LGG vs. HGG | Transfer learning with AlexNet | Multiple CV (K2, K5, K10) | RE = 94.85%, PRE = 94.75%, F1 score = 94.8% | 95.97 | ||||||||||||
| Chatterjee et al. [ | Normal vs. HGG vs. LGG | Transfer learning with ResNet | 3-fold CV | F1 score = 93.45% | 96.84% | ||||||||||||
| Ayadi et al. [ | 1. Normal vs. LGG vs. HGG | Custom CNN model | 5-fold CV | 95 | |||||||||||||
| 2. MEN vs. Glioma vs. PT | Custom CNN model | 5-fold CV | 94.74 | ||||||||||||||
| Guo et al. [ | GBM vs. AST vs. OLI | Custom CNN | 3-fold CV | SEN = 0.772, SPE = 93.0%, AUC = 0.902 | 87.8% | ||||||||||||
| Rizwan et al. [ | Grade I vs. Grade II vs. Grade III | Custom CNN | No info shared | 97.14% | |||||||||||||
| Tariciotti et al. [ | Metastasis vs. GBM vs. PCNSL | Resnet101 | Hold-out | PRE = 91.88%, SEN = 90.84%, SPE = 96.34%, F1 score = 91.0%, AUC = 0.92 | 94.72% | ||||||||||||
| 4 classes | |||||||||||||||||
| Ahammed et al. [ | Grade I vs. Grade II vs. Grade III vs. Grade IV | VGG19 | No info shared | PRE = 94.71%, SEN = 92.72%, SPE = 98.13%, F1 score = 93.71% | 98.25 | ||||||||||||
| Mohammed et al. [ | EP vs. MEN vs. MB vs. Normal | Custom CNN model | No info shared | SEN = 96%, PRE = 100% | 96 | ||||||||||||
| Gilanie et al. [ | AST-I vs. AST-II vs. AST-III vs. AST-IV | Custom CNN model | No info shared | 96.56 | |||||||||||||
| Artzi et al. [ | Normal vs. PA vs. MB vs. EP | Custom CNN model | 5-fold CV | 88 | |||||||||||||
| Nayak et al. [ | Normal vs. MEN vs. Glioma vs. PT | Transfer learning with EfficientNet | No info shared | PRE = 98.75%, F1 score = 98.75% | 98.78% | ||||||||||||
| Rajini [ | Normal vs. Grade II vs. Grade III vs. Grade IV | Custom CNN model | 5-fold CV | 96.77 | |||||||||||||
| Anaraki et al. [ | Normal vs. Grade II vs. Grade III vs. Grade IV | Custom CNN model | 5-fold CV | ||||||||||||||
| Sajjad et al. [ | Grade I vs. Grade II vs. Grade III vs. Grade IV | Transfer learning with VGG19 | No info shared | 90.67 | |||||||||||||
| Xiao et al. [ | MEN vs. Glioma vs. PT vs. Normal | Transfer learning with ResNet50 | 3-fold, 5-fold, 10-fold CV | PRE = 97.43%, RE = 97.67%, SPE = 99.24%, F1 score = 97.55% | 97.7 | ||||||||||||
| Tandel et al. [ | Normal vs. AST vs. OLI vs. GBM | Transfer learning with AlexNet | Multiple CV (K2, K5, K10) | RE = 94.17%, PRE = 95.41%, F1 score = 94.78% | 96.56 | ||||||||||||
| Ayadi et al. [ | 1. normal vs. AST vs. OLI vs. GBM | Custom CNN model | 5-fold CV | 94.41 | |||||||||||||
| 2. Grade I vs. Grade II vs. Grade III vs. Grade IV | Custom CNN model | 5-fold CV | 93.71 | ||||||||||||||
| 5 classes | |||||||||||||||||
| Tandel et al. [ | AST-II vs. AST-III vs. OLI-II vs. OLI-III vs. GBM-IV | Transfer learning with AlexNet | Multiple CV (K2, K5, K10) | RE = 84.4%, PRE = 89.57%, F1 score = 86.89% | 87.14 | ||||||||||||
| Ayadi et al. [ | AST-II vs. AST-III vs. OLI-II vs. OLI-III vs. GBM | Custom CNN model | 5-fold CV | 86.08 | |||||||||||||
| 6 classes | |||||||||||||||||
| Tandel et al. [ | Normal vs. AST-II vs. AST-III vs. OLI-II vs. OLI-III vs. GBM-IV | Transfer learning with AlexNet | Multiple CV (K2, K5, K10) | RE = 91.51%, PRE = 92.46%, F1 score = 91.97% | 93.74 | ||||||||||||
| Ayadi et al. [ | normal vs. AST-II vs. AST-III vs. OLI-II vs. OLI-III vs. GBM | Custom CNN model | 5-fold CV | 92.09 | |||||||||||||
Notes: 1 Rigid registration unless otherwise notes; 2 translation also referred to as shifting; 3 scaling also referred to as zooming; 4 reflection also referred to as flipping or mirroring; ** The Cancer Imaging Archive, https://www.cancerimagingarchive.net/ (accessed on 27 July 2022). 5 Referring to overall accuracy, mean accuracy, or highest accuracy depending on the information provided by the paper or the highest accuracy when multiple models are used.
Figure 7Classification accuracy by publication year.
Figure 8Classification accuracy by classification task.
Figure 9Classification accuracy by number of patients.
Figure 10Classification accuracy by number of images.
Figure 11Classification accuracy by CNN architecture.
Figure 12Classification accuracy by number of preprocessing operations.
Figure 13Classification accuracy by number of data augmentation operations.
Figure 14Histogram (left scale) and cumulative distribution (right scale) of factors not fully reported or considered in the studies reported in Table 4.