| Literature DB >> 34943580 |
Ahmed M Gab Allah1,2, Amany M Sarhan1, Nada M Elshennawy1.
Abstract
The wide prevalence of brain tumors in all age groups necessitates having the ability to make an early and accurate identification of the tumor type and thus select the most appropriate treatment plans. The application of convolutional neural networks (CNNs) has helped radiologists to more accurately classify the type of brain tumor from magnetic resonance images (MRIs). The learning of CNN suffers from overfitting if a suboptimal number of MRIs are introduced to the system. Recognized as the current best solution to this problem, the augmentation method allows for the optimization of the learning stage and thus maximizes the overall efficiency. The main objective of this study is to examine the efficacy of a new approach to the classification of brain tumor MRIs through the use of a VGG19 features extractor coupled with one of three types of classifiers. A progressive growing generative adversarial network (PGGAN) augmentation model is used to produce 'realistic' MRIs of brain tumors and help overcome the shortage of images needed for deep learning. Results indicated the ability of our framework to classify gliomas, meningiomas, and pituitary tumors more accurately than in previous studies with an accuracy of 98.54%. Other performance metrics were also examined.Entities:
Keywords: brain tumor; convolutional neural network; deep learning; generative adversarial network; magnetic resonance imaging
Year: 2021 PMID: 34943580 PMCID: PMC8700152 DOI: 10.3390/diagnostics11122343
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Summarization of the previous models.
| Ref. | Accuracy % | No. of Images |
|---|---|---|
| [ | 91.28 | 3064 |
| [ | 91.43 | 989 * |
| [ | 93.68 | 3064 |
| [ | 92.13 | 989 * |
| [ | 91.13 | 3064 |
| [ | 86.56 | 3064 |
| [ | 90.89 | 3064 |
| [ | 84.19 | 2100 |
| [ | 94.2 | 989 * |
| [ | 96.13 | 3064 |
| [ | 98 | 3064 |
| [ | 95.6 | 3064 |
* Only the axial plane is used.
Figure 1Block diagram of the proposed framework (DA: Data augmentation).
Figure 2Distribution of each brain tumor type by plane.
Figure 3Three planes for three samples of brain tumor MR images from the primary data set, (A–C) glioma, (D–F) meningioma, and (G–I) pituitary.
Figure 4Samples of brain tumor MRIs and results of augmentation: (A) primary MRI, (B) rotation, (C) left-right mirroring, and (D) up–down flipping.
Figure 5PGGAN architecture for the 256 × 256 pixel MR image brain tumor generator during training progress.
Architecture details of the VGG19+CNN model. The maxpooling layer is applied after each convolution layer.
| VGG19 + CNN Model | Output Shapes |
|---|---|
| [Conv (3 × 3) − 64] × 2 | 224 × 224 × 64 |
| [Conv (3 × 3) − 128] × 2 | 112 × 112 × 128 |
| [Conv (3 × 3) − 256] × 4 | 56 × 56 × 256 |
| [Conv (3 × 3) − 512] × 4 | 28 × 28 × 512 |
| [Conv (3 × 3) − 512] × 4 | 14 × 14 × 512 |
| Max pool | 7 × 7 × 512 |
| Flatten | 25,088 |
| Dense (relu) | 4096 |
| Dropout (0.5) | 4096 |
| Dense (relu) | 4096 |
| Dropout (0.5) | 4096 |
| Dense (SoftMax) | 3 |
Architecture details of the VGG19 + GRU model.
| VGG19 + GRU Model | Output Shapes |
|---|---|
| [Conv (3 × 3) − 64] × 2 | 224 × 224 × 64 |
| [Conv (3 × 3) − 128] × 2 | 112 × 112 × 128 |
| [Conv (3 × 3) − 256] × 4 | 56 × 56 × 256 |
| [Conv (3 × 3) − 512] × 4 | 28 × 28 × 512 |
| [Conv (3 × 3) − 512] × 4 | 14 × 14 × 512 |
| Max pool | 7 × 7 × 512 |
| Reshape | 7 × 7 × 512 |
| Time Distributed | 7 × 3584 |
| GRU (512) | 512 |
| Dense (relu) | 1024 |
| Dropout (0.5) | 1024 |
| Dense (relu) | 1024 |
| Dropout (0.5) | 1024 |
| Dense (SoftMax) | 3 |
Architecture details of the VGG19 + Bi-GRU model.
| VGG19 + Bi-GRU Model | Output Shapes |
|---|---|
| [Conv (3 × 3) − 64] × 2 | 224 × 224 × 64 |
| [Conv (3 × 3) − 128] × 2 | 112 × 112 × 128 |
| [Conv (3 × 3) − 256] × 4 | 56 × 56 × 256 |
| [Conv (3 × 3) − 512] × 4 | 28 × 28 × 512 |
| [Conv (3 × 3) − 512] × 4 | 14 × 14 × 512 |
| Max pool | 7 × 7 × 512 |
| Reshape | 7 × 7 × 512 |
| Time Distributed | 7 × 3584 |
| Bidirectional | 1024 |
| Dense (relu) | 1024 |
| Dropout (0.5) | 1024 |
| Dense (relu) | 1024 |
| Dropout (0.5) | 1024 |
| Dense (SoftMax) | 3 |
Figure 6Number of trainable parameter-wise distributions for each of the three models.
Comparison between different optimizers for the three proposed models.
| Models | Adam | Adamax | RMSprop | Nadam |
|---|---|---|---|---|
|
| 94.16 | 92.94 | 95.59 |
|
|
|
| 87.10 | 93.19 | 93.67 |
|
| 95.38 | 89.05 |
| 92.94 |
The bold values indicate the best accuracy value that was achieved.
All performance results for the three proposed models (where Men, Gli, and Pit refer to meningioma, glioma, and pituitary tumor, respectively).
| Model | Class | Accuracy | Precision | Sensitivity | F1-Score | Specificity | NPV | MCC |
|---|---|---|---|---|---|---|---|---|
|
| Gli. | 96.97 | 93.81 | 100 | 96.81 | 94.40 | 100 | 94.10 |
| Men. | 97.44 | 98.92 | 90.2 | 94.36 | 99.70 | 97.02 | 92.87 | |
| Pit. | 99.06 | 100 | 96.92 | 98.44 | 100 | 98.68 | 97.80 | |
|
| Gli. | 95.01 | 95.87 | 89.92 | 92.8 | 97.84 | 94.58 | 89.01 |
| Men. | 94.46 | 87.27 | 94.12 | 90.1 | 94.59 | 97.61 | 86.78 | |
| Pit. | 98.89 | 98.46 | 98.46 | 98.46 | 99.13 | 99.13 | 97.60 | |
|
| Gli. | 96.11 | 94.05 | 97.21 | 95.6 | 95.26 | 97.79 | 92.15 |
| Men. | 96.59 | 95.83 | 90.20 | 92.93 | 98.71 | 96.83 | 90.76 | |
| Pit. | 98.54 | 97.69 | 97.69 | 97.69 | 98.93 | 98.93 | 96.62 |
Figure 7Confusion matrix for: (A) VGG19 + CNN model, (B) VGG19 + GRU model, and (C) VGG19 + Bi-GRU model.
Figure 8Loss over the training process for: (A) VGG19, (B) VGG19 + GRU, and (C) VGG19 + Bi-GRU.
Figure 9Samples of ‘realistic’ synthetic MR images, in three planes, produced by PGGAN: (A–C) glioma (D–F) meningioma, and (G–I) pituitary tumor.
Figure 10Samples of ‘wrong’ synthetic MR images produced by PGGAN.
Comparison between different optimizers for the three proposed models using PGGAN DA.
| Models | Adam | Adamax | RMSprop | Nadam |
|---|---|---|---|---|
|
|
| 97.57 | 97.57 | 96.59 |
|
| 95.62 | 92.21 | 95.13 |
|
|
| 95.62 | 91.97 |
| 96.35 |
The bold values indicate the best accuracy value that was achieved.
All performance results for the three proposed models using PGGAN DA.
| Model | Class | Accuracy | Precision | Sensitivity | F1-Score | Specificity | NPV | MCC |
|---|---|---|---|---|---|---|---|---|
|
| Gli. | 98.54 | 98.87 | 97.77 | 98.31 | 99.14 | 98.29 | 97.03 |
| Men. | 98.54 | 96.15 | 98.04 | 97.09 | 98.71 | 99.35 | 96.12 | |
| Pit. | 100 | 100 | 100 | 100 | 100 | 100 | 100 | |
|
| Gli. | 97.08 | 95.63 | 97.77 | 96.69 | 96.55 | 98.25 | 94.1 |
| Men. | 97.23 | 95.05 | 94.12 | 94.58 | 98.38 | 98.06 | 92.81 | |
| Pit. | 98.78 | 99.21 | 96.92 | 98.05 | 99.64 | 98.59 | 97.18 | |
|
| Gli. | 96.84 | 94.62 | 98.32 | 96.44 | 95.69 | 98.67 | 93.65 |
| Men. | 97.08 | 96.88 | 91.18 | 93.94 | 99.03 | 97.14 | 92.09 | |
| Pit. | 99.76 | 100 | 99.23 | 99.61 | 100 | 99.64 | 99.44 |
Figure 11Confusion matrix for: (A) VGG19 + CNN model, (B) VGG19 + GRU model, and (C) VGG19 + Bi-GRU model.
Figure 12The losses of the three proposed models during training for: (A) VGG19, (B) VGG19 + GRU, and (C)VGG19 + Bi-GRU).
Figure 13Comparison of the (A) accuracy and (B) average losses for the three proposed models.
Comparison of our framework with other works based on classic augmentation.
| Ref. | Augmentation Operation | Data Set Size after Augmentation | Data Set Division | ACC% |
|---|---|---|---|---|
| [ | Rotation random angle between (0 and 360 angle). | Not mentioned | Use images for 191 patients and divided them according to patients as follows: training: 149 patients, validation: 21 patients, and testing: 21 patients. | 91.43 |
| [ | Rotation 10, 20, or 30 clockwise or counterclockwise. Mirroring, and translating 15 pixels to right or left. | Augmentation is done after the train and test images are randomly selected. | Divide the data set images into: training group, and testing group. | 94.2 |
| [ | Rotation image with angle 45. Mirroring right/left. Flipping up–down. Adding salt noise. | Images are shuffled, splitting, and finally, the author applies augmentation. The authors increased the original number to 15,320 images. | Divide the data set images into training group: 68%, validation group: 16%, and testing group: 16%. | 96.13 |
| [ | Rotation random angle between (0 and 359 angle). | Not mentioned | Not mentioned | 95.6 |
| Our framework (VGG19 + CNN and classic DA) | Rotation 90, 180, 270 angles. Mirroring right/left. | Images are randomly selected, then split, and finally, we augment the only training and validation groups. The original images increased to 19,215 images for the training group, and 4536 for the validation group. | Divide the data set images into: training group: 70%, validation group: 15%, and testing group: 15%. | 96.59 |
| Our framework (VGG19 + CNN and PGGAN DA) | Rotation 90, 180, 270 angles. | We increased the original images to 27,315 images for training, and 4536 images for validation. | Divide data set into: training: 70% + PGGAN-generated images, validation: 15%, and testing: 15%. | 98.54 |
Comparison of our framework with other works by precision, sensitivity, and specificity.
| Method | PPV | Sensitivity | SPC | ACC% | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Gli. | Men. | Pit. | Gli. | Men. | Pit. | Gli. | Men. | Pit. | ||
| [ | - | - | - | 96.4 | 86.0 | 87.3 | 96.3 | 95.5 | 95.3 | 91.28 |
| [ | 91.0 | 94.5 | 98.3 | 97.5 | 76.8 | 100 | - | - | - | 93.68 |
| [ | - | - | - | 95.1 | 86.97 | 91.24 | 96.29 | 96.0 | 95.66 | 91.13 |
| [ | 91.9 | 95.3 | 95.7 | 98.3 | 87.8 | 96.5 | 95.7 | 97.8 | 97.8 | 94.2 |
| [ | 97.2 | 95.8 | 95.2 | 94.4 | 95.5 | 93.4 | 95.1 | 98.7 | 97.0 | 96.13 |
| [ | 99.2 | 94.7 | 98.0 | 97.9 | 96.0 | 98.9 | 99.4 | 98.4 | 99.1 | 98.0 |
| [ | 95.89 | 92.43 | 95.29 | 96.83 | 89.98 | 97.93 | 96.38 | 97.79 | 97.54 | 95.6 |
| Our framework | 98.87 | 96.15 | 100 | 97.77 | 98.04 | 100 | 99.14 | 98.71 | 100 | 98.54 |