| Literature DB >> 36092017 |
Nadiah A Baghdadi1, Amer Malki2, Hossam Magdy Balaha3, Yousry AbdulAzeem4, Mahmoud Badawy3, Mostafa Elhosseini2,3.
Abstract
Due to its high prevalence and wide dissemination, breast cancer is a particularly dangerous disease. Breast cancer survival chances can be improved by early detection and diagnosis. For medical image analyzers, diagnosing is tough, time-consuming, routine, and repetitive. Medical image analysis could be a useful method for detecting such a disease. Recently, artificial intelligence technology has been utilized to help radiologists identify breast cancer more rapidly and reliably. Convolutional neural networks, among other technologies, are promising medical image recognition and classification tools. This study proposes a framework for automatic and reliable breast cancer classification based on histological and ultrasound data. The system is built on CNN and employs transfer learning technology and metaheuristic optimization. The Manta Ray Foraging Optimization (MRFO) approach is deployed to improve the framework's adaptability. Using the Breast Cancer Dataset (two classes) and the Breast Ultrasound Dataset (three-classes), eight modern pre-trained CNN architectures are examined to apply the transfer learning technique. The framework uses MRFO to improve the performance of CNN architectures by optimizing their hyperparameters. Extensive experiments have recorded performance parameters, including accuracy, AUC, precision, F1-score, sensitivity, dice, recall, IoU, and cosine similarity. The proposed framework scored 97.73% on histopathological data and 99.01% on ultrasound data in terms of accuracy. The experimental results show that the proposed framework is superior to other state-of-the-art approaches in the literature review. ©2022 N Baghdadi et al.Entities:
Keywords: Breast cancer; Convolutional neural network (CNN); Deep learning (DL); Manta-Ray foraging algorithm (MRFO); Metaheuristic optimization
Year: 2022 PMID: 36092017 PMCID: PMC9454783 DOI: 10.7717/peerj-cs.1054
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
Figure 1General deep learning architecture for image classification.
Figure 2The proposed hybrid breast cancer recognition framework.
The used datasets specifications summarization.
|
|
|
|
|
|
|---|---|---|---|---|
|
| 2 | “Benign” and “Malignant” | 7, 783 | 10, 608 |
|
| 3 | “Benign”, “Malignant”, and “Normal” | 780 | 1, 311 |
The different implemented data augmentation techniques and the corresponding configurations.
|
|
|
|---|---|
| Rotation | 30° |
| Shear ratio | 20% |
| Zoom ratio | 20% |
| Width shift ratio | 20% |
| Height shift ratio | 20% |
| Brightness change | [0.8:1.2] |
| Vertical flip | ✓ |
| Horizontal flip | ✓ |
Figure 3A flowchart summarization of the MRFO steps.
The solution indexing and the corresponding hyperparameters definitions and ranges.
|
|
|
|
|---|---|---|
| 1 | Training loss function | Categorical Crossentropy, Categorical Hinge, KLDivergence, Poisson, Squared Hinge, and Hinge |
| 2 | Training batch size | 4 → 48 (step = 4) |
| 3 | Model dropout ratio | [0 → 0.6] |
| 4 | Transfer learning freezing ratio | 1 → 100 (step = 1) |
| 5 | Weights (i.e., parameters) optimizer | Adam, NAdam, AdaGrad, AdaDelta, AdaMax, RMSProp, SGD, Ftrl, SGD Nesterov, RMSProp Centered, and Adam AMSGrad |
| 6 | Dimension scaling technique | Normalize, Standard, Min Max, and Max Abs |
| 7 | Utilize data augmentation techniques or not | [ |
| 8 | The value of rotation (In the case of data augmentation). | 0° → 45° (step = 1°) |
| 9 | In the case of data augmentation, width shift value. | [0 → 0.25] |
| 10 | The value of height shift if data augmentation is applied | [0 → 0.25] |
| 11 | Value of shear in case of data augmentation | [0 → 0.25] |
| 12 | Value of Zoom (if data augmentation is used) | [0 → 0.25] |
| 13 | Flag for horizontal flipping (if data augmentation is utilized) | [ |
| 14 | (If augmentation of data has been applied), the value of Vertical flipping flag | [ |
| 15 | Range of brightness changes (if data augmentation is applied) | [0.5 → 2.0] |
Common experiments configurations.
|
|
|
|---|---|
| Datasets | Table 1 |
| Apply dataset shuffling? | Yes (Random) |
| Input image size | (128 × 128 × 3) |
| Hyperparameters metaheuristic optimizer | Manta-ray foraging algorithm (MRFO) |
| Train split ratio | 85% to 15% (i.e., 85% for training (and validation) and 15% for testing) |
| MRFO population size | 10 |
| MRFO iterations count | 10 |
| Epochs number | 5 |
| O/P activation function | SoftMax |
| Pre-trained models | InceptionV3, Xception, EfficientNetB7, NASNetLarge, VGG19, SeNet154, DenseNet201, and ResNet152V2 |
| Pre-trained parameters initializers | ImageNet |
| Hyperparameters | Table 3 |
| Scripting language | Python |
| Python major packages | Tensorflow, Keras, NumPy, OpenCV, and Matplotlib |
| Working environment | Google Colab with GPU (i.e., Intel(R) Xeon(R) CPU @ 2.00 GHz, Tesla T4 16 GB GPU, CUDA v.11.2, and 12 GB RAM) |
Binary dataset specific experiments configurations.
|
|
|
|---|---|
| Dataset source |
|
| Number of classes | 2 |
| Classes | (‘Benign’ and ‘Malignant’) |
| Dataset size before data balancing | “Benign”: 2,479 and “Malignant”: 5,304 |
| Dataset size after data balancing | “Benign”: 5,304 and “Malignant”: 5,304 |
Confusion matrix results concerning the binary dataset.
|
|
|
|
|
|
|---|---|---|---|---|
| InceptionV3 | 10,288 | 10,288 | 320 | 320 |
| Xception | 10,363 | 10,363 | 241 | 241 |
| EfficientNetB7 | 7,916 | 7,916 | 2,668 | 2,668 |
| NASNetLarge | 9,817 | 9,817 | 791 | 791 |
| VGG19 | 10,308 | 10,308 | 300 | 300 |
| SeNet154 | 10,171 | 10,171 | 413 | 413 |
| DenseNet201 | 10,350 | 10,350 | 242 | 242 |
| ResNet152V2 | 10,225 | 10,225 | 383 | 383 |
Learning and optimization best solutions concerning the binary dataset.
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|
|
| KLDivergence | KLDivergence | Poisson | KLDivergence | Poisson | KLDivergence | Categorical Crossentropy | KLDivergence |
|
| 12 | 44 | 28 | 12 | 12 | 36 | 32 | 48 |
|
| 0.41 | 0.28 | 0.25 | 0.33 | 0.47 | 0.08 | 0.16 | 0.43 |
|
| 91 | 56 | 92 | 70 | 33 | 79 | 49 | 62 |
|
| AdaGrad | AdaMax | SGD Nesterov | SGD | SGD | SGD Nesterov | AdaGrad | SGD Nesterov |
|
| Normalization | Normalization | MinMax | Normalization | Standardization | MinMax | MaxAbs | MinMax |
|
| Yes | Yes | No | No | No | Yes | Yes | No |
|
| 1 | 26 | N/A | N/A | N/A | 11 | 24 | N/A |
|
| 0.22 | 0.17 | N/A | N/A | N/A | 0.18 | 0.03 | N/A |
|
| 0.22 | 0.25 | N/A | N/A | N/A | 0.06 | 0.03 | N/A |
|
| 0.25 | 0.08 | N/A | N/A | N/A | 0.14 | 0.12 | N/A |
|
| 0.24 | 0.13 | N/A | N/A | N/A | 0.16 | 0.17 | N/A |
|
| Yes | No | N/A | N/A | N/A | No | Yes | N/A |
|
| Yes | Yes | N/A | N/A | N/A | No | No | N/A |
|
| 1.18–1.24 | 1.28–1.32 | N/A | N/A | N/A | 1.64–1.78 | 0.65–1.72 | N/A |
The binary dataset experiments with the maxmimized metrics.
|
|
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|---|---|
| InceptionV3 | 96.98% | 96.98% | 96.98% | 96.98% | 96.98% | 96.98% | 99.14% | 97.07% | 97.47% | 97.41% |
| Xception | 97.73% | 97.73% | 97.73% | 97.73% | 97.73% | 97.73% | 99.50% | 97.90% | 98.17% | 98.04% |
| EfficientNetB7 | 74.79% | 74.79% | 74.79% | 74.79% | 74.79% | 74.79% | 81.85% | 74.95% | 78.83% | 80.26% |
| NASNetLarge | 92.54% | 92.54% | 92.54% | 92.54% | 92.54% | 92.54% | 98.00% | 91.55% | 93.01% | 94.02% |
| VGG19 | 97.17% | 97.17% | 97.17% | 97.17% | 97.17% | 97.17% | 99.65% | 95.53% | 96.45% | 97.57% |
| SeNet154 | 96.10% | 96.10% | 96.10% | 96.10% | 96.10% | 96.10% | 99.03% | 95.17% | 96.04% | 96.80% |
| DenseNet201 | 97.72% | 97.72% | 97.72% | 97.72% | 97.72% | 97.72% | 99.57% | 89.33% | 92.02% | 96.98% |
| ResNet152V2 | 96.39% | 96.39% | 96.39% | 96.39% | 96.39% | 96.39% | 99.19% | 94.92% | 95.90% | 96.91% |
The binary dataset experiments with the minimized metrics.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| InceptionV3 | 0.011 | 0.038 | 0.024 | 0.012 | 0.156 |
| Xception | 0.008 | 0.027 | 0.018 | 0.009 | 0.135 |
| EfficientNetB7 | 0.082 | 0.318 | 0.177 | 0.087 | 0.421 |
| NASNetLarge | 0.025 | 0.105 | 0.054 | 0.027 | 0.233 |
| VGG19 | 0.010 | 0.053 | 0.022 | 0.011 | 0.150 |
| SeNet154 | 0.014 | 0.059 | 0.030 | 0.015 | 0.172 |
| DenseNet201 | 0.015 | 0.120 | 0.031 | 0.016 | 0.177 |
| ResNet152V2 | 0.013 | 0.062 | 0.029 | 0.014 | 0.169 |
Three-classes specific experiments configurations.
|
|
|
|---|---|
| Dataset source |
|
| Number of classes | 3 |
| Classes | (‘Benign’, ‘Malignant’, and ‘Normal’) |
| Dataset size before data balancing | “Benign”: 437, “Malignant”: 210, and “Normal”: 133 |
| Dataset size after data balancing | “Benign”: 437, “Malignant”: 437, and “Normal”: 437 |
Confusion matrix results concerning the three-classes dataset.
|
|
|
|
|
|
|---|---|---|---|---|
| InceptionV3 | 1191 | 2521 | 55 | 97 |
| Xception | 1267 | 2584 | 32 | 41 |
| EfficientNetB7 | 875 | 2475 | 117 | 421 |
| NASNetLarge | 1184 | 2575 | 41 | 124 |
| VGG19 | 1053 | 2471 | 121 | 243 |
| SeNet154 | 1257 | 2572 | 36 | 47 |
| DenseNet201 | 1204 | 2491 | 61 | 72 |
| ResNet152V2 | 1294 | 2603 | 13 | 14 |
The promising solutions concerning the three-classes dataset.
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|
|
| KLDivergence | Categorical Crossentropy | KLDivergence | Categorical Crossentropy | Categorical Crossentropy | KLDivergence | Categorical Crossentropy | Categorical Crossentropy |
|
| 28 | 4 | 24 | 4 | 24 | 8 | 44 | 4 |
|
| 0.25 | 0 | 0.38 | 0 | 0.06 | 0.09 | 0.46 | 0 |
|
| 47 | 0 | 68 | 0 | 40 | 54 | 81 | 0 |
|
| RMSProp | Adam | AdaMax | Adam | AdaGrad | SGD Nesterov | SGD | Adam |
|
| Standardization | Normalization | Standardization | Normalization | Standardization | Normalization | Standardization | Normalization |
|
| Yes | Yes | Yes | Yes | Yes | No | Yes | Yes |
|
| 13 | 0 | 6 | 0 | 3 | N/A | 28 | 0 |
|
| 0.12 | 0 | 0.22 | 0 | 0.04 | N/A | 0.11 | 0 |
|
| 0.11 | 0 | 0.2 | 0 | 0.1 | N/A | 0.16 | 0 |
|
| 0.06 | 0 | 0.1 | 0 | 0.02 | N/A | 0.13 | 0 |
|
| 0.03 | 0 | 0.24 | 0 | 0.06 | N/A | 0.06 | 0 |
|
| No | Yes | Yes | Yes | Yes | N/A | No | Yes |
|
| Yes | Yes | No | Yes | Yes | N/A | Yes | Yes |
|
| 1.03–1.07 | 0.5–0.5 | 1.35–1.59 | 0.5–0.5 | 0.93–0.94 | N/A | 1.25–1.37 | 0.5–0.5 |
The three-classes dataset experiments with the maxmimized metrics.
|
|
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|---|---|
| InceptionV3 | 94.41% | 93.81% | 95.34% | 92.47% | 92.47% | 97.86% | 98.80% | 88.69% | 90.95% | 94.71% |
| Xception | 97.09% | 97.04% | 97.55% | 96.87% | 96.87% | 98.78% | 99.03% | 93.65% | 94.98% | 97.06% |
| EfficientNetB7 | 79.40% | 74.33% | 86.43% | 67.52% | 67.52% | 95.49% | 92.61% | 69.70% | 74.64% | 81.80% |
| NASNetLarge | 95.11% | 92.15% | 96.18% | 90.52% | 90.52% | 98.43% | 99.29% | 87.78% | 90.13% | 94.30% |
| VGG19 | 85.49% | 84.70% | 89.24% | 81.25% | 81.25% | 95.33% | 96.89% | 79.44% | 83.17% | 88.47% |
| SeNet154 | 96.93% | 96.70% | 97.16% | 96.40% | 96.40% | 98.62% | 99.48% | 92.61% | 94.19% | 96.76% |
| DenseNet201 | 95.06% | 94.73% | 95.13% | 94.36% | 94.36% | 97.61% | 98.84% | 94.31% | 95.20% | 95.76% |
| ResNet152V2 | 99.01% | 98.96% | 99.01% | 98.93% | 98.93% | 99.50% | 99.77% | 98.28% | 98.60% | 98.97% |
The three-classes dataset experiments with the minimized metrics.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| InceptionV3 | 0.016 | 0.090 | 0.033 | 0.016 | 0.183 |
| Xception | 0.009 | 0.050 | 0.019 | 0.009 | 0.139 |
| EfficientNetB7 | 0.051 | 0.254 | 0.106 | 0.052 | 0.326 |
| NASNetLarge | 0.017 | 0.099 | 0.036 | 0.018 | 0.190 |
| VGG19 | 0.033 | 0.168 | 0.069 | 0.034 | 0.263 |
| SeNet154 | 0.010 | 0.058 | 0.021 | 0.010 | 0.144 |
| DenseNet201 | 0.012 | 0.048 | 0.026 | 0.013 | 0.162 |
| ResNet152V2 | 0.003 | 0.014 | 0.007 | 0.003 | 0.083 |
Figure 4Hyperparameters selection and best combinations graphical summarization.
Figure 5Confusion matrix related to the two-classes dataset.
Figure 6Confusion matrix related to the three-classes dataset.
Figure 7Summarization of the learning and optimization experiments related to the two-classes dataset.
Figure 8Summarization of the learning and optimization experiments related to the three-classes dataset.
Comparison between the suggested approach and related studies.
|
|
|
|
|
|
|---|---|---|---|---|
|
| 2016 | Histopathological | GoogLeNet DL | 98.40% |
|
| 2017 | Histopathological | Structured DL | 93.20% |
|
| 2017 | Ultrasound | GoogLeNet DL | 90.00% |
|
| 2018 | Histopathological | DL | 96.40% |
|
| 2018 | Histopathological | DL | 93.00% |
|
| 2019 | Ultrasound | GoogLeNet DL | 92.50% |
|
| 2019 | Ultrasound | VGG16 DL | 97.00% |
|
| 2021 | Histopathological | VGG19 DL | 98.13% and 88.95% |
|
| 2021 | Histopathological | AlexNet DL | 95.00% |
| Current Study | 2022 | Histopathological | Hybrid (MRFO and CNN) | 97.73% |
| Current Study | 2022 | Ultrasound | Hybrid (MRFO and CNN) | 99.01% |