| Literature DB >> 33052172 |
Mehedi Masud1, Amr E Eldin Rashed1, M Shamim Hossain2,3.
Abstract
Breast cancer is the most prevailing cancer in the world and each year affecting millions of women. It is also the cause of largest number of deaths in women dying in cancers. During the last few years, researchers are proposing different convolutional neural network models in order to facilitate diagnostic process of breast cancer. Convolutional neural networks are showing promising results to classify cancers using image datasets. There is still a lack of standard models which can claim the best model because of unavailability of large datasets that can be used for models' training and validation. Hence, researchers are now focusing on leveraging the transfer learning approach using pre-trained models as feature extractors that are trained over millions of different images. With this motivation, this paper considers eight different fine-tuned pre-trained models to observe how these models classify breast cancers applying on ultrasound images. We also propose a shallow custom convolutional neural network that outperforms the pre-trained models with respect to different performance metrics. The proposed model shows 100% accuracy and achieves 1.0 AUC score, whereas the best pre-trained model shows 92% accuracy and 0.972 AUC score. In order to avoid biasness, the model is trained using the fivefold cross validation technique. Moreover, the model is faster in training than the pre-trained models and requires a small number of trainable parameters. The Grad-CAM heat map visualization technique also shows how perfectly the proposed model extracts important features to classify breast cancers. © Springer-Verlag London Ltd., part of Springer Nature 2020.Entities:
Keywords: Breast cancer; Convolutional neural network; Deep learning; Transfer learning
Year: 2020 PMID: 33052172 PMCID: PMC7545025 DOI: 10.1007/s00521-020-05394-5
Source DB: PubMed Journal: Neural Comput Appl ISSN: 0941-0643 Impact factor: 5.102
Datasets and distribution of classes
| Datasets | Benign | Malignant | Normal | Total samples |
|---|---|---|---|---|
| Dataset 1 | 100 | 150 | – | 250 |
| Dataset 2 | 437 | 210 | 133 | 780 |
| Total | 537 | 360 | 133 | 1030 |
Fig. 1Sample images of breast ultrasound images of different cases in the two datasets
Data distribution in fivefolds after applying fivefold cross validation
| Train | Test | |||||
|---|---|---|---|---|---|---|
| Benign | Malignant | Normal | Benign | Malignant | Normal | |
| Fold 1 | 429 | 288 | 107 | 108 | 72 | 26 |
| Fold 2 | 429 | 288 | 107 | 108 | 72 | 26 |
| Fold 3 | 430 | 288 | 106 | 107 | 72 | 27 |
| Fold 4 | 430 | 288 | 106 | 107 | 72 | 27 |
| Fold 5 | 430 | 288 | 106 | 107 | 72 | 27 |
Pre-trained models and their image input size
| Pre-Trained models | Number of convolutional layers | Input size |
|---|---|---|
| AlexNet | 8 | 227 × 227 |
| DarkNet19 | 19 | 256 × 256 |
| GoogleNet | 22 | 224 × 224 |
| MobileNet-v2 | 53 | 224 × 224 |
| ResNet18 | 18 | 224 × 224 |
| ResNet50 | 50 | 224 × 224 |
| VGG16 | 16 | 224 × 224 |
| Xception | 71 | 229 × 229 |
Fig. 2The architecture of the custom model
Training parameters of the pre-trained models
| Training options | Value |
|---|---|
| Initial learn rate | 1.0000e−04 or 10e−5 |
| Learn rate schedule | Constant or piece wise |
| Mini-batch size | 8 |
| Shuffle | Every epoch |
| Optimizer | Adam, RMSprop, SGDM |
| Cross validation | 5 |
| Max epochs | 20 |
| ExecutionEnvironment | gpu |
Evaluation results of the pre-trained models and the custom model
| Models | Optimizer | AUC | Accuracy | Sensitivity | Specificity | Precision | Recall |
|---|---|---|---|---|---|---|---|
| AlexNet | SGDM | 0.967 | 89.71 | 90.88 | 88.44 | 89.54 | 90.88 |
| Adam | 0.943 | 84.85 | 85.85 | 83.77 | 85.21 | 85.85 | |
| RMSprop | 0.919 | 82.62 | 81.38 | 83.98 | 84.69 | 81.38 | |
| Darknet19 | SGDM | 0.948 | 87.18 | 88.27 | 86 | 87.29 | 88.27 |
| Adam | 0.958 | 88.35 | 88.27 | 88.44 | 89.27 | 88.27 | |
| RMSprop | 0.959 | 88.93 | 89.57 | 88.24 | 89.24 | 89.57 | |
| GoogleNet | SGDM | 0.959 | 88.93 | 91.25 | 86.41 | 87.97 | 91.25 |
| Adam | 0.966 | 90.49 | 93.11 | 87.63 | 89.13 | 93.11 | |
| RMSprop | 0.96 | 91.26 | 93.85 | 88.44 | 89.84 | 93.85 | |
| MobileNet | SGDM | 0.944 | 85.73 | 85.1 | 86.41 | 87.21 | 85.1 |
| Adam | 0.969 | 90.19 | 89.01 | 91.48 | 91.92 | 89.01 | |
| RMSprop | 0.961 | 90.68 | 90.32 | 91.08 | 91.68 | 90.32 | |
| ResNet18 | SGDM | 0.948 | 87.67 | 89.94 | 85.19 | 86.87 | 89.94 |
| Adam | 0.967 | 90 | 90.13 | 89.86 | 90.64 | 90.13 | |
| RMSprop | 0.97 | 91.26 | 90.32 | 92.29 | 92.73 | 90.32 | |
| ResNet50 | SGDM | 0.952 | 87.77 | 83.46 | 88.41 | 51.63 | 83.46 |
| Adam | 0.971 | 92.04 | 93.23 | 91.86 | 62.94 | 93.23 | |
| RMSprop | 0.971 | 91.84 | 90.23 | 92.08 | 62.83 | 90.23 | |
| VGG16 | SGDM | 0.965 | 89.9 | 91.99 | 87.63 | 89.01 | 91.99 |
| Adam | 0.972 | 89.9 | 92.92 | 86.61 | 88.32 | 92.92 | |
| RMSprop | 0.955 | 90.29 | 92.92 | 87.42 | 88.95 | 92.92 | |
| Xception | SGDM | 0.924 | 83.11 | 83.8 | 82.35 | 83.8 | 83.8 |
| Adam | 0.968 | 90.68 | 89.94 | 91.48 | 92 | 89.94 | |
| RMSprop | 0.968 | 91.17 | 91.43 | 90.87 | 91.6 | 91.43 |
Comparison results of the best pre-trained models and the custom model
| Pre-trained models | Custom model | ||||
|---|---|---|---|---|---|
| Performance metric | Score | Models | Optimizers | Score | Optimizer |
| AUC | 0.972 | VGG16 | Adam | 1.0 | Adam, RMSprop |
| Accuracy | 92.04 | ResNet50 | Adam | 100 | Adam |
| Sensitivity | 93.85 | GoogleNet | RMSprop | 100 | SGDM, Adam, RMSprop |
| Specificity | 92.29 | ResNet18 | RMSprop | 100 | Adam |
| Precision | 92.73 | ResNet50 | RMSprop P | 100 | Adam |
| Recall | 93.85 | GoogleNet | RMSprop | 100 | Adam |
Fig. 3Confusion matrix of different pre-rained models and the custom model
Comparison of classification results (the custom model (Adam optimizer) and best pre-trained models)
| Models | Benign (537 original case) | Malignant (360 original case) | Normal (133 original case) |
|---|---|---|---|
| Custom model | 537 | 360 | 133 |
| GoogleNet | 504 | 319 | 117 |
| ResNet18 | 485 | 331 | 124 |
| ResNet50 | 494 | 330 | 124 |
| VGG16 | 499 | 310 | 121 |
Comparison of the pre-trained models and the custom model (accuracy, prediction time, and parameters)
| Performance comparison | |||||||||
|---|---|---|---|---|---|---|---|---|---|
| Model | Custom model | AlexNet | DarkNet19 | GoogleNet | MobileNet | ResNet50 | ResNet18 | VGG16 | Xception |
| Accuracy | 100 | 89.7 | 88.9 | 91.3 | 90.7 | 92 | 91.3 | 90.3 | 91.2 |
| Prediction time (second) | 0.00495 | 0.04564 | 0.06824 | 0.018441 | 0.245589 | 0.02194 | 0.007419 | 0.04667 | 0.13845 |
| Number of parameters | 2985263 | 56880515 | 20838179 | 5976627 | 4233133 | 23487619 | 11173251 | 1.36E + 08 | 20808883 |
Fig. 4Performance comparison of different pre-rained models and the custom model
Fig. 5Accuracy and loss of the custom model during training phase
Fig. 6Heat map visualization of same images using the custom model