| Literature DB >> 32764398 |
Zabit Hameed1, Sofia Zahia1, Begonya Garcia-Zapirain1, José Javier Aguirre2,3, Ana María Vanegas4.
Abstract
Breast cancer is one of the major public health issues and is considered a leading cause of cancer-related deaths among women worldwide. Its early diagnosis can effectively help in increasing the chances of survival rate. To this end, biopsy is usually followed as a gold standard approach in which tissues are collected for microscopic analysis. However, the histopathological analysis of breast cancer is non-trivial, labor-intensive, and may lead to a high degree of disagreement among pathologists. Therefore, an automatic diagnostic system could assist pathologists to improve the effectiveness of diagnostic processes. This paper presents an ensemble deep learning approach for the definite classification of non-carcinoma and carcinoma breast cancer histopathology images using our collected dataset. We trained four different models based on pre-trained VGG16 and VGG19 architectures. Initially, we followed 5-fold cross-validation operations on all the individual models, namely, fully-trained VGG16, fine-tuned VGG16, fully-trained VGG19, and fine-tuned VGG19 models. Then, we followed an ensemble strategy by taking the average of predicted probabilities and found that the ensemble of fine-tuned VGG16 and fine-tuned VGG19 performed competitive classification performance, especially on the carcinoma class. The ensemble of fine-tuned VGG16 and VGG19 models offered sensitivity of 97.73% for carcinoma class and overall accuracy of 95.29%. Also, it offered an F1 score of 95.29%. These experimental results demonstrated that our proposed deep learning approach is effective for the automatic classification of complex-natured histopathology images of breast cancer, more specifically for carcinoma images.Entities:
Keywords: breast cancer; deep learning; ensemble models; histopathology; image classification
Mesh:
Year: 2020 PMID: 32764398 PMCID: PMC7472736 DOI: 10.3390/s20164373
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The complete process of biopsy is depicted in Figure. Steps 01 and 02 are taken from [7] whereas steps 03 and 04 are retrieved from our own dataset.
Characteristics of our proposed dataset.
| Images | Quantity | Color Model | Staining |
|---|---|---|---|
| carcinoma | 437 | RGB | H & E |
| non-carcinoma | 408 | RGB | H & E |
| Total | 845 | RGB | H & E |
Figure 2The examples of original (A,C) and normalized (B,D) images of carcinoma and non-carcinoma cases.
Criteria for the selection of training, validation, and test images.
| No. of Images | Percentage | |
|---|---|---|
| Training | 540 | 64% |
| Validation | 135 | 16% |
| Test | 170 | 20% |
| Total | 845 | 100% |
Parameters of data augmentation.
| Parameters of Image Augmentation | Values |
|---|---|
| Zoom range | 0.2 |
| Rotation range | 40 |
| Width shift range | 0.2 |
| Height shift range | 0.2 |
| Horizontal flip | True |
| Fill mode | Reflect |
Figure 3Representation of fine-tuned VGG16 architecture [20]. In fine-tuned VGG16 and VGG19 models, the first block (comprising two convolutional layers and one max-pooling layer) is frozen whereas the rest of layers are trainable. However, in fully-trained VGG16 and VGG19 models, all the five blocks are trainable.
Figure 4The proposed ensemble architecture using the fine-tuned VGG16 and VGG19 models along with 5-fold cross-validation approach.
Hyperparameters used in the individual and an ensemble models.
| Hyperparameters | VGG16 with Data Augmentation | VGG19 with Data Augmentation |
|---|---|---|
| Train approach | 5-fold cross-validation | 5-fold cross-validation |
| Optimizer | Adam | Adam |
| Loss function | Binary cross-entropy | Binary cross-entropy |
| Learning rate |
|
|
| Batch size | 32 | 32 |
| Convolution | ||
| Padding | Same | Same |
| Pooling | ||
| Epochs | 200 | 200 |
| Drop out | 0.3 | 0.3 |
| Regularizer | N/A | N/A |
| Architecture | Fully-trained and Fine-tuned | Fully-trained and Fine-tuned |
Performance metrics of VGG16 architecture on our dataset.
| Architecture | Folds | Confusion Matrices | Performance Evaluation (%) | Average (%) | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Predict → | NC | C | Precision | Recall | F1 | Test | Acc. | F1 | ||
|
| Fold 1 | non-carcinoma | 75 | 7 | 97.40 | 91.46 | 94.34 | 82 |
|
|
| carcinoma | 2 | 86 | 92.47 | 97.73 | 95.03 | 88 | ||||
| Fold 2 | non-carcinoma | 65 | 17 | 90.28 | 79.27 | 84.42 | 82 |
|
| |
| carcinoma | 7 | 81 | 82.65 | 92.05 | 87.10 | 88 | ||||
| Fold 3 | non-carcinoma | 73 | 9 | 93.59 | 89.02 | 91.25 | 82 |
|
| |
| carcinoma | 5 | 83 | 90.22 | 94.32 | 92.22 | 88 | ||||
| Fold 4 | non-carcinoma | 70 | 12 | 95.89 | 85.37 | 90.32 | 82 |
|
| |
| carcinoma | 3 | 85 | 87.63 | 96.59 | 91.89 | 88 | ||||
| Fold 5 | non-carcinoma | 78 | 4 | 91.76 | 95.12 | 93.41 | 82 |
|
| |
| carcinoma | 7 | 81 | 95.29 | 92.05 | 93.64 | 88 | ||||
| Avg. | non-carcinoma | – | – | 93.78 | 88.05 | 90.75 | 82 |
|
| |
| carcinoma | – | – | 89.65 | 94.55 | 91.98 | 88 | ||||
|
| Fold 1 | non-carcinoma | 67 | 15 | 87.01 | 81.71 | 84.28 | 82 |
|
|
| carcinoma | 10 | 78 | 83.87 | 88.64 | 86.19 | 88 | ||||
| Fold 2 | non-carcinoma | 74 | 8 | 92.50 | 90.24 | 91.36 | 82 |
|
| |
| carcinoma | 6 | 82 | 91.11 | 93.18 | 92.13 | 88 | ||||
| Fold 3 | non-carcinoma | 76 | 6 | 95.00 | 92.68 | 93.83 | 82 |
|
| |
| carcinoma | 4 | 84 | 93.33 | 95.45 | 94.38 | 88 | ||||
| Fold 4 | non-carcinoma | 73 | 9 | 96.05 | 89.02 | 92.41 | 82 |
|
| |
| carcinoma | 3 | 85 | 90.43 | 96.59 | 93.41 | 88 | ||||
| Fold 5 | non-carcinoma | 75 | 7 | 96.15 | 91.46 | 93.75 | 82 |
|
| |
| carcinoma | 3 | 85 | 92.39 | 96.59 | 94.44 | 88 | ||||
| Avg. | non-carcinoma | – | – | 93.34 | 89.02 | 91.13 | 82 |
|
| |
| carcinoma | – | – | 90.23 | 94.09 | 92.11 | 88 | ||||
Figure 5The training and validation accuracy curves of fully-trained and fine-tuned VGG16 models.
Figure 6The training and validation loss curves of fully-trained and fine-tuned VGG16 models.
The training and prediction times of fully-trained and fine-tuned models.
| Model | Single Training Time | 5-Fold Training Time | Prediction Time |
|---|---|---|---|
| Fully-trained VGG16 | 17 min 50 s | 89 min | 30 s |
| Fine-tuned VGG16 | 17 min 25 s | 87 min | 31 s |
| Fully-trained VGG19 | 20 min 40 s | 103 min | 35 s |
| Fine-tuned VGG19 | 19 min 55 s | 99 min | 36 s |
Performance metrics of VGG19 architecture on our dataset.
| Architecture | Folds | Confusion Matrices | Performance Evaluation (%) | Average (%) | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Predict → | NC | C | Precision | Recall | F1 | Test | Acc. | F1 | ||
|
| Fold 1 | non-carcinoma | 66 | 16 | 98.51 | 80.49 | 88.59 | 82 |
|
|
| carcinoma | 1 | 87 | 84.47 | 98.86 | 91.10 | 88 | ||||
| Fold 2 | non-carcinoma | 71 | 11 | 94.67 | 86.59 | 90.45 | 82 |
|
| |
| carcinoma | 4 | 84 | 88.42 | 95.45 | 91.80 | 88 | ||||
| Fold 3 | non-carcinoma | 69 | 13 | 98.57 | 84.15 | 90.79 | 82 |
|
| |
| carcinoma | 1 | 87 | 87.00 | 98.86 | 92.55 | 88 | ||||
| Fold 4 | non-carcinoma | 69 | 13 | 90.79 | 84.15 | 87.34 | 82 |
|
| |
| carcinoma | 7 | 81 | 86.17 | 92.05 | 89.01 | 88 | ||||
| Fold 5 | non-carcinoma | 73 | 9 | 91.25 | 89.02 | 90.12 | 82 |
|
| |
| carcinoma | 7 | 81 | 90.00 | 92.05 | 91.01 | 88 | ||||
| Avg. | non-carcinoma | – | – | 94.76 | 84.88 | 89.46 | 82 |
|
| |
| carcinoma | – | – | 87.21 | 95.45 | 91.09 | 88 | ||||
|
| Fold 1 | non-carcinoma | 64 | 18 | 98.46 | 78.05 | 87.07 | 82 |
|
|
| carcinoma | 1 | 87 | 82.86 | 98.86 | 90.16 | 88 | ||||
| Fold 2 | non-carcinoma | 75 | 7 | 93.75 | 91.46 | 92.59 | 82 |
|
| |
| carcinoma | 5 | 83 | 92.22 | 94.32 | 93.26 | 88 | ||||
| Fold 3 | non-carcinoma | 75 | 7 | 97.40 | 91.46 | 94.34 | 82 |
|
| |
| carcinoma | 2 | 86 | 92.47 | 97.73 | 95.03 | 88 | ||||
| Fold 4 | non-carcinoma | 76 | 6 | 96.20 | 92.68 | 94.41 | 82 |
|
| |
| carcinoma | 3 | 85 | 93.41 | 96.59 | 94.97 | 88 | ||||
| Fold 5 | non-carcinoma | 71 | 11 | 89.87 | 86.59 | 88.20 | 82 |
|
| |
| carcinoma | 8 | 80 | 87.91 | 90.91 | 89.39 | 88 | ||||
| Avg. | non-carcinoma | – | – | 95.14 | 88.05 | 91.32 | 82 |
|
| |
| carcinoma | – | – | 89.77 | 95.68 | 92.56 | 88 | ||||
Figure 7The training and validation accuracy curves of fully-trained and fine-tuned VGG19 models.
Figure 8The training and validation loss curves of fully-trained and fine-tuned VGG19 models.
Performance metrics of ensemble VGG16 and VGG19 architectures.
| Ensemble Method | Confusion Matrices | Performance Evaluation (%) | Average (%) | ||||||
|---|---|---|---|---|---|---|---|---|---|
| Predict → | NC | C | Precision | Recall | F1 | Test | Accuracy | F1 | |
|
| non-carcinoma | 73 | 9 | 97.33 | 89.02 | 92.99 | 82 |
|
|
|
| carcinoma | 2 | 86 | 90.53 | 97.73 | 93.99 | 88 | ||
|
| non-carcinoma | 76 | 6 | 97.44 | 92.68 | 95.00 | 82 |
|
|
|
| carcinoma | 2 | 86 | 93.48 | 97.73 | 95.56 | 88 | ||