| Literature DB >> 35030211 |
Esraa A Mohamed1, Essam A Rashed1,2, Tarek Gaber3,4, Omar Karam5.
Abstract
Breast cancer is one of the most common diseases among women worldwide. It is considered one of the leading causes of death among women. Therefore, early detection is necessary to save lives. Thermography imaging is an effective diagnostic technique which is used for breast cancer detection with the help of infrared technology. In this paper, we propose a fully automatic breast cancer detection system. First, U-Net network is used to automatically extract and isolate the breast area from the rest of the body which behaves as noise during the breast cancer detection model. Second, we propose a two-class deep learning model, which is trained from scratch for the classification of normal and abnormal breast tissues from thermal images. Also, it is used to extract more characteristics from the dataset that is helpful in training the network and improve the efficiency of the classification process. The proposed system is evaluated using real data (A benchmark, database (DMR-IR)) and achieved accuracy = 99.33%, sensitivity = 100% and specificity = 98.67%. The proposed system is expected to be a helpful tool for physicians in clinical use.Entities:
Mesh:
Year: 2022 PMID: 35030211 PMCID: PMC8759675 DOI: 10.1371/journal.pone.0262349
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Breast thermography procedure (thermal image is aquired at room temperature = 22°C).
Fig 2Flowchart of the proposed method.
Fig 3Example of U-Net architecture [33].
Fig 4Example of breast area segmentation with U-Net.
Fig 5Architecture of the proposed deep learning model.
Fig 6Different cases of breast (a) small breast (b) large breast (c) asymmetric breast.
Dataset description.
| Dataset categories | Dimension | Training | Validation | Testing | Total |
|---|---|---|---|---|---|
| Normal | 640x480 | 350 | 75 | 75 | 500 |
| Abnormal | 640x480 | 350 | 75 | 75 | 500 |
Fig 7Breast area segmentation resuls (a) thermal image (b) ground truth (c) output.
Fig 8The training progress of the proposed deep learning model.
Fig 9The confusion matrix of the proposed model.
Comparison between solvers (initial learn rate = 2.0e−3, number of epochs = 30 and batch size = 60).
| Solver | Accuracy (%) | Sensitivity (%) | Specificity (%) |
|---|---|---|---|
|
| 99.33 | 100 | 98.67 |
|
| 84.17 | 100 | 68.33 |
|
| 50 | 100 | 0.0 |
The impact of using different number of epochs on the classification accuracy, sensitivity and specificity (solver = ADAM, initial learn rate = 2.0e−3, batch size = 60).
| Number of Epochs | Accuracy (%) | Sensitivity (%) | Specificity (%) |
|---|---|---|---|
|
| 88.67 | 94.67 | 82.67 |
|
| 97.33 | 100 | 94.67 |
|
| 99.33 | 100 | 98.67 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
Impact of using different batch size on the classification accuracy, sensitivity and specificity (solver = ADAM, initial learn rate = 2.0e−3, number of epochs = 30).
| Batch Size | Accuracy (%) | Sensitivity (%) | Specificity(%) |
|---|---|---|---|
|
| 50 | 100 | 0.0 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 99.33 | 100 | 98.67 |
|
| 98.67 | 100 | 97.33 |
|
| 95.83 | 100 | 91.67 |
|
| 91.33 | 100 | 82.66 |
|
| 83.33 | 100 | 66.67 |
Impact of starting the training process with different initial learn rate on the classification accuracy, sensitivity and specificity (solver = ADAM, batch size = 60 and number of epochs = 30).
| Initial learning rate | Accuracy(%) | Sensitivity(%) | specificity(%) |
|---|---|---|---|
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 100 | 0.0 |
|
| 50 | 0.0 | 100 |
|
| 50 | 0.0 | 100 |
|
| 83.33 | 100 | 66.67 |
|
| 41.07 | 10.0 | 73.33 |
|
| 51.67 | 5.0 | 98.33 |
|
| 56.67 | 13.33 | 100 |
|
| 83.33 | 100 | 66.67 |
|
| 99.33 | 100 | 98.67 |
|
| 99.33 | 100 | 98.67 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
|
| 100 | 100 | 100 |
Comparison between the performance metrics of different CNN models and the proposed model.
| CNN model | Accuracy | Sensitivity | Specificity |
|---|---|---|---|
|
| 93.3 | 88.0 | 98.7 |
|
| 79.33 | 84.00 | 74.67 |
|
| 50.0 | 0.0 | 100 |
|
| 100 | 100 | 100 |
|
| 99.33 | 100 | 98.67 |
Fig 10Evaluation metrics over different dataset size.
Comparison between the performance metrics of different machine learning classifier with texture features and the proposed model.
| Classifier | Accuracy(%) | Sensitivity(%) | specificity(%) |
|---|---|---|---|
|
| 89.33 | 86.67 | 92.0 |
|
| 53.33 | 64.0 | 42.67 |
|
| 82.67 | 96.00 | 96.0 |
|
| 99.33 | 100 | 98.67 |
Comparison between the performance metrics of different machine learning classifier with HOG features and the proposed model.
| Classifier | Accuracy(%) | Sensitivity(%) | specificity(%) |
|---|---|---|---|
|
| 78.28 | 73.33 | 86.67 |
|
| 20.0 | 20.0 | 20.0 |
|
| 40.0 | 58.67 | 21.33 |
|
| 99.33 | 100 | 98.67 |
Results of the ANOVA test of the proposed model and CNN models.
| Model | P-value |
|---|---|
|
| 0.0423 |
|
| 0.0173 |
|
| 0.0023 |
Comparison with other studies on breast cancer detection (n = normal, ab = abnormal, Ea = Early, Ac = Acute).
| Ref. | Segmentation method | #patients / Thermograms | Classification Method | Results |
|---|---|---|---|---|
| [ | an enhanced segmentation method based on both Neutrosophic sets (NS) and optimized Fast Fuzzy c-mean (F-FCM) algorithm. | 63 thermograms (29 N / 34 AB) | SVM Classifier | Accuracy = 92.06% |
| Precision = 87.5% | ||||
| Recall = 96.55% | ||||
| [ | Manual | 40 thermograms (26 N / 14 AB) | SVM, Naïve Bayes and KNN classifier | Accuracy = 92.5% and Sensitivity = 78.6% with KNN |
| Accuracy = 85% and Sensitivity = 85.7% with SVM | ||||
| Accuracy = 80% and Sensitivity = 85.7% with Naïve Bayes | ||||
| [ | Manual | 68 thermograms (26 Ea / 42 Ac) | DT, KNN, SVM and SVM-RBF | Accuracy = 95.59%, |
| Sensitivity = 96% and Specificity = 95.35% with SVM-RBF | ||||
| [ | Canny edge detection methods followed by gradient operators and Hough transform for boundary detection | Thermograms of 22 women (11 N / 11AB) | SVM Classifier | Accuracy = 90.91%, |
| Sensitivity = 81.82% | ||||
| Specificity = 100% | ||||
| [ | Otsu’s threshold to remove background followed by a reconstruction technique. | 306 thermograms (183 N / 123 AB) | Feed-forward artificial neural network with gradient decent | Accuracy = 90.48%, |
| Sensitivity = 87.6%, | ||||
| Specificity = 89.73% | ||||
| [ | Manual | 600 thermograms (300 N / 300 AB) | SVM-C | Accuracy = 93.5%, |
| Sensitivity = 93%, | ||||
| Specificity = 94% | ||||
| [ | Not defined | 282 thermograms (147 N / 135 AB) | CNN using transfer learning | Accuracy = 94.3% |
| Precision = 94.7% | ||||
| Recall = 93.3% | ||||
| [ | Projection profile analysis | 140 patients (98 N / 32 AB) | Convolutional Neural Networks optimized by Bayes algorithm | Accuracy = 98.95% |
|
| U-Net network | 1000 thermograms (500 N / 500 AB) | Two-class CNN-based deep learning model | Accuracy = 99.33%, |
| Sensitivity = 100%, | ||||
| Specificity = 98.67% |
Table of abbreviation.
| Abbreviation | Definition |
|---|---|
|
| Convolutional Neural Networks |
|
| Computer-Aided Detection |
|
| Region of Interest |
|
| Extended Hidden Markov Models |
|
| Neutrosophic Sets |
|
| Fast Fuzzy C-Mean |
|
| Support Vector Machine |
|
| K-Nearest Neighbor |
|
| Naïve Bayes |
|
| Decision Tree |
|
| Principal Component Analysis |
|
| Deep Neural Network |
|
| Deep-Wavelet Neural Networks |
|
| Rectified Linear Activation Function |
|
| Adaptive Moment Estimation |
|
| Stochastic Gradient Descent with Momentum |
|
| Root Mean Square propagation |
|
| True Positive |
|
| True Negative |
|
| False Positive |
|
| False Negative |
|
| Gray Level Co-occurrence Matrices |
|
| Histogram of Oriented Gradients |
|
| normal |
|
| abnormal |