| Literature DB >> 32258124 |
Habib Dhahri1,2, Ines Rahmany2, Awais Mahmood1, Eslam Al Maghayreh3, Wail Elkilani1.
Abstract
Breast cancer is the most diagnosed cancer among women around the world. The development of computer-aided diagnosis tools is essential to help pathologists to accurately interpret and discriminate between malignant and benign tumors. This paper proposes the development of an automated proliferative breast lesion diagnosis based on machine-learning algorithms. We used Tabu search to select the most significant features. The evaluation of the feature is based on the dependency degree of each attribute in the rough set. The categorization of reduced features was built using five machine-learning algorithms. The proposed models were applied to the BIDMC-MGH and Wisconsin Diagnostic Breast Cancer datasets. The performance measures of the used models were evaluated owing to five criteria. The top performing models were AdaBoost and logistic regression. Comparisons with others works prove the efficiency of the proposed method for superior diagnosis of breast cancer against the reviewed classification techniques.Entities:
Mesh:
Year: 2020 PMID: 32258124 PMCID: PMC7064857 DOI: 10.1155/2020/4671349
Source DB: PubMed Journal: Biomed Res Int Impact factor: 3.411
Summary of machine-learning algorithms for breast cancer diagnosis.
| Author | Feature | Algorithm | Accuracy (%) | Dataset |
|---|---|---|---|---|
| Asri et al. [ | FNA | SVM | 97.13 | UCI |
| Ivančáková et al. [ | FNA | SVM | 97.66 | WDBC |
| Mondal et al. [ | Entropy | SVM | 91.5 | Gene Expression Omnibus |
| Ghasemzadeh et al. [ | Gabor wavelet | SVM | 96 | Mammography (DDSM) |
| Ayoub Shaikh and Ali [ | Wrapper subset eval | SVM | 97 | Breast Cancer Digital Repository (BCDR) |
| Wang et al. [ | Full features | SVM | 33.34 | SEER |
| Mengjie Yu [ | Concave points | SVM | 99.77 | UCI |
| Wei et al. [ | BiCNN | CNN | 97 | BreaKHis |
| Bejnordi et al. [ | Morphology | CNN | 92 | WSIs |
| Arau et al. [ | Full features | CNN | 83 | Histology Dataset |
| Yap et al. [ | FCN-alexnet | CNN | 98 | B&K Medical Panther 2002 and B&K Medical Hawk 2102 US systems |
| 92 | ||||
| Ting et al. [ | Wise | CNN | 90.50 | Digital Mammogram |
| Zhou et al. [ | CNN | CNN | 95.8 | SWE data |
| Sun et al. [ | mRMR | Deep neural network | 18.7 | METABRIC/MDNNMD |
| Kaur et al. [ | CNN | MLP | 86 | Mini-MIAS |
| Joshi et al. [ | Scaling | NN | 96.47 | WDBC |
| Radiya-Dixit et al. [ | Computational method | LR | 91.8 | BIDMC-MGH |
| Tahmassebi et al. [ | Volume distribution | LR | 92 | WDBC |
| Braman et al. [ | Heterogeneity | LR | 93 | ISPY1-TRIAL |
| Maysanjaya et al. [ | Wrapper | NB | 99.27. | UCI |
| Chaurasia et al. [ | FNA | NB | 97.36 | WDBC |
| Tamilvanan [ | Dimensionality reduction | NB | 82 | WDBC |
| Qiao et al. [ | BI-RADS | AdaBoost | 93.48 | 138 pathologically proven breast tumors |
| Turkki et al. [ | Morphological | KNN | 95 | FinProg |
| Amrane et al. [ | FNA | KNN | 97.51 | WDBC |
Figure 1Sample of trial solution.
Figure 2Feature selection strategy.
Classifier results on BIDMC-MGH without Tabu feature selection.
| Model | Accuracy (%) | Precision (%) | Sensitivity (%) |
| AUC (%) |
|---|---|---|---|---|---|
| KNN | 72.54 | 75 | 73 | 73 | 74 |
| GNB | 68.62 | 70 | 69 | 69 | 69 |
| LR | 82.35 | 83 | 82 | 83 | 83 |
| ET | 72.54 | 72 | 73 | 72 | 69 |
| AB | 78.43 | 79 | 78 | 79 | 77 |
Classifier results on BIDMC-MGH with Tabu feature selection.
| Model | Accuracy (%) | Precision (%) | Sensitivity (%) |
| AUC (%) |
|---|---|---|---|---|---|
| KNN | 74.50 | 79 | 75 | 74 | 76 |
| GNB | 78 | 79 | 76 | 77 | 78 |
| LR | 82.35 | 82 | 82 | 82 | 82 |
| ET | 90.19 | 91 | 90 | 90 | 89 |
| AB | 96.07 | 96 | 96 | 96 | 95 |
Figure 3Analyzing the obtained results via different classifiers on BIDMC-MGH.
Comparison of proposed method and other methods on BIDMC-MGH.
| Model | AUC (%) |
|---|---|
| L1-regularized LR [ | 85.8 |
| L1-regularized LR with active feature [ | 89.7 |
| LR with early stopping and active features [ | 88.4 |
| CAFE model [ | 91.8 |
|
|
|
Figure 4The ROC curve of the applied machine-learning algorithms.
Figure 5The effect of feature selection on accuracy on BIDMC-MGH.
Classifier results on WDBC without Tabu feature selection.
| Model | Accuracy (%) | Precision (%) | Sensitivity (%) |
| AUC (%) |
|---|---|---|---|---|---|
| KNN | 97.07 | 97 | 97 | 97 | 97 |
| GNB | 92.98 | 93 | 93 | 93 | 93 |
| LR | 96.49 | 97 | 96 | 97 | 97 |
| ET | 95.32 | 96 | 95 | 95 | 96 |
| AB | 95.90 | 96 | 96 | 96 | 96 |
Classifier results on WDBC with Tabu feature selection.
| Model | Accuracy (%) | Precision (%) | Sensitivity (%) |
| AUC (%) |
|---|---|---|---|---|---|
| KNN | 98.24 | 98 | 98 | 98 | 97 |
| GNB | 95.32 | 95 | 95 | 95 | 95 |
| LR | 98 | 99 | 99 | 99 | 98 |
| ET | 97 | 97 | 97 | 97 | 97 |
| AB | 97.66 | 98 | 98 | 98 | 97 |
Figure 6Effect of feature selection on accuracy classifiers on WDBC.
Comparison of proposed method and other methods on WDBC.
| Model | Accuracy (%) |
|---|---|
| NN [ | 96.47 |
| KNN [ | 97.51 |
| LR [ | 92 |
| NB [ | 97.36 |
| NB [ | 82 |
|
|
|