| Literature DB >> 34055041 |
S Murugesan1, R S Bhuvaneswaran1, H Khanna Nehemiah1, S Keerthana Sankari2, Y Nancy Jane3.
Abstract
A computer-aided diagnosis (CAD) system that employs a super learner to diagnose the presence or absence of a disease has been developed. Each clinical dataset is preprocessed and split into training set (60%) and testing set (40%). A wrapper approach that uses three bioinspired algorithms, namely, cat swarm optimization (CSO), krill herd (KH) ,and bacterial foraging optimization (BFO) with the classification accuracy of support vector machine (SVM) as the fitness function has been used for feature selection. The selected features of each bioinspired algorithm are stored in three separate databases. The features selected by each bioinspired algorithm are used to train three back propagation neural networks (BPNN) independently using the conjugate gradient algorithm (CGA). Classifier testing is performed by using the testing set on each trained classifier, and the diagnostic results obtained are used to evaluate the performance of each classifier. The classification results obtained for each instance of the testing set of the three classifiers and the class label associated with each instance of the testing set will be the candidate instances for training and testing the super learner. The training set comprises of 80% of the instances, and the testing set comprises of 20% of the instances. Experimentation has been carried out using seven clinical datasets from the University of California Irvine (UCI) machine learning repository. The super learner has achieved a classification accuracy of 96.83% for Wisconsin diagnostic breast cancer dataset (WDBC), 86.36% for Statlog heart disease dataset (SHD), 94.74% for hepatocellular carcinoma dataset (HCC), 90.48% for hepatitis dataset (HD), 81.82% for vertebral column dataset (VCD), 84% for Cleveland heart disease dataset (CHD), and 70% for Indian liver patient dataset (ILP).Entities:
Year: 2021 PMID: 34055041 PMCID: PMC8149240 DOI: 10.1155/2021/6662420
Source DB: PubMed Journal: Comput Math Methods Med ISSN: 1748-670X Impact factor: 2.238
Abbreviations used.
| Abbreviation | Phrase |
|---|---|
| ABCO | Artificial bee colony optimization |
| ACO | Ant colony optimization |
| ANN | Artificial neural networks |
| BCS | Binary cuckoo search |
| BFA | Binary firefly algorithm |
| BFO | Bacterial foraging optimization |
| BP | Back propagation |
| BPNN | Back propagation neural network |
| CAD | Computer-aided diagnosis |
| CDC | Counts of dimension to change |
| CDSS | Clinical decision support system |
| CFCSA | Hybrid crow search optimization algorithm |
| CGA | Conjugate gradient algorithm |
| CHD | Cleveland heart disease |
| CMVO | Chaotic multiverse optimization |
| CSM | Cosine similarity measure |
| CSO | Cat swarm optimization |
| CT | Computed tomography |
| DE | Differential evolution |
| DGA | Distance-based genetic algorithm |
| DISON | Diverse intensified strawberry optimized neural network |
| DNN | Deep neural network |
| E.coli | Escherichia Coli Bacteria |
| ECSA | Enhanced crow search algorithm |
| ELM | Extreme learning machine |
| FBFO | Feature selected by bacterial foraging optimization |
| FCM | Fuzzy |
| FCSO | Feature selected by cat swarm optimization |
| FFO | Firefly optimization |
| FKH | Feature selected by krill herd |
| GA | Genetic algorithm |
| GSO | Glowworm swarm optimization |
| HCC | Hepatocellular carcinoma |
| HD | Hepatitis |
| IBPSO | Improved binary particle swarm optimization |
| ILP | Indian liver patient |
| ISSA | Improved Salp swarm algorithm |
| KH | Krill herd |
| k-NN |
|
| LO | Lion optimization |
| LR | Logistic regression |
| MCC | Mathew's correlation coefficient |
| MFO | Moth-flame optimization |
| ML | Machine learning |
| MPNN | Multilayer perceptron neural network |
| MR | Mixed ratio |
| NB | Naive Bayes |
| PCC | Pearson correlation coefficient |
| PID | Pima Indian diabetes |
| PSO | Particle swarm optimization |
| RD | Random diffusion |
| RDM | Rough dependency measure |
| RF | Random forest |
| RoIs | Regions of interest |
| SHD | Statlog heart disease |
| SMOTE | Synthetic minority oversampling technique |
| SMP | Seeking memory pool |
| SPC | Self-position consideration |
| SRD | Seeking range of the selected dimension |
| SVC | Support vector classification |
| SVM | Support vector machine |
| TS | Thoracic surgery |
| UCI | University of California Irvine |
| VCD | Vertebral column dataset |
| WBC | Wisconsin breast cancer |
| WDBC | Wisconsin diagnostic breast cancer |
| WOA | Whale optimization algorithm |
Outline of the datasets used.
| Dataset name | No. of instances | No. of features∗ | No. of missing values | Class labels with no. of instances associated with each class label | Interpretation of class labels |
|---|---|---|---|---|---|
| WDBC | 569 | 31 | Nil | M (212)/B (357) | M-malignant, B-benign |
| SHD | 270 | 13 | Nil | 2 (120)/1 (150) | 2-present, 1-absent |
| HCC | 165 | 49 | 826 | 0 (63)/1 (102) | 0-dies, 1-lives |
| HD | 155 | 18 | 167 | 1 (32)/2 (123) | 1-die, 2-live |
| VCD | 310 | 6 | Nil | 0 (210)/1 (100) | 0-abnormal, 1-normal |
| CHD | 303 | 13 | Nil | 1 (139)/2 (164) | 1-presence, 2-absence |
| ILP | 583 | 10 | Nil | 1 (416)/2 (167) | 1-diseased, 2-nondiseased |
∗ without class label.
Figure 1System framework.
Outline of training and testing instances of each C.
| Instances | WDBC dataset | SHD dataset | HCC dataset | HD dataset | VCD dataset | CHD dataset | ILP dataset |
|---|---|---|---|---|---|---|---|
| Total number of instances before SMOTE | 569 | 270 | 165 | 155 | 310 | 303 | 583 |
| Total number of instances after SMOTE | 780 | 270 | 228 | 251 | 410 | 303 | 750 |
| Number of training instances for FCSO/FKH/FBFO classifiers | 468 | 162 | 137 | 151 | 246 | 182 | 450 |
| 60% of the total number of instances after SMOTE | |||||||
| Number of testing instances for FCSO/FKH/FBFO classifiers | 312 | 108 | 91 | 100 | 164 | 121 | 300 |
| 40% of the total number of instances after SMOTE | |||||||
| Number of training instances for super learner | 250 | 86 | 73 | 80 | 131 | 97 | 240 |
| 80% of the total testing instances∗ for FCSO/FKH/FBFO classifiers | |||||||
| Number of testing instances for | 62 | 22 | 18 | 20 | 33 | 24 | 60 |
| Super learner | 20% of the total testing instances∗ for FCSO/FKH/FBFO classifiers | ||||||
∗Each instance refers to the classification result pertaining to each instance of the testing set for FCSO, FKH, and FBFO classifiers and the class label corresponding to each instance of the testing set.
Outline of the parameters used in CSO.
| Parameter | Description |
|---|---|
| SMP | SMP is used to define the size of the seeking memory of each cat. Each cat selects possible neighborhood position from a set of solutions. |
| SRD | SRD is used to define the seeking range of the selected dimension. |
| CDC | CDC is a count of dimensions to be changed in seeking mode. |
| SPC | SPC indicates whether the cat is in the current position or not. |
| N | Number of cats |
| MR | Mixed ratio of cats |
| C | Constant value |
| D | Size of dimension |
| R | Random number in the range of [0,1] |
Algorithm 1Outline of the parameters used in the KH algorithm.
| Parameter | Definition | Value |
|---|---|---|
|
| Maximum foraging speed |
|
| RDmax | Maximum random diffusion speed | RDmax ∈ (0.002 − 0.01 ) m/s−1 |
|
| Maximum induction speed |
|
|
| Inertia weight of the motion induced |
|
|
| Inertia weight of the foraging motion |
|
|
| Step-length scaling factor | Constant no.between [0, 2] |
|
| Random directional vector | Random numbers [−1, 1] |
Algorithm 2Outline of the parameters used in the BFO algorithm.
| Parameter | Description |
|---|---|
|
| Number of features |
|
| Number of bacteria |
|
| Number of bacteria in the reproduction steps |
|
| No. of reproductive steps |
|
| No. of elimination-dispersal steps |
|
| No. of chemotactic steps |
|
| No. of swimming steps |
|
| Bacteria step size length |
|
| Elimination probability |
| ∅( | Direction of |
|
| Index of the chemotactic process |
|
| Index of the reproduction process. |
|
| Index of the elimination-dispersal process |
|
| The |
|
| A bacterium on the optimization domain |
|
| The highest objective function value |
| ∆( | A random vector and its value lie between -1 and 1 |
|
| Cell-to-cell attractant effect to nutrient concentration |
Figure 2Classification using BPNN.
Parameter settings for BPNN.
| BPNN parameter | Bioinspired algorithm | WDBC dataset | SHD dataset | HCC dataset | HD dataset | VCD dataset | CHD dataset | ILP dataset |
|---|---|---|---|---|---|---|---|---|
| Number of input nodes | CSO | 15 | 9 | 20 | 16 | 3 | 6 | 5 |
| KH | 17 | 10 | 39 | 10 | 3 | 10 | 8 | |
| BFO | 18 | 9 | 35 | 19 | 2 | 11 | 5 | |
| Number of hidden nodes | CSO | 30 | 18 | 40 | 32 | 6 | 12 | 10 |
| KH | 34 | 20 | 78 | 20 | 6 | 20 | 16 | |
| BFO | 36 | 18 | 70 | 38 | 4 | 22 | 10 |
Figure 3Classifier testing.
Figure 4Super learner training and testing.
Parameter settings for super learner.
| Name of the parameter | WDBC dataset | SHD dataset | HCC dataset | HD dataset | VCD dataset | CHD dataset | ILP dataset |
|---|---|---|---|---|---|---|---|
| Initial population size | 250 | 86 | 73 | 80 | 131 | 97 | 240 |
| Number of input nodes | 3 | 3 | 3 | 3 | 3 | 3 | 3 |
| Number of hidden nodes | 6 | 6 | 6 | 6 | 6 | 6 | 6 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on WDBC dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 15 | 137 | 4 | 6 | 165 | 96.79 | 96.49 | 97.16 | 97.63 | 0.97 |
| KH | 17 | 139 | 2 | 5 | 166 | 97.76 | 97.08 | 98.58 | 98.81 | 0.98 |
| BFO | 18 | 139 | 2 | 8 | 163 | 96.79 | 95.32 | 98.58 | 98.79 | 0.97 |
| Super learner | — | 22 | 0 | 2 | 39 | 96.83 | 95.12 | 100.00 | 100.00 | 0.98 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on Statlog dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 9 | 53 | 8 | 9 | 38 | 84.26 | 80.85 | 86.89 | 82.61 | 0.82 |
| KH | 10 | 53 | 8 | 11 | 36 | 82.41 | 76.60 | 86.89 | 81.82 | 0.79 |
| BFO | 9 | 51 | 10 | 10 | 37 | 81.48 | 78.72 | 83.61 | 78.72 | 0.79 |
| Super learner | — | 10 | 1 | 2 | 9 | 86.36 | 81.82 | 90.91 | 90.00 | 0.86 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on HCC dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 20 | 43 | 9 | 9 | 31 | 80.43 | 77.50 | 82.69 | 77.50 | 0.78 |
| KH | 39 | 48 | 4 | 13 | 27 | 81.52 | 67.50 | 92.31 | 87.10 | 0.76 |
| BFO | 35 | 47 | 5 | 20 | 20 | 72.83 | 50.00 | 90.38 | 80.00 | 0.62 |
| Super learner | — | 10 | 1 | 0 | 8 | 94.74 | 100.00 | 90.91 | 88.89 | 0.94 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on hepatitis dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 16 | 47 | 2 | 10 | 42 | 88.12 | 80.77 | 95.92 | 95.45 | 0.88 |
| KH | 10 | 45 | 4 | 6 | 46 | 90.10 | 88.46 | 91.84 | 92.00 | 0.90 |
| BFO | 19 | 47 | 2 | 12 | 40 | 86.14 | 76.92 | 95.92 | 95.24 | 0.85 |
| Super learner | — | 8 | 1 | 1 | 11 | 90.48 | 91.67 | 88.89 | 91.67 | 0.92 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on vertebral column dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 3 | 74 | 10 | 17 | 63 | 83.54 | 78.75 | 88.10 | 86.30 | 0.82 |
| KH | 3 | 81 | 3 | 13 | 67 | 90.24 | 83.75 | 96.43 | 95.71 | 0.89 |
| BFO | 2 | 80 | 4 | 17 | 63 | 87.20 | 78.75 | 95.24 | 94.03 | 0.86 |
| Super learner | — | 19 | 2 | 4 | 8 | 81.82 | 66.67 | 90.48 | 80.00 | 0.73 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on Cleveland heart disease dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 6 | 60 | 8 | 12 | 42 | 83.61 | 77.78 | 88.24 | 84.00 | 0.81 |
| KH | 10 | 56 | 12 | 11 | 43 | 81.15 | 79.63 | 82.35 | 78.18 | 0.79 |
| BFO | 11 | 53 | 15 | 13 | 41 | 77.05 | 75.93 | 77.94 | 73.21 | 0.75 |
| Super learner | — | 13 | 2 | 2 | 8 | 84.00 | 80.00 | 86.67 | 80.00 | 0.80 |
Performance of FCSO, FKH, and FBFO classifiers and super learner on Indian liver patient dataset.
| Feature selection algorithm | Size of feature subset | TN | FP | FN | TP | Accuracy | Sensitivity | Specificity | Precision |
|
|---|---|---|---|---|---|---|---|---|---|---|
| CSO | 5 | 103 | 62 | 33 | 102 | 68.33 | 75.56 | 62.42 | 62.20 | 0.68 |
| KH | 8 | 104 | 61 | 40 | 95 | 66.33 | 70.37 | 63.03 | 60.90 | 0.65 |
| BFO | 5 | 101 | 64 | 34 | 101 | 67.33 | 74.81 | 61.21 | 61.21 | 0.67 |
| Super learner | — | 26 | 15 | 3 | 16 | 70.00 | 84.21 | 63.41 | 51.61 | 0.64 |
Comparison of the proposed work and existing work using clinical dataset.
| Author/year | Method/reference | Accuracy % | ||||||
|---|---|---|---|---|---|---|---|---|
| WDBC | SHD | HCC | HD | VCD | CHD | ILP | ||
| Ayon et al. (2020) | DNN [ | — | 98.15 | — | — | — | 94.39 | — |
| SVM [ | — | 97.41 | — | — | — | 97.36 | — | |
| Bai Ji et al. (2020) | IBPSO with | 96.14 | — | — | — | — | — | — |
| Elgin et al. (2020) | Cooperative coevolution and RF [ | 97.1 | 96.8 | 72.2 | 82.3 | 91.4 | 93.4 | — |
| Magesh et al. (2020) | Cluster-based decision tree [ | — | — | — | — | — | 89.30 | — |
| Rabbi et al. (2020) | PCC and AdaBoost [ | — | — | — | — | — | — | 92.19 |
| Rajesh et al. (2020) | RF classifier [ | — | — | 80.64 | — | — | — | — |
| Salima et al. (2020) | ECSA with | 95.76 | 82.96 | — | — | — | — | — |
| Singh J et al. (2020) | Logistic regression [ | — | — | — | — | — | — | 74.36 |
| Sreejith et al. (2020) | CMVO and RF [ | — | — | — | — | — | — | 82.46 |
| Sreejith et al. (2020) | DISON and ERT[ | — | 94.5 | — | — | 87.17 | 93.67 | — |
| Tougui et al. (2020) | ANN with Matlab [ | — | — | — | — | — | 85.86 | — |
| Tubishat et al. (2020) | ISSA with k-NN [ | — | 88.1 | — | — | 89.0 | — | — |
| Abdar et al. (2019) | Novel nested ensemble nu-SVC [ | — | — | — | — | — | 98.60 | — |
| Anter et al. (2019) | CFCSA with chaotic maps [ | 98.6 | — | — | 68.0 | — | 88.0 | 68.4 |
| Aouabed et al. (2019) | Nested ensemble nu-SVC, GA and multilevel balancing [ | — | — | — | — | — | 98.34 | — |
| Elgin et al. (2019) | DE, LO and GSO with Adaboost SVM [ | 98.73 | — | — | 93.9 | — | — | — |
| Książek et al. (2019) | SVM [ | — | 97.41 | — | — | — | 97.36 | — |
| Sayed et al. (2019) | Novel chaotic crow search algorithm with | 90.28 | 78.84 | — | 83.7 | — | — | 71.68 |
| Abdar et al. (2018) | MPNN and C5.0 [ | — | — | — | — | — | — | 94.12 |
| Abdullah et al. (2018) |
| — | — | — | — | 85.32 | — | — |
| RF [ | — | — | — | — | 79.57 | — | — | |
| Sawhney et al. (2018) | BFA and RF [ | — | — | 83.50 | — | — | — | — |
| Abdar et al. (2017) | Boosted C5.0 [ | — | — | — | — | — | — | 93.75 |
| CHAID [ | — | — | — | — | — | — | 65.0 | |
| Zamani et al. (2016) | WOA with | — | 77.05 | — | 87.10 | — | — | — |
| Abdar (2015) | SVM with rapid miner [ | — | — | — | — | — | — | 72.54 |
| C5.0 with IBM SPSS modeller [ | — | — | — | — | — | — | 87.91 | |
| Santos et al. (2015) | Neural networks and augmented set approach [ | — | — | 75.2 | — | — | — | — |
| Chiu et al. (2013) | ANN and LR [ | — | — | 85.10 | — | — | — | — |
| Mauricio et al. (2013) | ABCO with SVM [ | — | 84.81 | — | 87.10 | — | 83.17 | — |
| Proposed | CSO, KH, BFO, and super learner | 96.83 | 86.36 | 94.74 | 90.48 | 81.82 | 84.00 | 70.00 |