| Literature DB >> 35684587 |
Ivanoe De Falco1, Giuseppe De Pietro1, Giovanna Sannino1.
Abstract
The classification of images is of high importance in medicine. In this sense, Deep learning methodologies show excellent performance with regard to accuracy. The drawback of these methodologies is the fact that they are black boxes, so no explanation is given to users on the reasons underlying their choices. In the medical domain, this lack of transparency and information, typical of black box models, brings practitioners to raise concerns, and the result is a resistance to the use of deep learning tools. In order to overcome this problem, a different Machine Learning approach to image classification is used here that is based on interpretability concepts thanks to the use of an evolutionary algorithm. It relies on the application of two steps in succession. The first receives a set of images in the inut and performs image filtering on them so that a numerical data set is generated. The second is a classifier, the kernel of which is an evolutionary algorithm. This latter, at the same time, classifies and automatically extracts explicit knowledge as a set of IF-THEN rules. This method is investigated with respect to a data set of MRI brain imagery referring to Alzheimer's disease. Namely, a two-class data set (non-demented and moderate demented) and a three-class data set (non-demented, mild demented, and moderate demented) are extracted. The methodology shows good results in terms of accuracy (100% for the best run over the two-class problem and 91.49% for the best run over the three-class one), F_score (1.0000 and 0.9149, respectively), and Matthews Correlation Coefficient (1.0000 and 0.8763, respectively). To ascertain the quality of these results, they are contrasted against those from a wide set of well-known classifiers. The outcome of this comparison is that, in both problems, the methodology achieves the best results in terms of accuracy and F_score, whereas, for the Matthews Correlation Coefficient, it has the best result over the two-class problem and the second over the three-class one.Entities:
Keywords: Alzheimer’s disease; classification; evolutionary algorithm; interpretable machine learning; magnetic resonance imagery
Mesh:
Year: 2022 PMID: 35684587 PMCID: PMC9183018 DOI: 10.3390/s22113966
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1The approach. The left part of the image contains the filter (1st step), whereas the right part contains the classifier (2nd step) based on an Evolutionary Algorithm (EA). This latter also extracts explicit knowledge. MRI images contained in this figure are taken from the original Alzheimer’s data set [53] available on Kaggle.
Figure 2Example of items from the three classes. Left pane: non-demented. Center pane: mild demented. Right pane: moderate demented. MRI images contained in this figure are taken from the original Alzheimer’s data set [53] available on Kaggle.
The encoding with the 29 attributes and their positions.
|
|
|
|---|---|
| 1 | color correlogram: black bin |
| 2 to 8 | color correlogram: gray bins |
| 9 | color correlogram: white bin |
| 10–16 | mean values of the seven texture moment attributes |
| 17–23 | variation values of the seven texture moment attributes |
| 24 | first-order R color moment |
| 25 | first-order G color moment |
| 26 | first-order B color moment |
| 27 | second-order R color moment |
| 28 | second-order G color moment |
| 29 | second-order B color moment |
The Parameter Setting for DEREx for the 2-Class Problem.
|
|
|
|---|---|
| Pop_Size | 30 |
| Max_Gens | 500 |
| Cr_Ratio | 0.3 |
| Mut_F | 0.75 |
| DE_Algo | DE/rand-to-best/1/bin |
| N_Max_Rules | 2 |
| Rule_Thr | 0.0 |
| Lit_Thr | 0.90 |
| class 1 |
|
| class 2 |
|
Confusion Matrices of the Best Rule Set (2-Class Problem).
|
|
|
| ||||
|---|---|---|---|---|---|---|
|
|
|
| ||||
|
|
|
|
|
|
|
|
|
| 37 | 0 | 15 | 0 | 52 | 0 |
|
| 0 | 47 | 0 | 21 | 0 | 68 |
The classification algorithms contained in WEKA and used in this paper.
| Class | Algorithm | Acronym | Reference |
|---|---|---|---|
|
| Bayes Net | BN | [ |
| Naive Bayes | NB | [ | |
|
| MultiLayer Perceptron | MLP | [ |
| Radial Basis Function | RBF | [ | |
| Support Vector Machine | SVM | [ | |
|
| AdaBoost | AB | [ |
| Bagging | Bag | [ | |
|
| One Rule | OneR | [ |
| Repeated Incremental Pruning (JRip) | JRip | [ | |
| Partial Decision Tree (PART) | PART | [ | |
| Ripple-Down Rule | Ridor | [ | |
|
| C4.5 decision tree (J48) | J48 | [ |
| Random Forest | RF | [ | |
| REPTree | RT | [ |
The numerical results for the 2-Class Problem.
|
|
|
| ||||
|---|---|---|---|---|---|---|
| Average | Best | Average | Best | Average | Best | |
| Bayes Net | 97.22 % | 97.22% | 0.9721 | 0.9721 | 0.9439 | 0.9439 |
| Naive Bayes | 97.22% | 97.22% | 0.9721 | 0.9721 | 0.9439 | 0.9439 |
| MLP | 97.22% | 97.22% | 0.9724 | 0.9724 | 0.9439 | 0.9439 |
| RBF |
|
|
|
|
|
|
| SVM |
|
|
|
|
|
|
| Adaboost |
|
|
|
|
|
|
| Bagging | 94.44% | 94.44% | 0.9448 | 0.9448 | 0.8935 | 0.8935 |
| JRip | 94.44% | 94.44% | 0.9448 | 0.9448 | 0.8935 | 0.8935 |
| OneR | 94.44% | 94.44% | 0.9438 | 0.9438 | 0.8896 | 0.8896 |
| PART |
|
|
|
|
|
|
| Ridor | 88.89% | 88.89% | 0.8896 | 0.8896 | 0.7994 | 0.7994 |
| J48 |
|
|
|
|
|
|
| Random Forest |
|
|
|
|
|
|
| REPTree | 88.89% | 88.89% | 0.8896 | 0.8896 | 0.7994 | 0.7994 |
| DEREx | 96.89% |
| 0.9683 |
| 0.9385 |
|
The Parameter Setting for DEREx for the 3-Class Problem.
|
|
|
|---|---|
| Pop_Size | 50 |
| Max_Gens | 5000 |
| Cr_Ratio | 0.3 |
| Mut_F | 0.75 |
| DE_Algo | DE/rand-to-best/1/bin |
| N_Max_Rules | 3 |
| Rule_Thr | 0.00 |
| Lit_Thr | 0.90 |
| class 1 |
|
| class 2 |
|
| class 3 |
|
Confusion Matrices of the Best Rule Set (3-Class Problem).
|
|
|
| |||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
|
| |||||||
|
|
|
|
|
|
|
|
|
|
|
|
| 35 | 0 | 0 | 17 | 0 | 0 | 52 | 0 | 0 |
|
| 2 | 38 | 0 | 1 | 11 | 0 | 3 | 49 | 0 |
|
| 8 | 0 | 26 | 2 | 1 | 15 | 10 | 1 | 41 |
The numerical results for the 3-Class Problem.
|
|
|
| ||||
|---|---|---|---|---|---|---|
| Average | Best | Average | Best | Average | Best | |
| Bayes Net | 85.11 % | 85.11% | 0.8491 | 0.8491 | 0.7748 | 0.7748 |
| Naive Bayes | 74.47% | 74.47% | 0.7351 | 0.7351 | 0.6080 | 0.6080 |
| MLP |
|
|
|
|
|
|
| RBF | 76.60% | 76.60% | 0.7604 | 0.7604 | 0.6382 | 0.6382 |
| SVM | 87.23% | 87.23% | 0.8722 | 0.8722 | 0.8023 | 0.8023 |
| Adaboost | 85.11% | 85.11% | 0.8508 | 0.8508 | 0.7718 | 0.7718 |
| Bagging | 89.36% | 89.36% | 0.8936 | 0.8936 | 0.8312 | 0.8312 |
| JRip | 89.36% | 89.36% | 0.8950 | 0.8950 | 0.8389 | 0.8389 |
| OneR | 72.34% | 72.34% | 0.7021 | 0.7021 | 0.6003 | 0.6003 |
| PART | 82.98% | 82.98% | 0.8298 | 0.8298 | 0.7346 | 0.7346 |
| Ridor | 89.36% | 89.36% | 0.8928 | 0.8928 | 0.8384 | 0.8384 |
| J48 | 82.98% | 82.98% | 0.8261 | 0.8261 | 0.7350 | 0.7350 |
| Random Forest |
|
| 0.9137 | 0.9137 | 0.8682 | 0.8682 |
| REPTree | 87.23% | 87.23% | 0.8738 | 0.8738 | 0.8086 | 0.8086 |
| DEREx | 86.21% |
| 0.8675 |
| 0.7902 | 0.8763 |
Size of information extracted by the classifiers performing interpretable machine learning.
| Two-Class | Three-Class | ||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
| |
| JRip | 3 | 2 | 4 | 5 | 3.50 | 3.50 | 1.00 |
| OneR | 1 | 2 | 1 | 7 | 1.00 | 4.50 | 4.50 |
| PART | 3 | 2 | 8 | 16 | 5.50 | 9.00 | 1.64 |
| Ridor | 3 | 3 | 3 | 3 | 3.00 | 3.00 | 1.00 |
| J48 | 3 | 4 | 14 | 27 | 8.50 | 15.50 | 1.82 |
| Random Forest | 4 | 6 | 13 | 25 | 8.50 | 15.50 | 1.82 |
| RepTree | 2 | 2 | 4 | 9 | 3.00 | 5.50 | 1.83 |
| DEREx | 2 | 8 | 3 | 7 | 2.50 | 7.50 | 3.00 |