| Literature DB >> 35979449 |
Musatafa Abbas Abbood Albadr1, Masri Ayob1, Sabrina Tiun1, Fahad Taha Al-Dhief2, Mohammad Kamrul Hasan3.
Abstract
Many works have employed Machine Learning (ML) techniques in the detection of Diabetic Retinopathy (DR), a disease that affects the human eye. However, the accuracy of most DR detection methods still need improvement. Gray Wolf Optimization-Extreme Learning Machine (GWO-ELM) is one of the most popular ML algorithms, and can be considered as an accurate algorithm in the process of classification, but has not been used in solving DR detection. Therefore, this work aims to apply the GWO-ELM classifier and employ one of the most popular features extractions, Histogram of Oriented Gradients-Principal Component Analysis (HOG-PCA), to increase the accuracy of DR detection system. Although the HOG-PCA has been tested in many image processing domains including medical domains, it has not yet been tested in DR. The GWO-ELM can prevent overfitting, solve multi and binary classifications problems, and it performs like a kernel-based Support Vector Machine with a Neural Network structure, whilst the HOG-PCA has the ability to extract the most relevant features with low dimensionality. Therefore, the combination of the GWO-ELM classifier and HOG-PCA features might produce an effective technique for DR classification and features extraction. The proposed GWO-ELM is evaluated based on two different datasets, namely APTOS-2019 and Indian Diabetic Retinopathy Image Dataset (IDRiD), in both binary and multi-class classification. The experiment results have shown an excellent performance of the proposed GWO-ELM model where it achieved an accuracy of 96.21% for multi-class and 99.47% for binary using APTOS-2019 dataset as well as 96.15% for multi-class and 99.04% for binary using IDRiD dataset. This demonstrates that the combination of the GWO-ELM and HOG-PCA is an effective classifier for detecting DR and might be applicable in solving other image data types.Entities:
Keywords: Diabetic Retinopathy; Histogram of Oriented Gradients; Principal Component Analysis; extreme learning machine; gray wolf optimization
Mesh:
Year: 2022 PMID: 35979449 PMCID: PMC9376263 DOI: 10.3389/fpubh.2022.925901
Source DB: PubMed Journal: Front Public Health ISSN: 2296-2565
Illustrates the previous works of DR detection using ML and deep learning techniques.
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| Sridhar et al. ( | Kaggle dataset | Binary classification | CNN | 86% Accuracy | • The proposed system tested based on binary classification only and ignored the multi-class classification. |
| Gangwar and Ravi. ( | APTOS-19 and Messidor-1 | Multi-class classification | Hybrid CNN | 72.33% Accuracy on the Messidor-1 dataset and 82.18% accuracy on the APTOS-19 dataset. | • The evaluation of both systems considered only the multi-class classification and ignored the binary classification. |
| Reddy et al. ( | Messidor | Multi-class classification | SVM | 69.09% Accuracy | |
| Asha and Karpagavalli. ( | DIARETDB1 | Binary classification | ELM | 90% Accuracy | • The proposed system tested based on binary classification only and ignored the multi-class classification. |
| Zhang and An ( | Messidor | Binary classification | KELM | 88.60% Accuracy | |
| Punithavathi and Kumar ( | DIARETDB0 | Multi-class classification | ELM | 95.40% Accuracy | • The evaluation of both systems considered only the multi-class classification and ignored the binary classification. |
| Deepa et al. ( | 4 classes dataset | Multi-class classification | KELM | 93.20% |
Figure 1Block diagram of the proposed DR detection approach.
The description of the APTOS-2019 dataset.
|
|
|
|
| ||
|---|---|---|---|---|---|
|
| |||||
| No DR | 190 |
|
|
| 1 |
| Mild | 190 |
|
|
| 2 |
| Moderate | 190 |
|
|
| 3 |
| Severe | 190 |
|
|
| 4 |
| Proliferative | 190 |
|
|
| 5 |
The description of the IDRiD dataset.
|
|
|
|
| ||
|---|---|---|---|---|---|
|
| |||||
| No DR | 168 |
|
|
| 1 |
| Mild | 25 |
|
|
| 2 |
| Moderate | 168 |
|
|
| 3 |
| Severe | 93 |
|
|
| 4 |
| Proliferative | 62 |
|
|
| 5 |
Figure 2The pre-processing steps.
Figure 3Steps of the features extraction.
Elaborate the features extraction step dimensionality for single image and whole dataset images.
|
| ||
|---|---|---|
|
|
|
|
| First Step: HOG Features | (1 x 32400) | (950 x 23400) |
| Second Step: HOG-PCA Features | (1 x 949) | (950 x 949) |
|
| ||
|
|
|
|
| First Step: HOG Features | (1 x 32400) | (516 x 32400) |
| Second Step: HOG-PCA Features | (1 x 515) | (516 x 515) |
Figure 4Flowchart of the GWO algorithm.
Figure 5GWO-ELM algorithm flowchart.
The parameters settings for the ELM and GWO.
|
|
| ||
|---|---|---|---|
|
|
|
|
|
|
| assemble of the biases and input weights | Population (wolves or search agents) | Consists of the position of all search agents |
| ρ | Output-weights matrix | Position | Start stochastically generated within the range of [-1, 1] for the input-weights and [0, 1] for the biases |
| Input-weights ( | −1 to 1 | Population size or number of search agents (NSA) | 50 |
| Bias values ( | 0 to 1 | r1 and r2 | Stochastically generated with the range of [0, 1] |
| Input-nodes number ( | Input attributes | Number of iterations it | 100 |
| Hidden-nodes number ( | [100–300]; with a 25 increment step | C1, C2, and C3 | Randomly generated vectors based on Equation (4) |
| Output neurones number ( | Number of classes | A1, A2, and A3 | Randomly generated vectors using Equation (3) |
| Activation function | Sigmoid | Xα | Best position of all search agents. |
The highest experiment outcomes of the binary and multi-class classifications for GWO-ELM approach using APTOS-2019 and IDRiD datasets.
|
| |||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| 5 | 96.21 | 90.53 | 90.53 | 97.63 | 88.16 | 90.53 | 90.53 |
| 2 | 99.47 | 99.34 | 100.00 | 97.44 | 98.38 | 99.67 | 99.67 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| 5 | 96.15 | 90.38 | 90.38 | 97.60 | 87.98 | 90.38 | 90.38 |
| 2 | 99.04 | 100.00 | 98.59 | 100.00 | 97.82 | 99.29 | 99.29 |
Figure 6The confusion matrix of the highest multi-class classification outcome for the GWO-ELM approach using the APTOS-2019 dataset.
Figure 10The ROC of the highest binary classification outcome for the GWO-ELM approach using the APTOS-2019 dataset.
Figure 11The ROC of the highest binary classification outcome for the GWO-ELM approach using the IDRiD dataset.
The highest experiment outcomes of the binary and multi-class classifications for ELM approach using APTOS-2019 and IDRiD datasets.
|
| |||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| 5 | 80.21 | 50.53 | 50.53 | 87.63 | 38.16 | 50.53 | 50.53 |
| 2 | 92.63 | 93.42 | 93.42 | 77.27 | 78.60 | 95.30 | 95.32 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| 5 | 74.62 | 36.54 | 36.54 | 84.13 | 20.67 | 36.54 | 36.54 |
| 2 | 72.12 | 85.71 | 75.95 | 60.00 | 32.75 | 80.54 | 80.68 |
The highest experiments outcomes of the classification and detection for NN approach using APTOS-2019 and IDRiD datasets.
|
| |||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| 5 | 78.53 | 46.32 | 46.32 | 86.58 | 32.89 | 46.32 | 46.32 |
| 2 | 90.53 | 98.68 | 90.36 | 91.67 | 68.13 | 94.34 | 94.43 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| 5 | 72.31 | 30.77 | 30.77 | 82.69 | 13.46 | 30.77 | 30.77 |
| 2 | 71.15 | 97.14 | 70.83 | 75.00 | 26.04 | 81.93 | 82.95 |
The experiments outcomes of the binary and multi-class classification for SVM (linear kernel), SVM (precomputed kernel), and RF approaches using APTOS-2019 and IDRiD datasets.
|
| |||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| SVM (linear) | 79.58 | 48.95 | 48.95 | 87.24 | 36.18 | 48.95 | 48.95 |
| SVM (Precomputed Kernel) | 79.37 | 48.42 | 48.42 | 87.11 | 35.53 | 48.42 | 48.42 |
| RF | 79.37 | 48.42 | 48.42 | 87.11 | 35.53 | 48.42 | 48.42 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| SVM (linear) | 88.95 | 100.00 | 87.86 | 100.00 | 62.69 | 93.54 | 93.73 |
| SVM (Precomputed Kernel) | 88.95 | 100.00 | 87.86 | 100.00 | 62.69 | 93.54 | 93.73 |
| RF | 91.58 | 100.00 | 90.48 | 100.00 | 72.37 | 95.00 | 95.12 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| SVM (linear) | 73.85 | 34.62 | 34.62 | 83.65 | 18.27 | 34.62 | 34.62 |
| SVM (Precomputed Kernel) | 73.08 | 32.69 | 32.69 | 83.17 | 15.87 | 32.69 | 32.69 |
| RF | 74.23 | 35.58 | 35.58 | 83.89 | 19.47 | 35.58 | 35.58 |
|
| |||||||
|
|
|
|
|
|
|
|
|
| SVM (linear) | 68.27 | 98.57 | 68.32 | 66.67 | 12.48 | 80.70 | 82.06 |
| SVM (Precomputed Kernel) | 67.31 | 84.29 | 71.95 | 50.00 | 19.11 | 77.63 | 77.87 |
| RF | 69.23 | 100.00 | 68.63 | 100.00 | 20.09 | 81.40 | 82.84 |
The comparison of accuracy between the proposed GWO-ELM and other previous works.
|
|
| ||
|---|---|---|---|
|
|
|
|
|
| DNN ( | 81.70 | DNN ( | 97.41 |
| Hybrid model ( | 86.34 | DNN ( | 98.00 |
| DNN ( | 82.54 | Hybrid CNN-SVD and ELM ( | 99.32 |
| 77.90 | Ensemble (trimmed mean) ( | 98.60 | |
| MLP ( | 83.09 | ResNet34 ( | 96.35 |
| CNN512 ( | 89.00 | CNN ( | 91.00 |
| Tuned XGBoot ( | 94.20 | RA-EfficientNet ( | 98.36 |
| Proposed GWO-ELM | 96.21 | Proposed GWO-ELM | 99.47 |
|
|
| ||
|
|
|
|
|
| MLP ( | 92.01 | MLP ( | 98.87 |
| ResNet50 + J48 ( | 92.46 | CNN ( | 90.29 |
| XG-Boost ( | 88.20 | Coarse Network ( | 80.00 |
| Lesion(Semi + Adv) ( | 91.34 | HE-CNN ( | 96.76 |
| Proposed GWO-ELM | 96.15 | Proposed GWO-ELM | 99.04 |