| Literature DB >> 33286339 |
Aqib Ali1, Salman Qadri1, Wali Khan Mashwani2, Wiyada Kumam3, Poom Kumam4,5, Samreen Naeem1, Atila Goktas6, Farrukh Jamal7, Christophe Chesneau8, Sania Anam9, Muhammad Sulaiman10.
Abstract
The object of this study was to demonstrate the ability of machine learning (ML) methods for the segmentation and classification of diabetic retinopathy (DR). Two-dimensional (2D) retinal fundus (RF) images were used. The datasets of DR-that is, the mild, moderate, non-proliferative, proliferative, and normal human eye ones-were acquired from 500 patients at Bahawal Victoria Hospital (BVH), Bahawalpur, Pakistan. Five hundred RF datasets (sized 256 × 256) for each DR stage and a total of 2500 (500 × 5) datasets of the five DR stages were acquired. This research introduces the novel clustering-based automated region growing framework. For texture analysis, four types of features-histogram (H), wavelet (W), co-occurrence matrix (COM) and run-length matrix (RLM)-were extracted, and various ML classifiers were employed, achieving 77.67%, 80%, 89.87%, and 96.33% classification accuracies, respectively. To improve classification accuracy, a fused hybrid-feature dataset was generated by applying the data fusion approach. From each image, 245 pieces of hybrid feature data (H, W, COM, and RLM) were observed, while 13 optimized features were selected after applying four different feature selection techniques, namely Fisher, correlation-based feature selection, mutual information, and probability of error plus average correlation. Five ML classifiers named sequential minimal optimization (SMO), logistic (Lg), multi-layer perceptron (MLP), logistic model tree (LMT), and simple logistic (SLg) were deployed on selected optimized features (using 10-fold cross-validation), and they showed considerably high classification accuracies of 98.53%, 99%, 99.66%, 99.73%, and 99.73%, respectively.Entities:
Keywords: classification; clustering; diabetic retinopathy; hybrid features; segmentation
Year: 2020 PMID: 33286339 PMCID: PMC7517087 DOI: 10.3390/e22050567
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Typical retinal fundus images.
Figure 2Clustering-based segmentation and hybrid-feature analysis for diabetic retinopathy (DR) classification framework.
Figure 3Irregular polygonal seed selection proposed novel mechanism.
Figure 4Irregular polygonal seed-based improved region growing outcome using K-mean clustering.
Figure 5DR segmentation based on the watershed technique.
Post-optimize feature selection table (Fisher (F), probability of error (POE) plus average correlation (AC), mutual information (MI), and correlation-based feature selection (CFS)).
|
|
|
|
|
|
| 1. 45dgr_RLNonUni | 6. S(5,0)Entropy | 9. WavEnHL_s-4 | 11. Variance |
Figure 6Data visualization in 3D vector space.
The overall classification accuracy table of the employed machine learning (ML) classifiers on a histogram features-based dataset. ROC: receiver-operating characteristic; TP: true positive; FR: false positive; MAE: mean absolute errors; RMSE: root mean squared error; OA: overall accuracy; MLP: multi-layer perceptron; SLg: simple logistic; LMT: logistic model tree; Lg: logistic; and SMO: sequential minimal optimization.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| MLP | 0.6717 | 0.737 | 0.066 | 0.916 | 0.1485 | 0.2799 | 0.78 | 73.73% |
| SLg | 0.6633 | 0.731 | 0.067 | 0.921 | 0.1637 | 0.2813 | 0.59 | 73.07% |
| LMT | 0.6625 | 0.730 | 0.068 | 0.919 | 0.1534 | 0.2807 | 0.52 | 73.00% |
| Lg | 0.6617 | 0.729 | 0.068 | 0.923 | 0.1556 | 0.2789 | 0.37 | 72.07% |
| SMO | 0.6075 | 0.686 | 0.09 | 0.878 | 0.2602 | 0.3465 | 0.25 | 68.60% |
Figure 7The overall accuracy graph of the employed ML classifiers on a histogram features-based dataset.
The overall classification accuracy table of the employed ML classifiers on a wavelet features-based dataset.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| MLP | 0.7492 | 0.799 | 0.050 | 0.943 | 0.0894 | 0.2559 | 1.02 | 79.93% |
| LMT | 0.7225 | 0.778 | 0.056 | 0.933 | 0.1072 | 0.2571 | 1.98 | 77.80% |
| Lg | 0.7108 | 0.769 | 0.058 | 0.944 | 0.1299 | 0.2581 | 0.04 | 76.86% |
| SLg | 0.7075 | 0.766 | 0.059 | 0.945 | 0.1362 | 0.2582 | 0.98 | 76.60% |
| SMO | 0.6975 | 0.758 | 0.061 | 0.914 | 0.2533 | 0.3367 | 0.09 | 75.80% |
Figure 8The overall accuracy graph of the employed ML classifiers on a wavelet features-based dataset.
The overall classification accuracy table of the employed ML classifiers on co-occurrence matrix (COM) features-based dataset.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| LMT | 0.8742 | 0.899 | 0.025 | 0.982 | 0.057 | 0.1755 | 0.99 | 89.93% |
| MLP | 0.8733 | 0.899 | 0.025 | 0.891 | 0.046 | 0.1865 | 1.21 | 89.86% |
| SLg | 0.8483 | 0.879 | 0.030 | 0.981 | 0.075 | 0.1886 | 0.43 | 87.86% |
| SMO | 0.835 | 0.868 | 0.033 | 0.950 | 0.247 | 0.3269 | 0.12 | 86.80% |
| Lg | 0.8583 | 0.887 | 0.028 | 0.978 | 0.051 | 0.1896 | 0.09 | 86.67% |
Figure 9The overall accuracy graph of the employed ML classifiers on a COM features-based dataset.
The overall classification accuracy table of the employed ML classifiers on a run-length matrix (RLM) features-based dataset.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| SLg | 0.9492 | 0.959 | 0.010 | 0.996 | 0.0307 | 0.1167 | 0.62 | 95.93% |
| Lg | 0.9475 | 0.958 | 0.011 | 0.996 | 0.0247 | 0.1166 | 0.72 | 95.80% |
| LMT | 0.9467 | 0.957 | 0.011 | 0.994 | 0.0292 | 0.1186 | 0.82 | 95.73% |
| MLP | 0.9392 | 0.951 | 0.012 | 0.994 | 0.0292 | 0.1242 | 0.98 | 95.13% |
| SMO | 0.9208 | 0.937 | 0.016 | 0.978 | 0.243 | 0.3209 | 0.06 | 93.67% |
Figure 10The overall accuracy graph of the employed ML classifiers on an RLM features-based dataset.
Figure 11The overall accuracy graph of the employed ML classifiers on a post-optimized fused dataset.
The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused hybrid-features dataset.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| SLg | 0.9967 | 0.997 | 0.001 | 1.000 | 0.0051 | 0.0367 | 0.32 | 99.73% |
| LMT | 0.9967 | 0.997 | 0.001 | 1.000 | 0.0048 | 0.0353 | 0.58 | 99.73% |
| MLP | 0.9958 | 0.997 | 0.001 | 1.000 | 0.0038 | 0.0323 | 0.33 | 99.67% |
| Lg | 0.9892 | 0.991 | 0.002 | 0.999 | 0.0036 | 0.0566 | 0.42 | 99.13% |
| SMO | 0.9817 | 0.985 | 0.004 | 0.996 | 0.2406 | 0.3172 | 0.21 | 98.53% |
The confusion matrix of post-optimized, fused hybrid-feature dataset for the SLg classifier.
| Classified as | Healthy | Mild | Moderate | Non-Proliferative | Proliferative | Total |
|---|---|---|---|---|---|---|
|
|
| 0 | 0 | 0 | 0 |
|
|
| 0 |
| 0 | 0 | 0 |
|
|
| 1 | 0 |
| 0 | 0 |
|
|
| 0 | 0 | 0 |
| 1 |
|
|
| 0 | 1 | 0 | 1 |
|
|
Figure 12Classification accuracy graph of five DR stages using the SLg classifier on a post-optimized fused dataset.
Figure 13Comparative analysis of DR stage classification accuracies among different type of feature modalities.
The overall classification accuracy table of the employed ML classifiers on a post-optimized, fused public dataset.
| Classifiers | Kappa Statistics | TP Rate | FP Rate | ROC | MAE | RMSE | Time (s) | OA |
|---|---|---|---|---|---|---|---|---|
| SLg | 0.9767 | 0.988 | 0.012 | 1.000 | 0.0191 | 0.0914 | 0.38 | 98.83% |
| LMT | 0.9733 | 0.987 | 0.013 | 0.985 | 0.0165 | 0.1137 | 0.18 | 98.67% |
| MLP | 0.9667 | 0.983 | 0.017 | 0.998 | 0.0268 | 0.1089 | 0.28 | 98.33% |
| Lg | 0.9433 | 0.972 | 0.028 | 0.998 | 0.0313 | 0.1404 | 0.39 | 97.17% |
| SMO | 0.9367 | 0.968 | 0.032 | 0.997 | 0.0302 | 0.1542 | 0.11 | 96.83% |
Figure 14The overall accuracy graph of the employed ML classifiers on post-optimized, fused public dataset.
Figure 15A comparison accuracy graph between Bahawal Victoria Hospital (BVH) and public (RF image) dataset with the SLg classifier.
A comparison table among the proposed and current state-of-the-art techniques.
| Source/Reference | Methodology | Modality | Accuracy |
|---|---|---|---|
| Pires, R. et al. [ | Convolutional Neural Networks | RF Image | 99.0% |
| Zhang, W. et al. [ | Neural Networks | RF Image | 98.1% |
| Harun, N. H. et al. [ | MLP and Artificial Neural Network | RF Image | 72.11% |
| Verbraak, F. D. et al. [ | Hybrid Features and Deep Learning | RF Image | 93.8% |
| Afrin, R. and Shill, P. C. [ | Fused Feature and Fuzzy Logic | RF Image | 95.63% |
| Parmar, R. et al. [ | Neural Networks | RF Image | 85% |
| Xu, K. et al. [ | Neural Networks | RF Image | 94.5% |
| Gulshan, V. et al. [ | Deep Learning | RF Image | 97.5% |
| Gargeya R. [ | Data-Driven Deep Learning Algorithm | RF Image | 94% |
|
| CARGS, Post-Optimized, Fused Hybrid-Features, and Simple Logistic | BVH RF Image Dataset | 99.73% |
|
| CARGS, Post-Optimized, Fused Hybrid-Features, and Simple Logistic | Publicly | 98.83% |