| Literature DB >> 36010325 |
Sabiha Gungor Kobat1, Nursena Baygin2, Elif Yusufoglu3, Mehmet Baygin4, Prabal Datta Barua5,6, Sengul Dogan7, Orhan Yaman7, Ulku Celiker1, Hakan Yildirim1, Ru-San Tan8,9, Turker Tuncer7, Nazrul Islam10, U Rajendra Acharya11,12,13.
Abstract
Diabetic retinopathy (DR) is a common complication of diabetes that can lead to progressive vision loss. Regular surveillance with fundal photography, early diagnosis, and prompt intervention are paramount to reducing the incidence of DR-induced vision loss. However, manual interpretation of fundal photographs is subject to human error. In this study, a new method based on horizontal and vertical patch division was proposed for the automated classification of DR images on fundal photographs. The novel sides of this study are given as follows. We proposed a new non-fixed-size patch division model to obtain high classification results and collected a new fundus image dataset. Moreover, two datasets are used to test the model: a newly collected three-class (normal, non-proliferative DR, and proliferative DR) dataset comprising 2355 DR images and the established open-access five-class Asia Pacific Tele-Ophthalmology Society (APTOS) 2019 dataset comprising 3662 images. Two analysis scenarios, Case 1 and Case 2, with three (normal, non-proliferative DR, and proliferative DR) and five classes (normal, mild DR, moderate DR, severe DR, and proliferative DR), respectively, were derived from the APTOS 2019 dataset. These datasets and these cases have been used to demonstrate the general classification performance of our proposal. By applying transfer learning, the last fully connected and global average pooling layers of the DenseNet201 architecture were used to extract deep features from input DR images and each of the eight subdivided horizontal and vertical patches. The most discriminative features are then selected using neighborhood component analysis. These were fed as input to a standard shallow cubic support vector machine for classification. Our new DR dataset obtained 94.06% and 91.55% accuracy values for three-class classification with 80:20 hold-out validation and 10-fold cross-validation, respectively. As can be seen from steps of the proposed model, a new patch-based deep-feature engineering model has been proposed. The proposed deep-feature engineering model is a cognitive model, since it uses efficient methods in each phase. Similar excellent results were seen for three-class classification with the Case 1 dataset. In addition, the model attained 87.43% and 84.90% five-class classification accuracy rates using 80:20 hold-out validation and 10-fold cross-validation, respectively, on the Case 2 dataset, which outperformed prior DR classification studies based on the five-class APTOS 2019 dataset. Our model attained about >2% classification results compared to others. These findings demonstrate the accuracy and robustness of the proposed model for classification of DR images.Entities:
Keywords: deep feature extraction; diabetic retinopathy; neighborhood component analysis; patch division; support vector machine; transfer learning
Year: 2022 PMID: 36010325 PMCID: PMC9406859 DOI: 10.3390/diagnostics12081975
Source DB: PubMed Journal: Diagnostics (Basel) ISSN: 2075-4418
Summary of a nonsystematic review of recent studies on machine-learning-enabled automated classification of diabetic retinopathy using fundal photography.
| Study | Method | Dataset | Classes (Validation) | Results (%) |
|---|---|---|---|---|
| Arunkumar and Karthigaikumar 2017 [ | DBN-based feature extraction, SVM | ARIA [ | Normal; DR, age-related macular degeneration | Acc 96.73 |
| Abbas et al., 2017 [ | GLOH, principal component analysis, deep neural network | DIARETDB1 [ | Normal, mild NPDR, moderate NPDR, severe NPDR, PDR | AUC 92.4 |
| Krause et al., 2018 [ | Custom CNN | Own dataset, Messidor-2 [ | Not available | AUC 98.6 |
| Chetoui et al., 2018 [ | LTP, LESH, and SVM | Messidor [ | DR, non-DR | Acc 90.04 |
| Orlando et al., 2018 [ | CNN and handcrafted feature extraction, random forest | Messidor [ | Lesion detection | AUC 93.47 |
| Zeng et al., 2019 [ | InceptionV3-based CNN | Kaggle [ | DR, non-DR | AUC 95.1 |
| Ali et al., 2020 [ | Texture analysis (histogram, wavelet, co-occurrence, run-length matrix), logistic model tree | Own dataset | Normal, mild NPDR, moderate NPDR, severe NPDR, PDR | Acc 99.73 |
| Gayathri et al., 2021 [ | Multipath CNN, ResNet-50, and VGG-16-based feature extraction, classifiers (SVM, random forest, J48) | 3 datasets: | IDRiD: Normal, mild NPDR, moderate NPDR, severe NPDR, PDR; | Overall Acc 99.62 |
| Mahmoud et al., 2021 [ | HIMLA (preprocessing, segmentation, feature extraction, and classification) | Chase_DB1 [ | DR, non-DR | Acc 96.62 |
| Math and Fatima 2021 [ | Custom CNN | Kaggle [ | Normal, mild, moderate, NPDR, PDR | AUC 96.3 |
Definitions in the Table 1: Acc, accuracy; AUC, area-under-curve; CNN, convolutional neural network; CV, cross-validation; DBN, deep belief neural network; DR, diabetic retinopathy; GLOH, gradient location orientation histogram; HIMLA, hybrid inductive machine learning algorithm; LESH, local energy-based shape histogram; LTP, local ternary pattern; NPDR, non-proliferative diabetic retinopathy; PDR, proliferative diabetic retinopathy; Sen, sensitivity; Spe, specificity; SVM, support vector machine.
Figure 1Sample images from the new diabetic retinopathy dataset: (a) normal, (b) NPDR, (c) PDR.
Figure 2Sample images from APTOS 2019 dataset: (a) normal, (b) mild DR, (c) moderate DR, (d) severe DR, (e) PDR.
Figure 3Block diagram of the proposed patch-based model. Each input image is first resized to 256 × 256, then divided into four equal nonoverlapping quadrants. Each quadrant is further subdivided into horizontal or vertical patches. Pretrained DenseNet201 is used to generate 2920 features from the main image (M1), as well as each of the horizontal (P1, P2, P7, and P8) and vertical (P3, P4, P5, and P6) patches, which are all concatenated to form a final feature vector of length 26,280 (= 9 × 2920). Neighborhood component analysis (NCA) is used to select the top 500 features, which are fed to a support vector machine (SVM) for classification. APTOS, Asia Pacific Tele-Ophthalmology Society; DR, diabetic retinopathy; NPDR, non-proliferative diabetic retinopathy; PDR, proliferative diabetic retinopathy.
Figure 4Block diagram of the feature extraction process. The resized main image (MI) is divided into four quadrants. Depending on its relative position in the MI, the quadrants are divided into horizontal (P1, P2, P7, and P8) and vertical (P3, P4, P5, and P6) patches.
Figure 5Confusion matrixes for our new dataset obtained using 10-fold cross-validation (a) and 80:20 hold-out validation (b). Class names: 1, normal; 2, non-proliferative diabetic retinopathy; 3, proliferative diabetic retinopathy.
Overall three-class classification performance on our new diabetic retinopathy dataset.
| Results (%) | ||
|---|---|---|
| Performance Metric | 10-Fold CV | 80:20 Hold-Out Validation |
| Accuracy | 91.55 | 94.06 |
| Unweighted average recall | 92.67 | 94.45 |
| Unweighted average precision | 92.77 | 94.74 |
| Average F1 | 92.70 | 94.59 |
| Cohen’s kappa | 86.34 | 90.38 |
| Geometric mean | 92.61 | 94.45 |
Figure 6Confusion matrixes for our new dataset obtained using 10-fold cross-validation (a) and 80:20 hold-out validation (b). Class names: 1, normal; 2, non-proliferative diabetic retinopathy; 3, proliferative diabetic retinopathy.
Overall three-class classification performance on Case 1 of the APTOS dataset.
| Results (%) | ||
|---|---|---|
| Performance Metric | 10-Fold CV | 80:20 Hold-Out Validation |
| Accuracy | 92.60 | 93.85 |
| Unweighted average recall | 79.64 | 80.60 |
| Unweighted average precision | 86.63 | 90.90 |
| Average F1 | 82.06 | 83.78 |
| Cohen’s kappa | 86.74 | 88.94 |
| Geometric mean | 75.28 | 76.04 |
Figure 7Confusion matrixes for our new dataset obtained using 10-fold cross-validation (a) and 80:20 hold-out validation (b). Class names: 1, normal; 2, mild diabetic retinopathy; 3, moderate diabetic retinopathy; 4, severe diabetic retinopathy; 5, proliferative diabetic retinopathy.
Overall five-class classification performance on Case 2 of the APTOS dataset.
| Results (%) | ||
|---|---|---|
| Performance Metric | 10-Fold CV | 80:20 Hold-Out Validation |
| Accuracy | 84.90 | 85.93 |
| Unweighted average recall | 68.53 | 69.72 |
| Unweighted average precision | 74.32 | 77.11 |
| Average F1 | 70.75 | 72.51 |
| Cohen’s kappa | 76.91 | 78.37 |
| Geometric mean | 65.25 | 66.61 |
Figure 8Misclassified samples from our new DR dataset, Case 1 and Case 2 in the top, middle, and bottom rows, respectively.
Figure 9Comparison of the performance of six pretrained deep feature extractors on the ImageNet1k dataset [38]. Neighborhood component analysis was used for each run to select the 1000 most discriminative features, which were then fed to the support vector machine for classification. DenseNet201 yielded the best classification accuracy and was chosen as the deep-feature extractor in our proposed model.
Figure 10Comparison of classification accuracy by the number of top features selected by neighborhood component analysis (NCA) on the new diabetic retinopathy dataset. For each run, a support vector machine was used for classification. The top 500 NCA-selected features yielded the best classification accuracy, and the number was chosen for our proposed model.
Figure 11Classification accuracies calculated extractors on our new diabetic retinopathy dataset using six classifiers: fine tree (FT), linear discriminant (LD), Gaussian naïve Bayes (GNB), cubic support vector machine (CSVM), fine k-nearest neighbor (FKNN), and medium neural network (MNN). Neighborhood component analysis was used for each run to select the 500 most discriminative features, which were then fed to respective classifiers for classification. Cubic SVM yielded the best classification accuracy and was chosen as the classifier in our proposed model.
Figure 12Classification accuracies using Chi2, ReliefF, PCA, and NCA feature selectors on the APTOS dataset.
Comparison of model performance model on our new DR dataset and Case 1 dataset.
| Automated Diabetic Retinopathy Detection Model | ||||
|---|---|---|---|---|
| Our New DR Dataset | Case 1 Created Using APTOS 2019 | |||
| Dataset Information | 3 Class: PDR/NPDR/Normal | 3 Class: Normal/NPDR/PDR | ||
| Performance Metric | 10-Fold CV | 80:20 Hold-Out | 10-Fold CV | 80:20 Hold-Out |
| Accuracy | 91.55 | 94.06 | 92.60 | 93.85 |
| Unweighted average recall | 92.67 | 94.45 | 79.64 | 80.60 |
| Unweighted average precision | 92.77 | 94.74 | 86.63 | 90.90 |
| Average F1 | 92.70 | 94.59 | 82.06 | 83.78 |
| Cohen’s kappa | 86.34 | 90.38 | 86.74 | 88.94 |
| Geometric mean | 92.61 | 94.45 | 75.28 | 76.04 |
Comparison of model performance on Case 2 dataset with prior studies using the APTOS 2019 dataset.
| Author(s) | Method | Key Points | Results (%) |
|---|---|---|---|
| Majumder and Kehtarnavaz 2021 [ | Modified |
90:10 hold-out validation Data augmentation End-to-end learning | Acc 81.0 |
| Podapati et al., 2020 [ | Feature extraction using VGG16′s fc1 and fc2 layers and XCeption’s global average pooling layers, deep neural network |
80:20 hold-out validation Training using combined features | Acc 80.96 |
| Kassani et al., 2019 [ | Xception CNN architecture |
70:20:10 hold-out validation Min-pooling and normalization-based image pre-processing | Acc 83.09 |
| Taufiqurrahman et al., 2020 [ | MobileNetV2 CNN architecture, SVM |
10-fold cross validation Data augmentation | Acc 79.0 |
| Gangwar and Ravi 2021 [ | Inception and |
75:25 hold-out validation Data augmentation End-to-end learning | Acc 82.18 |
| Our model, Case 2 | Feature extraction with DenseNet, feature selection with neighborhood component analysis, cubic SVM classifier |
10-fold cross validation 80:20 hold-out validation No data augmentation Validation with three datasets (our new dataset, Case 1, and Case 2 derived from APTOS 2019 dataset) | 10-fold CV |
| Acc 84.90 | |||
| 80:20 hold-out | |||
| Acc 85.93 |
Acc, accuracy; CNN, convolutional neural network; F1, F1-score; GM, geometric mean; Kap, Cohen’s kappa; Pre, precision; Rec, recall; Sen, sensitivity; Spe, specificity; SVM, support vector machine; UAP, unweighted average precision; UAR, unweighted average recall.