| Literature DB >> 32855845 |
Vincent S Tseng1,2, Ching-Long Chen3, Chang-Min Liang3, Ming-Cheng Tai3, Jung-Tzu Liu4, Po-Yi Wu4, Ming-Shan Deng4, Ya-Wen Lee4, Teng-Yi Huang4, Yi-Hao Chen3.
Abstract
Purpose: To improve disease severity classification from fundus images using a hybrid architecture with symptom awareness for diabetic retinopathy (DR).Entities:
Keywords: convolutional neural network; diabetic retinopathy; fundus image; fusion architecture; object detection
Mesh:
Year: 2020 PMID: 32855845 PMCID: PMC7424907 DOI: 10.1167/tvst.9.2.41
Source DB: PubMed Journal: Transl Vis Sci Technol ISSN: 2164-2591 Impact factor: 3.283
Figure 1.Workflow diagram showing distribution of DR severity level and the incidence rate of DR lesions in a different data set.
Figure 2.Lesion location GT production process. (a) Lesion annotated by two ophthalmologists (D1 and D2). (b) Rule-based combination results.
Figure 3.Distribution of DR lesion types by DR severity level.
Figure 4.Workflow of the baseline model (M0) and four fusion models (M1–M4).
Figure 5.Input images: (a) raw image and (b) enhanced image with highlighted lesion locations (MA, H, and SE).
Figure 6.Strategies of the potential DR lesions extraction. Blue dots: H; yellow dots: HE or SE; red dots: MA.
Figure 7.Closeup of MA. Example of IoU smaller than 0.15. The larger bounding box is produced by GT; the smaller bounding box is produced by a prediction model.
Performance of Binary Lesion Type-Classification Model at the Image Level
| Lesion Type | Accuracy (%) | AUC (%) | Sensitivity (%) | Specificity (%) |
|---|---|---|---|---|
| MA | 77.04 | 81.32 | 69.90 | 80.14 |
| H | 87.08 | 90.06 | 79.14 | 90.23 |
| HE | 79.59 | 82.26 | 64.57 | 83.95 |
| SE | 81.79 | 87.90 | 77.44 | 82.09 |
Data are means.
Performance Comparison of the Baseline Model (M0) and the Proposed Fusion Models (M1–M4)
| Five-Class | Two-Class | |||||
|---|---|---|---|---|---|---|
| Model | Accuracy | Weighted κ | Accuracy | AUC | Sensitivity | Specificity |
| M0 | 81.60 | 80.09 | 92.12 | 94.19 | 80.98 | 94.92 |
| M1 | 81.69 | 81.19 | 92.22 | 95.08 | 82.20 | 94.73 |
| M2 | 84.24 | 83.86 | 92.27 | 95.06 | 90.49 | 92.71 |
| M3 | 85.12 | 84.43 | 91.09 | 94.21 | 90.98 | 91.12 |
| M4 | 84.29 | 84.01 | 92.95 | 95.51 | 86.83 | 94.49 |
Performance Comparison on Messidor-2 in Detecting Referable DR
| Training Data | Training with Mydriatic/ | Accuracy | AUC | Sensitivity | Specificity | |||
|---|---|---|---|---|---|---|---|---|
| Benchmark | Year | Set | Nonmydriatic Imaging | Approach | (%) | (%) | (%) | (%) |
| Abràmoff et al. | 2016 | 1,250,000 images (EyeCheck project and the University of Iowa) | Unknown | CNN | — | 98.0 | 96.8 | 87.0 |
| Gulshan et al. | 2016 | 128,175 images (EyePACS) | Mixed | Inception-v3 CNN | — | 99.0 | 87 | 98.5 |
| Gargeya and Leng | 2017 | 75,137 images (EyePACS) | Mixed | CNN + gradient boosting classifier | — | 94 | 93 | 87 |
| Pires et al. | 2019 | 35,126 images (part of Kaggle data set) | Mixed | Similar to VGG-16 | — | 98.2 | — | — |
| Voets et al. | 2019 | 88,702 images (part of Kaggle data set) | Mixed | Inception-v3 CNN | 84.21 | 85.30 | 68.70 | 88.50 |
| Li et al. | 2019 | 19,233 images (Chinese hospitals) | Unknown | Inception-v3 CNN | 93.49 | 99.05 | 96.93 | 93.45 |
| Zago et al. | 2020 | 28 images with 262,144 patches per image (DiaretDB1) | Mydriatic | Patch-based CNN | — | 94.4 | 90.0 | 87.0 |
| Proposed model M4 | 2020 | 22,617 images (Taiwanese hospitals) | Mixed | Fusion CNN architecture | 91.99 | 97.09 | 93.68 | 91.52 |
Any DR.
Only 800 images were selected from Messidor-2 to create the testing set.
Figure 8.(a) Raw images. (b) Enhanced images.