| Literature DB >> 35807135 |
José Camara1,2, Roberto Rezende1,2, Ivan Miguel Pires2,3, António Cunha2,4.
Abstract
Public databases for glaucoma studies contain color images of the retina, emphasizing the optic papilla. These databases are intended for research and standardized automated methodologies such as those using deep learning techniques. These techniques are used to solve complex problems in medical imaging, particularly in the automated screening of glaucomatous disease. The development of deep learning techniques has demonstrated potential for implementing protocols for large-scale glaucoma screening in the population, eliminating possible diagnostic doubts among specialists, and benefiting early treatment to delay the onset of blindness. However, the images are obtained by different cameras, in distinct locations, and from various population groups and are centered on multiple parts of the retina. We can also cite the small number of data, the lack of segmentation of the optic papillae, and the excavation. This work is intended to offer contributions to the structure and presentation of public databases used in the automated screening of glaucomatous papillae, adding relevant information from a medical point of view. The gold standard public databases present images with segmentations of the disc and cupping made by experts and division between training and test groups, serving as a reference for use in deep learning architectures. However, the data offered are not interchangeable. The quality and presentation of images are heterogeneous. Moreover, the databases use different criteria for binary classification with and without glaucoma, do not offer simultaneous pictures of the two eyes, and do not contain elements for early diagnosis.Entities:
Keywords: databases; glaucoma; glaucoma screening; machine learning; retinal images
Year: 2022 PMID: 35807135 PMCID: PMC9267177 DOI: 10.3390/jcm11133850
Source DB: PubMed Journal: J Clin Med ISSN: 2077-0383 Impact factor: 4.964
Figure 1Example of an OCT scan.
Figure 2Illustration of a typical deep learning network with multiple layers between the input and output levels (acquired from [4]).
Comparison of the dataset with other publicly available databases of background images. Question marks indicate missing information, and N/A means “not applicable”. Adapted from [8].
| Dataset | Num. of Images | Ground Truth labels | Different Cameras | Training/ | Diagnosis from | Evaluation Framework | ||||
|---|---|---|---|---|---|---|---|---|---|---|
| g+ | g- | Total | Class of Glaucoma | Segmentation Disc/ | Location of the Fovea? | |||||
| 0 | 143 | 143 | on | yes/no | yes | on | on | ? | on | |
| - | - | 110 | on | yes/no | on | ? | on | N/A | on | |
| 70 | 31 | 101 | yes | yes/no | on | on | yes | images | on | |
| 10 | 29 | 39 | yes | no/no | on | yes | on | clinical | on | |
| 0 | 516 | 516 | on | yes/no | yes | on | yes | ? | yes | |
| 15 | 30 | 45 | yes | no/no | on | on | on | clinical | on | |
| 11 | 11 | 22 | yes | no/no | on | on | on | clinical | on | |
| - | - | 99 | on | yes/no | on | on | on | N/A | on | |
| 168 | 482 | 650 | yes | yes/no | on | ? | on | ? | on | |
| 172 | 313 | 485 | yes | yes/no | on | yes | yes | images | on | |
| - | - | 750 | on | yes/no | on | yes | on | ? | on | |
| 120 | 1080 | 1200 | yes | yes/no | yes | yes | yes | clinical | yes | |
g+ glaucoma, g- normal, N/A “not applicable,” (?) missing information.
Summarizes the characteristics of gold standard public databases.
| Format | Normal and Glaucoma Eyes | Training Group | Test Group | Segmentation | Diagnostic Elements | Simultaneous Imaging OD/OE | ||
|---|---|---|---|---|---|---|---|---|
| Right | Left | |||||||
|
| .png | 313-/172+ | 195-/116+ | 118-/56+ | (+) | (+) | clinical | not |
|
| .png | 31-/70+ | 50 | 51 | (+) | (+) | image | not |
|
| .jpeg | 1080-/120+ | 360-/40+ | 400 offline400 online | (+) | (+) | clinical | not |
* Presence or absence of the notch is analyzed on a DRISHTI-GS1 basis. Normal eyes (-) with glaucoma (+) optic disc (OD), cupping (ESC), segmentation present (+), and absent (-).
Presentation of the RIM-ONE DL database datasets.
| TEST_SET | TRAINING_SET | |||
|---|---|---|---|---|
| Normal | Glaucoma | Normal | Glaucoma | |
|
| 118 | 56 | 195 | 116 |
|
| 52 | 94 | 219 | 120 |
Figure 3The images T0002.jpg (left) with normal excavation and T0010.jpg (right) with enlarged excavation were obtained from the test group at 25% of the original size.
Figure 4Photos n0002.jpg of the group without glaucoma (left) and g0002.jpg of the group with glaucoma (right), both reduced to 25% of the original size.
Figure 5Photo V0002.jpg reduced to 25% of the original size.
Evaluation of different networks using the randomized test set.
| Network | AUC | Se | Acc. |
|---|---|---|---|
|
| 0.9867 | 1.0000 | 0.9315 |
|
| 0.9834 | 0.9615 | 0.9247 |
|
| 0.8771 | 0.9808 | 0.9178 |
|
| 0.9755 | 0.9808 | 0.9110 |
|
| 0.9738 | 0.9423 | 0.9041 |
|
| 0.9726 | 0.9615 | 0.9041 |
|
| 0.9712 | 0.9615 | 0.9315 |
|
| 0.9685 | 0.9808 | 0.9110 |
|
| 0.9597 | 0.9423 | 0.8904 |
|
| 0.9290 | 0.9231 | 0.7534 |
Figure 6ROC curves for all networks using the Madrid and Zaragoza test set.
Evaluation of networks using the Madrid and Zaragoza test suite.
| Network | AUC | Se | Acc. |
|---|---|---|---|
|
| 0.9272 | 0.8750 | 0.8563 |
|
| 0.9177 | 0.8214 | 0.8506 |
|
| 0.9015 | 0.7500 | 0.8046 |
|
| 0.8982 | 0.7500 | 0.7989 |
|
| 0.8919 | 0.7143 | 0.7816 |
|
| 0.8912 | 0.7500 | 0.8276 |
|
| 0.8855 | 0.7321 | 0.8333 |
|
| 0.8396 | 0.625 | 0.7644 |
|
| 0.7969 | 0.6071 | 0.7989 |
|
| 0.7765 | 0.4464 | 0.5287 |
Distribution of normal/glaucomatous eye images and notch cases in the training and test sets—adapted from [39].
| Notching | Diagnosis | |||
|---|---|---|---|---|
| Absent | Present | Glaucomatous | Normal | |
| 31 | 19 | 32 | 18 | Train |
| 25 | 26 | 38 | 13 | Test |
Optical disc segmentation results—adapted from [39].
| Test | Train | ||
|---|---|---|---|
| Boundary Localization Error (Pixels) | F-Score | Boundary Localization Error (Pixels) | F-Score |
| 8.93 ± 2.96 | 0.96 ± 0.02 | 8.61 ± 8.89 | 0.96 ± 0.05 |
The table entries represent means ± standard deviation obtained in images.
Excavation segmentation results—adapted from [39].
| Test | Train | ||
|---|---|---|---|
| Boundary Localization Error (Pixels) | F-Score | Boundary Localization Error (Pixels) | F-Score |
| 30.51 ± 24.80 | 0.77 ± 0.20 | 33.91 ± 25.14 | 0.74 ± 0.20 |
| 25.28 ± 18.00 | 0.79 ± 0.18 | 24.24 ± 16.90 | 0.77 ± 0.17 |
| 21.21 ± 15.09 | 0.81 ± 0.16 | 22.10 ± 19.47 | 0.80 ± 0.18 |
The table entries represent means ± standard deviation obtained in images.
Error in CDR estimation in OD segmentation and excavation (deviation ± mean) evaluated against different experts—adapted from [39].
| Test | Train | |
|---|---|---|
| 0.18 ± 0.14 | 0.15 ± 0.12 | Expert 1 |
| 0.17 ± 0.11 | 0.13 ± 0.10 | Expert 2 |
| 0.13 ± 0.12 | 0.10 ± 0.10 | Expert 3 |
| 0.14 ± 0.12 | 0.11 ± 0.11 | Expert 4 |
| 0.16 ± 0.02 | 0.12 ± 0.02 | Average |
Performance of the proposed method for notch detection—adapted from [39].
|
|
|
| |
|
| 0.81 | 0.84 | 0.71 |
|
| 0.79 | 0.81 | 0.72 |
Ranking results of the participating teams in the REFUGE test set. The last row corresponds to the results obtained using the vCDR.
| Rank | Team | AUC | Reference Sensitivity |
|---|---|---|---|
|
|
|
| 0.9752 |
| 2 | SDSAIRC | 0.9817 |
|
| 3 | CUHKMED | 0.9644 | 0.9500 |
| 4 | NKSG | 0.9587 | 0.8917 |
| 5 | Mammoth | 0.9555 | 0.8918 |
| 6 | Masker | 0.9524 | 0.8500 |
| 7 | SMILEDeepDR | 0.9508 | 0.8750 |
| 8 | BUCT | 0.9348 | 0.8500 |
| 9 | WinterFell | 0.9327 | 0.9250 |
| 10 | NightOwl | 0.9101 | 0.9000 |
| 11 | Cvblab | 0.8806 | 0.7318 |
| 12 | AIML | 0.8458 | 0.7250 |
| Ground truth vCDR | 0.9471 | 0.8750 | |
Figure 7ROC curves and AUC values corresponding to the three top-rated glaucoma classification methods (solid lines) and the vertical cup-to-disc ratio (green dotted line). Crosses indicate the operating points of two experts.
Main advantages and disadvantages of gold standard databases.
| Databases | Advantages | Disadvantages |
|---|---|---|
|
|
Segmentation by five specialists Classified from clinical data Division into test/training groups |
It does not show symmetrical images between the two eyes. Image is cut around the optical disc. There are no clinical data or examinations corresponding to each dataset. |
|
|
Segmentation by five specialists Classification based on clinical notch findings, CDR, and examinations -Divided between training and test group |
It does not show symmetrical images between the two eyes. There is a small number of data and experts. Glaucoma/non-glaucoma classification is conducted based on image feature analysis. There are no clinical data and examinations corresponding to each dataset. |
|
|
A larger number of images Includes segmentation by experts Classification based on clinical data Divided between training and test group |
Sampling is limited to a specific population. It does not show symmetrical images between the two eyes. The image is not centered on the papilla. There is no access to the patient’s clinical data with prejudice to the access of other comorbidities. |