| Literature DB >> 35858986 |
Sebastian Otálora1, Maria A Zuluaga2, Lucas Pascal3,4, Oscar J Perdomo5, Xavier Bost4, Benoit Huet6.
Abstract
Glaucoma is an eye condition that leads to loss of vision and blindness if not diagnosed in time. Diagnosis requires human experts to estimate in a limited time subtle changes in the shape of the optic disc from retinal fundus images. Deep learning methods have been satisfactory in classifying and segmenting diseases in retinal fundus images, assisting in analyzing the increasing amount of images. Model training requires extensive annotations to achieve successful generalization, which can be highly problematic given the costly expert annotations. This work aims at designing and training a novel multi-task deep learning model that leverages the similarities of related eye-fundus tasks and measurements used in glaucoma diagnosis. The model simultaneously learns different segmentation and classification tasks, thus benefiting from their similarity. The evaluation of the method in a retinal fundus glaucoma challenge dataset, including 1200 retinal fundus images from different cameras and medical centers, obtained a [Formula: see text] AUC performance compared to an [Formula: see text] obtained by the same backbone network trained to detect glaucoma. Our approach outperforms other multi-task learning models, and its performance pairs with trained experts using [Formula: see text] times fewer parameters than training each task separately. The data and the code for reproducing our results are publicly available.Entities:
Mesh:
Year: 2022 PMID: 35858986 PMCID: PMC9300731 DOI: 10.1038/s41598-022-16262-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Results for the test set in the REFUGE dataset over the four tasks using 5-fold cross-validation.
| Models | Tasks | |||||
|---|---|---|---|---|---|---|
| Glaucoma | OD | OC | Fovea | |||
| Model | #P | time | AUC | DSC | Fovea Error | |
| STL | 61.2 | 0.686 | ||||
| MTL-IO (ours) | 0.557 | |||||
| Vanilla MTL | ||||||
| GradNorm | 0.260 | |||||
| PCGrad | 0.556 | |||||
Best values are in bold.
Results for the test set in the REFUGE dataset over the four different tasks using 5-fold cross-validation.
| Model | Tasks | |||
|---|---|---|---|---|
| Glaucoma | OD | OC | Fovea | |
| AUC | DSC | Fovea Error | ||
| STL | ||||
| MTL-IO (ours) | ||||
| Vanilla MTL | ||||
| Res34-Unet | − | − | ||
| GradNorm | ||||
| PCGrad | ||||
The results in this table correspond to the models trained with transfer learning. Res34-Unet contains parameters. Given the multiple stages of this method, we roughly estimate an iteration time of 0.375s consisting of the training phases through deep networks and excluding any pre/post-processing stages.
Best values are in bold.
Figure 1Receiver operating characteristic (ROC) curve for the glaucoma detection task for the single task learning model (STL) and our multi-task learning (MTL-IO) approach.
Figure 2Performance versus learning rate. Per task performance as a function of the learning rate () for the single task learning model (STL) and our multi-task learning (MTL-IO) approach. Standard deviation in shaded colour.
Figure 3Loss versus learning rate. Final loss values as a function of the learning rate () for the single task learning model (STL) and our multi-task learning (MTL-IO) approach. MTL-IO tends to have lower loss values, suggesting the benefit of learning also from other tasks. Standard deviation in shaded colour.
Figure 4Glaucomatous image from the test set (first image), where the three dots represent the fovea location’s ground truth (red), the MTL-IO prediction (blue) and the STL prediction (green), followed by OD (green) and OC (yellow) ground truth (second), MTL-IO (third) and STL (fourth) segmentation masks.
Results for the test set in the REFUGE dataset over the four tasks using the challenge’s official splits for training, validation and testing.
| Model | Tasks | |||
|---|---|---|---|---|
| Glaucoma | OD | OC | Fovea | |
| AUC | DSC | Fovea Error | ||
| STL | ||||
| MTL-IO (ours) | ||||
Best values are in bold.
Results for the test set in the REFUGE dataset over the four different tasks with transfer learning using the challenge’s official splits for training, validation and testing.
| Model | Tasks | |||
|---|---|---|---|---|
| Glaucoma | OD | OC | Fovea | |
| AUC | DSC | Fovea Error | ||
| STL | ||||
| MTL-IO | ||||
Best values are in bold.
Figure 5Multi-task learning framework for glaucoma detection, OD and OC segmentation, and fovea localization. The framework uses a U-net as its backbone architecture.
Figure 6Example of a retinal fundus image (left and the correspondent saliency map centered on the fovea coordinates (right).