| Literature DB >> 33133774 |
Jason Kugelman1, David Alonso-Caneiro1,2, Yi Chen2, Sukanya Arunachalam2, Di Huang2,3,4, Natasha Vallis2, Michael J Collins1, Fred K Chen2,5,6.
Abstract
Purpose: To use a deep learning model to develop a fully automated method (fully semantic network and graph search [FS-GS]) of retinal segmentation for optical coherence tomography (OCT) images from patients with Stargardt disease. <br> Methods: Eighty-seven manually segmented (ground truth) OCT volume scan sets (5171 B-scans) from 22 patients with Stargardt disease were used for training, validation and testing of a novel retinal boundary detection approach (FS-GS) that combines a fully semantic deep learning segmentation method, which generates a per-pixel class prediction map with a graph-search method to extract retinal boundary positions. The performance was evaluated using the mean absolute boundary error and the differences in two clinical metrics (retinal thickness and volume) compared with the ground truth. The performance of a separate deep learning method and two publicly available software algorithms were also evaluated against the ground truth. <br> Results: FS-GS showed an excellent agreement with the ground truth, with a boundary mean absolute error of 0.23 and 1.12 pixels for the internal limiting membrane and the base of retinal pigment epithelium or Bruch's membrane, respectively. The mean difference in thickness and volume across the central 6 mm zone were 2.10 µm and 0.059 mm3. The performance of the proposed method was more accurate and consistent than the publicly available OCTExplorer and AURA tools. Conclusions: The FS-GS method delivers good performance in segmentation of OCT images of pathologic retina in Stargardt disease. Translational Relevance: Deep learning models can provide a robust method for retinal segmentation and support a high-throughput analysis pipeline for measuring retinal thickness and volume in Stargardt disease. Copyright 2020 The Authors.Entities:
Keywords: ABCA4; OCT; artificial intelligence; image segmentation; inherited retinal diseases; machine learning; trial end point
Mesh:
Year: 2020 PMID: 33133774 PMCID: PMC7581491 DOI: 10.1167/tvst.9.11.12
Source DB: PubMed Journal: Transl Vis Sci Technol ISSN: 2164-2591 Impact factor: 3.283
Figure 1.Overview of the proposed method, including (1) an example raw OCT image, (2) corresponding contrast enhanced image, (3) neural network architecture with 4 pooling layers incorporating squeeze and excite blocks, (4) example layer probability map (pink, vitreous + padding; blue, retina; red, choroid + sclera) and, (5) example of OCT image with boundary predictions marked (solid lines, truths; dotted lines, predictions; blue lines, ILM; red lines, RPE). For the neural network architecture, #F, filters, #S, stride, BR, batch normalization + ReLU activation, while all convolutional layer inputs are padded such that input size is equal to output size.
Summary of Training, Validation, and Testing Sets Used
| Set | No. of Images | No. of Participants | No. of Volumes |
|---|---|---|---|
| Training | 2424–2429 | 10 | 40 |
| Validation | 483–488 | 2 | 8 |
| Testing (subset, all) | 2259 (7845) | 10 | 39 (129) |
| Total (subset, all) | 5171 (10,762) | 22 | 87 (177) |
The number of images for training and validation varies very slightly between folds owing to the handful of truncated scans that were discarded. For evaluating boundary error performance, the subset of cleaned test data is used (2259 images), whereas all testing data (uncleaned) is used for volume and thickness calculations (7845 images).
Boundary MAE Results for Inner and Outer Retinal Boundaries
| Average | Ensemble | |||
|---|---|---|---|---|
| Method | ILM MAE (SD) [px] | RPE MAE (SD) [px] | ILM MAE (SD) [px] | RPE MAE (SD) [px] |
| ON 4 | 0.32 (0.68) | 1.37 (2.42) | 0.23 (0.20) | 1.22 (1.74) |
| OFF 4 | 0.36 (1.10) | 1.47 (3.03) | 0.25 (0.45) | 1.37 (2.54) |
| ON 5 | 0.32 (0.81) | 1.39 (2.96) | 0.23 (0.30) | 1.25 (2.28) |
| ON 4 scSE | 0.35 (1.28) | 1.31 (2.15) | 0.23 (0.23) | 1.12 (1.41) |
| ON 4 scSE NA | 0.34 (0.88) | 1.41 (2.50) | 0.22 (0.24) | 1.17 (1.60) |
Each method was run five times with the average error and standard deviation across the five runs presented here. 4/5, number of pooling layers used; NA, no augmentations were used; ON/OFF, whether contrast enhancement was used; px, pixels; scSE, squeeze + excitation blocks were incorporated.
Boundary Mean Error Results for Inner and Outer Retinal Boundaries
| Average | Ensemble | |||
|---|---|---|---|---|
| Method | ILM ME (SD) [px] | RPE ME (SD) [px] | ILM ME (SD) [px] | RPE ME (SD) [px] |
| ON 4 | 0.02 (1.32) | 0.15 (2.15) | ‒0.02 (0.20) | 0.14 (1.79) |
| OFF 4 | 0.14 (1.08) | 0.24 (2.84) | 0.05 (0.51) | 0.18 (2.96) |
| ON 5 | 0.05 (0.79) | 0.23 (2.82) | ‒0.02 (0.32) | 0.21 (2.62) |
| ON 4 scSE | 0.01 (1.26) | 0.15 (2.16) | ‒0.02 (0.20) | 0.17 (1.46) |
| ON 4 scSE NA | 0.09 (0.86) | 0.16 (2.51) | ‒0.01 (0.22) | 0.10 (1.61) |
Each method was run five times with the average error and standard deviation across the five runs presented here. Positive values indicate that predictions are lower on the image than the corresponding truths. 4/5, number of pooling layers used; ME, mean error; NA, no augmentations were used; ON/OFF, whether contrast enhancement was used; px, pixels; scSE, squeeze + excitation blocks were incorporated.
Comparison of Various Methods with the Best Performing Semantic Segmentation Method
| Method | ILM MAE (SD) [px] | RPE MAE (SD) [px] |
|---|---|---|
| FS-GS [Semantic ON 4 scSE] (ensemble) | 0.23 (0.23) | 1.12 (1.41) |
| Patch-based 64 × 32 (ensemble) | 0.26 (0.17) | 1.15 (1.13) |
| OCTExplorer | 1.32 (1.41) | 4.95 (5.33) |
| AURA | 1.26 (0.92) | 7.17 (6.87) |
| Spectralis (automatic) | 1.07 (1.63) | 6.50 (3.27) |
Some volume segmentations failed (18% of all testing volumes) and could not be included in error calculations.
Sections of some scans (1.7% of compared columns) returned undefined segmentations. px, pixels; scSE, squeeze + excitation blocks were incorporated.
Figure 2.Example segmentations of Stargardt OCT images from FS-GS (the proposed ML-based semantic segmentation) method (left) and those provided by OCTExplorer (right). Blue, ILM; Red, RPE/Bruch's membrane. Solid lines indicate the ground truth boundary locations and the dotted lines correspond with the predicted locations.
Figure 3.Mean absolute difference (in bold) and standard deviation in (parentheses) of the thickness (in µm) (subplot A) and volume in cubic millimeters (subplot B). Mean (in bold) and limits of agreement (+1.96 SD above, –1.96 SD below) from Bland–Altman analysis between FS-GS and the ground truth for each subfield for thickness (in µm) (subplot C) and volume cubic millimeters (subplot D). Measurements are performed across the entire testing dataset for all nine Early Treatment of Diabetic Retinopathy Study subfields across the central 6 mm of all 129 testing volumes. Here the circles (from inner to outer, respectively) represent 1, 3, and 6 mm in diameter.