Literature DB >> 32821471

A Deep-Learning Approach for Automated OCT En-Face Retinal Vessel Segmentation in Cases of Optic Disc Swelling Using Multiple En-Face Images as Input.

Mohammad Shafkat Islam1, Jui-Kai Wang1,2, Samuel S Johnson1, Matthew J Thurtell3, Randy H Kardon2,3, Mona K Garvin1,2.   

Abstract

Purpose: In cases of optic disc swelling, segmentation of projected retinal blood vessels from optical coherence tomography (OCT) volumes is challenging due to swelling-based shadowing artifacts. Based on our hypothesis that simultaneously considering vessel information from multiple projected retinal layers can substantially increase vessel visibility, in this work, we propose a deep-learning-based approach to segment vessels involving the simultaneous use of three OCT en-face images as input.
Methods: A human expert vessel tracing combining information from OCT en-face images of the retinal pigment epithelium (RPE), inner retina, and total retina as well as a registered fundus image served as the reference standard. The deep neural network was trained from the imaging data from 18 patients with optic disc swelling to output a vessel probability map from three OCT en-face input images. The vessels from the OCT en-face images were also manually traced in three separate stages to compare with the performance of the proposed approach.
Results: On an independent volume-matched test set of 18 patients, the proposed deep-learning-based approach outperformed the three OCT-based manual tracing stages. The manual tracing based on three OCT en-face images also outperformed the manual tracing using only the traditional RPE en-face image. Conclusions: In cases of optic disc swelling, use of multiple en-face images enables better vessel segmentation when compared with the traditional use of a single en-face image. Translational Relevance: Improved vessel segmentation approaches in cases of optic disc swelling can be used as features for an improved assessment of the severity and cause of the swelling. Copyright 2020 The Authors.

Entities:  

Keywords:  U-Net; deep learning; multiple en-face images; optic disc swelling; optical coherence tomography; papilledema; retinal blood vessels; vessel segmentation

Mesh:

Year:  2020        PMID: 32821471      PMCID: PMC7401896          DOI: 10.1167/tvst.9.2.17

Source DB:  PubMed          Journal:  Transl Vis Sci Technol        ISSN: 2164-2591            Impact factor:   3.283


Introduction

Retinal blood vessel attributes, such as the width, location, obscuration, integrity, and tortuosity, are commonly considered important features in assessments of optic disc swelling. For example, in the modified Frisén grading system, the number of obscured vessel segments leaving the optic disc, is considered one of the key features to distinguish papilledema from mild to severe levels using fundus photographs.– Echegaray et al. have also shown that the measurement of vessel discontinuity can be helpful for a machine-learned Frisén grading system to achieve a substantial agreement between its output and the human expert's decision. Figure 1a shows three example fundus photographs with mild, moderate, and severe optic disc swelling (from the top to the bottom row in the figure). As indicated with yellow arrows, the vessel attributes change substantially on the swollen optic disc among these cases.
Figure 1.

Comparisons of the fundus photograph and OCT pairs with mild optic disc swelling (top row: a1, b1, c1), moderate swelling (middle row: a2, b2, c2), and severe swelling (bottom row: a3, b3, c3). Left column (a1, a2, a3) shows fundus photographs. Middle column (b1, b2, b3) shows the OCT central B-scans with automated layer segmentation. Right column (c1, c2, c3) shows the OCT RPE en-face images. Note: In case of swelling, the yellow arrows indicate vessel attributes changes in (a), the cyan lines in (c) represent the location of the central B-scans, and the green arrows in (b) and (c) indicate the matched shadow regions.

Comparisons of the fundus photograph and OCT pairs with mild optic disc swelling (top row: a1, b1, c1), moderate swelling (middle row: a2, b2, c2), and severe swelling (bottom row: a3, b3, c3). Left column (a1, a2, a3) shows fundus photographs. Middle column (b1, b2, b3) shows the OCT central B-scans with automated layer segmentation. Right column (c1, c2, c3) shows the OCT RPE en-face images. Note: In case of swelling, the yellow arrows indicate vessel attributes changes in (a), the cyan lines in (c) represent the location of the central B-scans, and the green arrows in (b) and (c) indicate the matched shadow regions. Spectral-domain optical coherence tomography– (OCT) is another imaging modality that is regularly used for assessing optic disc swelling. To date, most OCT-based measurements that have been used in the clinic and in research studies in cases of optic disc swelling (such as the retinal nerve fiber layer as well as the total retinal layer thicknesses, the optic nerve head [ONH] volume, and Bruch's membrane shapes),,– do not incorporate vessel information. However, especially for purposes of developing automated systems for assessing the severity and causes of optic disc swelling, having robust automated approaches for the OCT-based segmentation of retinal vessels is needed not only for computation of vessel-based features but also as one of the preprocessing steps for computation of other features. For example, removing retinal vessels is often involved as a part of preprocessing for further retinal texture analyses, such as retinal fold analysis., Having an accurate vessel tree location map can substantially reduce the false-positive rate for an automated method to detect retinal folds in OCT. Furthermore, vessels are often an important structure used for the alignment of images (e.g., color fundus to OCT or OCT images over time), which can be used for multimodal analyses and regional longitudinal analyses. Thus, the motivation for an OCT-based vessel segmentation in cases of optic disc swelling includes the need for the direct computation of vessel-related features in OCT (especially in cases where fundus photography is not available) for direct measures of severity or for differentiation, the need for additional contextual information in the development of techniques for the automated segmentation and analysis of other structures (e.g., for fold/wrinkle detection) that may help in differentiation, and the need for an alignment technique for region-based longitudinal analyses. Although OCT has been widely used for capturing cross-sectional information of the retina, observing the vessels in the common B-scan orientation is not straightforward (as shown in Fig. 1b). A common method to display the vessels in OCT is to create an en-face view by projecting the pixel intensity values within the retinal pigment epithelium (RPE) complex along each A-scan.– In cases without optic disc swelling, projection at the level of the RPE works well given the high contrast between the bright RPE and inner-retinal vessel shadows. However, in cases of optic disc swelling, the presence of swelling can cause image shadows, making the task of vessel segmentation much more challenging. Figure 1c continues to show the RPE en-face images from the same three patients; it is noticeable that the challenge of delineating the vessels for both manual and automated approaches increases when the image shadow (from the swollen disc) grows. However, based on our prior preliminary experience with the segmentation of vessels in the OCT scans of mice whereby using multiple en-face images was advantageous over a single projection image as well as our observation that the vessels could sometimes be seen more prominently in layers other than the RPE layer in cases of optic disc swelling in humans, we hypothesized that simultaneously considering vessel information from various projected retinal layers in cases of optic disc swelling would substantially increase the vessel visibility and enable a better segmentation. Thus, instead of relying on a single projection image at the level of the RPE, we have developed a deep-learning approach (using a modification of a U-Net architecture) to simultaneously input three OCT en-face images from the RPE complex, inner retina, and total retina and to output an OCT vessel probability map (Fig. 2). Although current deep neural networks have shown prominent performance among the frequent implementations of vessel segmentation algorithms,– there is no study specifically focusing on the OCT cases with optic disc swelling. Both quantitative and qualitative comparisons between manual tracings in en-face images from various retinal layers and the automated segmentation results are performed.
Figure 2.

Architecture of the proposed deep-learning approach. Three image patches (32 × 32 pixels) are separately extracted from the OCT en-face images of the RPE complex, the inner retina, and the total retina. Next, these three patches are concatenated to each other at the first layer in the network. The numbers in black and in gray at each block represent the number of channels and dimensions at the current network layer, and the colors of the arrows represent different network operations.

Architecture of the proposed deep-learning approach. Three image patches (32 × 32 pixels) are separately extracted from the OCT en-face images of the RPE complex, the inner retina, and the total retina. Next, these three patches are concatenated to each other at the first layer in the network. The numbers in black and in gray at each block represent the number of channels and dimensions at the current network layer, and the colors of the arrows represent different network operations.

Methods

Training/Testing Data

From 122 patients with various causes of optic disc swelling who had been recruited for research use of their clinical imaging data from the Neuro-Ophthalmology Clinic at the University of Iowa, we had previously analyzed the volumetric OCT imaging data of 22 of these patients for a preliminary fold detection image analysis approach. To best ensure a true separation of training and testing sets (whereby evaluation on the testing set is limited to untouched data), our training set was selected from the 22 previously analyzed images. More specifically, of the 22 patients previously analyzed, we selected the 18 patients who had (1) both volumetric ONH-centered OCT scans (Zeiss Cirrus, Carl Zeiss Meditec, Inc., Dublin, CA, USA) and the corresponding fundus photographs (Topcon Medical Systems, Inc.) available at the same visit and (2) an intact retinal structure in the OCT scans to allow the automated retinal layer segmentation to process correctly. The training data set was used for the purposes of designing the neural network architecture, deciding the hyperparameters, and training the neuron weights. For the independent testing data set, an additional 18 pairs of OCT scans and fundus photographs of 18 patients with optic disc swelling collected from the same set of 122 patients having optic disc swelling were included by matching the total retinal volume distribution with the training data set. The reason for this volume-matching process was to maintain a similar vessel visibility (which can be substantially affected by the degree of optic disc swelling, as shown in Fig. 1) in the OCT en-face images within different patients between the training and testing data sets. Figure 3 shows the data distributions (by ONH volume) of the training and testing data sets. For the causes of optic disc swelling in the training data set, among the 18 training patients, 13 had papilledema, 1 had nonarteritic anterior ischemic optic neuropathy (NAION), and 4 had other causes of optic disc swelling. For the 18 testing participants, 15 had papilledema, 2 had NAION, and 1 had another cause of optic disc swelling. The training data set consisted of 16 women and 2 men with a mean ± standard deviation (SD) age of 36.4 ± 11.4 years while the volume-matched testing data set consisted of 17 women and 1 man with a mean ± SD age of 31.6 ± 12.9 years. In total, the 36 patients had a mean ± SD age of 34 ± 12.4 years.
Figure 3.

Data distribution (by ONH volume) of the training and testing data sets (shown in pink and green bars, respectively). There are 36 patients in total: 18 in the training data set and the other 18 in the testing data set. The severity of the disc swelling has been matched between both data sets based on the ONH volumes.

Data distribution (by ONH volume) of the training and testing data sets (shown in pink and green bars, respectively). There are 36 patients in total: 18 in the training data set and the other 18 in the testing data set. The severity of the disc swelling has been matched between both data sets based on the ONH volumes. All the fundus photographs were obtained using a retinal camera (TRC-50DX; Topcon Medical Systems, Inc.) with 2392 × 2048 pixels. Each OCT scan (Cirrus; Carl Zeiss Meditec, Dublin, CA, USA) was centered at the ONH and had 200 × 200 × 1024 voxels covering (approximately) 6 × 6 × 2 mm3. Note that the fundus photographs in this study were only used for helping to create ground truth images (as discussed in the section on the manual tracing process) and were registered and cropped with respect to the OCT en-face images. The study protocol was approved by the University of Iowa's Institutional Review Board and adhered to the tenets of the Declaration of Helsinki.

En-Face Images from Multiple Retinal Layers (Inputs to Deep-Learning Approach)

A customized three-dimensional graph-based algorithm was utilized to segment the swollen retinal layers (Fig. 1b) as a part of the preprocessing. Then, based on the segmentation results, en-face images were generated of the RPE complex (between cyan and green surfaces in Fig. 1b), the inner retina (between red and yellow surfaces in Fig. 1b), and the total retina (between red and green surfaces in Fig. 1b) by averaging the pixel intensities within the interested layers along each A-scan in the OCT. Figure 4 demonstrates the vessel visibility of the en-face images from these three layers with mild, moderate, and severe optic disc swelling. As shown in Figure 4, retinal vessels away from the swollen optic disc appear more distinct in the RPE en-face image given the high contrast between the RPE and the vessel shadow, whereas the retinal vessels in the swollen regions appear more distinct in the en-face images of the inner retina and total retina. All three of these images were simultaneously used as inputs to the deep-learning network, as discussed in the next section. Furthermore, all three of these images (as well as the registered fundus photograph) were used to create the reference “ground truth” image used for training and evaluation.
Figure 4.

Demonstrations of vessel visibility of en-face images from different retinal layers in mild (top row), moderate (middle row), and severe (bottom row) optic disc swelling (continued from Fig. 1): left column, the RPE en-face image; middle column, the inner retina en-face image; and right column, the total retina en-face image. The yellow arrows indicate the vessel visibility changes.

Demonstrations of vessel visibility of en-face images from different retinal layers in mild (top row), moderate (middle row), and severe (bottom row) optic disc swelling (continued from Fig. 1): left column, the RPE en-face image; middle column, the inner retina en-face image; and right column, the total retina en-face image. The yellow arrows indicate the vessel visibility changes.

Architecture of the Proposed Deep-Learning Neural Network

As was shown in Figure 2, our proposed deep-learning network is designed to take (patches of) the three en-face images described above as input and output a pixel-based vessel probability map (with values close to 1 indicating a high vesselness probability and values close to 0 indicating a low vesselness probability). The high-level architecture of our proposed deep-learning network is based on a well-known U-shaped deep neural network (U-Net), with modifications to allow for three inputs rather than one and modifications to the number of layers. More specifically, the architecture of the proposed deep-learning approach (Fig. 2) contains a total of 16 neural layers, including 1 concatenation layer, 13 convolution layers, and 2 max-pooling layers. The proposed approach is designed to obtain image features in different resolutions by passing the concatenated input image patches through a contracting path (i.e., the first half; repeatedly uses a combination of convolutional layers, rectified linear units [ReLU], and a max-pooling layer] and then an “up-sampling” path (i.e., the second half; repeatedly uses a combination of convolutional layers, ReLU, and “up-convolutional” layers). Moreover, the corresponding feature maps between both paths are also concatenated in different resolutions. For the input of the proposed approach, location-matched image patches (size: 32 × 32 pixels) were first extracted from the three input en-face images to reduce the computational time as well as computer memory. At the end of the proposed approach, a soft-max layer, which is a 1 × 1 convolution layer, was applied to compute the probability value of the retinal vessel at each corresponding pixel location in the input image patch coordinates. For each patient, these location-matched image patches slid (one pixel each time) through the entire en-face image dimensions, and the outputted small vessel probability maps from all the image patches were stitched together (by averaging all the overlapping regions) to form a complete vessel probability map with the same coordinates as the input en-face images. More details about the deep neural network and its hyperparameters are described in the Appendix.

Manual Tracings of Retinal Blood Vessels and Ground Truth Images

In order to compare how a human expert would segment the vessels with access to various combinations of the input en-face images with the results from the proposed deep-learning-based approach, all the en-face images in both training and testing data sets were independently traced in three separate stages (by J-KW): stage I, referring only to the RPE en-face image (Table 1, column 1); stage II, referring to the combination of the RPE and inner-retina en-face images (Table 1, columns 1–2); and stage III, referring to the combination of the RPE, inner-retinal, and total-retinal en-face images (Table 1, columns 1–3). Furthermore, to serve as the overall reference standard (also known as the “ground truth”) for training and evaluation purposes, the images of the retinal vessels were again separately created by the same expert not only using all the RPE + inner-retinal + total-retinal en-face images but also referring to the registered ONH-centered fundus photographs to obtain the most vessel information (Table 1, columns 1–4). Also note that the proposed deep-learning approach output is also shown (Table 1, row 5) for a comparison. More specific details regarding this manual-tracing process are provided in the Appendix.
Table 1.

Inputs and Outputs for Manual Tracing (MT) Stages I, II, and III and the Proposed Deep-Learning Approach

The green, yellow, pink, red, and cyan vessels overlaid images show the outputs of MT stages I, II, and III; the ground truth; and the proposed approach. The black and white binary images are also shown at the next column to help visualization.

Inputs and Outputs for Manual Tracing (MT) Stages I, II, and III and the Proposed Deep-Learning Approach The green, yellow, pink, red, and cyan vessels overlaid images show the outputs of MT stages I, II, and III; the ground truth; and the proposed approach. The black and white binary images are also shown at the next column to help visualization.

Overview of Evaluation Approach

In the training process, leave-one-subject-out cross-validation (for a total of 18 training participants, 17 participants were used for training, 1 patient for validation, and then rotating the selected patient until all the patients had been tested) was used to help design and decide the architecture (using area under the receiver operating characteristic curve, AUC, as the evaluation metric) and hyperparameters in the proposed method. Next, based on the decided network, all the 18 patients were trained together to result in the final proposed deep neural network. The separate testing data set of 18 patients with matched ONH volumes was used with the quantitative measurements (described below) to compare the performances among the manual tracing stages I, II, and III and the proposed deep-learning approach.

Quantitative Evaluation Measurements

Pixel-based evaluation metrics were used to compare each ground truth binary image (0 = background and 1 = vessel object) to the binary results from each of the three OCT-based manual tracing stages, to the output probability map from the proposed deep-learning method, and to a binarized version (using Otsu's thresholding algorithm) of the proposed deep-learning method. More specifically, given a ground truth image and a corresponding binary map from another approach (e.g., an OCT-based manual tracing stage or a thresholded version of the output probability map), the true positive (TP), false positive (FP), true negative (TN), false negative (FN), and the total number of pixels (K = TP + FP + TN + FN) can be computed. We correspondingly computed the area under the receiver operating characteristic curves (AUC), for all approaches (manual tracings from the three stages, the original probability map, and the Otsu-thresholded binary map), to measure the TP rate against the FP rate across all possible threshold values. We also computed the average precision (AP) (for all approaches) to quantify the relationship between the precision (P) and recall (R) across all possible threshold values; AP =, where R = TP/(TP + FN), P = TP/(TP + FP), and n is the nth threshold value. For the binary-only results (manual tracings and Otsu-thresholded probability map but not the original probability map), we also computed the accuracy (ACC) to compute the ratio of correctly classified pixels to the total number of pixels in the OCT en-face image; ACC = (TP + TN)/K. In addition, for all approaches, we computed the mean squared error (MSE) as a measure of the label distance between the approaches and the ground truth; , where and represent the predicted label value at pixel location from the approach and the truth label, respectively; and the coefficient of determination (R2 score) to estimate how well the approach is with respect to the ground truth in the sense of regression; , where .

Results

The mean area under the ROC curves (AUC) of probability maps from a leave-one-subject-out cross-validation over the 18 patients in the training data set was 0.93. Using the independent imaging data from the test set, the mean AUCs for the manual tracing stages I, II, and III were 0.79, 0.83, and 0.85, respectively; for the proposed deep-learning approach, the mean AUC was 0.96 for the direct output of the vessel probability map, but the mean AUC was 0.83 when the probability maps were converted to binary maps using the Otsu algorithm. For the AP, the results of the manual tracing stages I, II, and III were 0.73, 0.77, and 0.78, respectively; the results of the proposed method were 0.84 and 0.77 for the probability map and binary map, respectively. Other results among the manual tracing stages I, II, and III and the probability map as well as the binary map from the proposed method were as follows: MSE, 0.071, 0.061, 0.061, 0.047, and 0.061; mean coefficient of determination (R2), 0.38, 0.46, 0.47, 0.59, and 0.46; and mean accuracy (ACC), 0.93, 0.94, 0.94, N/A (Not applicable), and 0.94, respectively. Table 2 shows the summary of the quantitative results.
Table 2.

Qualitative Measurements among the Manual Tracing Stages and Proposed Deep Neural Network in the Testing Data Set (Including the Mean Processing Time per Patient)

Testing Data Set (18 Patients)
MethodsMean AUCAverage Precision (AP)Mean Square Error (MSE)Mean R2ScoreMean Accuracy (ACC)Mean Processing Time*
Manual tracing, stage I0.79 ± 0.040.73 ± 0.040.071 ± 0.010.38 ± 0.080.93 ± 0.0111 min 19 s
Manual tracing, stage II0.83 ± 0.030.77 ± 0.030.061 ± 0.010.46 ± 0.060.94 ± 0.0113 min 51 s
Manual tracing, stage III0.85 ± 0.030.78 ± 0.040.061 ± 0.010.47 ± 0.080.94 ± 0.0114 min 7 s
Proposed deep-learning approachProbability map Binarymap0.96 ± 0.020.83 ± 0.050.84 ± 0.070.77 ± 0.050.047 ± 0.010.061 ± 0.010.59 ± 0.090.46 ± 0.10N/A0.94 ± 0.011 min 5 s

Hardware details: A Linux machine is used with single GPU (NVIDIA GeForce GTX 1080 Ti) and 128 GB RAM. The time to train the proposed neural network, including all the 18 patients in the training data set, was 2 hours 2 minutes 12 seconds.

The probability map is the direct output from the proposed deep neural network; the binary map is automatically obtained using the Otsu thresholding algorithm. (Note: The outputs from manual tracings are inherently binary maps.)

Qualitative Measurements among the Manual Tracing Stages and Proposed Deep Neural Network in the Testing Data Set (Including the Mean Processing Time per Patient) Hardware details: A Linux machine is used with single GPU (NVIDIA GeForce GTX 1080 Ti) and 128 GB RAM. The time to train the proposed neural network, including all the 18 patients in the training data set, was 2 hours 2 minutes 12 seconds. The probability map is the direct output from the proposed deep neural network; the binary map is automatically obtained using the Otsu thresholding algorithm. (Note: The outputs from manual tracings are inherently binary maps.) Figure 5 displays all the data points from the 18 testing participants with a 95% confidence interval of the mean for the quantitative results in the testing data set. The direct output (i.e., the vessel probability maps) of the proposed deep-learning approach shows the best performance in all five quantitative measurements. Furthermore, for all the testing participants, the thresholded binary map of the proposed deep-learning approach provides better vessel segmentation results than manual tracing stage I (tracing on only the RPE en-face image) for all five measurements of all 18 testing participants (additional details in Appendix). The results from manual tracing stages II and III and that of the thresholded binary map of the proposed deep-learning approach were similar. More specific details regarding the subject-wise comparison between different approaches are provided in the Appendix.
Figure 5.

Data dot plots for the measurements of area under ROC curve (AUC), average precision (AP), mean square error (MSE), mean coefficient of determination (R2), and mean accuracy (ACC) with 95% confidence intervals: green, manual tracing (MT) stage I; dark yellow, MT stage II; purple, MT stage III; blue, the probability map from the proposed deep learning (DL) approach; and red, the binary map, which is obtained using Ostu algorithm) from the DL approach.

Data dot plots for the measurements of area under ROC curve (AUC), average precision (AP), mean square error (MSE), mean coefficient of determination (R2), and mean accuracy (ACC) with 95% confidence intervals: green, manual tracing (MT) stage I; dark yellow, MT stage II; purple, MT stage III; blue, the probability map from the proposed deep learning (DL) approach; and red, the binary map, which is obtained using Ostu algorithm) from the DL approach. The mean processing times per patient for each method in the testing data set were also recorded (Table 2). The three manual tracing stages using Adobe Illustrator Draw (version 4.6.1; Adobe Systems, Inc., San Jose, CA, USA) on an iPad (Apple, Inc., Cupertino, CA, USA) required at least 11 minutes for each patient in the testing data set (precisely, 11 minutes 19 seconds, 13 minutes 51 seconds, and 14 minutes 7 seconds for stages I, II, and III, respectively). The computation for the proposed deep neural network was performed using a Linux machine with single GPU, NVIDIA GeForce GTX 1080 Ti (NVIDIA Corporation, Santa Clara, CA, USA) and 128 GB RAM. The mean processing time for each patient in the testing data set for the proposed deep neural network was 1 minute 5 seconds. (Note: The total time of training the proposed neural network, including all the 18 patients in the training data set, was 2 hours 2 minutes 12 seconds.) Six patients with various levels of optic disc swelling (the ONH volume range is from 11.46 mm3 [the top row] to 26.45 mm3 [the bottom row]) are shown as qualitative results in Table 3. The en-face images of the RPE complex, the inner retina, and the total retina are listed in the table to show the shadow region growth in different degrees of swelling. The manual tracing stage III (highlighted in purple), the proposed deep-learning approach binary maps (highlighted in cyan), and the ground truth (highlighted in red) are displayed in the next three columns, respectively. The corresponding ONH-registered fundus photographs are also added at the last column for reference.
Table 3.

Examples of Vessel Segmentation in Six Patients with Various Levels of Optic Disc Swelling

The patients are arranged based on the ONH volumes from the top to the bottom rows.

Examples of Vessel Segmentation in Six Patients with Various Levels of Optic Disc Swelling The patients are arranged based on the ONH volumes from the top to the bottom rows.

Discussion

Although a preliminary OCT-based vessel segmentation for swollen optic disc had been utilized in our previous studies as a part of the preprocessing, this is the first time that we directly focus on OCT to reveal the obscured vessels from the image shadow by using a modified deep neural network that simultaneously considers multiple OCT en-face images from various retinal layers as its inputs. Furthermore, while use of an RPE en-face image would traditionally be considered sufficient for visualization of projected retinal vessels in OCT (especially in nonswollen cases), our results also show that having access to multiple en-face views is important for even a human expert to properly visualize the vessels. Overall, our deep-learning approach (even when using a nonoptimized thresholding approach to generate the binary maps) performs at least as well as one would expect from a human expert having access to multiple en-face OCT views and better than what one would expect from a human expert having access to only the RPE en-face image (the traditional approach). In fact, when using the probability map itself (rather than the binary map), perhaps surprisingly, our results suggest that the deep-learning approach even outperforms the human expert in segmenting the retinal vessels with access to only OCT en-face images (discussed further below). Effectively, having multiple simultaneous input images enabled the proposed deep-learning network to learn to extract the vessel information from the region at the retinal layer with better signal response to compensate the regions that are eclipsed. Figure 4 shows examples that the en-face images around the optic disc of the inner retina contain substantially more vessel information than the one from the RPE complex in cases with optic disc swelling; however, the vessel visibility at the peripheral region is still the clearest in the RPE en-face image compared with the others. Since the vessels shown in the en-face images are the mean intensity values of the shadow from the superficial vessels, it is not surprising that the vessel visibility in the RPE en-face image is considerably deteriorated around the optic disc in which the OCT signal is greatly weakened by passing through the swollen inner retinal layers. Overall, the total-retinal en-face image displays in-between vessel clarity, which is potentially helpful for the proposed deep-learning approach having an extra reference when the vessel patterns are inconsistent between the RPE and inner-retinal en-face images. When the optic disc swelling is severe, in addition to the optic disc elevation, other pathogenic conditions, such as hemorrhage and nerve fiber layer infarcts (cotton-wool spots), may appear., Meanwhile, the vessel appearance may also be affected: the retinal vessels can be seen extremely blurred, discontinuous, and/or covered by the cotton-wool spots. Under these circumstances, it is sometimes tricky to clearly define the boundary of the vessels in the OCT en-face images. The bottom row in Table 3 illustrates the difficulty of the process of tracing complete vessel trees even with the extra information from the fundus photograph (i.e., the ground truth). Figure 6 shows that the performances of both manual tracing and proposed method gradually decline when the ONH volume increases; however, the proposed method still seems visually more robust than the manual tracing stage III in the testing data set.
Figure 6.

Scatterplots of 18 testing participants for displaying the relationships between the AUC and ONH volume from the manual tracing (MT) stage III (the best performance in all three manual tracing stages; shown as magenta crosses) and proposed deep-learning (DL) approach probability map (shown as cyan triangles).

Scatterplots of 18 testing participants for displaying the relationships between the AUC and ONH volume from the manual tracing (MT) stage III (the best performance in all three manual tracing stages; shown as magenta crosses) and proposed deep-learning (DL) approach probability map (shown as cyan triangles). Because the degree of optic swelling influences the difficulty of the vessel segmentation, it is important to keep in mind our overall reported results are, in part, reflective of the distribution of swelling levels tested. Our training set (and, correspondingly, our test set because of the volume-matching process) likely reflected a higher proportion of cases with moderate-to-severe optic disc swelling than one might encounter in clinical practice. If a larger percentage of cases with milder swelling were evaluated, we would actually expect to obtain better overall performance numbers. However, while we evaluated the approach on a reasonably balanced data set of cases with optic disc swelling, one limitation of our work is that we have not quantitively evaluated the result of the proposed approach on a separate normative data set. Nevertheless, based on visually assessing the results from separately applying our trained neural network on eyes with no apparent swelling (having approximate optic nerve head volume of 8–10 mm3), we found that the proposed approach still successfully segments the major vessels in such eyes. Thus, while not quantitatively evaluated on eyes without optic disc swelling, we still expect the approach to be robust in these cases as well (where traditional approaches involving only use of an RPE en-face image may already work). Also note that in computing the ONH volumes used for estimating the degree of optic disc swelling, we were unable to correct for ocular magnification due to the lack of axial length information, and thus the reported volumetric measures are technically approximations. However, while we did use the estimated measures to provide a similar distribution of swelling severity in the training and testing data sets as well as to provide some insight into the dependence of the performance on degree of optic disc swelling, correcting for ocular magnification is not needed for appropriately training the algorithm. As previously mentioned, the quantitative results in Table 2 and Figure 5 demonstrate that the probability map from the proposed deep neural network has the best performance compared with all the other methods. However, after thresholding the probability map (by the Otsu algorithm) into a binary map, the performance from the neural network declines to a similar level of manual tracing stages II and III. This is because of a part of vessel information gets lost in the process of thresholding. For example, in the probability map, some thinner vessels may have a smaller probability, which could be thresholded to become background in the binary map. The performance gap between the probability and binary maps can possibly be shortened by developing more sophisticated thresholding methods. Using regionally adaptive threshold values based on vessel continuity rather than one global threshold value is one of the options to improve. Regarding the stages of manual tracing, it is worthy to note that we strictly followed the order that we described in the Appendix to avoid the extra information from the higher stages, especially for the ground truth images, to prejudice the tracing in the lower stages. Table 2 and Figure 5c have shown that, for the same human expert, only using the RPE en-face image to trace the retinal vessels (i.e., the traditional method) provides the worst performance among all three stages for all the measurements. After adding consideration of the inner-retinal and total-retinal en-face images, the performance in stages II and III noticeably increased. Also, our manual tracings are performed using Adobe Illustrator Draw (Adobe Systems, Inc., version 4.6.1) on an iPad, which allows us to overlay all input en-face images on each other so that the vessel information can be intuitively accumulated from all the input retinal layers. The image-overlay method is potentially a more robust method than separately tracing these en-face images and then adding the results together. Also note that the time-consuming multistage manual tracing stages were one factor that limited the size of our training and test sets (18 cases each). While 18 cases in a training set would likely be considered too small for an image-level classification task for a deep-learning approach (e.g., determining the cause of the swelling), since our work focused on a pixel-level task in determining the probability of each pixel being a vessel (in combination with our use of a U-Net-based architecture), we were able to have sufficient data to be able to train the approach to provide a good performance overall on the independent test set. However, confirming the performance on a larger data set (perhaps using less time-consuming multistage reference standard) would be useful in future work. It is conceivable that our proposed deep-learning approach can be extended to detect other tubular objects in the OCT volumes. True three-dimensional vessel segmentation (instead of on the projected en-face planes) could be a subject of future study. Also, analyses of retinal folds,, are believed to be one of the key features to scrutinize the mechanisms of stress/strain at the ONH region. Automatically detecting the retinal folds and further quantifying them can be another possible extension for our proposed neural network.
  19 in total

1.  Baseline OCT measurements in the idiopathic intracranial hypertension treatment trial, part I: quality control, comparisons, and variability.

Authors:  Peggy Auinger; Mary Durbin; Steven Feldon; Mona Garvin; Randy Kardon; John Keltner; Mark Kupersmith; Patrick Sibony; Kim Plumb; Jui-Kai Wang; John S Werner
Journal:  Invest Ophthalmol Vis Sci       Date:  2014-11-04       Impact factor: 4.799

Review 2.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

3.  Diagnosis and grading of papilledema in patients with raised intracranial pressure using optical coherence tomography vs clinical expert assessment using a clinical staging scale.

Authors:  Colin J Scott; Randy H Kardon; Andrew G Lee; Lars Frisén; Michael Wall
Journal:  Arch Ophthalmol       Date:  2010-06

4.  Swelling of the optic nerve head: a staging scheme.

Authors:  L Frisén
Journal:  J Neurol Neurosurg Psychiatry       Date:  1982-01       Impact factor: 10.154

Review 5.  Blood vessel segmentation methodologies in retinal images--a survey.

Authors:  M M Fraz; P Remagnino; A Hoppe; B Uyyanonvara; A R Rudnicka; C G Owen; S A Barman
Journal:  Comput Methods Programs Biomed       Date:  2012-04-22       Impact factor: 5.428

6.  Shape analysis of the peripapillary RPE layer in papilledema and ischemic optic neuropathy.

Authors:  Patrick Sibony; Mark J Kupersmith; F James Rohlf
Journal:  Invest Ophthalmol Vis Sci       Date:  2011-10-10       Impact factor: 4.799

Review 7.  Optic disc edema.

Authors:  Gregory P Van Stavern
Journal:  Semin Neurol       Date:  2007-07       Impact factor: 3.420

Review 8.  Blood vessel segmentation algorithms - Review of methods, datasets and evaluation metrics.

Authors:  Sara Moccia; Elena De Momi; Sara El Hadji; Leonardo S Mattos
Journal:  Comput Methods Programs Biomed       Date:  2018-02-10       Impact factor: 5.428

9.  Retinal and Choroidal Folds in Papilledema.

Authors:  Patrick A Sibony; Mark J Kupersmith; Steven E Feldon; Jui-Kai Wang; Mona Garvin
Journal:  Invest Ophthalmol Vis Sci       Date:  2015-09       Impact factor: 4.799

10.  Optic nerve head quantification in idiopathic intracranial hypertension by spectral domain OCT.

Authors:  Falko Kaufhold; Ella Maria Kadas; Christoph Schmidt; Hagen Kunte; Jan Hoffmann; Hanna Zimmermann; Timm Oberwahrenbrock; Lutz Harms; Konrad Polthier; Alexander U Brandt; Friedemann Paul
Journal:  PLoS One       Date:  2012-05-15       Impact factor: 3.240

View more
  3 in total

Review 1.  The use of deep learning technology for the detection of optic neuropathy.

Authors:  Mei Li; Chao Wan
Journal:  Quant Imaging Med Surg       Date:  2022-03

2.  Feasibility study to improve deep learning in OCT diagnosis of rare retinal diseases with few-shot classification.

Authors:  Tae Keun Yoo; Joon Yul Choi; Hong Kyu Kim
Journal:  Med Biol Eng Comput       Date:  2021-01-25       Impact factor: 3.079

3.  Diagnosis of Retinal Diseases Based on Bayesian Optimization Deep Learning Network Using Optical Coherence Tomography Images.

Authors:  Malliga Subramanian; M Sandeep Kumar; V E Sathishkumar; Jayagopal Prabhu; Alagar Karthick; S Sankar Ganesh; Mahseena Akter Meem
Journal:  Comput Intell Neurosci       Date:  2022-04-15
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.