| Literature DB >> 35459040 |
Mengchen Lin1, Guidong Bao1, Xiaoqian Sang1, Yunfeng Wu1.
Abstract
With non-invasive and high-resolution properties, optical coherence tomography (OCT) has been widely used as a retinal imaging modality for the effective diagnosis of ophthalmic diseases. The retinal fluid is often segmented by medical experts as a pivotal biomarker to assist in the clinical diagnosis of age-related macular diseases, diabetic macular edema, and retinal vein occlusion. In recent years, the advanced machine learning methods, such as deep learning paradigms, have attracted more and more attention from academia in the retinal fluid segmentation applications. The automatic retinal fluid segmentation based on deep learning can improve the semantic segmentation accuracy and efficiency of macular change analysis, which has potential clinical implications for ophthalmic pathology detection. This article summarizes several different deep learning paradigms reported in the up-to-date literature for the retinal fluid segmentation in OCT images. The deep learning architectures include the backbone of convolutional neural network (CNN), fully convolutional network (FCN), U-shape network (U-Net), and the other hybrid computational methods. The article also provides a survey on the prevailing OCT image datasets used in recent retinal segmentation investigations. The future perspectives and some potential retinal segmentation directions are discussed in the concluding context.Entities:
Keywords: machine learning; neural networks; ophthalmic diseases; optical coherence tomography; retinal fluid segmentation
Mesh:
Year: 2022 PMID: 35459040 PMCID: PMC9029682 DOI: 10.3390/s22083055
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
Figure 1The OCT retinal images from left to right are from the three vendors of Cirrus, Spectralis, and Topcon in the RETOUCH dataset [2]. The images in the first row are not manually labeled, while the red, blue, and green segmentations on the second row represent the intraretinal fluid (IRF), subretinal fluid (SRF), and epithelial detachment (PED), respectively.
Figure 2Statistics of the papers and their sources on the deep learning methods for retinal fluid segmentation published in recent years. (a) Number of publications on retinal fluid segmentation based on deep learning from 2016 to 2020. (b) Percentage of papers published by conferences and journal publishers. (c) Keywords listed in the recent publications.
Figure 3Two different types of retinal fluid segmentation on the same optical coherence tomography (OCT) image from the RETOUCH dataset [2]. (a) Semantic segmentation: all of the retinal fluids are segmented into yellow regions; (b) instance segmentation: the IRF (red), SRF (green), and PED (blue) are segmented separately.
Figure 4General structures of (a) the fully convolution network (FCN) and (b) the U-Net.
Public benchmark optical coherence tomography (OCT) datasets for retinal fluid segmentation. AMD: Age-related macular degeneration, CNV: choroidal neovascularization, DME: diabetic macular edema, IRF: intraretinal fluid, MS: multiple sclerosis, PED: pigment epithelial detachment, RVO: retinal vein occlusion, SRF: subretinal fluid. Dataset web pages last accessed on 9 February 2022.
| Dataset | Data Size | Manual Labeling | Disease | Web Page |
|---|---|---|---|---|
| RETOUCH | 70 volumes | IRF, SRF, PED | AMD, RVO |
|
| UMN | 600 B-scan images | IRF, SRF, PED | AMD |
|
| OPTIMA | 30 volumes | IRF | AMD, RVO, DME |
|
| Duke | 110 B-scan images | Fluid regions | DME |
|
| HCMS | 49 B-scan images | Fluid regions | MS |
|
| Kermany | 108,312 B-scan images | Fluid regions | CNV, DME, drusen |
|
Deep learning architectures and paradigms for retinal fluid segmentation. ASPP: atrous spatial pyramid pooling, CNN: convolutional neural network, DAC: dense atrous convolution, FCN: fully convolutional network.
| Deep Learning | Methodology | Research Works | Paradigms |
|---|---|---|---|
| Schlegl et al. [ | 8-layer FCN | ||
| Chen et al. [ | Faster R-CNN | ||
| Modified CNN | Sanchez et al. [ | SEUNet model | |
| Wang et al. [ | Deeplab model | ||
| Sappa et al. [ | CNN (skip-connect operations + atrous spatial pyramid pooling) | ||
| FCN backbones | Girish et al. [ | Enhanced semantics preprocessing + CNN with a depthwise separable convolution filter | |
| Liu and Wang [ | Two (Student + Teacher) FCNs | ||
| Pawan et al. [ | DRIP-Caps (SegCaps with dilation, residual connections, inception blocks, and capsule pooling) | ||
| Modified FCN | Hu et al. [ | ResNet added with ASPP and modified stochastic ASPP | |
| Gao et al. [ | FCN with modified loss funciton | ||
| Xing et al. [ | FCN architecture with attention gate and spatial pyramid pooling module | ||
| Roy et al. [ | ReLayNet optimized with the weighted logistic regression and Dice loss | ||
| Fine-tuning | Kang et al. [ | A cascade of two fine-tuned U-Nets (the second U-Net with 2-channel inputs) | |
| Lee et al. [ | U-Net with 18 convolutional layers and sigmoid activation function for probability distribution mapping | ||
| Tennakoon et al. [ | Adversarial loss | ||
| Modified loss function | Liu et al. [ | Semi-supervised loss | |
| U-Net backbones | Wei and Peng [ | Mutex Dice loss | |
| Chen et al. [ | Squeeze-and-Excitation block (SE-block) | ||
| Hassan et al. [ | Multiscale feature extractor module (DAC + ASPP) | ||
| Functional modules | Ma et al. [ | U-Net + FCN with ASPP module | |
| Ye et al. [ | Context pyramid guide + Context shrink-age encode | ||
| Yang et al. [ | Six-string convolutions + Multiscale pyramid pooling module | ||
| Bao et al. [ | Channel multiscale module + Spatial multiscale module | ||
| 3D U-Net | Lin et al. [ | Batch normalization + Focal loss | |
| Shortest path methods | Rashno et al. [ | Graph shortest path + CNN | |
| Liu et al. [ | BM3D method + Graph shortest path + CNN | ||
| Graph-based methods | Rao et al. [ | IOWA reference algorithm + U-Net | |
| Hybrid methods | Lu et al. [ | Graph-cut algorithm + FCN + Random forest | |
| Montuoro et al. [ | Unsupervised representation + Auto-context loop + Graph-theoretic segmentation | ||
| Unsupervised learning algorithms | Gopinath and Sivaswamy [ | CNN trained with generalized motion patterns + Clustering | |
| He et al. [ | Intra- and inter-slice contrastive learning network (ISCLNet) + CNN |
Summary of important results of the IRF, SRF, and PED segmentations. The Dice results of IRF, SRF, and PED were the mean value of OCT image segmentation provided by different OCT vendors. “N/A” indicates that this type of fluid is not segmented in the paper or is not segmented by the method of deep learning.
| Research Work | Dataset | Backbone | Loss Function | IRF | SRF | PED |
|---|---|---|---|---|---|---|
| Dice (%) | Dice (%) | Dice (%) | ||||
| Hassan et al. [ | RETOUCH | U-Net | Dice loss | 90.90 | 91.30 | 91.80 |
| Ye et al. [ | RETOUCH | U-Net | Combination of cross-entropy and Dice loss | 73.17 | 79.70 | 71.06 |
| Schlegl et al. [ | Private | FCN | Softmax loss | 77.00 | 79.00 | N/A |
| Gao et al. [ | Private | FCN | Combination of area and softmax loss | N/A | 95.30 | 91.90 |
| Lee et al. [ | Private | U-Net | ReLu, sigmoid loss | 72.90 | N/A | N/A |
| Rao et al. [ | Private | FCN | ReLu, sigmoid loss | N/A | 91.00 | N/A |
| Yang et al. [ | Private | U-Net | Combination of binary cross-entropy, Dice, and diff loss | N/A | 96.20 | N/A |
| Bao et al. [ | Private | U-Net | Combination of binary cross-entropy and Dice loss | N/A | N/A | 79.17 |
| Pawan et al. [ | Private | SegCaps | Margin loss | N/A | 94.04 | N/A |
| Hu et al. [ | Private | ResNet50 | Cross-entropy loss | N/A | 87.59 | 73.71 |
| Venhuizen et al. [ | Private | FCN | Weight map loss | 75.40 | N/A | N/A |
| Chen et al. [ | RETOUCH | Faster R-CNN | Softmax and smooth L1 loss | 70.44 | 58.83 | 70.31 |
| Sappa et al. [ | RETOUCH | FCN | Combination of softmax and cross entropy loss | 78.95 | 90.90 | 95.78 |
| Girish et al. [ | OPTIMA | DSCN | Weight map loss | 74.00 | N/A | N/A |
| Liu et al. [ | RETOUCH | FCN | Combination of Dice loss, cross-entropy, regression and consistency loss | 73.60 | 75.57 | 73.93 |
| Xing et al. [ | RETOUCH | FCN | Shape prior loss function | 79.00 | 74.00 | 77.00 |
| Kang et al. [ | RETOUCH | U-Net | Maxout loss | 86.00 | 92.67 | 91.33 |
| Tennakoon et al. [ | RETOUCH | U-Net | Combination of cross-entropy, Dice loss, and adversarial loss | 69.00 | 67.00 | 85.00 |
| Rashno et al. [ | RETOUCH | CNN | ReLu, tanh loss | 85.45 | 81.30 | 84.08 |
| Lu et al. [ | RETOUCH | FCN | Combination of softmax and cross entropy loss | 71.90 | 75.53 | 72.06 |
| Montuoro et al. [ | Duke | Unsupervised 3D-layer | Probability map | 76.00 | 76.00 | N/A |
| Gopinath et al. [ | OPTIMA | CNN | Weight binary cross entropy loss | 71.00 | N/A | N/A |
| He et al. [ | RETOUCH | ISCLNet | Inter-slice loss, cross-entropy loss, and intra-slice loss | N/A | 73.16 | 69.42 |
Summary of important results of retinal fluid segmentation, without further segmentation of IRF, SRF, or PED regions. The Dice results are sorted in ascending order. SEUNet: U-Net with squeeze-and-excitation blocks, DRIU: deep retinal understanding.
| Research Work | Dataset | Backbone | Loss Function | Fluid in Dice (%) |
|---|---|---|---|---|
| Ma et al. [ | Private | U-Net+FCN | Combination of weighted Dice loss and the weighted logistic loss | 51.32 |
| Sanchez et al. [ | Duke | Ensemble network (SEUNet, DRIU) | Cross-entropy, Dice, Jaccard loss | 62.45 |
| Roy et al. [ | Duke | U-Net | Combination of weighted logistic regression and Dice loss | 77.00 |
| Liu et al. [ | Duke | U-Net | Combination of cross entropy, Dice, adversarial, and semi-supervised loss | 80.00 |
| Liu et al. [ | Private | CNN | Relu and softmax loss | 81.10 |
| Hao et al. [ | Duke | U-Net | Combination of weighted cross-entropy, Dice and mutex Dice loss | 81.00 |
| Chen et al. [ | Public | U-Net | Dice loss | 94.21 |
| Wang et al. [ | Private | DeepLab | Cross-entropy loss | 95.43 |
| Li et al. [ | Private | 3D U-Net | Combination of weighted loss and focal loss | 95.50 |