Literature DB >> 27555965

Classification of SD-OCT Volumes Using Local Binary Patterns: Experimental Validation for DME Detection.

Guillaume Lemaître1, Mojdeh Rastgoo1, Joan Massich1, Carol Y Cheung2, Tien Y Wong3, Ecosse Lamoureux3, Dan Milea3, Fabrice Mériaudeau4, Désiré Sidibé1.   

Abstract

This paper addresses the problem of automatic classification of Spectral Domain OCT (SD-OCT) data for automatic identification of patients with DME versus normal subjects. Optical Coherence Tomography (OCT) has been a valuable diagnostic tool for DME, which is among the most common causes of irreversible vision loss in individuals with diabetes. Here, a classification framework with five distinctive steps is proposed and we present an extensive study of each step. Our method considers combination of various preprocessing steps in conjunction with Local Binary Patterns (LBP) features and different mapping strategies. Using linear and nonlinear classifiers, we tested the developed framework on a balanced cohort of 32 patients. Experimental results show that the proposed method outperforms the previous studies by achieving a Sensitivity (SE) and a Specificity (SP) of 81.2% and 93.7%, respectively. Our study concludes that the 3D features and high-level representation of 2D features using patches achieve the best results. However, the effects of preprocessing are inconsistent with different classifiers and feature configurations.

Entities:  

Year:  2016        PMID: 27555965      PMCID: PMC4983398          DOI: 10.1155/2016/3298606

Source DB:  PubMed          Journal:  J Ophthalmol        ISSN: 2090-004X            Impact factor:   1.909


1. Introduction

Eye diseases such as Diabetic Retinopathy (DR) and Diabetic Macular Edema (DME) are the most common causes of irreversible vision loss in individuals with diabetes. Just in United States alone, health care and associated costs related to eye diseases are estimated at almost $500 M [1]. Moreover, the prevalent cases of DR are expected to grow exponentially affecting over 300 M people worldwide by 2025 [2]. Given this scenario, early detection and treatment of DR and DME play a major role in preventing adverse effects such as blindness. DME is characterized as an increase in retinal thickness within 1-disk diameter of the fovea center with or without hard exudates and sometimes associated with cysts [3]. Fundus images which have proven to be very useful in revealing most of the eye pathologies [4, 5] are not as good as OCT images which provide information about cross-sectional retinal morphology [6]. Many of the previous works on OCT image analysis have focused on the problem of retinal layers segmentation, which is a necessary step for retinal thickness measurements [7, 8]. However, few have addressed the specific problem of DME and its associated features detection from OCT images. Figure 1 shows one normal B-scan and two abnormal B-scans.
Figure 1

Example of SD-OCT images for normal (a) and DME patients (b)-(c) with cyst and exudate, respectively.

A summary of the existing work can be found in Table 1. Srinivasan et al. [9] proposed a classification method to distinguish DME, Age-Related Macular Degeneration (AMD), and normal SD-OCT volumes. The OCT images are preprocessed by reducing the speckle noise by enhancing the sparsity in a transform-domain and flattening the retinal curvature to reduce the interpatient variations. Then, Histograms of Oriented Gradients (HOG) are extracted for each slice of a volume and linear Support Vector Machine (SVM) is used for classification. On a dataset of 45 patients equally subdivided into the three aforementioned classes, this method leads to a correct classification rate of 100%, 100%, and 86.67% for normal, DME, and AMD patients, respectively. The images that have been used in their paper are publicly available but are already preprocessed (i.e., denoised), have different sizes for the OCT volumes, and do not offer a huge variability in terms of DME lesions, and some of them, without specifying which, have been excluded for the training phase; all these reasons prevent us from using this dataset to benchmark our work.
Table 1

Summary of the state-of-the-art methods.

Reference DiseasesData sizePreprocessingFeaturesRepresentationClassifierEvaluationResults
AMDDMENormalDenoiseFlattenAligningCropping
[9]45HOGLinear SVMACC86.7%, 100%, and 100%
[10]384TextonBoW, PCARFAUC0.984
[11]326Edge, LBPPCASVM-RBFAUC0.93
[12]62LBP-LBP-TOPPCA, BoW, and histogramRFSE, SP87.5%, 75%
Venhuizen et al. proposed a method for OCT images classification using the Bag-of-Words (BoW) model [10]. The method starts with the detection and selection of key points in each individual B-scan, by keeping the most salient points corresponding to the top 3% of the vertical gradient values. Then, a texton of size 9 × 9 pixels is extracted around each key point, and Principal Component Analysis (PCA) is applied to reduce the dimension of every texton to get a feature vector of size 9. All extracted feature vectors are used to create a codebook using k-means clustering. Then, each OCT volume is represented in terms of this codebook and is characterized as a histogram that captures the codebook occurrences. These histograms are used as feature vector to train a Random Forest (RF) with a maximum of 100 trees. The method was used to classify OCT volumes between AMD and normal cases and achieved an Area Under the Curve (AUC) of 0.984 with a dataset of 384 OCT volumes. Liu et al. proposed a methodology for detecting macular pathology in OCT images using LBP and gradient information as attributes [11]. The method starts by aligning and flattening the images and creating a 3-level multiscale spatial pyramid. The edge and LBP histograms are then extracted from each block of every level of the pyramid. All the obtained histograms are concatenated into a global descriptor whose dimensions are reduced using PCA. Finally a SVM with a Radial Basis Function (RBF) kernel is used as classifier. The method achieved good results in detection OCT scan containing different pathologies such as DME or AMD, with an AUC of 0.93 using a dataset of 326 OCT scans. Lemaitre et al. [12] proposed using 2D and 3D LBP features extracted from denoised volumes and dictionary learning using the BoW models [13]. In the proposed method all the dictionaries are learned with the same size of “visual words” (k = 32) and final descriptors are classified using RF classifier. The work described in this paper is an extension of our previous work [12]. In this research, beside the comparison of 2D and 3D features, we explore different possible representations of the features and different preprocessing steps for OCT data (i.e., aligning, flattening, and denoising). We also compare the performances of different classifiers. This paper is organized as follows: the proposed framework is explained in Section 2, while the experiments and results are discussed through Sections 3 and 4. Finally, the conclusion and avenue for future directions are drawn in Section 5.

2. Materials and Methods

The proposed method, as well as its experimental setup, for OCT volume classification is outlined in Figure 2. The methodology is formulated as a standard classification procedure which consists of five steps. First, the OCT volumes are preprocessed as presented in detail in Section 2.1. Then, LBP and LBP-TOP features are detected, mapped, and represented as discussed in depth in Sections 2.2, 2.3, and 2.4, respectively. Finally, the classification step is presented in Section 2.5.
Figure 2

Our proposed classification pipeline.

2.1. Image Preprocessing

This section describes the set of preprocessing techniques which aim at enhancing the OCT volume. The influences of these preprocessing methods and their possible combinations are extensively studied in Section 3.

2.1.1. Non-Local Means (NLM)

OCT images suffer from speckle noise, like other image modalities such as Ultrasound (US) [14]. The OCT volumes are enhanced by denoising each B-scan (i.e., each (x-z) slice) using the NLM [15], as shown in Figure 3. NLM has been successfully applied to US images to reduce speckle noise and outperforms other common denoising methods [16]. NLM filtering preserves fine structures as well as flat zones, by using all the possible self-predictions that the image can provide rather than local or frequency filters such as Gaussian, anisotropic, or Wiener filters [15].
Figure 3

OCT: (a) organization of the OCT data, (b) original image, and (c) NLM filtering. Note that the images have been negated for visualization purposes.

2.1.2. Flattening

Textural descriptors characterize spatial arrangement of intensities. However, the OCT scans suffer from large type of variations: inclination angles, positioning, and natural curvature of the retina [11]. Therefore, these variations have to be taken into account to ensure a consistent characterization of the tissue disposition, regardless of the location in the retina. This invariance can be achieved in different manners: (i) using a rotation invariant descriptor (cf. Section 2.2) or (ii) unfolding the curvature of the retina. This latter correction is known as image flattening which theoretically consists of two distinct steps: (i) estimate and fit the curvature of the Retinal Pigment Epithelium (RPE) and (ii) warp the OCT volume such that the RPE becomes flat. Our correction is similar to the one of Liu et al. [11]: each B-scan is thresholded using Otsu's method followed by a median filtering to detect the different retina layers (see Figures 4(c) and 4(d)). Then, a morphological closing and opening is applied to fill the holes and the resulting area is fitted using a second-order polynomial (see Figure 4(d)). Finally, the scan is warped such that the curve becomes a line as presented in Figures 4(e) and 4(f).
Figure 4

Flattening procedure: (a) original image, (b) thresholding, (c) median filtering, (d) curve fitting, (e) warping, and (f) flatten image.

2.1.3. Slice Alignment

The flattening correction does not enforce an alignment through the OCT volume. Thus, in addition to the flattening correction, the warped curves of each B-scan are positioned at the same altitude in the z-axis.

2.2. Feature Detection

In this research, we choose to detect simple and efficient LBP texture features with regard to each OCT slice and volume. LBP is a texture descriptor based on the signs of the differences of a central pixel with respect to its neighboring pixels [17]. These differences are encoded in terms of binary patterns as follows:where g , g are the intensities of the central pixel and a given neighbor pixel, respectively, and P is the number of sampling points in the circle of radius R. Ojala et al. further extended the original LBP formulation to achieve rotation invariance at the expense of limiting the texture description to the notion of circular “uniformity” [17]. Referring to the coordinate system defined in Figure 3(a), the LBP codes are computed on each (x-z) slice, leading to a set of LBP maps, a map for each (x-z) slice. Volume encoding is later proposed by Zhao et al. by computing LBP descriptors in three orthogonal planes, so-called LBP-TOP [18]. More precisely, the LBP codes are computed considering the (x-z) plane, (x-y) plane, and (y-z) plane, independently. Thus, three sets of LBP maps are obtained, one for each orthogonal plane. In this work, we consider rotation invariant and uniform LBP and LBP-TOP features with various sampling points (i.e., {8,16,24}) with respect to different radius (i.e., {1,2, 3}). The number of patterns (LBP#pat) in regard to each configuration is reported in Table 2.
Table 2

Number of patterns (LBP#pat) for different sampling points and radius ({P, R}) of the LBP descriptor.

 Sampling point for a radius ({P, R})
{8, 1}{16, 2}{24, 3}
LBP#pat 101826

2.3. Mapping

The mapping stage is used to partition the previously computed LBP maps; for this work, two mapping strategies are defined: (i) global and (ii) local mapping. The size of the feature descriptor is summarized in Table 3.
Table 3

Size of a descriptor for an SD-OCT volume. d denotes the number of slices in the volume, N the number of 2D windows, and N′ the number of 3D subvolumes, respectively.

Global mappingLocal mapping
LBP d × LBP#pat (N × d) × LBP#pat
LBP-TOP1 × (3 × LBP#pat) N′ × (3 × LBP#pat)

2.3.1. Global

Global mapping extracts the final descriptors from the 2D feature image for LBP and 3D volume for LBP-TOP. Therefore, for a volume with d slices, the global-LBP mapping will lead to the extraction of d elements, while the global-LBP-TOP represents the whole volume as a single element. The global mapping for 2D images and 3D volume is shown in Figures 5(a) and 5(b).
Figure 5

Graphical representation of the feature extraction: (a) extraction of LBP for global mapping, (b) extraction of LBP-TOP for global mapping, (c) extraction of LBP for local mapping, and (d) extraction of LBP-TOP for local mapping.

2.3.2. Local

Local mapping extracts the final descriptors from a set of (m × m) 2D patches for LBP and a set of (m × m × m) subvolumes for LBP-TOP. Given N and N′ as the total number of 2D patches and 3D subvolumes, respectively, the local-LBP approach provides N × d elements, while local-LBP-TOP provides N′ elements. This mapping is illustrated in Figures 5(c) and 5(d).

2.4. Feature Representation

Two strategies are used to describe each OCT volume's texture.

2.4.1. Low-Level Representation

The texture descriptor of an OCT volume is defined as the concatenation of the LBP histograms with the global mapping. The LBP histograms are extracted from the previously computed LBP maps (see Section 2.2). Therefore, the LBP-TOP final descriptor is computed through the concatenation of the LBP histograms of the three orthogonal planes with the final size of 3 × LBP#pat. More precisely, an LBP histogram is computed for each set of LBP maps (x-z) plane, (x-y) plane, and (y-z) plane, respectively. Similarly, the LBP descriptor is defined through concatenation of the LBP histograms per each (x-z) slice with the final size of d × LBP#pat.

2.4.2. High-Level Representation

The concatenation of histograms employed in the low-level representation in conjunction with either global or local mapping can lead to a high-dimensional feature space. For instance, local mapping results in a size of N × d × LBP#pat for the final LBP descriptor and N′ × LBP#pat for the final LBP-TOP descriptor, where N and N′ are the total number of 2D patches and 3D subvolumes, respectively. High-level representation simplifies this high-dimensional feature space into a more discriminant lower space. BoW approach is used for this purpose [13]. This model represents the features by creating a codebook or visual dictionary, from the set of low-level features. The set of low-level features are clustered using k-means to create the codebook with k clusters or visual words. After creating the codebook from the training set, the low-level descriptors are replaced by their closest word within the codebook. The final descriptor is a histogram of size k which represents the codebook occurrences for a given mapping.

2.5. Classification

The last step of our framework consists in the classification of SD-OCT volumes as normal or DME. For that matter, five different classifiers are used: (i) k-Nearest Neighbor (NN), (ii) Logistic Regression (LR) [19], (iii) Random Forest (RF) [20], (iv) Gradient Boosting (GB) [21, 22], and (v) Support Vector Machines (SVM) [23, 24]. Details regarding the parameters used in our experiments are provided in Section 3.

3. Experiments

A set of three experiments is designed to test the influence of the different blocks of the proposed framework in comparison to our previous work [12]. These experiments are designed as follows: Table 4 reports the experiments which have been carried out in [12] as a baseline and outlines the complementary experimentation here proposed. The reminder of this section details the common configuration parameters across the experiments, while the detailed explanations are presented in the following subsections.
Table 4

The outline and summary of the performed experiments.

CommonDatasetPreprocessingFeaturesMappingRepresentationClassificationEvaluation
SERINLMLBP, LBP-TOP   P = {8,16,24}   R = {1,2, 3}Leave-One-Patient Out Cross-Validation (LOPO-CV) SE, SP
Baseline [12]Goal: evaluation of features, mapping, and representation+ Duke~~ Global Local BoW HistogramRF+ comparison with [10]

Experiment 1 Goal: finding the optimum number of words~+F +F+A~ Global Local BoW   kK LR+ACC, F1 score (F1)

Experiment 2 Goal: evaluation of different preprocessing and classifiers for high-level features~+F +F+A~ Global Local BoW  Optimal k 3-NN RF SVM GB~

Experiment 3 Goal: evaluation of different preprocessing and classifiers for low-level features~+F +F+A~ Global Histogram3-NN RF SVM GB~

 ~ indicates that common configuration applies.

Experiment 1 evaluates the effects of number of words used in BoW (high-level representation). Experiment 2 evaluates the effects of different preprocessing steps and classifiers on high-level representation. Experiment 3 evaluates the effects of different preprocessing steps and classifiers on low-level representation. All the experiments are performed using a private dataset (see Section 3.1) and are reported as presented in Section 3.2. In all the experiments, LBP and LBP-TOP features are extracted using both local and global mapping for different sampling points of 8, 16, and 24 for radius of 1, 2, and 3 pixels, respectively. The partitioning for local-mapping is set to (7 × 7)-pixel patch for 2D LBP and (7 × 7 × 7)-pixel subvolume for LBP-TOP.

3.1. SERI Dataset

This dataset was acquired by the Singapore Eye Research Institute (SERI), using CIRRUS™ (Carl Zeiss Meditec, Inc., Dublin, CA) SD-OCT device. The dataset consists of 32 OCT volumes (16 DME and 16 normal cases). Each volume contains 128 B-scan with resolution of 512 × 1024 pixels. All SD-OCT images are read and assessed by trained graders and identified as normal or DME cases based on evaluation of retinal thickening, hard exudates, intraretinal cystoid space formation, and subretinal fluid.

3.2. Validation

All the experiments are evaluated in terms of Sensitivity (SE) and Specificity (SP) using the LOPO-CV strategy, in line with [12]. SE and SP are statistics driven from the confusion matrix as depicted in Figure 6. The SE evaluates the performance of the classifier with respect to the positive class, while the SP evaluates its performance with respect to negative class. The use of LOPO-CV implies that, at each round, a pair of DME-normal volumes is selected for testing while the remaining volumes are used for training. Subsequently, no SE or SP variance can be reported. However, LOPO-CV strategy has been adopted despite this limitation due to the reduced size of the dataset.
Figure 6

Evaluation metrics: (a) confusion matrix and (b) SE-SP.

3.3. Experiment 1

This experiment intends to find the optimal number of words and its effect on the different configurations (i.e., preprocessing and feature representation), on the contrary to [12], where the codebook size was arbitrarily set to k = 32. Several preprocessing strategies are used: (i) NLM, (ii) a combination of NLM and flattening (NLM+F), and (iii) a combination of NLM, flattening, and aligning (NLM+F+A). LBP and LBP-TOP descriptors are detected using the default configuration. Volumes are represented using BoW, where the codebook size ranges within k ∈ {10,20,30,…, 100,200,…, 500, 1000}. Finally, the volumes are classified using LR. The choice of this linear classifier avoids the case that the results get boosted by the classifier. In this manner, any improvement would be linked to the preprocessing and the size of the codebook. The usual build of the codebook consists of clustering the samples in the feature space using k-means (see Section 2.4). However, this operation is rather computationally expensive and the convergence of the k-means algorithm for all codebook sizes is not granted. Nonetheless, Nowak et al. [25] pointed out that randomly generated codebooks can be used at the expense of accuracy. Thus, the codebook is randomly generated since the final aim is to assess the influence of the codebook size and not the performance of the framework. For this experiment, the codebook building is carried out using random initialization using k-means++ algorithm [26], which is usually used as a k-means initialization algorithm. For this experiment, SE and SP are complemented with ACC and F1 score (see (2)). ACC offers an overall sense of the classifier performance, and F1 illustrates the trade-off between SE and precision. Precision or positive predictive value is a measure of algorithm exactness and is defined as a ratio of True Positive over the total predicted positive samples: Table 6 in Appendix shows the results obtained for the optimal dictionary size while the complete set of all ACC and F1 graphics can be found at [27]. According to the obtained results, it is observed that the optimum number of words is smaller for local-LBP features in comparison to local-LBP-TOP and global-LBP, respectively. Using LR classifier, the best performances were achieved using local-LBP with 70 words (SE and SP of 75.0%) and local-LBP-TOP with 500 words (SE and SP of 75.0% as well). These results are shown in bold in Table 6 in Appendix.
Table 6

Experiment 1—optimum number of words for each configuration as a result of LR classification, for high-level feature extraction of global and local-LBP, and local-LBP-TOP features with different preprocessing. The preprocessing includes NF, F, and F+A. The achieved performance is indicated in terms of ACC, F1, SE, and SP.

FeaturesPreprocessing{8,1}{16,2}{24,3}
ACC%F1%SE%SP%Number of wordsACC%F1%SE%SP%Number of wordsACC%F1%SE%SP%Number of words
Global-LBPNF81.278.568.793.750062.558.056.262.58062.562.562.562.580
F71.971.068.775.040068.766.762.575.030068.766.762.575.0300
F+A71.971.068.775.050071.971.068.775.020075.068.768.768.7500

Local-LBPNF 75.0 75.0 75.0 75.0 70 65.664.562.568.79062.560.056.268.730
F75.073.368.781.23071.861.068.775.07062.562.562.562.5100
F+A75.069.062.581.24071.971.068.775.020068.766.768.762.510

Local-LBP-TOPNF68.768.768.768.7400 75.0 75.0 75.0 75.0 500 71.971.068.775.060
F68.768.768.768.730068.766.762.575.05075.076.581.268.780
F+A75.073.368.781.210075.073.368.781.29075.069.062.581.270

3.4. Experiment 2

This experiment explores the improvement associated with (i) different preprocessing methods and (ii) using larger range of classifiers (i.e., linear and nonlinear) on the high-level representation. All the preprocessing stages are evaluated (NLM, NLM+F, and NLM+F+A). In this experiment, the codebooks for the BoW representation of LBP and LBP-TOP features are computed using regular k-means algorithm which is initialized using k-means++, where k is chosen according to the findings of Experiment 1. Finally, the volumes are classified using k-NN, RF, GB, and SVM. The k-NN classifier is used in conjunction with the 3 nearest neighbors rule to classify the test set. The RF and GB classifiers are trained using 100 unpruned trees, while SVM classifier is trained using an RBF kernel and its parameters C and γ are optimized through grid-search. Complete list of the obtained results from this experiment is shown in Table 7 in Appendix. Despite the fact that highest performances are achieved when NLM+F or NLM+F+A is used, most configurations decline when applied with extra preprocessing stages. The best results are achieved using SVM followed by RF.

3.5. Experiment 3

This experiment replicates Experiment 2 for the case of low-level representation of LBP and LBP-TOP features extracted using global mapping. The obtained results from this experiment are listed in Table 8 in Appendix. In this experiment, flattening the B-scan boosts the results of the best performing configuration. However, its effects is not consistent across all the configurations. RF has a better performance by achieving better SE (81.2%, 75.0%, and 68.7%), while SVM achieves the highest SP (93.7%), see Table 8 in the Appendix. In terms of classifier, RF has a better performance than the others despite the fact that the highest SP is achieved using SVM.

4. Results and Discussion

Table 5 combines the obtained results from Section 3 with those reported by Lemaitre et al. [12], while detailing the frameworks configurations. This table shows the achieved performances with SE higher than 55%.
Table 5

Summary of all the results in descending order.

LineExperimentEvaluationPre-processingFeat. DetectionMappingFeat. RepresentationClassifierBoW
SESPType{8,1}{16,2}{24,3}
1281.293.7NLM+FLBPLocalHighSVM
2275.093.7NLM+F+ALBPLocalHighSVM
3275.093.7NLMLBPLocalHighSVM
4275.0100NLMLBP-TOPLocalHighSVM
5281.287.5NLMLBP-TOPLocalHighSVM
6281.287.5NLM+F+ALBP-TOPLocalHighRF
7281.281.2NLMLBPLocalHighRF
8381.281.2NLMLBP-TOPGlobalLowRF
9281.281.2NLM+FLBP-TOPLocalHighSVM
10381.281.2NLM+F+ALBP-TOPGlobalLowGB
11381.281.2NLM+FLBP-TOPGlobalLowRF
12275.087.5NLMLBPLocalHigh k-NN
13Lemaitre et al. [12]75.087.5NLMLBPLocalHighRF
14Lemaitre et al. [12]75.087.5NLMLBP-TOPGlobalLowRF
15268.793.7NLMLBPGlobalHighRF
1637581.2NLM+F+ALBP-TOPGlobalLowRF
17268.781.2NLMLBP-TOPLocalHighRF
18362.593.7NLMLBP-TOPGlobalLowSVM
19368.787.5NLMLBP-TOPGlobalLowRF
20368.781.2NLMLBP-TOPGlobalLowRF
21375.075.0NLMLBP-TOPGlobalLowRF
22368.775.0NLM+FLBP-TOPGlobalLowSVM
23356.275.0NLMLBPGlobalLowRF
24356.275.0NLM+FLBPGlobalLow k-NN
25356.275.0NLM+F+ALBPGlobalLow k-NN
26Venhuizen et al. [10]61.558.8
The obtained results indicate that expansion and tuning of our previous framework improve the results. Tuning the codebook size, based on the finding of Experiment 1, leads to an improvement of 6% in terms of SE (see Table 5 at lines 7 and 13). Furthermore, the fine-tuning of our framework (see Section 2) also leads to an improvement of 6% in both SE and SP (see Table 5 at lines 1 and 13). Our framework also outperforms the proposed method of [10] with an improvement of 20% and 36% in terms of SE and SP, respectively. Note that although the effects of preprocessing are not consistent through all the performances, the best results are achieved with NLM+F and NLM+F+A configurations as preprocessing stages. In general, the configurations presented in Experiment 2 outperform the others, in particular the high-level representation of locally mapped features with an SVM classifier. Focusing on the most desirable radius and sampling point configuration, smaller radius and sampling points are more effective in conjunction with local mapping, while global mapping benefits from larger radius and sampling points.

5. Conclusions

The work presented here addresses automatic classification of SD-OCT volumes as normal or DME. In this regard, an extensive study is carried out covering the (i) effects of different preprocessing steps, (ii) influence of different mapping and feature extraction strategies, (iii) impact of the codebook size in BoW, and (iv) comparison of different classification strategies. While outperforming the previous studies [10, 12], the obtained results in this research showed the impact and importance of optimal codebook size, the potential of 3D features, and high-level representation of 2D features while extracting from local patches. The strengths of SVM while being used along with BoW approach and RF classifier while being used with global mapping were shown. In terms of preprocessing steps, although the highest performances are achieved while alignment and flattening were used in the preprocessing, it was shown that the effects of these extra steps are not consistent for all the cases and do not guarantee a better performance. Several avenues for future directions can be explored. The flattening method proposed by Liu et al. flattens roughly the RPE due to the fact that the RPE is not segmented. Thus, in order to have a more accurate flattening preprocessing, the RPE layer should be presegmented as proposed by Garvin et al. [28]. In this work, the LBP invariant to rotation was used and the number of patterns encoded is reduced. Once the data are flattened, the nonrotation invariant LBP could be studied since this descriptor encodes more patterns. In addition to LBP, other feature descriptors can be included in the framework.
(a)
FeaturesPreprocessing k-NNSVM
{8,1}{16,2}{24,3}{8,1}{16,2}{24,3}
SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%
Global-LBPNF43.793.743.787.543.762.568.787.562.562.550.056.2
F43.756.250.075.062.556.256.256.256.275.056.268.7
FA56.262.543.781.268.756.256.268.768.768.756.275.0

Local-LBPNF 75.0 87.5 50.068.743.743.7 75.0 93.7 50.075.056.256.2
F 56.2 56.2 50.050.050.043.7 81.2 93.7 68.768.768.775.0
FA 56.2 43.7 50.075.050.062.5 75.0 93.7 75.068.768.768.7

Local-LBP-TOPNF56.275.056.275.062.556.2 81.2 87.5 75.0 100 56.275.0
F62.543.737.568.743.762.5 81.2 81.2 75.0 68.7 81.268.7
F+A56.256.268.750.043.762.5 62.5 75.0 68.7 75.0 62.581.2
(b)
FeaturesPreprocessingRFGB
8riu2 16riu2 24riu2 8riu2 16riu2 24riu2
SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%
Global-LBPNF 68.7 93.7 43.762.550.068.756.250.037.531.250.043.7
F 56.2 50.0 56.275.050.075.050.056.256.275.043.762.5
FA 68.7 50.0 56.262.562.556.256.250.068.750.043.775.0

Local-LBPNF 81.2 81.2 62.556.256.256.275.062.568.787.550.075.0
F 56.2 81.2 62.568.768.762.568.775.050.075.050.062.5
FA 68.7 62.5 62.668.743.743.756.250.068.756.250.050.0

Local-LBP-TOPNF68.762.5 68.7 81.2 68.768.737.568.762.581.262.550.0
F50.062.5 62.5 62.5 43.775.050.056.243.762.550.062.5
F+A50.062.5 81.2 87.5 50.068.756.262.581.268.775.068.7

(a)  

FeaturesPreprocessing k-NNSVM
{8,1}{16,2}{24,3}{8,1}{16,2}{24,3}
SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%
Global-LBPNF37.550.025.050.037.568.756.262.556.243.756.268.7
F62.550.056.275.062.568.775.068.762.562.562.568.7
FA56.250.056.275.062.568.775.068.762.562.562.568.7

Global-LBP-TOPNF31.293.737.5100.037.581.262.575.0 62.5 93.7 56.287.5
F50.056.256.275.056.262.568.775.0 43.7 68.7 68.756.2
F+A75.043.756.243.768.750.068.762.5 62.5 56.2 56.268.7

(b)  

FeaturesPreprocessingRFGB
8riu2 16riu2 24riu2 8riu2 16riu2 24riu2
SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%SE%SP%
Global-LBPNF43.762.543.762.556.27543.743.743.737.537.531.25
F56.256.268.762.562.568.72556.250.043.725.043.7
F+A65.256.250.050.056.268.743.7562.562.550.031.231.2

Global-LBP-TOPNF56.268.7 68.7 87.5 68.7 81.2 68.768.775.050.0 56.2 43.7
F56.262.5 81.2 68.7 81.2 81.2 56.262.562.568.7 68.7 81.2
F+A68.762.5 75.0 68.7 75.0 81.2 56.243.762.562.5 75.0 75.0
  14 in total

1.  Rotation-invariant image and video description with local binary pattern features.

Authors:  Guoying Zhao; Timo Ahonen; Jiří Matas; Matti Pietikäinen
Journal:  IEEE Trans Image Process       Date:  2011-11-11       Impact factor: 10.856

2.  Validating retinal fundus image analysis algorithms: issues and a proposal.

Authors:  Emanuele Trucco; Alfredo Ruggeri; Thomas Karnowski; Luca Giancardo; Edward Chaum; Jean Pierre Hubschman; Bashir Al-Diri; Carol Y Cheung; Damon Wong; Michael Abràmoff; Gilbert Lim; Dinesh Kumar; Philippe Burlina; Neil M Bressler; Herbert F Jelinek; Fabrice Meriaudeau; Gwénolé Quellec; Tom Macgillivray; Bal Dhillon
Journal:  Invest Ophthalmol Vis Sci       Date:  2013-05-01       Impact factor: 4.799

3.  Speckle in optical coherence tomography.

Authors:  J M Schmitt; S H Xiang; K M Yung
Journal:  J Biomed Opt       Date:  1999-01       Impact factor: 3.170

4.  Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images.

Authors:  Pratul P Srinivasan; Leo A Kim; Priyatham S Mettu; Scott W Cousins; Grant M Comer; Joseph A Izatt; Sina Farsiu
Journal:  Biomed Opt Express       Date:  2014-09-12       Impact factor: 3.732

5.  Automated macular pathology diagnosis in retinal OCT images using multi-scale spatial pyramid and local binary patterns in texture and shape encoding.

Authors:  Yu-Ying Liu; Mei Chen; Hiroshi Ishikawa; Gadi Wollstein; Joel S Schuman; James M Rehg
Journal:  Med Image Anal       Date:  2011-06-22       Impact factor: 8.545

6.  Comparison of Prevalence of Diabetic Macular Edema Based on Monocular Fundus Photography vs Optical Coherence Tomography.

Authors:  Yu T Wang; Mongkol Tadarati; Yulia Wolfson; Susan B Bressler; Neil M Bressler
Journal:  JAMA Ophthalmol       Date:  2016-02       Impact factor: 7.389

Review 7.  The impact of diabetic retinopathy on health-related quality of life.

Authors:  Sanjay Sharma; Alejandro Oliver-Fernandez; Wei Liu; Patricia Buchholz; John Walt
Journal:  Curr Opin Ophthalmol       Date:  2005-06       Impact factor: 3.761

8.  Photocoagulation for diabetic macular edema. Early Treatment Diabetic Retinopathy Study report number 1. Early Treatment Diabetic Retinopathy Study research group.

Authors: 
Journal:  Arch Ophthalmol       Date:  1985-12

9.  Global prevalence of diabetes: estimates for the year 2000 and projections for 2030.

Authors:  Sarah Wild; Gojka Roglic; Anders Green; Richard Sicree; Hilary King
Journal:  Diabetes Care       Date:  2004-05       Impact factor: 19.112

10.  Automatic segmentation of seven retinal layers in SDOCT images congruent with expert manual segmentation.

Authors:  Stephanie J Chiu; Xiao T Li; Peter Nicholas; Cynthia A Toth; Joseph A Izatt; Sina Farsiu
Journal:  Opt Express       Date:  2010-08-30       Impact factor: 3.894

View more
  14 in total

1.  Transfer learning based classification of optical coherence tomography images with diabetic macular edema and dry age-related macular degeneration.

Authors:  S P K Karri; Debjani Chakraborty; Jyotirmoy Chatterjee
Journal:  Biomed Opt Express       Date:  2017-01-04       Impact factor: 3.732

2.  Deep learning-based automated detection of retinal diseases using optical coherence tomography images.

Authors:  Feng Li; Hua Chen; Zheng Liu; Xue-Dian Zhang; Min-Shan Jiang; Zhi-Zheng Wu; Kai-Qian Zhou
Journal:  Biomed Opt Express       Date:  2019-11-11       Impact factor: 3.732

3.  AOCT-NET: a convolutional network automated classification of multiclass retinal diseases using spectral-domain optical coherence tomography images.

Authors:  Ali Mohammad Alqudah
Journal:  Med Biol Eng Comput       Date:  2019-11-14       Impact factor: 2.602

4.  Self-supervised patient-specific features learning for OCT image classification.

Authors:  Leyuan Fang; Jiahuan Guo; Xingxin He; Muxing Li
Journal:  Med Biol Eng Comput       Date:  2022-08-05       Impact factor: 3.079

5.  Deep learning is effective for the classification of OCT images of normal versus Age-related Macular Degeneration.

Authors:  Cecilia S Lee; Doug M Baughman; Aaron Y Lee
Journal:  Ophthalmol Retina       Date:  2017-02-13

6.  Machine learning techniques for diabetic macular edema (DME) classification on SD-OCT images.

Authors:  Khaled Alsaih; Guillaume Lemaitre; Mojdeh Rastgoo; Joan Massich; Désiré Sidibé; Fabrice Meriaudeau
Journal:  Biomed Eng Online       Date:  2017-06-07       Impact factor: 2.819

7.  Automatic diagnosis of macular diseases from OCT volume based on its two-dimensional feature map and convolutional neural network with attention mechanism.

Authors:  Yankui Sun; Haoran Zhang; Xianlin Yao
Journal:  J Biomed Opt       Date:  2020-09       Impact factor: 3.170

8.  Precise higher-order reflectivity and morphology models for early diagnosis of diabetic retinopathy using OCT images.

Authors:  A Sharafeldeen; M Elsharkawy; F Khalifa; A Soliman; M Ghazal; M AlHalabi; M Yaghi; M Alrahmawy; S Elmougy; H S Sandhu; A El-Baz
Journal:  Sci Rep       Date:  2021-02-25       Impact factor: 4.379

9.  A Multitask Deep-Learning System to Classify Diabetic Macular Edema for Different Optical Coherence Tomography Devices: A Multicenter Analysis.

Authors:  Fangyao Tang; Xi Wang; An-Ran Ran; Carmen K M Chan; Mary Ho; Wilson Yip; Alvin L Young; Jerry Lok; Simon Szeto; Jason Chan; Fanny Yip; Raymond Wong; Ziqi Tang; Dawei Yang; Danny S Ng; Li Jia Chen; Marten Brelén; Victor Chu; Kenneth Li; Tracy H T Lai; Gavin S Tan; Daniel S W Ting; Haifan Huang; Haoyu Chen; Jacey Hongjie Ma; Shibo Tang; Theodore Leng; Schahrouz Kakavand; Suria S Mannil; Robert T Chang; Gerald Liew; Bamini Gopinath; Timothy Y Y Lai; Chi Pui Pang; Peter H Scanlon; Tien Yin Wong; Clement C Tham; Hao Chen; Pheng-Ann Heng; Carol Y Cheung
Journal:  Diabetes Care       Date:  2021-07-27       Impact factor: 17.152

10.  Classification of healthy and diseased retina using SD-OCT imaging and Random Forest algorithm.

Authors:  Md Akter Hussain; Alauddin Bhuiyan; Chi D Luu; R Theodore Smith; Robyn H Guymer; Hiroshi Ishikawa; Joel S Schuman; Kotagiri Ramamohanarao
Journal:  PLoS One       Date:  2018-06-04       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.