Literature DB >> 34945957

Multilevel Deep Feature Generation Framework for Automated Detection of Retinal Abnormalities Using OCT Images.

Prabal Datta Barua1,2,3, Wai Yee Chan4, Sengul Dogan5, Mehmet Baygin6, Turker Tuncer5, Edward J Ciaccio7, Nazrul Islam8, Kang Hao Cheong9, Zakia Sultana Shahid10, U Rajendra Acharya11,12,13.   

Abstract

Optical coherence tomography (OCT) images coupled with many learning techniques have been developed to diagnose retinal disorders. This work aims to develop a novel framework for extracting deep features from 18 pre-trained convolutional neural networks (CNN) and to attain high performance using OCT images. In this work, we have developed a new framework for automated detection of retinal disorders using transfer learning. This model consists of three phases: deep fused and multilevel feature extraction, using 18 pre-trained networks and tent maximal pooling, feature selection with ReliefF, and classification using the optimized classifier. The novelty of this proposed framework is the feature generation using widely used CNNs and to select the most suitable features for classification. The extracted features using our proposed intelligent feature extractor are fed to iterative ReliefF (IRF) to automatically select the best feature vector. The quadratic support vector machine (QSVM) is utilized as a classifier in this work. We have developed our model using two public OCT image datasets, and they are named database 1 (DB1) and database 2 (DB2). The proposed framework can attain 97.40% and 100% classification accuracies using the two OCT datasets, DB1 and DB2, respectively. These results illustrate the success of our model.

Entities:  

Keywords:  OCT image classification; diabetic macular edema (DME); digital image processing; hybrid deep feature generation; iterative feature selection

Year:  2021        PMID: 34945957      PMCID: PMC8700736          DOI: 10.3390/e23121651

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

Optical coherence tomography (OCT) is an imaging technique using low coherence light sources to produce high-resolution cross-sectional images of the retina and optic nerve. It is useful to diagnose various pathologies causing optic atrophy or optic nerve swelling [1]. It displays the retinal layers in three dimensions and allows the evaluation of changes in the macula. It is widely used in the detection of diabetic macular edema (DME) and age-related macular degeneration (AMD) [2,3]. DME is an eye disease caused by damage or proliferative blood vessels in the retina, particularly at the macula [4]. DME, which is an important complication of diabetes in the eyes, can lead to irreversible vision loss. On the other hand, AMD is degeneration at the macula region caused by various external risk factors, such as age, genetic variants, family history, smoking, etc. It usually occurs in individuals over the age of 60 years [5]. The prevalence rate of AMD increases with age and is the leading cause of blindness in the elderly [6]. Vision loss is not observed in the early stage of DME and AMD. However, it causes vision loss and blindness in advanced stages [7]. Therefore, early diagnosis of DME and AMD is very important. However, analyzing each section in OCT individually is time consuming and increases the workload of clinicians [8]. Many machine learning techniques have been proposed for the automatic classification and abnormality detection of OCTs [9,10,11]. In these techniques, segmentation of the retinal layer or classification of images with the same characteristics is generally studied [4]. These techniques offer high-performance analysis support at a low cost. Deep learning (DL) methods, which work effectively using large databases, are of immense interest in the literature [12,13,14,15]. In this work, a novel intelligent system is proposed using the most effective DL model. One way to achieve this is to use multiple deep network based feature extractors, and in this study, we present a new hybrid and multileveled deep feature generator. Moreover, we employ hybrid pooling methods to generate compressed images. The utilized pooling methods are maximum, max-min, and max-mean. The routing problem is solved by using these three pooling methods together, and multileveled deep features are generated with three simple pooling functions. The main purpose of the IRF selector is to automatically choose the most appropriate feature vector, and the chosen optimal feature vector is fed as input of the quadratic SVM classifier [16]. As discussed above, our work presents a deep feature engineering framework developed with two widely used public OCT image datasets, and the key contributions of our model are the following: An intelligent deep feature generator is presented using transfer learning. Using transfer learning, 1000 features are generated from each pre-trained CNN, 18 pre-trained networks are involved in this framework, and an 18-feature generation function is proposed using these pre-trained CNNs and three maximum pooling methods. The proposed framework generates the best deep features to attain the best classification rates. An effective learning model is presented by deploying the proposed multiple CNNs based on a deep feature generator, iterative feature selector (IRF), and classification with SVM. This learning model is developed using two public OCT image datasets. It attained the highest classification performance using both OCT datasets. The rest of this paper is organized as follows. The literature review is tabulated and discussed in Section 2. The material (datasets) and the proposed DL-based feature engineering method are presented in Section 3. The calculated classification results and performance analysis are given in Section 4. The results and findings are discussed in Section 5. Section 6 ends with the main conclusions of the research and gives an outlook on future directions.

2. Literature Review

Deep learning methods are widely used to diagnose many different diseases and have attained high performance [17,18,19,20]. The convolutional neural network (CNN) is one of the most widely used models in deep learning and is often preferred for OCT image analysis [21]. The classification of OCT images has also been conducted using handcrafted methods. Rajagopalan et al. [22] proposed a deep CNN framework with OCT images for the diagnosis and classification of drusen macular degeneration (DMD), DME and normal. The framework achieved a classification accuracy of 95.7%. Alsaih et al. [23] proposed a classification (DME vs. normal) pipeline with spectral domain optical coherence tomography (SD-OCT) images. The pipeline comprised pre-processing, feature extraction, feature representation, and feature classification. With principal component analysis and a linear-support vector machine (SVM), they achieved a sensitivity of 87.5% and a specificity of 87.5%. Sunija et al. [24] proposed a deep CNN method that has six convolutional blocks for the classification of DME, drusen, choroidal neovascularization (CNV) from normal OCT images. The proposed method achieved an accuracy of 99.69%. Das et al. [25] proposed a method that introduces a multi-scale deep feature fusion-based classification approach using CNN for classification of DME and two stages of AMD (drusen and CNV) from healthy OCT images. The proposed method achieved an average sensitivity, specificity, and accuracy of 99.6%, 99.87% and 99.6% respectively. In elsewhere, Lemaitre et al. proposed a method for automatic identification of patients with DME versus normal subjects based on local binary patterns features to describe the texture of OCT images and they compared different local binary pattern feature extraction approaches to compute a single signature for the whole OCT volume [26]. The proposed method achieved a specificity of 75.00% and sensitivity of 87.50%. Rong et al. [2] proposed a surrogate-assisted classification method to classify retinal OCT images based on CNNs. The proposed method achieved a classification (AMD, DME and normal) accuracy of 100%. Tayal et al. [27] presented an automatic diagnostic tool based on a deep-learning framework (three different CNN models) for the classification of CNV, DME, drusen and normal based on images of OCT scans. The diagnostic tool obtained a classification accuracy of 96.5%. Srinivasan et al. [28] presented an automated algorithm that utilizes multiscale histograms of oriented gradient descriptors as feature vectors of a support vector machine-based classifier for the detection of retinal diseases via OCT imaging. Their classifier correctly identified 100% with AMD, 100% with DME and 86.67% normal. Hussain et al. [29] proposed a model with random forest for the detection of DME, AMD and normal, using retinal features from SD-OCT images. The classification method uses features such as the thickness of the retina and the thickness of the individual retinal layers. They obtained an accuracy of more than 96%. The studies described above are summarized in Table 1.
Table 1

Summary of recent studies involving OCT images.

StudyMethodPurposeResults (%)
Rajagopalan et al. [22]CNNDetecting DMD, DME and normal using OCT images Acc: 95.70
Alsaih et al. [23]Local binary patterns and histograms of oriented gradientsClassification of DME and normal using SD-OCT imagesSpe: 87.50Sen: 87.50
Sunija et al. [24]CNNClassification of CNV, DME, Drusen and normal using OCT imagesAcc: 99.69
Das et al. [25]CNNClassification of DME, Drusen, CNV and normal using OCT imagesAcc: 99.60
Lemaitre et al. [26]Local binary patternsIdentification of patients with DME versus normal subjects with SD-OCT imagesSpe: 75.00Sen: 87.50
Rong et al. [2]CNNClassification of AMD, DME and normal using OCT imagesAcc: 100.0
Tayal et al. [27]CNNIdentification of CNV, DME, Drusen and normal using OCT imagesAcc: 96.50
Srinivasan et al. [28]CNNClassification of normal, AMD and DME with SD-OCT imagesAcc:100.0 AMD100.0 DME86.67 normal
Hussain et al. [29]Random forest techniqueClassification of normal, AMD and DME with SD-OCT imagesAcc: 97.33 for two classes case (DME and normal)Acc: 95.58 for three classes case (DME, AMD, and normal)
DL models have generally been used to develop automatic classification methods [30,31,32]. The deep models have unique benefits and have achieved good results for computer vision problems. As seen from Table 1, CNNs are the flagship of OCT image classification, but there are various types of CNNs in the literature, and each CNN has its own peculiarity, resulting in variable performance on OCT image datasets. The main goal of hand-crafted methods is to create discriminative features for attaining high classification performance with low time complexity. However, they cannot attain high performance on large/complex datasets. To overcome this problem, deep learning models have been used to classify OCT images. Deep models can attain high performance but there are many deep networks, and each of them has individual performance. The primary purpose of our framework is to use the activity of 18 widely known pre-trained CNNs. In addition, a package learning model is presented without using the trial-and-error method. With this framework, there is no longer the need to propose many methods based on the various CNN models for the classification of OCT images. A general image classification model is now proposed by using these networks together. This model can also be used to solve other computer vision problems, as it is a self-organizing framework. This framework can select the most suitable CNNs based on the context of the problem.

3. Material and Method

3.1. Material

Two public image datasets (DB1 and DB2) were used to develop the proposed retinal abnormality classification system using OCT images. The details of these databases are given below.

3.1.1. First OCT Image Dataset (DB1)

The used OCT image dataset comprised 11,000 OCT images with four classes, named choroidal neovascularization (CNV), diabetic macular edema (DME), Drusen, and normal (https://data.mendeley.com/datasets/rscbjbr9sj/3, accessed on 10 October 2021). There were 2750 OCT images in each of the four categories [33,34]. In this OCT database, 10,000 images out of 11,000 were used for training, and 1000 were used for testing. In the test database, 250 images were used in each category. The dimension of each image was 496 × 1024, and they were stored in jpeg format. Sample images of this dataset are shown in Figure 1.
Figure 1

Sample OCT images of the used OCT dataset (CNV—choroidal neovascularization, DME—diabetic macular edema). Images reproduced from ref. [33]. (a) CNV sample image, (b) DME sample image, (c) a sample image Drusen class, (d) a sample of healthy OCT.

3.1.2. Second Image Dataset (DB2)

The utilized second database is named DB2, and it contained 3194 images belonging to three categories: age-related macular degeneration (AMD), DME, and healthy classes (https://people.duke.edu/~sf59/Srinivasan_BOE_2014_dataset.htm, accessed on 10 October 2021). It contained 686 AMD, 1101 DME, and 1407 healthy images [28]. The images were stored in tiff format. Sample images of this dataset are depicted in Figure 2.
Figure 2

Sample OCT images of DB2 dataset (AMD—age related macular degeneration; DME—diabetic macular edema). Images reproduced from ref. [28]. (a) AMD disorder, (b) DME disorder, (c) healthy.

3.2. The Proposed Framework

A new framework is presented to select the best deep features and correctly classify OCT images. The proposed framework consists of three fundamental phases: deep feature extraction using intelligent deep transfer learning, iterative feature choosing, and classification. Pseudocode for this framework is provided in Algorithm 1. The MATLAB (2020b) programming environment was used to implement this algorithm, and 18 pre-trained networks were included, using Get Add-Ons options. A graphical summary of this framework is presented in Figure 3. More explanations about this framework are provided in Section 3.2.1, Section 3.2.2 and Section 3.2.3.
Figure 3

Graphical illustration of the proposed multilevel fused/hybrid deep feature extraction-based OCT image classification model. Maximum, max-min, and max-mean pooling algorithms were used to generate decomposed images (c1, c2, …, c9). By employing transfer learning, 10,000 features were generated from each pre-trained CNN. These networks were trained on the ImageNet dataset. This dataset contained about 1.2 million images belonging to 1000 classes. In this work, we have used the last fully connected layer of each network. Thus, we generated 1000 features for each image. An original and nine compressed images are fed to each pre-trained network. Thus, 10,000 features are generated from an OCT image. One thousand features are selected from the generated 10,000 features utilizing ReliefF, and 18 loss values are calculated in the misclassification rate calculation block. The top five feature vectors were selected using calculated loss values, and the last feature vector with a length of 5000 is determined using the selected feature vectors. The IRF function selected the top features for classification, and results are obtained from SVM with a 10-fold cross-validation strategy. The parameters used in each framework are tabulated in Table 2.

More details about the suggested model and steps involved in various steps are given below.

3.2.1. Deep Feature Extraction

The most important phase of the proposed framework is deep feature extraction, and this phase is a novel component of the model. For our study, 18 pre-trained commonly used CNNs, three pooling functions, ReliefF [35] and SVM [16] were used together to create the best feature vector. As can be seen from the used methods, the proposed feature extraction model contains machine learning components. This model’s main feature generation functions are pre-trained CNNs; the utilized pre-trained CNNs are listed in Table 3.
Table 3

Deep CNNs used for deep feature generation.

No.CNNFE LayerNo.CNNFE Layer
1ResNet18 fc100010NasNetMobilepredictions
2ResNet50 fc100011NasNetLargepredictions
3ResNet101 fc100012DenseNet201fc1000
4DarkNet19avg113InceptionV3predictions
5MobileNetV2Logits14InceptionResNetV2predictions
6DarkNet53conv5315GoogLeNetloss3-classifier
7Xceptionpredictions16AlexNetfc8
8EfficientNet b0MatMul17VGG16fc8
9ShuffleNetnode_20218VGG19fc8
In Table 3, the FE layer is the feature extraction layer, the last fully connected layer used for feature generation. In this work, we use the MATLAB programming environment and we use 18 pre-trained networks. Our main aim is to create a general feature generation framework. The results of all pre-trained networks are obtained, and this framework is used to choose the most valuable ones to solve computer vision problems. The general steps of this generator are summarized below. Read each OCT image. Decompose OCT images using multileveled and multiple pooling-based methods. In our study, maximum, max-min, and max-mean pooling algorithms are utilized as compression methods. The mathematical definitions of this compression method are given below. Furthermore, this compression method is defined in lines 07–15 of Algorithm 1. where is the ith compressed image, defines image, and are the maximum, max-mean, and max-min pooling functions, respectively. Moreover, the mathematical definitions of and are given below. Herein, and are arrays with a length of and they store average and minimum values of a matrix () with a size of . The function is the average value calculation function, and is the minimum value calculation function. Equations (5) and (6) define the proposed function, and Equations (7) and (8) explain function. A graphical representation of the three pooling functions used is depicted in Figure 4.
Figure 4

Graphical representation of the used pooling functions with a 3 × 3 dimension sample block. Maximum pooling selects the maximum value. Max-mean and max-min pooling functions select the maximum column, and according to this example, the maximum column is [8,9,25]. By using [8,9,25] vector, max-mean pooling finds , and max-min pooling selects 8 as a compressed value.

The general problem of the pooling models is the routing problem [36]. An example of the routing problem caused by pooling is given as follows. Only the peak value can be routed by using maximum pooling. In order to solve this problem using the available pooling methods, our proposed multiple pooling function is presented and by using our proposed pooling function, the average, minimum and maximum values are routed together. The utilized pooling functions take two parameters: input image and used size of the non-overlapping blocks. Nine compressed images are calculated using Equations (1)–(4). This compression model uses three different pooling methods to solve the routing problem, and it has three levels created, using maximum pooling with 2 × 2 sized non-overlapping blocks. Generate deep features from the compressed nine images and original OCT image with 18 pre-trained CNNs as a deep feature generator. where is the jth feature in the deep feature vector (this extractor generates 18 feature vectors with a length of 10,000), defines the number of used OCT images, and represents jth pre-trained deep feature generator. In Equations (5) and (6), the deep feature generator is defined using 18 pre-trained CNNs (see Table 3). Each pre-trained CNN is employed for the original OCT and compressed images to extract features. Each CNN generates 1000 features, and 10 images are utilized as input for each CNN. Therefore, 10,000 features are extracted in total from an OCT image. Select the most informative 1000 features from the extracted 10,000 features using the ReliefF selector. where is the sorted indexes of the jth feature vector using ReliefF () selector, is the actual output and is the selected 1000 features. Calculate the loss value using SVM classifier with 10-fold CV. Herein, is loss value, defines the SVM classifier and it requires three parameters. These parameters are feature vector, labels, and k value of the used cross-validation method. Select the best five feature vectors using the vector and merge the selected feature vectors. where is the last feature vector with a length of 5000, and is the index of the sorted loss values arranged by ascending terms. These six steps defined our presented intelligent deep feature generator.

3.2.2. Feature Selection Using Iterative ReliefF

In this phase, an iterative selector is preferred to select the most informative features. The feature selection aims to increase the classification ability and decrease the time complexity of the used classifier. To achieve both of these aims, various feature selectors are presented in the literature. ReliefF [37,38] is a commonly used feature selector, and it is weight based. It generates both negative and positive features. Negative weights are calculated from redundant features. Positive and larger weights are calculated from informative/discriminative features. The indexes of the feature are calculated using ReliefF weights. However, ReliefF cannot select the best feature vector without a trial-and-error method. Therefore, IRF was presented by Tuncer et al. [39] in 2021. IRF can select the best feature vector automatically, and it is a parametric selector. The parameters used are a number of features range and loss generators. These parameters are chosen as [100, 1000] and SVM with 10-fold CV, respectively. IRF selects 794 features from the extracted 5000 features as the length of an optimal feature vector. Steps of the used IRF are given below. Apply ReliefF to generate 5000 features and generate qualified indexes (). Select features iteratively by using feature range. The length of the first feature vector is chosen as 100, and the length of the last feature vector is selected as 1000. Therefore, 901 feature vectors are selected. Calculate the loss value of all selected feature vectors using the SVM classifier. Choose the best feature vector.

3.2.3. Classification

In this section, the last phase, classification, is performed using quadratic kernelled SVM with a 10-fold cross-validation strategy. The SVM classifier is one of the widely preferred traditional classifiers and has many kernels [16,40]. The Bayesian optimization technique is used to select the best kernel for this OCT image classification problem. The hyperparameters search range of the Bayesian optimizer is given as follows: multiclass method, One-vs.-One, One-vs.-All; box constrains, 1–1000; kernel scale, 0.001–1000; kernel, Gaussian, quadratic, linear, and cubic; and standardize, false or true. Options of this optimizer are as follows: a maximum number of iterations is 30, and the fitness function is the minimum misclassification ration. The utilized optimizer selects quadratic SVM as the optimum SVM, and this classifier is used in feature extraction, IRF, and classification. The selected hyper-parameters of the used classifier are as follows: Kernel: Quadratic (2nd degree polynomial), Kernel scale: Auto, Box constraint: 1, Standardize: True. The last step (classification step) is denoted below. Classify the selected features employing quadratic SVM with 10-fold CV (for DB2) or hold-out validation (for DB1).

4. Results

The proposed retinal abnormality classification using OCT images based on an intelligent hybrid deep feature generator and IRF was implemented on a simple configured computer. The 18 CNNs were trained on the ImageNet dataset, and each CNN generated 1000 features. This framework was implemented on the MATLAB (2020b) platform. The proposed framework is a parametric framework, and the parameters used are presented in Table 2.
Table 2

Phases and parameters used in our proposed method.

PhaseMethodParameter
Feature extractionMultiple multilevel pooling decompositionNumber of level: 3Pooling methods: maximum, max-mean and max-minNumber of compressed image: 9
Deep feature generation and feature merging18 pre-trained convolutional neural networks are used to extract deep features from fully connected layers of these networks.18 feature vectors with a length of 10,000 are created
Feature selection using ReliefFThe top 1000 features of 10,000 features generated are chosen.
Loss value calculationQuadratic SVM
Top feature vectors selectionThe top five feature vectors have been selected.
Feature selectionIterative ReliefFRange of iteration: [100, 1000]Loss value generator: Quadratic SVM
ClassificationSVMKernel function: PolynomialPolynomial order: 2Kernel scale: AutoBox constraint: 1Standardize: True
In this work, two databases were used to test our proposed transfer learning and the ReliefF-based framework. We included 18 CNNs in this model, and the best five CNNs were selected to create the final features. The performance parameters, namely accuracy (Acc), precision (Pre), Cohen Kappa (CK), F1-score (F1), Matthew coefficient correlation (MCC), and recall (Rec), were used to evaluate the performance of the developed model. The first dataset (DB1) dataset comprises 11,000 images and is a homogenous dataset. The confusion matrix of the proposed classification framework for DB1 is given in Figure 5.
Figure 5

Confusion matrix obtained using our proposed framework for DB1.

It can be noted from Figure 5 that the best results are obtained for DME and normal classes (our proposal reached 98.80% class-wise accuracy on these classes), and our framework attained 97.40% classification accuracy with this dataset (DB1). The DB2 dataset has 3194 images with three categories, and the results obtained using DB2 dataset are calculated, and the confusion matrix for the DB2 is denoted in Figure 6.
Figure 6

Confusion matrix obtained using our proposed framework for DB2.

It can be noted from Figure 6 that the proposed framework attained high accuracy (100%) with the DB2 dataset. The summary of the results obtained using our proposed model with DB1 and DB2 datasets is listed in Table 4.
Table 4

Summary of results obtained using two datasets.

Overall ResultDB1DB2
Accuracy97.40100
Precision97.40100
Cohen Kappa96.40100
F1-score97.40100
MCC96.53100
Recall96.53100
The time complexity of the proposed framework is calculated using theta () notation and is given below. Feature extraction: Feature selection: Classification: Total: It can be noted from the above expressions that the proposed model used multiple pooling and pre-trained networks-based deep feature extraction. Herein, defines the number of the used pre-trained networks, and is the size of the used images. The used multiple pooling function creates decomposed images, and the decomposed images have lower sizes. Therefore, the complexity is equal to . Furthermore, we used the best feature vector selection process in the feature extraction phase, and the complexity of this step is equal to , where defines the complexity of the used loss value (classifier). is the time complexity of the used pre-trained network. An iterative feature selector was used. Herein, is the number of iterations, and is the time complexity of the feature selector. These results demonstrate that our proposed framework has linear time complexity.

5. Discussion

This work proposes a new framework for retinal disorder detection using two OCT image datasets (DB1 and DB2) consisting of 11,000 (4 classes) and 3194 (3 classes) images, respectively. The proposed framework is comprised of an intelligent deep feature generator using 18 pre-trained CNNs and multilevel multiple pooling decomposition, IRF selector, and classification. The proposed intelligent feature extractor utilized 18 pre-trained networks as the feature generator. The graph of accuracy rates obtained using 18 pre-trained CNNs is denoted in Figure 7a, and the iterative feature selection process is shown in Figure 7b.
Figure 7

Accuracies and iterative feature selection. (a) Classification accuracies obtained using various deep pre-trained CNNs and (b) iterative feature selection process using IRF.

It can be noted from the results that the proposed framework selected deep features of 6th (DarkNet53), 5th (MobileNetV2), 4th (DarkNet19), 8th (EfficientNet b0), and 12th (DenseNet201) CNNs for DB1. DarkNet53 attained 94.84% classification accuracy with DB1. The selected CNNs for DB2 are 4th (DarkNet19), 5th (MobileNetV2), 3rd (ResNet101), 6th (DarkNet53), and 8th (EfficientNet b0) deep feature generators. It can be noted from Figure 7a that we achieved an accuracy of 90.63% to 94.50% for DB1 and 99.22% to 99.97% for DB2 using QSVM. By merging five deep features, a feature vector of length 5000 is created for classification. These results denote that DarkNet19, DarkNet53, and MobileNetV2 are the top deep feature generators for both OCT datasets. Thereafter, IRF is applied to the feature vector. The lengths of the selected optimal feature vectors are 935 and 178 for the two OCT datasets, respectively. Figure 7b shows the error rates obtained for various features with two datasets. These graphs clearly indicate the error rate via the number of features. The error rates become 0 for DB2; hence, the classification accuracy is 100%. The error rates are low for DB1, and hence, the accuracy is high. Student’s t-test is applied to the generated and selected 935 features of DB1 to validate these classification results. There are four categories in this dataset. Therefore, couples were used to calculate p-values, and obtained p-values are shown in Figure 8.
Figure 8

t-test results: (a) boxplot of the calculated p-values, (b) number of observations with p-values are smaller than 0.05 via couple.

Figure 8 shows the statistical properties of our features, and it validates the classification accuracies calculated using a conventional classifier (QSVM). According to Figure 7b, there are 250 observations in each class, and the number of observations with p-values smaller than 0.05 is calculated as 216, 234, 233, 234, 226, and 220 for 1, 2, 3, 4, 5, and 6 respectively. On the other hand, our framework attained 100% accuracy on the DB2, and the minimum p-values of all couples are calculated as 0. Furthermore, comparison results are listed in Table 5 to denote the high classification success of the proposed framework.
Table 5

Summary of state-of-the-art retinal disorder classification models developed using OCT images.

StudyMethodClassifierDatasetSplit RatioNumber of ClassThe Results (%)
Rong et al. [2] Convolution neural networkConvolution neural network45 subjects 195 Test 1195 Test 2207 Test 3267 Test 4207 Test 572:10:183Acc: 100.0for volume level
Rasti et al. [3] Multi-Scale Convolutional Neural Network EnsembleSoftmaxDataset 1862 DME, 969 AMD2311 normalDataset 2856 DME711 AMD1707 normal5-fold cross validation3AUC: 99.80Rec: 99.36F1:99.34for Dataset 1AUC: 99.9Rec: 97.78F1:97.71 for Dataset2
Fang et at. [41]Lesion-aware convolution neural networkSoftmax500 CNV500 DME500 Drusen500 Normal10-fold cross validation4Acc: 90.10Sen: 86.80Pre: 86.20
He et al. [42]Label smoothing generative adversarial networkConvolution neural network1.37.455 CNV11.598 DME8866 Drusen26.565 Normal1.581 CNV4.592 DME1.563 Drusen1.168 NormalLeave-p-out cross- validation1. 42. 41. Pre: 87.25Sen: 87.21Spe: 95.09F1: 87.112. Pre: 68.36Sen: 66.68Spe: 86.73F1: 67.14
Seeböck et al. [43]Unsupervised deep learningRandom forest268 AMD (early AMD, late AMD)115 control218 AMD65 control for training 50 AMD50 control for testing3Acc: 81.40
Alqudah [44]Automated convolutional neural networkSoftmax250 CNV250 DME250 Drusen250 Normal250 AMD95.331 training40.856 validation1250 testing5Acc: 97.10
Huang et al. [45]Layer guided convolutional neuralnetworkConvolutional neuralnetwork1.37.455 CNV11.598 DME8866 Drusen26.565 Normal2. 1.581 CNV4.592 DME1.563 Drusen1.168 Normal100:11. 42. 41. Acc: 88.402. Acc: 89.90
Fang et al. [46]Iterative fusion convolutional neural networkConvolutional neuralnetwork37.455 CNV11.598 DME8866 Drusen26.565 Normal10-fold cross validation4Acc: 87.30
Saraiva et al. [47]Convolutional neuralnetworkConvolutional neuralnetwork5.313 CNV7.491 DME1.773 Drusen2.319 Normal100:14Acc: 94.35
Our methodConvolutional neural networks, iterative ReliefFSupport vector machine2750 CNV2750 DME2750 Drusen2750 Normal10,000 train and 1000 test(10:1)4Acc: 97.30Pre: 97.32F1: 97.30Rec: 97.30CK: 96.40MCC: 96.41
686 AMD1101 DME1407 Healthy10-fold cross-validation3Acc: 100Pre: 100F1: 100Rec: 100CK: 100MCC: 100
Table 5 shows that the proposed framework achieved the highest retinal disorder classification performance using both OCT image datasets. We used a maximum number of images and yet achieved optimal performance for both datasets. The model used transfer learning models; hence, the time burden is reasonable. The benefits of the proposed framework are given below. A cognitive transfer learning-based image classification framework is presented. An intelligent feature generator is described using 18 pre-trained CNNs and novel multilevel and multiple pooling-based compression methods. Moreover, this feature generator is designed as a learning model. Therefore, it has the best feature vector selection ability. The proposed framework is a simple and parametric classification model. It can be extended using more feature extractors, other classifiers, and feature selectors. A general computer vision framework is presented with a ten-fold cross-validation strategy. Hence, our developed model is accurate and robust. This framework is an extendable framework. By using other effective methods, new-generation image classification methods can be proposed. This framework is a fast-learning model since the used CNNs are used in the feedforward mode to extract the features. Two OCT image datasets are employed to verify general image classification capability.

6. Conclusions

This study has proposed a retinal disorder detection framework using transfer learning, in addition to multilevel multiple pooling decomposition, IRF, and tuned SVM with Bayesian optimization. The main aim of the proposed framework is to select the best pre-trained CNNs to solve the classification problem. We used two public OCT image datasets to evaluate the accuracy and robustness of our developed model. In the proposed framework, DarkNet53, MobileNetV2, DarkNet19, Efficient-Net b0, and DenseNet201 CNNs are selected as the top five deep feature generators for DB1. The selected top five CNNs for DB2 are DarkNet19, MobileNetV2, ResNet101, DarkNet53, and EfficientNet b0. The generated features from these networks are merged, and the best feature vector (the most valuable features) is selected using IRF. We obtained an accuracy of 97.40% and 100% with DB1 and DB2 datasets, respectively. Our proposed framework can potentially be used to detect early stages of retinal disorders. By adopting this framework, screening for retinal disorder in an appropriate patient cohort can be conducted more effectively and enable early treatment. Ultimately, we hope that with this work, irreversible vision loss can be prevented by early diagnosis and prompt medical intervention. This framework can select the most suitable CNNs based on the context of the problem, and new generation CNNs can be included in this framework as part of future work.
  19 in total

1.  Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images.

Authors:  Pratul P Srinivasan; Leo A Kim; Priyatham S Mettu; Scott W Cousins; Grant M Comer; Joseph A Izatt; Sina Farsiu
Journal:  Biomed Opt Express       Date:  2014-09-12       Impact factor: 3.732

2.  Attention to Lesion: Lesion-Aware Convolutional Neural Network for Retinal Optical Coherence Tomography Image Classification.

Authors:  Leyuan Fang; Chong Wang; Shutao Li; Hossein Rabbani; Xiangdong Chen; Zhimin Liu
Journal:  IEEE Trans Med Imaging       Date:  2019-02-08       Impact factor: 10.048

3.  Residual LSTM layered CNN for classification of gastrointestinal tract diseases.

Authors:  Şaban Öztürk; Umut Özkaya
Journal:  J Biomed Inform       Date:  2020-12-01       Impact factor: 6.317

4.  Unsupervised Identification of Disease Marker Candidates in Retinal OCT Imaging Data.

Authors:  Philipp Seebock; Sebastian M Waldstein; Sophie Klimscha; Hrvoje Bogunovic; Thomas Schlegl; Bianca S Gerendas; Rene Donner; Ursula Schmidt-Erfurth; Georg Langs
Journal:  IEEE Trans Med Imaging       Date:  2018-10-22       Impact factor: 10.048

5.  Surrogate-Assisted Retinal OCT Image Classification Based on Convolutional Neural Networks.

Authors:  Yibiao Rong; Dehui Xiang; Weifang Zhu; Kai Yu; Fei Shi; Zhun Fan; Xinjian Chen
Journal:  IEEE J Biomed Health Inform       Date:  2018-02-12       Impact factor: 5.772

6.  AOCT-NET: a convolutional network automated classification of multiclass retinal diseases using spectral-domain optical coherence tomography images.

Authors:  Ali Mohammad Alqudah
Journal:  Med Biol Eng Comput       Date:  2019-11-14       Impact factor: 2.602

Review 7.  Immunology of age-related macular degeneration.

Authors:  Jayakrishna Ambati; John P Atkinson; Bradley D Gelfand
Journal:  Nat Rev Immunol       Date:  2013-06       Impact factor: 53.106

8.  Deep learning is effective for the classification of OCT images of normal versus Age-related Macular Degeneration.

Authors:  Cecilia S Lee; Doug M Baughman; Aaron Y Lee
Journal:  Ophthalmol Retina       Date:  2017-02-13

Review 9.  Automated Detection of Hypertension Using Physiological Signals: A Review.

Authors:  Manish Sharma; Jaypal Singh Rajput; Ru San Tan; U Rajendra Acharya
Journal:  Int J Environ Res Public Health       Date:  2021-05-29       Impact factor: 3.390

10.  Classification of healthy and diseased retina using SD-OCT imaging and Random Forest algorithm.

Authors:  Md Akter Hussain; Alauddin Bhuiyan; Chi D Luu; R Theodore Smith; Robyn H Guymer; Hiroshi Ishikawa; Joel S Schuman; Kotagiri Ramamohanarao
Journal:  PLoS One       Date:  2018-06-04       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.