Literature DB >> 35624922

Transfer Learning-Based Model for Diabetic Retinopathy Diagnosis Using Retinal Images.

Muhammad Kashif Jabbar1, Jianzhuo Yan1, Hongxia Xu1, Zaka Ur Rehman2, Ayesha Jabbar3.   

Abstract

Diabetic retinopathy (DR) is a visual obstacle caused by diabetic disease, which forms because of long-standing diabetes mellitus, which damages the retinal blood vessels. This disease is considered one of the principal causes of sightlessness and accounts for more than 158 million cases all over the world. Since early detection and classification could diminish the visual impairment, it is significant to develop an automated DR diagnosis method. Although deep learning models provide automatic feature extraction and classification, training such models from scratch requires a larger annotated dataset. The availability of annotated training datasets is considered a core issue for implementing deep learning in the classification of medical images. The models based on transfer learning are widely adopted by the researchers to overcome annotated data insufficiency problems and computational overhead. In the proposed study, features are extracted from fundus images using the pre-trained network VGGNet and combined with the concept of transfer learning to improve classification performance. To deal with data insufficiency and unbalancing problems, we employed various data augmentation operations differently on each grade of DR. The results of the experiment indicate that the proposed framework (which is evaluated on the benchmark dataset) outperformed advanced methods in terms of accurateness. Our technique, in combination with handcrafted features, could be used to improve classification accuracy.

Entities:  

Keywords:  annotated data insufficiency; computer-aided diagnosis; convolutional neural network; diabetic retinopathy; fundus images; transfer learning

Year:  2022        PMID: 35624922      PMCID: PMC9139157          DOI: 10.3390/brainsci12050535

Source DB:  PubMed          Journal:  Brain Sci        ISSN: 2076-3425


1. Introduction

Diabetic retinopathy is directly associated with the prevalence of diabetes, which is now at epidemic proportions worldwide [1]. Currently, about 463 million diabetes patients are present, and approximately one-third of them have some form of diabetic retinopathy [2]. The international diabetes federation (IDF) reported that diabetes patients are expected to reach about 552 million by 2035 and 642 million by 2040 [3,4]. There are more than 158.2 million people currently suffering from DR, and this number is estimated to increase to about 191 million by 2030 [5]. DR is an ocular impediment caused by diabetic disease; it has held its position as one of the main factors behind the occurrence of blindness globally [6,7,8]. DR develops because of the long-standing occurrence of diabetes mellitus. The risks of disease are more common in patients with uncontrolled blood sugar. Generally, DR develops gradually and may not cause any symptoms or only mild vision loss in the primary stages. Eventually, if the treatment and diagnosis are not timely, then it tends to cause blindness [9]. DR is generally categorized into three categories, normal, non-proliferative DR (NPDR), and proliferative DR (PDR), based on the progression of the diabetic retinopathy [10]. NPDR develops when new retinal blood vessels are not growing, and the blood vessel walls become faded. NPDR is divided further into mild, moderate, or severe stages of disease. In PDR, the retinal area is occupied by new blood vessels and obstructs the blood supply to the retina. Figure 1 shows samples of NPDR, PDR, and its subdivided classes of retinopathy discussed above.
Figure 1

These images show the different types of retinopathies in the fundus images. (a) Normal, (b) mild, (c) moderate, (d) severe, and (e) proliferative.

DR detection and grading at initial periods is laborious, time-intensive, and requires domain experts [11]. Moreover, the manual screening of DR patients indorses ex- tensive inconsistency among various clinicians. Approximately 79% of DR patients belong to underdeveloped or developing nations, which are deficient in ophthalmologists and a basic setup for DR detection [12]. With the rapid prevalence of DR worldwide, manual screening techniques are unable to keep pace with the demand for diagnosis methods [13]. Due to the developments in computer vision techniques, numerous automatic techniques have been projected by researchers for the diagnosis of DR. There are various challenges associated with enhancements in computer-aided diagnosis (CAD) systems, such as identification of lesions from a retinal image, subdivision of optic disc, segmentation of blood vessel, etc. [14,15]. Although machine learning-based systems have shown resilient performance in DR detection, their efficacy is highly dependent on handcrafted features, which are very difficult to generalize [16]. To overcome such limitations, deep learning (DL) methods provide automatic feature extraction and classification from fundus images. The major determination of this study is to explain an efficient method for categorizing early-stage DR to assist ophthalmologists. The key contribution of this study is highlighted in the following points: A VGGNet model based on transfer learning is proposed for detecting and classifying diabetic retinopathy. Implementation of various preprocessing techniques such as interpolation image resizing, weighted Gaussian blur, and CLAHE for improving the value and visibility of retinal images. Performed data augmentation operations on each grade of DR individually to overcome annotated data insufficiency and to make a balanced dataset. A comprehensive DR classification system is developed with accuracy and robustness. Evaluation of the proposed model is performed on a large dataset, EyePACS, with 35,126 retinal fundus images. Various performance measures are implemented, such as accuracy, sensitivity, specificity, and the AUC to verify the analytical skill. The paper is compiled in the following manner: The dataset details and the proposed framework is elaborated in Section 2. Results and Discussions are described in Section 3 and the Conclusions in Section 4.

2. Methodology

The proposed framework of DR classification was mainly categorized into the following steps: image preprocessing, data augmentation, feature extraction, and classification. The graphical illustration of the proposed work is presented in Figure 2.
Figure 2

Proposed framework for detection and classification of diabetic retinopathy. In the first phase, retinal images are preprocessed, and data augmentation operations are performed individually on each grade of DR to improve classification accuracy. In the last, model-based on transfer learning is used for automatic features extraction and classification of DR into different stages.

2.1. The Kaggle EyePACS Dataset

The public dataset EyePACS was used for acquiring fundus images of the retina via Kaggle.com (accessed on 24 March 2021). These images were labeled by the ophthalmologists and subdivided into five categories: normal, mild, moderate, severe, and proliferative DR, as shown in Figure 1. The details related to this dataset are stated in Table 1.
Table 1

Kaggle EyePACS dataset details.

GradeSeverityNo. of Images% of Total Images
0Normal25,81073.15
1Mild24436.96
2Moderate529215.07
3Severe8732.81
4PDR7082.01
Total 35,126100%

2.2. Image Preprocessing

The EyePACS dataset is a heterogeneous dataset containing images from various smaller datasets captured with different cameras, under different adjustments, of different sizes, and with a lot of brightness and illumination differences; therefore, we adopted various preprocessing steps to standardize these images. First, we resized the fundus images to a uniform size by using bicubic interpolation over a 4 × 4-pixel neighborhood. We preferred this interpolation over simple resizing because it resizes the image by sustaining quality and locking the aspect ratio. Generally, the retinal images were yellowish with a dark background. The fundus details did not overlap with the background and thus can be abolished to reduce noise. Equalizing the black background of the fundus images resulted in darkness being expanded into the image’s details [17]. Concerning this matter, we agreed with preprocessing to delete the black background by setting the pixel value to zero and non-zero for all bright regions. After thresholding was performed, the extraction of the green channel was performed. This green channel conserves more retinal data in comparison to the red or blue channels. The implementation of CLAHE, which is the contrast limited adaptive histogram equalization, took place for improving the quality of the retinal image as well as for enhancing the small areas. Then, the weighted Gaussian blur was applied to images to reduce noise and increase image structure [18]. The mathematical expression to calculate the Gaussian function in two dimensions (x, y) along with σ standard deviation is given in Equation (1).

2.3. Data Augmentation

The training dataset size is one of the key features for the efficient performance of the DL models. Therefore, it is mandatory to have a larger dataset for the training of deep learning architecture to prevent generalization and overfitting issues. Although the Kaggle EyePACS dataset size is sufficient, it is considered very small as compared to the ImageNet [19] dataset. The dataset distribution over the classes was highly imbalanced, as shown in Figure 3, where the maximum images were from grade 0. This highly imbalanced dataset was in the ratio of (36:3:7:1:1), which may cause incorrect classification. We performed data augmentation operations to amplify the retinal dataset at various scales and to eliminate the noise in fundus images.
Figure 3

Dataset distribution over DR severity. There were about 73% of images in the normal category, while only 2% of them from the proliferative DR category. Thus, it was an imbalanced dataset with 36:1 for normal and proliferative DR.

We employed various data augmentation operations on each grade differently because of the highly imbalanced nature of the dataset. The visual exemplification of some augmentation techniques performed on preprocessed images is given. This augmentation operation includes cropping, flipping, translating, shearing, rotating, zooming, Gaussian scale-space theory (GST) augmentation [20], and Krizhevsky augmentation [21]. To produce visually appealing images, the clipping limit of CLAHE was set to 2 along the tile grid and 8, as shown in Figure 4.
Figure 4

The preprocessed retinal images after applying contrast limited adaptive histogram equalization to adjust contrast in images. (a,c) Original retinal fundus images, (b,d) preprocessed images.

For improvement of image quality and enhancement of image structure, we imployed weighted gaussian blur to the retinal fundus images. Figure 5 shows the effects of gaussian blur on the retinal fundus images. The representation of images before and after applying this method of image processing is shown in the Figure 5.
Figure 5

Some examples of adding weighted Gaussian blur to the retinal images, which is employed to reduce noise and increase image structure. First row are original fundus images, and second images are output of the preprocessed images.

Cropping: Images were randomly cropped from the corner and center to 60–75% of the original image. Flipping: Images were flipped on both the X and Y axes. Translating: Images were shifted between 0 and 30 pixels. Shearing: Images were sheared between 0 and 180 degrees randomly. Rotating: Images were rotated randomly in the range of 0 to 360 degrees. Zooming: Images were zoomed in the range of (0.7, 1.3). GST Augmentation: The GST-based augmentation was performed for a two-dimensional image. Krizhevsky Augmentation: Krizhevsky color augmentation technique was used for dataset augmentation. The representation of images after applying various data augemntation operation is displayed in the Figure 6.
Figure 6

The visual exemplification of some augmentation operations performed on preprocessed images to augment the retinal dataset (a) Original image (b) Cropping (c) Shearing (d) Flipping (e) Rotating (f) Zooming (g) Translating (h) All augmentation.

The details of augmentation operations after applying them to the training dataset are given in Table 2. Different augmentation operations were performed on each grade of fundus images. The augmented dataset was 3.6 times greater than the original dataset and was highly balanced with 1:1 for all grades of DR.
Table 2

The dataset statistics by using data augmentation operations.

GradeSeverityNo. of ImagesOperationsAugmented Images
0Normal25,810025,810
1Mild24431124,430
2Moderate5292526,460
3Severe8732925,317
4PDR7083625,488
Total 35,126 127,505

2.4. Proposed Architecture

Although deep learning algorithms are good enough to solve various classification problems, the main problem with the classification of medical images is the unavailability of labeled data. Transfer learning is widely adopted to overcome annotated data insufficiency by reusing already trained deep convolutional neural networks for another identical task. Thus, it may be employed for reducing the training overhead as well as for training with a smaller dataset. There arise questions about the implementation of transfer learning to overwhelm annotated data availability in medical image classification based on deep learning. In this study, we opted for pre-trained VGG 16 to take advantage of fixed size 3 × 3 kernel filters for detection. The architecture of the proposed model is shown in Table 3. This architecture consists of 16 layers with an input layer is 512 × 512. A 3 × 3 kernel size filter, united bias, and stride of 1 were utilized for all convolutional layers other than the first layer with a stride of 2. A 2 × 2 kernel size filter along with a stride of 2 was selected for all the max-pooling layers; extracted features were flattened before forwarding to the connected layers. The usage of the activation function rectified linear unit (ReLU) was made in all layers, and a drop of 0.5 was applied before the first two fully connected layers to evade overfitting. After convolution layers, two fully connected layers were added, with each having 1024 neurons. Finally, a softmax function was used as a single neuron output layer for the early-stage detection and classifying the diabetic retinopathy.
Table 3

The proposed architecture for early-stage detection and classification of diabetic retinopathy comprising of 16 layers.

Sr. No.Layer TypeKernel Size andNumberStrideOutput
1input(512,512,3)
2conv 1_13 × 3 × 322(256,256,32)
3conv 1_23 × 3 × 321(255,255,32)
4max-pooling2 × 22(127,127,32)
5conv 2_13 × 3 × 641(63,63,64)
6conv 2_23 × 3 × 641(64,64,64)
7max-pooling2 × 22(32,32,64)
8conv 3_13 × 3 × 1281(32,32,128)
9conv 3_23 × 3 × 1281(33,33,128)
10conv 3_33 × 3 × 1281(34,34,128)
11max-pooling2 × 22(17,17,128)
12conv 4_13 × 3 × 2561(9,9,256)
13conv 4_23 × 3 × 2561(10,10,256)
14conv 4_33 × 3 × 2561(11,11,256)
15max-pooling2 × 22(6,6,256)
16conv 5_13 × 3 × 5121(6,6,512)
17conv 5_23 × 3 × 5121(5,5,512)
18conv 5_33 × 3 × 5121(4,4,512)
19max-pooling2 × 22(2,2,512)
20fully connected1024−1024
21fully connected1024−1024
22fully connected1−1

2.5. Training

Training of transfer learning-based VGGNet was performed through the process of fine-tuning the hyperparameters. The model was trained through Adam’s optimization function, and the learning rate was set to 0.0001. The weights of the network were initialized randomly through 32 batch-sizes and then trained on a network for 300 epochs. The momentum size was set at 0.9, and categorical cross-entropy was considered as an objective function. Data augmentation operations were performed separately on each grade to overcome data misbalancing issues. The details of these hyperparameters are given in Table 4.
Table 4

Diagnostic test for the prevalence of the disease.

Sr. No.HyperparametersValue
1Learning Rate0.0001
2Batch Size32
3Activation FunctionReLU
4Epochs300
5OptimizerAdam
6Momentum0.9
7Loss FunctionCategorical Cross-Entropy

3. Results and Discussion

By utilizing a transfer learning technique, a detailed analysis of the working of the proposed scheme for diabetic retinopathy was evaluated. The Kaggle EyePACS dataset after data augmentation operations were segmented into 80%, 20% for the training and testing datasets.

3.1. The System Configurations

A system with the following specifications was used for the methodology. OS: Linux Mint 18.1 operating system. CPU: Intel(R) 3rd Generation Core (TM) i5-3470 CPU @ 3.20 GHz @ 3.20 GHz; CPU Ram: 16 GB; GPU: Quadro K620; Graphics card Ram size: 12 GB.

3.2. The Performance Metrics

For evaluation of the overall working of the presented scheme, we analyzed the performance of abnormal human activity recognition from a confusion matrix [22] and then computed the following performance metrics. Precision: Expressed as the ratio of a total number of TP with respect to the total number of component tags as appropriate to the positive class (i.e., the sum of TP and FP). Positive Predictive Value (PPV) is used to denote the precision. Precision can be measured as follows: F1 Score: The harmonic mean of recall and precision is used to measure it as in Equation (3): Accuracy: The ratio of a number of components TP and TN to the total number of components TP, TN, FP, and FN is used to measure the accuracy. The accuracy can be calculated using the following Equation (4). Specificity: It can be expressed in the ratio of the number of TN with respect to the total number of components that are appropriate to the negative class (i.e., the sum of TN and FP). The mathematical expression is shown in Equation (5). The Area Under the Curve: It can be obtained by performing a definite integral between the two points. This can be measured by using the following mathematical Equation (6).

3.3. Result Analysis

After the training of the proposed model, the learning data was passed to the combined extracted features by using transfer learning. Besides the proposed VGGNet approach, different transfer learning models were also implemented, as shown in Table 5.
Table 5

The performance comparison of the proposed model on an original and augmented dataset.

Sr. NoArchitectureSensitivitySpecificity AccuracyPrecision F1-ScoreAUC
1ResNet0.8540.9430.9240.9770.9180.924
2GoogLeNet0.8950.9890.9370.990.9390.935
3AlexNet0.9530.9380.9460.9370.9450.949
4VGGNet0.9490.9840.9660.9850.9670.971
From Table 5, we can make observations that the ResNet, GoogLeNet, and AlexNet gave average accuracies of 92.40%, 93.75%, and 94.62%, respectively. Meanwhile, the proposed model stated an accuracy level of about 96.61%. Thus, it is an efficient method for detecting and classifying diabetic retinopathy. Furthermore, the data is split into the training and testing data. For training purposes, 80% of the data were used; for testing the proposed architecture, the rest of the 20% of unseen data were used. The proposed model was assessed through both the original and augmented dataset, as presented in Table 6.
Table 6

The performance comparison of the proposed model and different existing transfer learning models on an original and augmented dataset.

Sr. No.DatasetAccuracyAUC
1Original0.9360.954
2Augmented0.9660.971
The proposed model showed higher accuracy for detecting and classifying diabetic retinopathy on a balanced augmented dataset. In Figure 7a,b, it can be observed that the training plots and validation accuracy improved constantly.
Figure 7

(a) Shows the training and validation accuracy of the proposed framework and (b) shows the training and validation loss of the proposed fine-tuned VGGNet framework.

The fine-tuned version of VGGNet displayed exceedingly acceptable performance. The rate of training and validation accuracy continuously improved with an increase in each epoch. The loss curve indicates that the loss of training and validation decreased with each epoch. The implemetation of various data augmentation techniques, the augmented data was obtained. The acquired augmented data was distributed into training and testing data as shown in the Table 7.
Table 7

The distribution of augmented data into training and testing data.

Sr. No.GradeSeverityOriginal ImagesAugmented ImagesTrainingValidation
10Normal25,81025,81020,6485162
21Mild244324,43019,5444886
32Moderate529226,46021,1685292
43Severe87325,31720,2545063
54PDR70825,48820,3905098
Total 35,126127,505102,00425,501
In addition to average results, the best performance results were also important and retained in this paper because the validation accuracy can be altered with each epoch during the CNN training. Table 8 shows the prevalence of the disease on the basis of the diagnosis test.
Table 8

Diagnostic test for the prevalence of the disease.

DRNormal
Identified as having DR TP FP
Identified as having no DR FN TN
The comparative analysis of the proposed architecture was carried out with five up-to-date methods to relate its strength, as shown in Table 9.
Table 9

Performance comparison of the proposed framework with baseline methods.

Reference #AuthorYearModelTargetAccuracy
[23]Rakhlin2018CNNDR85.3
[24]Sengupta et al.2019CNNDR90.4
[25]Gulshan et al.2016CNNDR91.7
[26]Chang2018CNNDR78.7
[27]Wan2018CNNrDR95.68
[28]Gargeya2017CNNDR95.03
[29]Zeng2019CNNrDR82.2
[30]Zhang2019CNNrDR87.06
[31]Lin2018CNNDR86.10
[32]Li2018DCNNvtDR86.04
[33]Seth2018CNNrDR84.36
[34]Keel2018 Third-party DL algorithmrDR89.08
Proposed Method CNNDR96.6
Observations made by the methods in [23,24,25,26,27,28,29,30,31,32,33,34] gave average accuracies of 82.3, 85.3, 90.4, 78.7, 95.68, 95.03%, and so on, as shown in the figure. This method yielded an accuracy of 96.6%, which is much higher than baseline methods. The experimental outputs indicate the strength of the proposed architecture with respect to the accuracy as compared to existing methods. Figure 8 shows the receiver the operating characteristic (ROC) and AUC curve of the proposed model.
Figure 8

The receiver operator characteristics curve of the proposed framework.

4. Conclusions

We proposed a framework for the automatic detection and classification of diabetic retinopathy by using the transfer learning concept in this paper. For the preprocessing, we employed effective preprocessing techniques such as interpolation image resizing, weighted Gaussian blur, and non-local mean denoising (NLMD) for improving the visibility of an image. After preprocessing, transfer learning-based VGGNet architecture was used for classifying DR into the normal, mild, moderate, severe, and proliferative classes. The experimental studies were conducted on the Kaggle EyePACS public dataset. The effectiveness of the method was evaluated using numerical procedures such as the sensitivity, specificity, accuracy, and area under the curve. We explored the extraction of features in fundus images using VGGNet architecture and united using the concept of transfer learning to improve classification performance. Moreover, we also employed various data augmentation operations on each grade of DR differently to make a balanced dataset and improve the efficiency of architecture. Finally, the results of the proposed study were compared to many deep learning architectures, and a comparison was made with the baseline methods. It is concluded that this proposed method displays exceptional presentation regarding various statistical measures. In the future, we intend to use hand-engineered features along with CNN to further improve classification accuracy.
  19 in total

1.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

Review 2.  Diabetic retinopathy - ocular complications of diabetes mellitus.

Authors:  Martin M Nentwich; Michael W Ulbig
Journal:  World J Diabetes       Date:  2015-04-15

Review 3.  Diabetic retinopathy: global prevalence, major risk factors, screening practices and public health challenges: a review.

Authors:  Daniel Shu Wei Ting; Gemmy Chui Ming Cheung; Tien Yin Wong
Journal:  Clin Exp Ophthalmol       Date:  2016-02-17       Impact factor: 4.207

4.  Automated Identification of Diabetic Retinopathy Using Deep Learning.

Authors:  Rishab Gargeya; Theodore Leng
Journal:  Ophthalmology       Date:  2017-03-27       Impact factor: 12.079

5.  Mortality attributable to diabetes in 20-79 years old adults, 2019 estimates: results from the International Diabetes Federation Diabetes Atlas, 9th edition.

Authors:  Pouya Saeedi; Paraskevi Salpea; Suvi Karuranga; Inga Petersohn; Belma Malanda; Edward W Gregg; Nigel Unwin; Sarah H Wild; Rhys Williams
Journal:  Diabetes Res Clin Pract       Date:  2020-02-14       Impact factor: 5.602

6.  An Automated Grading System for Detection of Vision-Threatening Referable Diabetic Retinopathy on the Basis of Color Fundus Photographs.

Authors:  Zhixi Li; Stuart Keel; Chi Liu; Yifan He; Wei Meng; Jane Scheetz; Pei Ying Lee; Jonathan Shaw; Daniel Ting; Tien Yin Wong; Hugh Taylor; Robert Chang; Mingguang He
Journal:  Diabetes Care       Date:  2018-10-01       Impact factor: 19.112

7.  Trends in diabetes: sounding the alarm.

Authors:  Etienne G Krug
Journal:  Lancet       Date:  2016-04-06       Impact factor: 79.321

8.  Inflammation and diabetic retinal microvascular complications.

Authors:  Wenbo Zhang; Hua Liu; Mohamed Al-Shabrawey; Robert W Caldwell; Ruth B Caldwell
Journal:  J Cardiovasc Dis Res       Date:  2011-04

9.  Feasibility and patient acceptability of a novel artificial intelligence-based screening model for diabetic retinopathy at endocrinology outpatient services: a pilot study.

Authors:  Stuart Keel; Pei Ying Lee; Jane Scheetz; Zhixi Li; Mark A Kotowicz; Richard J MacIsaac; Mingguang He
Journal:  Sci Rep       Date:  2018-03-12       Impact factor: 4.379

10.  Transforming Retinal Photographs to Entropy Images in Deep Learning to Improve Automated Detection for Diabetic Retinopathy.

Authors:  Gen-Min Lin; Mei-Juan Chen; Chia-Hung Yeh; Yu-Yang Lin; Heng-Yu Kuo; Min-Hui Lin; Ming-Chin Chen; Shinfeng D Lin; Ying Gao; Anran Ran; Carol Y Cheung
Journal:  J Ophthalmol       Date:  2018-09-10       Impact factor: 1.909

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.