Literature DB >> 29168750

Deep Convolutional Neural Network-Based Early Automated Detection of Diabetic Retinopathy Using Fundus Image.

Kele Xu1, Dawei Feng2, Haibo Mi3.   

Abstract

The automatic detection of diabetic retinopathy is of vital importance, as it is the main cause of irreversible vision loss in the working-age population in the developed world. The early detection of diabetic retinopathy occurrence can be very helpful for clinical treatment; although several different feature extraction approaches have been proposed, the classification task for retinal images is still tedious even for those trained clinicians. Recently, deep convolutional neural networks have manifested superior performance in image classification compared to previous handcrafted feature-based image classification methods. Thus, in this paper, we explored the use of deep convolutional neural network methodology for the automatic classification of diabetic retinopathy using color fundus image, and obtained an accuracy of 94.5% on our dataset, outperforming the results obtained by using classical approaches.

Entities:  

Keywords:  deep convolutional neural network; diabetic retinopathy; image classification

Mesh:

Year:  2017        PMID: 29168750      PMCID: PMC6149821          DOI: 10.3390/molecules22122054

Source DB:  PubMed          Journal:  Molecules        ISSN: 1420-3049            Impact factor:   4.411


1. Introduction

The number of people diagnosed as having diabetes has increased dramatically over the last several decades, and diabetes increases the risk of a range of eye diseases, of which diabetic retinopathy is one of the most severe [1]. Moreover, diabetic retinopathy is the main cause of blindness in the mid-age population [2]. Despite sustained efforts having been made, early detection of diabetic retinopathy is a time-consuming process even for a well–trained clinician, which may result in delayed treatment, miscommunication, etc. The importance of an automatic method for diabetic retinopathy detection has been recognized. In our study, we focus on the classification of retinal images into normal images and diabetic retinopathy images (sample frames used for our classification problem are given in Figure 1). Previous efforts using image feature extraction and machine learning methods made good progress. The features used for the classifiers include hard exudates [3,4], red lesions [5], micro-aneurysms and blood vessel detection [6], etc., while the classifiers used for the task contain neural networks, sparse representation classifiers, linear discriminant analysis (LDA), support vector machine (SVM), k-nearest neighbors (KNN) algorithm and so on. However, none of the handcrafted features can cover all the symptoms of diabetic retinopathy in the images, and a large fraction of cases turn out to be normal while much time has been spent diagnosing normal cases. Consequently, the practical clinical applications of the automatic diagnosing system are limited.
Figure 1

Sample frames of the retina images. The first two frames in the top row come from normal subjects, while the two frames in the bottom row come from the patients who have diabetic retinopathy.

Recent advances in convolutional neural networks (CNNs) have made it a state-of-the-art technique in image classification tasks [7], and its variants have begun to dominate many fields in computer vision, such as object detection [8], image classification [9], object tracking [10], edge detection [11]. Instead of making essential use of handcrafted features, CNN can learn a hierarchy of features, which can be used for image classification purposes. As the hierarchy approach is available to learn more complex features, as well as translation and distortion features in higher layers, the accuracy of the CNN-based image classification method can be higher. Based on this assumption, we explore the use of the CNN-based method for the diabetic retinopathy test in this work. Moreover, a specific multi-layer CNN architecture is designed, and experiments are conducted on real retina data. In addition, the results obtained demonstrate our assumption as we achieve 94.5% accuracy, which ranks as the highest in comparison with previous handcrafted feature-based classifiers. This paper is organized as follows: Section 2 describes the preprocessing method and gives an overview of the CNN architecture used in this work, while Section 3 presents the results to demonstrate the performance of the proposed approach compared to previous handcrafted feature-based classification methods. The conclusion and future work are discussed in Section 4.

2. Methodology

2.1. Data Augmentation

Until recently, datasets of labeled images with diabetic retinopathy were relatively small. Here, we used the data provided by Kaggle Community [12]. Indeed, the shortcomings of small image datasets have been widely recognized, thus, data augmentation is needed to artificially enlarge the datasets using label-preserving transformation, which can reduce overfitting on the image data and increase the performance of the algorithm [13]. In our experiment, we apply translation, stretching, rotation and flipping to the labeled dataset. A summary of the transformations is given in Table 1, while the sample transformed frames are presented in Figure 2. Five different transformation types are used in our experiment, including rotation, flipping, shearing, rescaling and translation. For each type, the parameter details are given in Table 1.
Table 1

Data augmentation parameters.

Transformation TypeDescription
Rotation0–360
Flipping0 (without flipping) or 1 (with flipping)
ShearingRandomly with angle between −15 and 15
RescalingRandomly with scale factor between 1/1.6 and 1.6
TranslationRandomly with shift between −10 and 10 pixels
Figure 2

Samples of the transformed frames.

2.2. Convolutional Neural Network-Based Image Classification

The convolutional neural network (CNN) belongs to the feed-forward artificial neural network (ANN), which is very similar to ordinary neural networks. The CNN is a well-known deep learning architecture, in which individual neurons are tiled in such a way that they respond to overlapping regions in the visual fields [14]. CNNs are an important class of learnable representation applications, and they were inspired by biological neural networks. Numerous variants have been proposed over last several years. However, the basic components are very similar. CNNs consist of alternating convolution and pooling operations [15]. Typically, the convolutional layers are interspersed with pooling layers to reduce computation time, and build up further spatial and configuration invariance [16]; the last few layers (close to the outputs) will be fully connected 1-dimensional layers. In more detail, a feed-forward neural network can be viewed as a function f of mapping data x: Each function takes ( is the input data x) as input with a learnable parameter vector . L denotes the depth of the neural network. Although the type and sequence of functions are usually handcrafted, the parameters can be discriminatively learned from example data such that the resulting function f realizes a useful mapping. Formally, in a CNN, each will be a M×N×C array. As our problem can be simplified as a binary-classification problem, we can define the loss function of the CNN as: where n is the number of samples, is the true label of sample i. The training problem can be converted to training a neural network to minimize the loss function L. Figure 3 gives the general architecture of a CNN network, which consists of multiple layers of small neurons. The results of these collections are then tiled so that they overlap to obtain a better representation of the original image (such as edges in the image). Convolutional layers consist of a rectangle grid of neurons, which takes a rectangle region of the previous layer as input. Moreover, there may be several grids in each convolutional layer, using potentially different filters. Typically, there is a pooling layer after each convolutional layer, which are subsampled from the previous convolutional layer. This pooling can be carried out in several ways, such as the average, maximum, etc. Finally, after several convolutional layers and max pooling layers, a fully connected layer (or several layers) will be built using outputs from previous layers (maybe the fully connected, pooling or convolutional layer), which is used as a compact feature to describe the whole input image. The network is optimized by backpropagation and stochastic gradient descent. Note that the forward and backward propagations may differ depending on the type of the layer.
Figure 3

An exemplary architecture of the convolutional neural network.

Several different CNN architectures have been proposed and evaluated in our experiments. The depth of the tested neural network ranges from 9–18, and the convolution kernel size ranges from 1 to 5. To fit the input size of the CNN, we resize the image size to 224 × 224 × 3. The final architecture of the network used in our work is given in Table 2. For a given input, the network outputs two probabilities that sum up to 1, one for each class (our problem is a binary classification problem). In our experiment, 800 labeled images are used to train the neural network, while 200 images are used to evaluate the performance of the trained neural network.
Table 2

Convolutional neural network (CNN) architecture used in our experiment.

Output ShapeDescription
224 × 224 × 3input
222 × 222 × 323 × 3 convolution, 32 filter
220 × 220 × 323 × 3 convolution, 32 filter
110 × 110 × 322 × 2 max-pooling
108 × 108 × 643 × 3 convolution, 64 filter
106 × 106 × 643 × 3 convolution, 64 filter
53 × 53 × 642 × 2 max-pooling
53 × 53 × 1283 × 3 convolution, 128 filter
51 × 51 × 1283 × 3 convolution, 128 filter
49 × 49 × 1282 × 2 max-pooling
24 × 24 × 2563 × 3 convolution, 256 filter
22 × 22 × 2563 × 3 convolution, 256 filter
11 × 11 × 2562 × 2 max-pooling
4096flatterned and fully connected
1024fully connected
2softmax

3. Experimental Results

To evaluate the performance of the proposed method, the classification task was conducted using the CNN and Gradient boosting machines. Moreover, to compare the results obtained by automatic classification algorithms with the performance of human judgement, a human specialist is introduced to label the images as ground-truth. In more detail, four different feature extraction approaches have been employed: hard exudates, red lesions, micro-aneurysms and blood vessel detection. Two different kinds of classifiers were trained for the classification task: one kind of which combines the aforementioned extracted features and gradient boosting trees-based (GBM) classification method (Hard exudates + GBM, Red lesions + GBM, Micro-aneurysms + GBM and Blood vessel detection + GBM shown in Table 3.), and the other kind is the CNN-based methods (with or without data augmentation). Specifically, we use the default hyper parameters for GBM, with the number of classes set to 2, and the maximum depth set to 6. The GBM package used in this paper is the eXtreme Gradient Boosting method (XGBoost) [17], as it shows superior performance in our experiments when compared to other approaches (i.e., Support Vector Machine, Random Forest). Regarding implementation of the CNN, we made use of the R package named “MXNet” [18]. A visualization of the trained neural networks is given in Figure 4.
Table 3

Performance comparison with different approaches.

MethodAccuracy
Hard exudates + GBM89.4%
Red lesions + GBM88.7%
Micro-aneurysms + GBM86.2%
Blood vessel detection + GBM79.1%
CNN without data augmentation91.5%
CNN with data augmentation94.5%
Figure 4

Visualization of the trained neural networks. Each image represents the activations of the first layer during the forward pass.

The classification accuracies are given in Table 3. As can be seen from the table, the CNN-based method provides superior performance compared to other methods, which supports the assumption made in the introduction section. Also, as can be seen from the table, the results obtained using the CNN with data augmentation are better than the CNN without data augmentation, and the reason may be that the data augmentation can be helpful for the CNN to deal with small rotation or translations during the data recording. Our experiment was conducted on a Windows 8 operation system with Intel 4-Core 3.7GHz CPU, 16GB RAM, Dual AMD Filepro 512GB PCIe-based flash storage, and a Geforce 1070GPU. Although the training process of the CNN required 2 days, the trained network can provide the probability of diabetic retinopathy in less than 1 second, which can be used by clinicians in practice [19].

4. Discussion and Conclusions

With a limited number of medical staff, an automated system can significantly decrease the tedious manual labor involved in diagnosing large quantities of retinal images. Feature extraction-based diabetic retinopathy diagnosis has played a dominant role in previous studies. However, the progress made in deep convolutional neural networks has led to them becoming a state-of-the-art technique in optical image classification. In this paper, we explored the potential usage of the CNN in retinal image classification. The contribution of this paper is two-fold: firstly, we proposed a special neural network architecture for the diabetic retinopathy image classification task, which demonstrates superior performance over conventional feature extraction-based methods. Moreover, a data augmentation method was introduced for the proposed algorithm, which also improves the algorithm’s performance. The results are encouraging compared to the reports of human grading [20], thus a clinical evaluation will be undertaken in order to be able to integrate the presented algorithm into a tool to diagnose diabetic retinopathy [21].
  10 in total

1.  A contribution of image processing to the diagnosis of diabetic retinopathy--detection of exudates in color fundus images of the human retina.

Authors:  Thomas Walter; Jean-Claude Klein; Pascale Massin; Ali Erginay
Journal:  IEEE Trans Med Imaging       Date:  2002-10       Impact factor: 10.048

2.  A novel automatic image processing algorithm for detection of hard exudates based on retinal image analysis.

Authors:  Clara I Sánchez; Roberto Hornero; María I López; Mateo Aboy; Jesús Poza; Daniel Abásolo
Journal:  Med Eng Phys       Date:  2007-06-06       Impact factor: 2.242

3.  Retinal image analysis based on mixture models to detect hard exudates.

Authors:  Clara I Sánchez; María García; Agustín Mayo; María I López; Roberto Hornero
Journal:  Med Image Anal       Date:  2009-05-28       Impact factor: 8.545

Review 4.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

5.  Receptive fields and functional architecture of monkey striate cortex.

Authors:  D H Hubel; T N Wiesel
Journal:  J Physiol       Date:  1968-03       Impact factor: 5.182

6.  An observational study to assess if automated diabetic retinopathy image assessment software can replace one or more steps of manual imaging grading and to determine their cost-effectiveness.

Authors:  Adnan Tufail; Venediktos V Kapetanakis; Sebastian Salas-Vega; Catherine Egan; Caroline Rudisill; Christopher G Owen; Aaron Lee; Vern Louw; John Anderson; Gerald Liew; Louis Bolter; Clare Bailey; SriniVas Sadda; Paul Taylor; Alicja R Rudnicka
Journal:  Health Technol Assess       Date:  2016-12       Impact factor: 4.014

7.  Exudate-based diabetic macular edema detection in fundus images using publicly available datasets.

Authors:  Luca Giancardo; Fabrice Meriaudeau; Thomas P Karnowski; Yaqin Li; Seema Garg; Kenneth W Tobin; Edward Chaum
Journal:  Med Image Anal       Date:  2011-07-23       Impact factor: 8.545

8.  Results of Automated Retinal Image Analysis for Detection of Diabetic Retinopathy from the Nakuru Study, Kenya.

Authors:  Morten B Hansen; Michael D Abràmoff; James C Folk; Wanjiku Mathenge; Andrew Bastawrous; Tunde Peto
Journal:  PLoS One       Date:  2015-10-01       Impact factor: 3.240

Review 9.  The worldwide epidemic of diabetic retinopathy.

Authors:  Yingfeng Zheng; Mingguang He; Nathan Congdon
Journal:  Indian J Ophthalmol       Date:  2012 Sep-Oct       Impact factor: 1.848

Review 10.  Automated detection of diabetic retinopathy in retinal images.

Authors:  Carmen Valverde; Maria Garcia; Roberto Hornero; Maria I Lopez-Galvez
Journal:  Indian J Ophthalmol       Date:  2016-01       Impact factor: 1.848

  10 in total
  13 in total

1.  Deep CNN with Hybrid Binary Local Search and Particle Swarm Optimizer for Exudates Classification from Fundus Images.

Authors:  J Ramya; M P Rajakumar; B Uma Maheswari
Journal:  J Digit Imaging       Date:  2022-01-07       Impact factor: 4.056

2.  An Intelligent Model for Blood Vessel Segmentation in Diagnosing DR Using CNN.

Authors:  S N Sangeethaa; P Uma Maheswari
Journal:  J Med Syst       Date:  2018-08-15       Impact factor: 4.460

3.  A Deep Neural Network-Based Method for Early Detection of Osteoarthritis Using Statistical Data.

Authors:  Jihye Lim; Jungyoon Kim; Songhee Cheon
Journal:  Int J Environ Res Public Health       Date:  2019-04-10       Impact factor: 3.390

4.  Study of the Application of Deep Convolutional Neural Networks (CNNs) in Processing Sensor Data and Biomedical Images.

Authors:  Weijun Hu; Yan Zhang; Lijie Li
Journal:  Sensors (Basel)       Date:  2019-08-17       Impact factor: 3.576

5.  The Use of Deep Learning to Predict Stroke Patient Mortality.

Authors:  Songhee Cheon; Jungyoon Kim; Jihye Lim
Journal:  Int J Environ Res Public Health       Date:  2019-05-28       Impact factor: 3.390

Review 6.  Different fundus imaging modalities and technical factors in AI screening for diabetic retinopathy: a review.

Authors:  Gilbert Lim; Valentina Bellemo; Yuchen Xie; Xin Q Lee; Michelle Y T Yip; Daniel S W Ting
Journal:  Eye Vis (Lond)       Date:  2020-04-14

7.  Machine Learning Based Automated Segmentation and Hybrid Feature Analysis for Diabetic Retinopathy Classification Using Fundus Image.

Authors:  Aqib Ali; Salman Qadri; Wali Khan Mashwani; Wiyada Kumam; Poom Kumam; Samreen Naeem; Atila Goktas; Farrukh Jamal; Christophe Chesneau; Sania Anam; Muhammad Sulaiman
Journal:  Entropy (Basel)       Date:  2020-05-19       Impact factor: 2.524

8.  Precise higher-order reflectivity and morphology models for early diagnosis of diabetic retinopathy using OCT images.

Authors:  A Sharafeldeen; M Elsharkawy; F Khalifa; A Soliman; M Ghazal; M AlHalabi; M Yaghi; M Alrahmawy; S Elmougy; H S Sandhu; A El-Baz
Journal:  Sci Rep       Date:  2021-02-25       Impact factor: 4.379

9.  Based on improved deep convolutional neural network model pneumonia image classification.

Authors:  Lingzhi Kong; Jinyong Cheng
Journal:  PLoS One       Date:  2021-11-04       Impact factor: 3.240

10.  Deep Learning-Based Diabetic Retinopathy Severity Grading System Employing Quadrant Ensemble Model.

Authors:  Charu Bhardwaj; Shruti Jain; Meenakshi Sood
Journal:  J Digit Imaging       Date:  2021-03-08       Impact factor: 4.056

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.