| Literature DB >> 29168750 |
Kele Xu1, Dawei Feng2, Haibo Mi3.
Abstract
The automatic detection of diabetic retinopathy is of vital importance, as it is the main cause of irreversible vision loss in the working-age population in the developed world. The early detection of diabetic retinopathy occurrence can be very helpful for clinical treatment; although several different feature extraction approaches have been proposed, the classification task for retinal images is still tedious even for those trained clinicians. Recently, deep convolutional neural networks have manifested superior performance in image classification compared to previous handcrafted feature-based image classification methods. Thus, in this paper, we explored the use of deep convolutional neural network methodology for the automatic classification of diabetic retinopathy using color fundus image, and obtained an accuracy of 94.5% on our dataset, outperforming the results obtained by using classical approaches.Entities:
Keywords: deep convolutional neural network; diabetic retinopathy; image classification
Mesh:
Year: 2017 PMID: 29168750 PMCID: PMC6149821 DOI: 10.3390/molecules22122054
Source DB: PubMed Journal: Molecules ISSN: 1420-3049 Impact factor: 4.411
Figure 1Sample frames of the retina images. The first two frames in the top row come from normal subjects, while the two frames in the bottom row come from the patients who have diabetic retinopathy.
Data augmentation parameters.
| Transformation Type | Description |
|---|---|
| Rotation | 0 |
| Flipping | 0 (without flipping) or 1 (with flipping) |
| Shearing | Randomly with angle between −15 |
| Rescaling | Randomly with scale factor between 1/1.6 and 1.6 |
| Translation | Randomly with shift between −10 and 10 pixels |
Figure 2Samples of the transformed frames.
Figure 3An exemplary architecture of the convolutional neural network.
Convolutional neural network (CNN) architecture used in our experiment.
| Output Shape | Description |
|---|---|
| 224 × 224 × 3 | input |
| 222 × 222 × 32 | 3 × 3 convolution, 32 filter |
| 220 × 220 × 32 | 3 × 3 convolution, 32 filter |
| 110 × 110 × 32 | 2 × 2 max-pooling |
| 108 × 108 × 64 | 3 × 3 convolution, 64 filter |
| 106 × 106 × 64 | 3 × 3 convolution, 64 filter |
| 53 × 53 × 64 | 2 × 2 max-pooling |
| 53 × 53 × 128 | 3 × 3 convolution, 128 filter |
| 51 × 51 × 128 | 3 × 3 convolution, 128 filter |
| 49 × 49 × 128 | 2 × 2 max-pooling |
| 24 × 24 × 256 | 3 × 3 convolution, 256 filter |
| 22 × 22 × 256 | 3 × 3 convolution, 256 filter |
| 11 × 11 × 256 | 2 × 2 max-pooling |
| 4096 | flatterned and fully connected |
| 1024 | fully connected |
| 2 | softmax |
Performance comparison with different approaches.
| Method | Accuracy |
|---|---|
| Hard exudates + GBM | 89.4% |
| Red lesions + GBM | 88.7% |
| Micro-aneurysms + GBM | 86.2% |
| Blood vessel detection + GBM | 79.1% |
| CNN without data augmentation | 91.5% |
| CNN with data augmentation | 94.5% |
Figure 4Visualization of the trained neural networks. Each image represents the activations of the first layer during the forward pass.