| Literature DB >> 33110484 |
Hao Liu1, Keqiang Yue1, Siyi Cheng1, Chengming Pan1, Jie Sun1, Wenjun Li1.
Abstract
Diabetic retinopathy (DR) is one of the most common complications of diabetes and the main cause of blindness. The progression of the disease can be prevented by early diagnosis of DR. Due to differences in the distribution of medical conditions and low labor efficiency, the best time for diagnosis and treatment was missed, which results in impaired vision. Using neural network models to classify and diagnose DR can improve efficiency and reduce costs. In this work, an improved loss function and three hybrid model structures Hybrid-a, Hybrid-f, and Hybrid-c were proposed to improve the performance of DR classification models. EfficientNetB4, EfficientNetB5, NASNetLarge, Xception, and InceptionResNetV2 CNNs were chosen as the basic models. These basic models were trained using enhance cross-entropy loss and cross-entropy loss, respectively. The output of the basic models was used to train the hybrid model structures. Experiments showed that enhance cross-entropy loss can effectively accelerate the training process of the basic models and improve the performance of the models under various evaluation metrics. The proposed hybrid model structures can also improve DR classification performance. Compared with the best-performing results in the basic models, the accuracy of DR classification was improved from 85.44% to 86.34%, the sensitivity was improved from 98.48% to 98.77%, the specificity was improved from 71.82% to 74.76%, the precision was improved from 90.27% to 91.37%, and the F1 score was improved from 93.62% to 93.9% by using hybrid model structures.Entities:
Year: 2020 PMID: 33110484 PMCID: PMC7579670 DOI: 10.1155/2020/8840174
Source DB: PubMed Journal: J Healthc Eng ISSN: 2040-2295 Impact factor: 2.682
Classification results from the basic models and the hybrid model structures.
| Loss | Accuracy | Sensitivity | Specificity | Precision | F1 score | ||
|---|---|---|---|---|---|---|---|
| Basic models | EfficientNetB4 | CE | 0.8158 | 0.9442 | 0.7182 | 0.9027 | 0.9230 |
| E-CE | 0.8544 | 0.9736 | 0.7061 | 0.9017 | 0.9362 | ||
| EfficientNetB5 | CE | 0.7932 | 0.9254 | 0.6782 | 0.8884 | 0.9065 | |
| E-CE | 0.8488 | 0.9809 | 0.6549 | 0.8872 | 0.9317 | ||
| NASNetLarge | CE | 0.7828 | 0.9151 | 0.7031 | 0.8951 | 0.9050 | |
| E-CE | 0.8470 | 0.9845 | 0.6353 | 0.8820 | 0.9304 | ||
| InceptionResNetV2 | CE | 0.7888 | 0.9657 | 0.5177 | 0.8471 | 0.9025 | |
| E-CE | 0.8502 | 0.9739 | 0.6963 | 0.8987 | 0.9348 | ||
| Xception | CE | 0.8100 | 0.9706 | 0.5742 | 0.8632 | 0.9138 | |
| E-CE | 0.8476 | 0.9848 | 0.6217 | 0.8781 | 0.9284 | ||
|
| |||||||
| Hybrid model | Hybrid-model-a | CE | 0.8584 |
| 0.6481 | 0.8860 | 0.9341 |
| Hybrid-model-f | CE | 0.8626 | 0.9652 |
|
| 0.9387 | |
| Hybrid-model-c | CE |
| 0.9706 | 0.7325 | 0.9094 |
| |
CE indicates cross-entropy loss function; E-CE indicates enhance cross-entropy loss function; and the bold values indicate the best results.
Figure 1Graph of the proposed algorithm architecture.
Figure 2Examples of different severity of DR fundus images. (a) No DR. (b) Mild DR. (c) Moderate DR. (d) Severe DR. (e) Proliferative DR.
The DR grade distribution of the dataset.
| Datasets | DR grade | |||||
|---|---|---|---|---|---|---|
| 0 | 1 | 2 | 3 | 4 | Total number | |
| DeepDR | 540 | 140 | 234 | 214 | 72 | 1200 |
| APTOS | 1805 | 370 | 999 | 193 | 295 | 3662 |
| EyePACS | 25810 | 2443 | 5292 | 873 | 708 | 35126 |
| Total number | 28155 | 2953 | 6525 | 1280 | 1075 | 39988 |
| Percentage (%) | 70.41 | 7.38 | 16.32 | 3.2 | 2.69 | — |
Figure 3The process of removing the black border of fundus images. (a) The unprocessed fundus image and (b) the processed fundus image.
Figure 4The structure of Hybrid-f.
Figure 5The structure of Hybrid-c.
The details of the Hybrid-c structure.
| Layer | Units | Filters | Kernel size | Padding | Output shape |
|
| |||||
| Input | — | — | — | — | 5 × 5 × 1 |
| Conv2d_1 | — | 256 | 3 | 1 | 5 × 5 × 256 |
| Conv2d_2 | — | 256 | 3 | 1 | 5 × 5 × 256 |
| Conv2d_3 | — | 256 | 3 | 0 | 3 × 3 × 256 |
| Flatten | — | — | — | — | 2304 |
| Dense_1 | 2048 | — | — | — | 2048 |
| Dense_2 | 5 | — | — | — | 5 |
Figure 6The epochs of training the basic models using E-CE loss and CE loss. (a) Training accuracy varies with epochs for basic models from EfficientNetB4 E-CE to Xception CE. (b) Training loss varies with epochs for basic models from EfficientNetB4 E-CE to Xception CE.
The confusion matrix of Hybrid-c.
| Predicted DR grade | 0 | 1 | 2 | 3 | 4 |
|---|---|---|---|---|---|
| 0 | 3565/97.06 | 36/0.98 | 68/1.85 | 1/0.03 | 3/0.08 |
| 1 | 204/58.96 | 87/25.14 | 54/15.61 | 0/0 | 1/0.29 |
| 2 | 140/18.62 | 47/6.25 | 540/71.81 | 15/1.99 | 10/1.33 |
| 3 | 7/5.69 | 0/0 | 55/44.72 | 54/43.9 | 7/5.69 |
| 4 | 4/3.78 | 1/0.94 | 18/16.98 | 12/11.32 | 71/66.98 |
The first item in each grid cell is the number of fundus images. The second item is the percentage of the images in the DR grade.