| Literature DB >> 33807545 |
Yu-Hsuan Li1,2, Wayne Huey-Herng Sheu1,3,4, Chien-Chih Chou5, Chun-Hsien Lin5, Yuan-Shao Cheng5, Chun-Yuan Wang5, Chieh Liang Wu6,7, I-Te Lee1,3,8,9.
Abstract
Deep learning-based software is developed to assist physicians in terms of diagnosis; however, its clinical application is still under investigation. We integrated deep-learning-based software for diabetic retinopathy (DR) grading into the clinical workflow of an endocrinology department where endocrinologists grade for retinal images and evaluated the influence of its implementation. A total of 1432 images from 716 patients and 1400 images from 700 patients were collected before and after implementation, respectively. Using the grading by ophthalmologists as the reference standard, the sensitivity, specificity, and area under the receiver operating characteristic curve (AUC) to detect referable DR (RDR) were 0.91 (0.87-0.96), 0.90 (0.87-0.92), and 0.90 (0.87-0.93) at the image level; and 0.91 (0.81-0.97), 0.84 (0.80-0.87), and 0.87 (0.83-0.91) at the patient level. The monthly RDR rate dropped from 55.1% to 43.0% after implementation. The monthly percentage of finishing grading within the allotted time increased from 66.8% to 77.6%. There was a wide range of agreement values between the software and endocrinologists after implementation (kappa values of 0.17-0.65). In conclusion, we observed the clinical influence of deep-learning-based software on graders without the retinal subspecialty. However, the validation using images from local datasets is recommended before clinical implementation.Entities:
Keywords: area under the curve; deep learning; diabetes; image; retinopathy
Year: 2021 PMID: 33807545 PMCID: PMC8035657 DOI: 10.3390/life11030200
Source DB: PubMed Journal: Life (Basel) ISSN: 2075-1729
Figure 1Distribution of diabetic retinopathy severity graded based on VeriSeeTM and ophthalmologists.
The performance of the software for diagnosing referable diabetic retinopathy (RDR) at the image level.
| VeriSeeTM | |
|---|---|
| Number of patients | - |
| Number of images | 981 |
| Sensitivity (95% CI) | 0.91 (0.83–0.96) |
| Specificity (95% CI) | 0.90 (0.87–0.92) |
| AUC (95% CI) | 0.90 (0.87–0.93) |
| F1 score | 0.62 (0.58–0.65) |
| Balanced accuracy | 0.90 (0.87–0.91) |
AUC = area under the receiver operating characteristic curve; CI = confidence interval.
The performance of software and the endocrinologists for diagnosing referable diabetic retinopathy at the patient level.
| VeriSeeTM | Endocrinologists | |
|---|---|---|
| Number of patients | 468 | 468 |
| Sensitivity (95% CI) | 0.91 (0.81–0.97) | 0.91 (0.81–0.97) |
| Specificity (95% CI) | 0.84 (0.80–0.87) | 0.50 (0.45–0.55) |
| AUC (95% CI) | 0.87 (0.83–0.91) | 0.70 (0.66–0.74) |
| F1 score (95% CI) | 0.58 (0.54–0.63) | 0.33 (0.28–0.37) |
| Balanced accuracy (95% CI) | 0.87 (0.83–0.89) | 0.70 (0.65–0.74) |
AUC = area under the receiver operating characteristic curve; CI = confidence interval.
Figure 2The area under the curve of the receiver operating characteristic curve (AUC) calculated at the patient level for VeriSeeTM and the endocrinologists.
Figure 3The rate of referable diabetic retinopathy (RDR) graded according to the endocrinologists, VeriSeeTM, and ophthalmologists.
The performance of endocrinologists before and after implementation of the software.
| Before | After | |
|---|---|---|
| Monthly RDR rate | 55.1% (258/468) | 42.9% (216/503) |
| Monthly rate of finishing grading on time * | 66.8% (478/716) | 77.6% (543/700) |
CI = confidence interval, RDR = referable diabetic retinopathy. * Within three days after fundus examination.
The characteristics and Kappa coefficients of the five endocrinologists.
| Experience * | Accuracy † | Images ‡ | Kappa § | |||
|---|---|---|---|---|---|---|
| Before | After | Change | ||||
| 1 | 2 | 0.71 | 209 | 0.17 | 0.50 | 0.33 |
| 2 | 8 | 0.72 | 257 | 0.16 | 0.43 | 0.27 |
| 3 | 11 | 0.7 | 230 | 0.06 | 0.31 | 0.25 |
| 4 | 13 | 0.61 | 189 | 0.05 | 0.17 | 0.12 |
| 5 | 17 | 0.77 | 121 | 0.37 | 0.65 | 0.28 |
* Experience: years of working as endocrinologist; † Accuracy: the diagnostic accuracy of referable diabetic retinopathy by the endocrinologists, and the grading by the ophthalmologists as the reference standard. ‡ Image: the number of images graded by each endocrinologist. § Kappa coefficient: the agreement between the software and endocrinologists.
Figure 4The kappa value for referable diabetic retinopathy (RDR) between each endocrinologist and VeriSeeTM before and after software implementation.