Literature DB >> 26425849

Results of Automated Retinal Image Analysis for Detection of Diabetic Retinopathy from the Nakuru Study, Kenya.

Morten B Hansen1, Michael D Abràmoff2, James C Folk2, Wanjiku Mathenge3, Andrew Bastawrous4, Tunde Peto5.   

Abstract

OBJECTIVE: Digital retinal imaging is an established method of screening for diabetic retinopathy (DR). It has been established that currently about 1% of the world's blind or visually impaired is due to DR. However, the increasing prevalence of diabetes mellitus and DR is creating an increased workload on those with expertise in grading retinal images. Safe and reliable automated analysis of retinal images may support screening services worldwide. This study aimed to compare the Iowa Detection Program (IDP) ability to detect diabetic eye diseases (DED) to human grading carried out at Moorfields Reading Centre on the population of Nakuru Study from Kenya. PARTICIPANTS: Retinal images were taken from participants of the Nakuru Eye Disease Study in Kenya in 2007/08 (n = 4,381 participants [NW6 Topcon Digital Retinal Camera]).
METHODS: First, human grading was performed for the presence or absence of DR, and for those with DR this was sub-divided in to referable or non-referable DR. The automated IDP software was deployed to identify those with DR and also to categorize the severity of DR. MAIN OUTCOME MEASURES: The primary outcomes were sensitivity, specificity, and positive and negative predictive value of IDP versus the human grader as reference standard.
RESULTS: Altogether 3,460 participants were included. 113 had DED, giving a prevalence of 3.3% (95% CI, 2.7-3.9%). Sensitivity of the IDP to detect DED as by the human grading was 91.0% (95% CI, 88.0-93.4%). The IDP ability to detect DED gave an AUC of 0.878 (95% CI 0.850-0.905). It showed a negative predictive value of 98%. The IDP missed no vision threatening retinopathy in any patients and none of the false negative cases met criteria for treatment.
CONCLUSIONS: In this epidemiological sample, the IDP's grading was comparable to that of human graders'. It therefore might be feasible to consider inclusion into usual epidemiological grading.

Entities:  

Mesh:

Year:  2015        PMID: 26425849      PMCID: PMC4591009          DOI: 10.1371/journal.pone.0139148

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Regular screening for diabetic retinopathy (DR) for at risk populations has been shown to be an effective public health intervention for reducing the burden of disease in people living with diabetes mellitus (DM), as DR is still one of the leading causes of visual impairment in industrialised countries. [1, 2] The latest global estimates show that there are 285 million people with visual impairment, of whom 39 million are blind, with DR contributing to 1% of both to blindness and visual impairment. [3] The low-income countries in Asia and Africa have both the highest prevalence of DM and the highest expected rise in diseased population[4], but these resource-poor communities lack infrastructure to implement organised DR screening programs (DRSP) where regular eye exams [5, 6] can minimise the risk of visual loss. For example with only 2.7 ophthalmologists per million people in Sub-Saharan Africa, recommending the addition of DRSP in the format it is currently delivered in more developed countries is extremely challenging. [7, 8] Even in developed countries, such as the US, less than 60% of the 23 million people with DM had an eye examination in 2010. [7, 9] On the contrary, in the UK, where the DRSP is publicly funded, nearly 80% of DM patients are screened annually, [5], and as a result of this, DR is no longer the leading cause of blindness in the working age-group. [6] Potentially, algorithms performing automated analysis of retinal images may address the need for more affordable DRSP in certain settings. The Iowa Detection Program (IDP) was originally designed to meet the rising demand for DRSP [10, 11]. IDP evaluates digital retinal images in an automated fashion for the presence of moderate or more severe DR as well as diabetic macular edema (DME) [10]. The IDP has been validated in independent cohorts of people with DM, using internationally recognized DR grading standards [9, 12] and was reported to have diagnostic accuracy comparable to that of fellowship-trained retinal specialists’. [10, 11]However, the performance of IDP is yet to be evaluated on populations from Sub-Saharan Africa, especially in the context of a population-based study where majority of the subjects are without DM. The aim of this study is to determine the sensitivity, specificity, positive and negative predictive value of IDP in detecting Diabetic Eye Disease (DED) and non-DED in the population based Nakuru Study from Kenya. Human grading for DR used in this study was that carried out in the Moorfields Reading Centre, London, UK.

Methods

Subject recruitment and imaging

Images and clinical data of participants of the Nakuru Eye Study, Kenya, were used for the automated image analysis via the IDP software. The Nakuru Eye Study started in 2007/08 where a total of 4,381 participants underwent complete ophthalmic examination, and in 3460 subjects, fundus imaging was performed and subsequently graded for DR. [13] In summary, for fundus photography participants were dilated with Mydriacyl drops (Alcon®) and retinal photographs were obtained using a Topcon® NW6S Non-Mydriatic Retinal Camera. Two fields were imaged, one centred on the disc, one on the fovea. [13] All fundus photographs were graded by human graders at Moorfields Eye Hospitals Reading Centre, UK. [14] The grading protocol was as follows:

The grading protocol and ICDR reference standard

The human graders were masked to the patient’s status to having or not having DM. Table 1 Shows details of the grading scheme for human graders, in short, no retinopathy, mild, moderate and severe non-proliferative DR (NPDR) and proliferative DR (PDR) characteristics were determined by identifying individual lesions in the retina and the macular region. These lesions were microaneurysm, haemorrhages, exudates, characteristics of clinically significant macular edema (CSME) without being able to determine thickness due to lack of stereo imaging, cotton wool spots (CWS), intraretinal microvascular abnormalities (IRMA) and new vessels on the disc or elsewhere. In addition vascular abnormalities resembling new vessels characteristic of DR were graded as referable retinopathy as well.
Table 1

Grading protocol ICDR and ETDRS Severity Levels.

MeasureScoreObservable Findings
ICDR severity level
No apparent retinopathy0No abnormalities (Level 10 ETDRS)
Mild non-proliferative diabetic retinopathy1Microaneurysm(s) only (Level 20 ETDRS)
Moderate non-proliferative diabetic retinopathy2More than just microaneurysm(s) but less than severe non-proliferative diabetic retinopathy (Level 35, 43, 47 ETDRS)
Severe non-proliferative diabetic retinopathy3Any of the following: > 20 intra-retinal haemorrhages in each of 4 quadrants, definite venous beading in ≥2 quadrants, prominent intra-retinal microvascular abnormalities in ≥1 quadrant, or no signs of proliferative retinopathy. (Level 53 ETDRS: 4-2-1 rule)
Proliferative diabetic retinopathy4One or more of the following: neovascularization and/or vitreous or preretinal haemorrhages. (Levels 61, 65, 71, 75, 81, 85 ETDRS)
Macular oedema severity level
No macular oedema0No exudates and no apparent thickening within 1 disc diameter from fovea
Macular oedema1Exudates or apparent thickening within 1 disc diameter from fovea

Abbreviations: ETDRS, Early Treatment Diabetic Retinopathy study; ICDR, International Clinical Diabetic Retinopathy

Abbreviations: ETDRS, Early Treatment Diabetic Retinopathy study; ICDR, International Clinical Diabetic Retinopathy

Iowa Diabetic Retinopathy (IDP) detection software

The details of the Iowa Detection Program (IDP) algorithm have been published in previous papers. [9-11] IDP examines and analyses every pixel in an image to detect microaneurysms, haemorrhages, exudates and cotton wool spots and has the ability to detect retinal neovascularization. It combines image quality with its detection of lesions and makes a numerical output between 0 and 1, called the dr-index. The closer the number to 1 the more likely that the patient has DED, or that the image was ungradable. Both of these categories (DED and ungradable) then is required to be regarded by a human grader. The dr-index is compared to the so-called set-point to give the final result of the analysis. The lower the set-point of the IDP, the higher the sensitivity, with increases the likelihood of false positives (overcalls), and in turn lowers specificity. Based on previous work[9, 11], 0.04 as a setpoint was found to be ideal to strike a balance between sensitivity and specificity as at this level a sensitivity of 91.6% was established. [11]

Methods of comparing human grading to IDP output

In order to be able to compare human grading output to DED output of IDP, the human grading had to be re-categorized as following: no-DED includes none/mild DR with no macular oedema (denoted with 0); while DED includes levels of DR that are either currently sight threatening (severe NPDR and above) or that has a higher chance of developing into sight threatening disease within a year, such as moderate NPDR. In summary, the algorithm is set to detect moderate to severe non-proliferative DR (NPDR), proliferative DR (PDR), or Diabetic Macular Edema (DME), denoted as 1 for statistical purposes. The IDP processed all available images. Once IDP results were locked using the predefined set-point the human grading results were re-coded to reflect the IDP’s outcome as above, so statistical analysis and like-for-like comparison could be made. Once the grading results were brought into a comparable outcome, we reviewed the results and statistical analysis took place. As mentioned the IDP categorises results as DED (moderate NPDR or above) present or absent. Ungradeable images also obtain the value of 1, meaning DED due to the fact that these images require human grading intervention, therefore for fair comparison, the image sets were also compared including the ungradables.

Primary outcome measure

The primary outcomes were sensitivity, specificity, positive and negative predictive value of IDP versus the human grading.

Statistical analysis

Statistical analysis was performed using SPSS, version 22.0.0.0 (IBM Corporation, Worldwide). Sensitivity, specificity, negative predictive value (NPV), positive predictive value (PPV) of IDP compared to the reference standard of human grading were calculated. Two methods were used to calculate these values. First we used the IDP’s output of DED and ungradeable/missing images were classified as a positive outcome and compared to the same setting for the human grading. Second method we used only results from gradable images to be analysed, both by the IDP and by the human grading for comparison.

Ethics approval

The Nakuru Eye study was granted ethical approval for their work granted by the London School of Hygiene & Tropical Medicine, the Kenya Medical Research Institute (KEMRI) Ethical Committee and African Medical and Research Foundation (AMREF). Informed consent was obtained from the participating subjects. [13]

Results

All available baseline images’ from the Nakuru study (6788 fundus photos of 3,460 participants) were graded both by human graders and IDP. Altogether, 132 eyes could not be photographed at all, and these were therefore not available for grading and subsequent statistical analysis. Of the 3,460 participants, the human graders determined that 113 subjects had DED, giving a prevalence of 3.3% (95% CI, 2.7–3.9%). There were 56 cases with moderate NPDR, 20 with severe NPDR and 37 cases with PDR. In the Nakuru study altogether 6.5% of patients had diabetes, giving a 50.8% prevalence of DR in this cohort. [14] Of the 3,460 participants, there were 91 true positive DED, 900 false positive DED, 2093 true negative DED and 22 false negative DED results. In 334 cases both the IDP and human graded the image to be ungradable and therefor as a positive outcome. In 20 cases the human graded the images as ungradable were the IDP analysed the images as sufficient for grading but with no-DED. First we analysed the total agreement between IDP and human grading including ungradable results. Sensitivity of the IDP to the RC was 91.0% (95% CI, 88.0–93.4%), and specificity was 69.9% (95% CI, 68.3–71.6%). The PPV was 32.1% (95% CI, 29.6–34.7%), whereas the NPV was 98.0% (95% CI, 97.4–98.6%). (Table 2)
Table 2

Results—Comparison of the grading of DED from the IDP and the Reading Centre.

Reading Centre
DEDNo-DEDUngradable
DED919000991
IDP No-DED222093202135
Ungradable00334334
11329933543460
The IDP ability to detect DED gave in Fig 1 an AUC of 0.878 (95% CI 0.850–0.905) using maximum likelihood estimation. [15]
Fig 1

–The ROC.

Then we analysed the results without the ungradable images. We used the 91 true positive, 900 false positive, 2093 true negative and 22 false negative. The sensitivity was 80.5% (95% CI, 72.0–87.4%) and the specificity was 69.9% (95% CI, 68.3–71.6%). The PPV was 9.2% (95% CI, 7.5–11.2%) whereas the NPV was 99.0% (95% CI, 98.4–99.4%). (Table 2) Of the 22 false negative cases (42 originally but 20 cases due to ungradable images), the human graders classified five cases as PDR, two as severe NPDR, and 15 as moderate NPDR. A retinal specialist, TP, reviewed these images a second time. The five cases read as PDR are shown in Fig 2. Review by the retinal specialist found none to have DR, but all were vascular abnormalities on the optic disc resembling PDR. Of the 2 cases graded as severe NPDR, both were confirmed to have haemorrhages and vascular abnormalities as IRMA, while of the 15 cases of moderate NPDR per the RC, three cases were re-adjudicated as mild DR while 12 remained at moderate level. The 12 cases of moderate NPDR were confirmed to be accurate with five showing only haemorrhages, three with IRMA and six cases showing a combination of small haemorrhages, IRMA and CWS.
Fig 2

–Five false negative cases.

Table 3 shows the results re-calculated once the adjudication results were taken into account. The modified sensitivity was 92.6% (95% CI, 89.7–94.7%), specificity 70.0% (95% CI, 68.3–71.6%), PPV 32.1% (95% CI, 29.6–34.7%) and NPV was 98.4% (95% CI, 97.8–98.9%).
Table 3

Results—Comparison of the grading of DED from the IDP and the Reading Centre–After review grading by retinal specialist.

Reading Centre
DEDNo-DEDUngradable
DED919000991
IDP No-DED142101202135
Ungradable00334334
10530013543460
Again after adjudication we performed analysis without the ungradable participants. This analysis resulted in 91 true positive, 1234 false positive, 2121 true negative and 14 false negative. This meaning a sensitivity of 86.7% (95% CI, 78.6–92.5%), and the specificity was 70.0% (95% CI, 68.3–71.7%). The PPV was 9.2% (95% CI, 7.5–11.2%) and the NPV was 99.3% (95% CI, 98.9–99.6%).

Discussion

Overall, using automated image analysis software designed to detect sight threatening DR or ungradable images on a population based sample at Nakuru district of Kenya, sensitivity of 91% and specificity of 69.9% were achieved with excellent NPV value of over 99%. No vision threatening DR was missed in any patients. None of the false negative cases were found to have vision-threatening DED by the human graders or met criteria for treatment. Delayed treatment could have resulted from these missed cases; it would have been unlikely to lead to visual loss and high-risk PDR within the recommended year of follow-up. [16] This is the first time IDP has been used on an epidemiologic population sample from Africa, i.e. not a set of people previously diagnosed with diabetes. With dark fundi in Sub-Saharan populations there has been concern that this would lead to underestimation of DED by the algorithms because lesions might have lower contrast with the retinal background. Our analysis did not show any such underestimation, and results were comparable to earlier studies in predominantly white / Caucasian populations. [9] The IDP performed well and it showed good results in differentiating participants into having no DED and having DED as a primary population screening tool and achieved this within the similar timeframe of one minute per eye as previously documented. [9] With DM becoming a major public health issue worldwide and a 69% rise estimated in the number of adults with DM in low-income countries [17], the need for cost-effective DRSP is essential. This opens up the opportunity for population based screening programs all over the world speed up the disease detection process from image acquisition to diagnosis and potentially reporting back to the outcome for timely treatment to be given. By reliably identifying patients who are in need of further image analysis by human graders and who potentially require treatment for DR conserves resources to those truly needing it. This could be of value in low-income countries where ophthalmic image grading is regularly carried out by the small number of ophthalmologists who are already overstretched coping with the demand. [18] Software that automatically detects sight threatening DR and ungradeable images in at risk patients could be part of the solution to that. As if IDP can cut down on the number of images to be seen by a human grading (in many cases an ophthalmologist) by over 60% as it did in this study, it would reduce the burden on health systems. Even though the IDP today delivers good results there are ways to improve. The IDP focuses on detecting microaneurysms, haemorrhages, and exudates. Better detection of subtle changes such as IRMA or beading and CWS may improve on the detection level between DED and non-DED. This should be a focus point for future improvements to the algorithm. At the moment the software only focuses on DR and no other eye diseases such as Age-Related Macular Degeneration (AMD) and glaucoma are covered and no other major blinding diseases are classified as abnormal by the software. For a complete clinical evaluation, the patients’ retinal images still need to be seen by trained graders or ophthalmologist. Combining software that screens for both DR, AMD and glaucoma would be the ideal solution for automated image analysis to lower the burden on healthcare worldwide. This should be an aim for the future. In conclusion, IDP performed well in detecting and differentiating between participants with or without sight threatening DED and identified those with too poor quality images appropriately. The IDP showed its ability to cut down on the images needed to be seen by a grader by over 60%, provided only the device’s positive DED outputs were to be reviewed by a human grader. Further improvements are required to enable automated image analysis for all common diseases so large scale epidemiological studies, such as the Nakuru Eye Study, can be safely graded with minimal human intervention.
  16 in total

1.  The number of ophthalmologists in practice and training worldwide: a growing gap despite more than 200,000 practitioners.

Authors:  Serge Resnikoff; William Felch; Tina-Marie Gauthier; Bruce Spivey
Journal:  Br J Ophthalmol       Date:  2012-03-26       Impact factor: 4.638

2.  The global inverse care law: a distorted map of blindness.

Authors:  Andrew Bastawrous; Benjamin D Hennig
Journal:  Br J Ophthalmol       Date:  2012-06-27       Impact factor: 4.638

3.  Automated early detection of diabetic retinopathy.

Authors:  Michael D Abràmoff; Joseph M Reinhardt; Stephen R Russell; James C Folk; Vinit B Mahajan; Meindert Niemeijer; Gwénolé Quellec
Journal:  Ophthalmology       Date:  2010-06       Impact factor: 12.079

4.  Fundus photographic risk factors for progression of diabetic retinopathy. ETDRS report number 12. Early Treatment Diabetic Retinopathy Study Research Group.

Authors: 
Journal:  Ophthalmology       Date:  1991-05       Impact factor: 12.079

Review 5.  Global estimates of visual impairment: 2010.

Authors:  Donatella Pascolini; Silvio Paolo Mariotti
Journal:  Br J Ophthalmol       Date:  2011-12-01       Impact factor: 4.638

Review 6.  Diabetic retinopathy.

Authors:  Ning Cheung; Paul Mitchell; Tien Yin Wong
Journal:  Lancet       Date:  2010-06-26       Impact factor: 79.321

Review 7.  Proposed international clinical diabetic retinopathy and diabetic macular edema disease severity scales.

Authors:  C P Wilkinson; Frederick L Ferris; Ronald E Klein; Paul P Lee; Carl David Agardh; Matthew Davis; Diana Dills; Anselm Kampik; R Pararajasegaram; Juan T Verdaguer
Journal:  Ophthalmology       Date:  2003-09       Impact factor: 12.079

8.  Global estimates of the prevalence of diabetes for 2010 and 2030.

Authors:  J E Shaw; R A Sicree; P Z Zimmet
Journal:  Diabetes Res Clin Pract       Date:  2009-11-06       Impact factor: 5.602

Review 9.  Current epidemiology of diabetic retinopathy and diabetic macular edema.

Authors:  Jie Ding; Tien Yin Wong
Journal:  Curr Diab Rep       Date:  2012-08       Impact factor: 4.810

10.  Evaluation of a system for automatic detection of diabetic retinopathy from color fundus photographs in a large population of patients with diabetes.

Authors:  Michael D Abràmoff; Meindert Niemeijer; Maria S A Suttorp-Schulten; Max A Viergever; Stephen R Russell; Bram van Ginneken
Journal:  Diabetes Care       Date:  2007-11-16       Impact factor: 19.112

View more
  18 in total

1.  Artificial Intelligence Screening for Diabetic Retinopathy: the Real-World Emerging Application.

Authors:  Valentina Bellemo; Gilbert Lim; Tyler Hyungtaek Rim; Gavin S W Tan; Carol Y Cheung; SriniVas Sadda; Ming-Guang He; Adnan Tufail; Mong Li Lee; Wynne Hsu; Daniel Shu Wei Ting
Journal:  Curr Diab Rep       Date:  2019-07-31       Impact factor: 4.810

2.  Diabetic Retinopathy Diagnosis from Retinal Images Using Modified Hopfield Neural Network.

Authors:  D Jude Hemanth; J Anitha; Le Hoang Son; Mamta Mittal
Journal:  J Med Syst       Date:  2018-10-31       Impact factor: 4.460

Review 3.  The Role of Retinal Imaging and Portable Screening Devices in Tele-ophthalmology Applications for Diabetic Retinopathy Management.

Authors:  Delia Cabrera DeBuc
Journal:  Curr Diab Rep       Date:  2016-12       Impact factor: 4.810

Review 4.  Diabetic Eye Screening: Knowledge and Perspectives from Providers and Patients.

Authors:  Yao Liu; Rebecca Swearingen
Journal:  Curr Diab Rep       Date:  2017-08-31       Impact factor: 4.810

Review 5.  A review on automatic analysis techniques for color fundus photographs.

Authors:  Renátó Besenczi; János Tóth; András Hajdu
Journal:  Comput Struct Biotechnol J       Date:  2016-10-06       Impact factor: 7.271

6.  Sensitivity and specificity of automated analysis of single-field non-mydriatic fundus photographs by Bosch DR Algorithm-Comparison with mydriatic fundus photography (ETDRS) for screening in undiagnosed diabetic retinopathy.

Authors:  Pritam Bawankar; Nita Shanbhag; S Smitha K; Bodhraj Dhawan; Aratee Palsule; Devesh Kumar; Shailja Chandel; Suneet Sood
Journal:  PLoS One       Date:  2017-12-27       Impact factor: 3.240

7.  Validation of automated screening for referable diabetic retinopathy with the IDx-DR device in the Hoorn Diabetes Care System.

Authors:  Amber A van der Heijden; Michael D Abramoff; Frank Verbraak; Manon V van Hecke; Albert Liem; Giel Nijpels
Journal:  Acta Ophthalmol       Date:  2017-11-27       Impact factor: 3.761

Review 8.  Impact and Trends in Global Ophthalmology.

Authors:  Lloyd B Williams; S Grace Prakalapakorn; Zubair Ansari; Raquel Goldhardt
Journal:  Curr Ophthalmol Rep       Date:  2020-06-22

9.  Detecting Visual Function Abnormality with a Contrast-Dependent Visual Test in Patients with Type 2 Diabetes.

Authors:  Li-Ting Tsai; Kuo-Meng Liao; Yuh Jang; Fu-Chang Hu; Wei-Chi Wu
Journal:  PLoS One       Date:  2016-09-09       Impact factor: 3.240

Review 10.  The Evolution of Diabetic Retinopathy Screening Programmes: A Chronology of Retinal Photography from 35 mm Slides to Artificial Intelligence.

Authors:  Josef Huemer; Siegfried K Wagner; Dawn A Sim
Journal:  Clin Ophthalmol       Date:  2020-07-20
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.