Varun Gulshan1, Lily Peng1, Marc Coram1, Martin C Stumpe1, Derek Wu1, Arunachalam Narayanaswamy1, Subhashini Venugopalan2, Kasumi Widner1, Tom Madams1, Jorge Cuadros3, Ramasamy Kim4, Rajiv Raman5, Philip C Nelson1, Jessica L Mega6, Dale R Webster1. 1. Google Inc, Mountain View, California. 2. Google Inc, Mountain View, California2Department of Computer Science, University of Texas, Austin. 3. EyePACS LLC, San Jose, California4School of Optometry, Vision Science Graduate Group, University of California, Berkeley. 4. Aravind Medical Research Foundation, Aravind Eye Care System, Madurai, India. 5. Shri Bhagwan Mahavir Vitreoretinal Services, Sankara Nethralaya, Chennai, Tamil Nadu, India. 6. Verily Life Sciences, Mountain View, California8Cardiovascular Division, Department of Medicine, Brigham and Women's Hospital and Harvard Medical School, Boston, Massachusetts.
Abstract
Importance: Deep learning is a family of computational methods that allow an algorithm to program itself by learning from a large set of examples that demonstrate the desired behavior, removing the need to specify rules explicitly. Application of these methods to medical imaging requires further assessment and validation. Objective: To apply deep learning to create an algorithm for automated detection of diabetic retinopathy and diabetic macular edema in retinal fundus photographs. Design and Setting: A specific type of neural network optimized for image classification called a deep convolutional neural network was trained using a retrospective development data set of 128 175 retinal images, which were graded 3 to 7 times for diabetic retinopathy, diabetic macular edema, and image gradability by a panel of 54 US licensed ophthalmologists and ophthalmology senior residents between May and December 2015. The resultant algorithm was validated in January and February 2016 using 2 separate data sets, both graded by at least 7 US board-certified ophthalmologists with high intragrader consistency. Exposure: Deep learning-trained algorithm. Main Outcomes and Measures: The sensitivity and specificity of the algorithm for detecting referable diabetic retinopathy (RDR), defined as moderate and worse diabetic retinopathy, referable diabetic macular edema, or both, were generated based on the reference standard of the majority decision of the ophthalmologist panel. The algorithm was evaluated at 2 operating points selected from the development set, one selected for high specificity and another for high sensitivity. Results: The EyePACS-1 data set consisted of 9963 images from 4997 patients (mean age, 54.4 years; 62.2% women; prevalence of RDR, 683/8878 fully gradable images [7.8%]); the Messidor-2 data set had 1748 images from 874 patients (mean age, 57.6 years; 42.6% women; prevalence of RDR, 254/1745 fully gradable images [14.6%]). For detecting RDR, the algorithm had an area under the receiver operating curve of 0.991 (95% CI, 0.988-0.993) for EyePACS-1 and 0.990 (95% CI, 0.986-0.995) for Messidor-2. Using the first operating cut point with high specificity, for EyePACS-1, the sensitivity was 90.3% (95% CI, 87.5%-92.7%) and the specificity was 98.1% (95% CI, 97.8%-98.5%). For Messidor-2, the sensitivity was 87.0% (95% CI, 81.1%-91.0%) and the specificity was 98.5% (95% CI, 97.7%-99.1%). Using a second operating point with high sensitivity in the development set, for EyePACS-1 the sensitivity was 97.5% and specificity was 93.4% and for Messidor-2 the sensitivity was 96.1% and specificity was 93.9%. Conclusions and Relevance: In this evaluation of retinal fundus photographs from adults with diabetes, an algorithm based on deep machine learning had high sensitivity and specificity for detecting referable diabetic retinopathy. Further research is necessary to determine the feasibility of applying this algorithm in the clinical setting and to determine whether use of the algorithm could lead to improved care and outcomes compared with current ophthalmologic assessment.
Importance: Deep learning is a family of computational methods that allow an algorithm to program itself by learning from a large set of examples that demonstrate the desired behavior, removing the need to specify rules explicitly. Application of these methods to medical imaging requires further assessment and validation. Objective: To apply deep learning to create an algorithm for automated detection of diabetic retinopathy and diabetic macular edema in retinal fundus photographs. Design and Setting: A specific type of neural network optimized for image classification called a deep convolutional neural network was trained using a retrospective development data set of 128 175 retinal images, which were graded 3 to 7 times for diabetic retinopathy, diabetic macular edema, and image gradability by a panel of 54 US licensed ophthalmologists and ophthalmology senior residents between May and December 2015. The resultant algorithm was validated in January and February 2016 using 2 separate data sets, both graded by at least 7 US board-certified ophthalmologists with high intragrader consistency. Exposure: Deep learning-trained algorithm. Main Outcomes and Measures: The sensitivity and specificity of the algorithm for detecting referable diabetic retinopathy (RDR), defined as moderate and worse diabetic retinopathy, referable diabetic macular edema, or both, were generated based on the reference standard of the majority decision of the ophthalmologist panel. The algorithm was evaluated at 2 operating points selected from the development set, one selected for high specificity and another for high sensitivity. Results: The EyePACS-1 data set consisted of 9963 images from 4997 patients (mean age, 54.4 years; 62.2% women; prevalence of RDR, 683/8878 fully gradable images [7.8%]); the Messidor-2 data set had 1748 images from 874 patients (mean age, 57.6 years; 42.6% women; prevalence of RDR, 254/1745 fully gradable images [14.6%]). For detecting RDR, the algorithm had an area under the receiver operating curve of 0.991 (95% CI, 0.988-0.993) for EyePACS-1 and 0.990 (95% CI, 0.986-0.995) for Messidor-2. Using the first operating cut point with high specificity, for EyePACS-1, the sensitivity was 90.3% (95% CI, 87.5%-92.7%) and the specificity was 98.1% (95% CI, 97.8%-98.5%). For Messidor-2, the sensitivity was 87.0% (95% CI, 81.1%-91.0%) and the specificity was 98.5% (95% CI, 97.7%-99.1%). Using a second operating point with high sensitivity in the development set, for EyePACS-1 the sensitivity was 97.5% and specificity was 93.4% and for Messidor-2 the sensitivity was 96.1% and specificity was 93.9%. Conclusions and Relevance: In this evaluation of retinal fundus photographs from adults with diabetes, an algorithm based on deep machine learning had high sensitivity and specificity for detecting referable diabetic retinopathy. Further research is necessary to determine the feasibility of applying this algorithm in the clinical setting and to determine whether use of the algorithm could lead to improved care and outcomes compared with current ophthalmologic assessment.
Authors: Bibo Shi; Lars J Grimm; Maciej A Mazurowski; Jay A Baker; Jeffrey R Marks; Lorraine M King; Carlo C Maley; E Shelley Hwang; Joseph Y Lo Journal: J Am Coll Radiol Date: 2018-02-02 Impact factor: 5.532
Authors: Cathy A Goldstein; Richard B Berry; David T Kent; David A Kristo; Azizi A Seixas; Susan Redline; M Brandon Westover Journal: J Clin Sleep Med Date: 2020-04-15 Impact factor: 4.062
Authors: Diane C Lim; Diego R Mazzotti; Kate Sutherland; Jesse W Mindel; Jinyoung Kim; Peter A Cistulli; Ulysses J Magalang; Allan I Pack; Philip de Chazal; Thomas Penzel Journal: Sleep Med Rev Date: 2020-03-20 Impact factor: 11.609
Authors: Julian Betancur; Frederic Commandeur; Mahsaw Motlagh; Tali Sharir; Andrew J Einstein; Sabahat Bokhari; Mathews B Fish; Terrence D Ruddy; Philipp Kaufmann; Albert J Sinusas; Edward J Miller; Timothy M Bateman; Sharmila Dorbala; Marcelo Di Carli; Guido Germano; Yuka Otaki; Balaji K Tamarappoo; Damini Dey; Daniel S Berman; Piotr J Slomka Journal: JACC Cardiovasc Imaging Date: 2018-03-14
Authors: Peter A Noseworthy; Zachi I Attia; LaPrincess C Brewer; Sharonne N Hayes; Xiaoxi Yao; Suraj Kapa; Paul A Friedman; Francisco Lopez-Jimenez Journal: Circ Arrhythm Electrophysiol Date: 2020-02-16
Authors: K G van Leeuwen; H Sun; M Tabaeizadeh; A F Struck; M J A M van Putten; M B Westover Journal: Clin Neurophysiol Date: 2018-11-17 Impact factor: 3.708