Literature DB >> 28842613

Accuracy of deep learning, a machine-learning technology, using ultra-wide-field fundus ophthalmoscopy for detecting rhegmatogenous retinal detachment.

Hideharu Ohsugi1, Hitoshi Tabuchi2, Hiroki Enno3, Naofumi Ishitobi2.   

Abstract

Rhegmatogenous retinal detachment (RRD) is a serious condition that can lead to blindness; however, it is highly treatable with timely and appropriate treatment. Thus, early diagnosis and treatment of RRD is crucial. In this study, we applied deep learning, a machine-learning technology, to detect RRD using ultra-wide-field fundus images and investigated its performance. In total, 411 images (329 for training and 82 for grading) from 407 RRD patients and 420 images (336 for training and 84 for grading) from 238 non-RRD patients were used in this study. The deep learning model demonstrated a high sensitivity of 97.6% [95% confidence interval (CI), 94.2-100%] and a high specificity of 96.5% (95% CI, 90.2-100%), and the area under the curve was 0.988 (95% CI, 0.981-0.995). This model can improve medical care in remote areas where eye clinics are not available by using ultra-wide-field fundus ophthalmoscopy for the accurate diagnosis of RRD. Early diagnosis of RRD can prevent blindness.

Entities:  

Mesh:

Year:  2017        PMID: 28842613      PMCID: PMC5573327          DOI: 10.1038/s41598-017-09891-x

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Rhegmatogenous retinal detachment (RRD) is a highly curable condition if properly treated early[1, 2]; however, if it is left untreated and develops proliferative changes, it becomes an uncontrollable condition called proliferative vitreoretinopathy (PVR). PVR is a serious condition that can result in blindness regardless of repeated treatments[3-5]. It is important, therefore, for patients to be seen and treated at a vitreoretinal centre at the early RRD stage to preserve visual function. However, establishing such vitreoretinal centres that provide advanced ophthalmological procedures is not practical because of rising social security costs, a problem that is troubling many nations around the world[6]. On the other hand, medical equipment has made remarkable advances, and one such advancement is the ultra–wide-field scanning laser ophthalmoscope (Optos 200Tx; Optos PLC, Dunfermline, United Kingdom). The Optos can provide nonmydriatic, noninvasive, wide-field fundus images easily (Fig. 1) and has been used to diagnose or follow-up various fundus conditions and treatment evaluation[7]. Because pupillary block and elevated intraocular pressure due to dilation can be avoided, examiners who are not qualified to perform ophthalmologic surgeries can safely capture images, which makes it ideal especially for telemedicine applications in areas where ophthalmologists are not available. In recent years, image processing technology using deep learning, a machine-learning technology, has attracted attention for its extremely high classification performance, and there have been a few studies regarding its applications to medical imaging[8-12]. To the best of our knowledge, there have been no studies on automatic diagnosis of retinal detachment nor studies on deep learning using Optos images.
Figure 1

Representative fundus images obtained by ultra–wide-field scanning laser ophthalmoscopy. Ultra–wide-field fundus images of right eye without rhegmatogenous retinal detachment (RRD) (a) and with RRD (b). The arrow indicates the retinal break, and the arrowheads indicate the areas of RRD.

Representative fundus images obtained by ultra–wide-field scanning laser ophthalmoscopy. Ultra–wide-field fundus images of right eye without rhegmatogenous retinal detachment (RRD) (a) and with RRD (b). The arrow indicates the retinal break, and the arrowheads indicate the areas of RRD. In this study, we assessed the ability of a deep learning technology to detect RRD using Optos images.

Results

A total of 420 non-RRD images from 238 patients (mean age, 69.8 ± 9.8 years; 103 men and 135 women) and 411 RRD images from 407 patients (mean age, 54.4 ± 15.2 years; 257 men and 150 women) were analysed. The deep learning model’s sensitivity was 97.6% [95% confidence interval (CI), 94.2–100%] and specificity was 96.5% (95% CI, 90.2–100%), and the area under the curve (AUC) was 0.988 (95% CI, 0.981–0.995). On the other hand, the support vector machine (SVM) model’s sensitivity was 97.5% (95% CI, 94.1–100%), specificity was 89.3% (95% CI, 82.6–96.0%) and AUC was 0.976 (95% CI, 0.957–0.996) (Fig. 2).
Figure 2

Representative receiver operating characteristic curves (AUC) of the deep learning model and support vector machine (SVM) model. The AUC of the deep learning model was 0.988 (95% CI, 0.981–0.995) and AUC of the SVM model was 0.976 (95% CI, 0.957–0.996). The AUC was better for the deep learning model than for the SVM model.

Representative receiver operating characteristic curves (AUC) of the deep learning model and support vector machine (SVM) model. The AUC of the deep learning model was 0.988 (95% CI, 0.981–0.995) and AUC of the SVM model was 0.976 (95% CI, 0.957–0.996). The AUC was better for the deep learning model than for the SVM model.

Discussion

Our results showed that the deep learning technology for detecting RRD on the Optos fundus photographs had high sensitivity and high specificity. The results indicate the great potential for RRD diagnosis in areas where ophthalmologists are not available. If a community has Optos, noninvasive wide-field fundus images can be captured easily without pupillary mydriasis and does not cause medical complications. A deep learning technology can provide accurate RRD classification at a high rate and patients classified as RRD by the deep learning technology are at first remotely reviewed by ophthalmologists. The patients diagnosed as having RRD in this way can be encouraged to see a specialist before the condition progresses to difficult-to-treat PVR, which shows the potential of this model for ophthalmic care. Generally, RRD does not immediately deteriorate to PVR; therefore, if air travel is available for patients, this model could be used to cover considerably wide areas. Because RRD is a highly curable condition if patients receive timely appropriate surgical treatment, the telemedicine system using Optos that we suggest here could significantly reduce the risk of blindness in RRD patients who live in areas where ophthalmic care is not available. In this study, we used a deep learning technology to classify the Optos images based on the presence of retinal detachment. Our analysis showed that the deep learning model using convolutional neural network (CNN) achieved a better AUC than that of the SVM model; however, slight difference was noted. Although the advantage of deep learning was slight, it was considered clinically important for achieving higher accuracy. It is suggested that multiple layers of non-linear processing in a deep learning model can learn features from a broader perspective with greater flexibility relative to what can be achieved by using SVM, which maps data to a feature space and determines identification boundary. Our results demonstrated classification performance that was almost equal to an ophthalmologist’s judging ability. This study had several limitations. When clarity of the eye is reduced because of severe cataract or dense vitreous haemorrhage, capturing images with Optos becomes challenging; thus, such cases were not included in this study. Additionally, this study only compared the images of normal eyes and eyes with RRD and did not include eyes with any other fundus diseases. We also need to point out that use of deep learning requires a large number of data sets. In the future, further studies using larger samples and including eyes with other fundus diseases are necessary to more broadly assess the performance and versatility of deep learning.

Methods

Data Set

Of the patients diagnosed with RRD by ophthalmologists at Tsukazaki Hospital since December 2011, those who had taken a fundus photograph with Optos were included. In addition, the patients without fundus diseases were extracted from the clinical database of the Ophthalmology Department of Tsukazaki Hospital. These images were reviewed by two ophthalmologists to confirm the presence of RRD and were registered in an analytical database. Eyes that contained vitreous haemorrhage to the extent that ophthalmologists could not determine the area of RRD with Optos image were excluded. Of 831 fundus images, 420 images were from non-RRD patients and 411 images were from RRD patients. According to ophthalmologists who use Optos images, the area of detachment of RRD eyes was less than one quadrant for 93 eyes (22.6%), one to two quadrants for 163 eyes (39.7%), two to three quadrants for 120 eyes (29.2%), and three or more quadrants for 35 eyes (8.5%). In addition, the number of retinal breaks was one for 292 eyes (71.0%), two for 46 eyes (11.2%), and three or more for 14 eyes (3.4%); however, breaks could not be confirmed in 59 eyes (14.4%). Of these, 665 images (336 non-RRD and 329 RRD images) were used for training, and 166 images (84 non-RRD and 82 RRD images) were used for validation of the trained model. The image data set for training was augmented to 11,970 images (6,048 non-RRD and 5,922 RRD images) by performing contrast adjustment for brightness, gamma correction, smoothing, noise imparting and reversal processing on the original images. This research adhered to the Declaration of Helsinki and was approved by the ethics committee of Tsukazaki Hospital. Since this was a study that retrospectively reviewed the Optos images and there were no anonymous issues involved, the Institutional Review Board of Tsukazaki Hospital waived the need for consent.

Deep Learning Model

We implemented a deep learning model that uses a CNN shown in Fig. 3 for use in our classification system. This type of multi-layer CNNs is known to automatically learn local features in images and generate a classification model[13-15]. The first layer is called Convolutional Layer 1 (Conv1) and obtains the feature quantities of the target through convolutional filters. Each of the convolutional layers (Conv1, 2 and 3) are followed by rectified linear unit[16] and max pooling layers (MP1, 2 and 3) to decrease position sensitivity and allow for more generic recognition. The last two layers (FC1, FC2) are fully connected, remove spatial information from extracted feature quantities and statistically recognise the target from other feature vectors. The last layer is a classification layer, which uses feature vectors of target images acquired in previous layers and the softmax function for binary classification. This network architecture is ideal for learning and recognition of local features of complex image data with individual differences.
Figure 3

Overall architecture of the model. The data set for the retinal fundus images (96 × 96 pixels) is labelled as Input. Each of the convolutional layers (Conv1–3) is followed by an activation function (ReLU) layer, pooling layers (MP1–3) and two fully connected layers (FC1, FC2). The final output layer performs binary classification by using a softmax function.

Overall architecture of the model. The data set for the retinal fundus images (96 × 96 pixels) is labelled as Input. Each of the convolutional layers (Conv1–3) is followed by an activation function (ReLU) layer, pooling layers (MP1–3) and two fully connected layers (FC1, FC2). The final output layer performs binary classification by using a softmax function.

Training the Deep Convolutional Neural Network

The aspect ratio of the Optos original image is 3900 × 3072 pixels, which is not square; but for analysis, we changed the aspect ratio of all input images and resized them to 96 × 96 pixels. For training, 100 images were mini-batch processed. The initial value of the network weight was randomly given as the Gaussian distribution with a zero mean and a standard deviation of 0.05. To avoid over-fitting during the training, a dropout technique was applied to two fully connected layers (FC1, FC2) to mask out with 70% probability and improve the performance[17]. The optimisation algorithm called AdaGrad (learning rate = 0.001), one of the stochastic gradient descent methods[18], was used to train the network weights. We trained the model by using the GPU of GeForce GTX970M by NVIDIA installed on a commercially available computer.

Support Vector Machine (SVM)

The SVM model was used as a reference to evaluate our deep learning recognition model[19]. For binary classification SVM, LIBSVM[20] from the Scikit–Learn library[21] with the Radial Basis Function (RBF) kernel was used. Other feature extraction or feature quantity selection was not performed so that we matched the criteria of the input data used in the deep learning training. The cost parameter (C = 10) and the RBF kernel parameter (γ = 0.001) were used, which were the results of the optimisation by grid search (i.e. C = 1, 10, 100, 1,000, 10,000; gamma = 0.000001, 0.00001, 0.0001, 0.001). Scoring for the grid search was performed on the basis of the mean AUC-receiver operating characteristic obtained by 5-fold cross validation of the grading data. The data set used in the training and validation were the same as the set used in the deep learning model.

Evaluation of Deep Learning and SVM Models

The deep learning and SVM models were operated one hundred times each. Subsequently, the average value and 95% CI for AUC were calculated.

Data availability

The Optos image datasets analysed during the current study are available with the corresponding author on reasonable request.
  12 in total

1.  Comparison of scleral buckling and vitrectomy for retinal detachment resulting from flap tears in superior quadrants.

Authors:  D Miki; T Hida; K Hotta; K Shinoda; A Hirakata
Journal:  Jpn J Ophthalmol       Date:  2001 Mar-Apr       Impact factor: 2.447

2.  Vitrectomy with silicone oil or sulfur hexafluoride gas in eyes with severe proliferative vitreoretinopathy: results of a randomized clinical trial. Silicone Study Report 1.

Authors: 
Journal:  Arch Ophthalmol       Date:  1992-06

3.  Support vector machines for classification and regression.

Authors:  Richard G Brereton; Gavin R Lloyd
Journal:  Analyst       Date:  2009-12-23       Impact factor: 4.616

4.  Scleral buckling versus primary vitrectomy in rhegmatogenous retinal detachment study (SPR Study): predictive factors for functional outcome. Study report no. 6.

Authors:  Nicole Heussen; Nicolas Feltgen; Peter Walter; Hans Hoerauf; Ralf-Dieter Hilgers; Heinrich Heimann
Journal:  Graefes Arch Clin Exp Ophthalmol       Date:  2011-02-09       Impact factor: 3.117

Review 5.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

Review 6.  ULTRA-WIDEFIELD FUNDUS IMAGING: A Review of Clinical Applications and Future Trends.

Authors:  Aaron Nagiel; Robert A Lalane; SriniVas R Sadda; Steven D Schwartz
Journal:  Retina       Date:  2016-04       Impact factor: 4.256

7.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

8.  Outcomes of surgery for retinal detachment associated with proliferative vitreoretinopathy using perfluoro-n-octane: a multicenter study.

Authors:  Ingrid U Scott; Harry W Flynn; Timothy G Murray; William J Feuer
Journal:  Am J Ophthalmol       Date:  2003-09       Impact factor: 5.258

9.  Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis.

Authors:  Geert Litjens; Clara I Sánchez; Nadya Timofeeva; Meyke Hermsen; Iris Nagtegaal; Iringo Kovacs; Christina Hulsbergen-van de Kaa; Peter Bult; Bram van Ginneken; Jeroen van der Laak
Journal:  Sci Rep       Date:  2016-05-23       Impact factor: 4.379

10.  Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia.

Authors:  Walter H L Pinaya; Ary Gadelha; Orla M Doyle; Cristiano Noto; André Zugman; Quirino Cordeiro; Andrea P Jackowski; Rodrigo A Bressan; João R Sato
Journal:  Sci Rep       Date:  2016-12-12       Impact factor: 4.379

View more
  26 in total

1.  Accuracy of ultrawide-field fundus ophthalmoscopy-assisted deep learning for detecting treatment-naïve proliferative diabetic retinopathy.

Authors:  Toshihiko Nagasawa; Hitoshi Tabuchi; Hiroki Masumoto; Hiroki Enno; Masanori Niki; Zaigen Ohara; Yuki Yoshizumi; Hideharu Ohsugi; Yoshinori Mitamura
Journal:  Int Ophthalmol       Date:  2019-02-23       Impact factor: 2.031

2.  Automatic screening of tear meniscus from lacrimal duct obstructions using anterior segment optical coherence tomography images by deep learning.

Authors:  Hitoshi Imamura; Hitoshi Tabuchi; Daisuke Nagasato; Hiroki Masumoto; Hiroaki Baba; Hiroki Furukawa; Sachiko Maruoka
Journal:  Graefes Arch Clin Exp Ophthalmol       Date:  2021-02-12       Impact factor: 3.117

3.  Developing an iOS application that uses machine learning for the automated diagnosis of blepharoptosis.

Authors:  Hitoshi Tabuchi; Daisuke Nagasato; Hiroki Masumoto; Mao Tanabe; Naofumi Ishitobi; Hiroki Ochi; Yoshie Shimizu; Yoshiaki Kiuchi
Journal:  Graefes Arch Clin Exp Ophthalmol       Date:  2021-11-04       Impact factor: 3.117

4.  Deep-Learning-Based Hemoglobin Concentration Prediction and Anemia Screening Using Ultra-Wide Field Fundus Images.

Authors:  Xinyu Zhao; Lihui Meng; Hao Su; Bin Lv; Chuanfeng Lv; Guotong Xie; Youxin Chen
Journal:  Front Cell Dev Biol       Date:  2022-05-19

5.  Real-time artificial intelligence evaluation of cataract surgery: A preliminary study on demonstration experiment.

Authors:  Hitoshi Tabuchi; Shoji Morita; Masayuki Miki; Hodaka Deguchi; Naotake Kamiura
Journal:  Taiwan J Ophthalmol       Date:  2022-04-13

Review 6.  Understanding required to consider AI applications to the field of ophthalmology.

Authors:  Hitoshi Tabuchi
Journal:  Taiwan J Ophthalmol       Date:  2022-04-13

7.  Current status of artificial intelligence for medicine.

Authors:  Hitoshi Tabuchi
Journal:  Taiwan J Ophthalmol       Date:  2022-06-01

8.  Automated detection of retinal exudates and drusen in ultra-widefield fundus images based on deep learning.

Authors:  Zhongwen Li; Chong Guo; Danyao Nie; Duoru Lin; Tingxin Cui; Yi Zhu; Chuan Chen; Lanqin Zhao; Xulin Zhang; Meimei Dongye; Dongni Wang; Fabao Xu; Chenjin Jin; Ping Zhang; Yu Han; Pisong Yan; Haotian Lin
Journal:  Eye (Lond)       Date:  2021-08-03       Impact factor: 4.456

9.  Development and validation of an offline deep learning algorithm to detect vitreoretinal abnormalities on ocular ultrasound.

Authors:  Venkatesh Krishna Adithya; Prabu Baskaran; S Aruna; Arthi Mohankumar; Jean Pierre Hubschman; Aakriti Garg Shukla; Rengaraj Venkatesh
Journal:  Indian J Ophthalmol       Date:  2022-04       Impact factor: 2.969

Review 10.  The Use of Fundus Autofluorescence in Dry Age-Related Macular Degeneration

Authors:  Nedime Şahinoğlu Keşkek; Figen Şermet
Journal:  Turk J Ophthalmol       Date:  2021-06-29
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.