Literature DB >> 30333258

Application of a deep convolutional neural network in the diagnosis of neonatal ocular fundus hemorrhage.

Binbin Wang1, Li Xiao2, Yang Liu2, Jing Wang3, Beihong Liu4, Tengyan Li4, Xu Ma4, Yi Zhao2.   

Abstract

There is a disparity between the increasing application of digital retinal imaging to neonatal ocular screening and slowly growing number of pediatric ophthalmologists. Assistant tools that can automatically detect ocular disorders may be needed. In present study, we develop a deep convolutional neural network (DCNN) for automated classification and grading of retinal hemorrhage. We used 48,996 digital fundus images from 3770 newborns with retinal hemorrhage of different severity (grade 1, 2 and 3) and normal controls from a large cross-sectional investigation in China. The DCNN was trained for automated grading of retinal hemorrhage (multiclass classification problem: hemorrhage-free and grades 1, 2 and 3) and then validated for its performance level. The DCNN yielded an accuracy of 97.85 to 99.96%, and the area under the receiver operating characteristic curve was 0.989-1.000 in the binary classification of neonatal retinal hemorrhage (i.e., one classification vs. the others). The overall accuracy with regard to the multiclass classification problem was 97.44%. This is the first study to show that a DCNN can detect and grade neonatal retinal hemorrhage at high performance levels. Artificial intelligence will play more positive roles in ocular healthcare of newborns and children.
© 2018 The Author(s).

Entities:  

Keywords:  artificial intelligence; convolutional neural network; fundus images; retinal hemorrhage

Mesh:

Year:  2018        PMID: 30333258      PMCID: PMC6435455          DOI: 10.1042/BSR20180497

Source DB:  PubMed          Journal:  Biosci Rep        ISSN: 0144-8463            Impact factor:   3.840


Introduction

Retinal hemorrhage (RH) is one of the most common ocular abnormalities in newborns and has a wide prevalence ranging from 2.6 to 50.0% [1-3]. The delivery mode, examination technique and time of examination after birth are considered as important factors that influence the rate of RH in newborns [4]. Vacuum extraction delivery was reportedly associated with the highest rate of RH, followed by spontaneous vaginal delivery and cesarean delivery [5,6]. Direct or indirect ophthalmoscopy is usually used for primary ocular screening of newborns. However, both techniques have some limitations. When performing direct ophthalmoscopy, some ocular abnormalities or RH in the peripheral regions may be missed due to the limited inspection range. Indirect ophthalmoscopy requires experienced and skilled experts for both performance and diagnosis, which may limit the widespread use and screening accuracy of this technique [7]. It was reported that birth-related hemorrhages may resolve quickly and can be of ‘small size and number’ [8], and those hemorrhages seen at 1 month of age were less likely birth-related [2]. But the other difference between NH and other hemorrhages found in other ocular or retinal disease was rarely reported. In recent years, a wide-field digital imaging system (RetCam; Clarity Medical Systems, Pleasanton, CA, U.S.A.) has been applied with increasing frequency for regular ocular examination of newborns. This system can provide and record high-resolution, objective and wide-angle images of ocular diseases [4,9,10]. The system can provide a 130-degree field of view so that lesions in the peripheral regions, including RH, can be detected easily and quickly. Although the RetCam can provide quick and precise information regarding RH, there is a growing disparity between the increasing need for neonatal ocular examinations and the limited number of pediatric ophthalmologists. This problem is more severe in China, where there are more than 16 million newborns born each year. Therefore, an automated detection system for ocular abnormalities, including RH of different grades, based on fundus images of newborns is an urgent need. Recent developments in deep learning techniques have provided evidence that artificial intelligence can help to diagnose ocular diseases, with an accuracy near or exceeding that attained by experienced experts, with much greater speed [11-15]. Gulshan et al. [11] developed a deep convolutional neural network (DCNN) to classify referable diabetic retinopathy using retinal fundus photographs with a sensitivity ranging from 87.0 to 97.5% and a specificity ranging from 93.4 to 98.5%. Similarly, Burlina et al. [12] applied a DCNN in the automated detection of age-related macular degeneration using color fundus images. Their DCNN yielded an accuracy of 88.4 to 91.6% with an area under the receiver operating characteristic curve (AUC) of 0.94 to 0.96, which are close to human performance levels. The goal of the present study is to develop a DCNN model that can automatically determine the presence and severity of RH in newborns based on fundus images.

Methods

Data source

We used color fundus images of neonates from the National Science and Technology Basic Work program, a multi-center cross-sectional investigation involving the performance of fundus examinations by a wide-field digital imaging system (RetCam) on consecutive series of newborns with written informed consent. The digital images of the posterior pole and the superior, inferior, nasal and temporal retinal fields of each participant were captured and labeled by professional ophthalmologists. Neonatal RH was diagnosed and graded according to Egge’s classification [16]: (1) Grade 1: small retinal hemorrhage confined to the area around the optic nerve, associated with dot or fine linear bleeding; (2) Grade 2: slightly larger amount of retinal hemorrhage than Grade 1; patchy, dot, blot or flame-shaped hemorrhage, size does not exceed the optic disc diameter; (3) Grade 3: retinal hemorrhage more than the diameter of the optic disc area, a line of flame-shaped hemorrhage along vessels, macular hemorrhage The images of graded RH and normal fundi were used for training and testing of the DCNN. This research was approved by the ethics committee of National Research Institute for Family Planning.

Model

The architecture of the DCNN model is shown in Figure 1. The network comprised the following types of hidden layers:
Figure 1

Framework of the deep convolution neural network

Convolutional layer (Conv): Performance of convolutional operation. Pooling layer (Max Pool): Performance of pooling operation. Maxout layer (Maxout): Use of a maxout activation function for non-linear changes, down-sampling. Dropout layer (Dropout): Prevention of over-fitting. Concatenate layer (Concat): Concatenation of input data. Dense layer (Dense): All nodes in the fully connected layer are connected to all nodes in the previous layer and are used to combine the features extracted from the front layers. Softmax layer (Softmax): Use of the softmax activation function to predict the results. The details of the layers are shown in Table 1. All images were resized to 512 × 512, with three channels representing RGB. The image first went through the convolutional layer and max-pooling layer, and 7 × 7 feature maps of 256 channels were obtained. This part contained 16 convolutional layers and 5 max pool layers, similar to VGG19. The feature maps then went through the maxout layer, dropout layer and dense layer to improve the performance [17,18]. Finally, we applied the softmax function to predict the probability of each type (0, normal; 1, mild; 2, moderate; 3, severe) and used the type with the highest probability as the predicted result. We used Theano and Lasagna [19] to perform the DCNN model.
Table 1

Details of each layer in the convolution neural network

KernelStridesNumber of filtersFeature map (height,width,channels)Function
Conv7*72*232(256, 256, 32)Extract features
Conv5*52*232(128, 128, 32)
Max-pooling3*32*2(63, 63, 32)Fuse feature and reduce feature map
Conv3*31*132(63, 63, 32)
Conv3*31*132(63, 63, 32)
Max-pooling3*32*2(31, 31, 32)
Conv3*31*164(31, 31, 64)
Conv3*31*164(31, 31, 64)
Conv3*31*164(31, 31, 64)
Conv3*31*164(31, 31, 64)
Max-pooling3*32*2(15, 15, 64)
Conv3*31*1128(15, 15, 128)
Conv3*31*1128(15, 15, 128)
Conv3*31*1128(15, 15, 128)
Conv3*31*1128(15, 15, 128)
Max-pooling3*32*2(7, 7, 128)
Conv3*31*1256(7, 7, 256)
Conv3*31*1256(7, 7, 256)
Conv3*31*1256(7, 7, 256)
Conv3*31*1256(7, 7, 256)
Max-pooling3*32*2(3, 3, 256)
Dropout(3, 3, 256)Prevent over fitting
Dense1024Purify features for classification
Maxout
Dropout
Dense1024
Maxout
Dense16
Dense8
Softmaxclassifier

Data preparation

Our study included 48,996 fundus images collected from 3770 newborns (5–20 imagines for each one). All samples were carefully classified by experts as 765 normal fundi, 1023 grade 1 RH, 975 grade 2 RH and 1007 grade 3 RH. Each sample donated 5–20 eligible images after assessment in the present study. We divided each classification of newborns into a training set and testing set in a 7:3 ratio (Table 2).
Table 2

Summary of the training and testing dataset

DatasetNumber of imagesNumber of patients
Training
  Normal6590708
  Grade 1 RH7816519
  Grade 2 RH11,470685
  Grade 3 RH8210727
Testing
  Normal2762299
  Grade 1 RH3678246
  Grade 2 RH5130290
  Grade 3 RH3340296

Abbreviation: RH, retinal hemorrhage.

Abbreviation: RH, retinal hemorrhage. Because the external illumination conditions can be quite different when performing fundus examinations, all images went through a normalization step at the preprocessing stage to remove the illumination effect. For each pixel in the image, assuming the values on the three channels are r, g and b, the normalization changed the values on the three channels as follows: Figure 2 shows the effect of normalizing the fundus images under different levels of illumination. After normalization (Figure 2B,D), the illumination effect was greatly reduced.
Figure 2

Normalization effect of digital fundus images

(A and C) Fundus images taken under bright and dark illumination conditions. (B and D) Images after normalization.

Normalization effect of digital fundus images

(A and C) Fundus images taken under bright and dark illumination conditions. (B and D) Images after normalization.

Training

We adopted stochastic gradient descent and the Nesterov momentum method to train the model. The loss function was a combination of a continuous kappa loss together with the cross-entropy (or log) loss [20] as defined in (Copyright © 2015 Jeffrey De Fauw). We gradually decreased the learning rate to balance the learning speed and accuracy. The learning rate of each step is listed in Table 3 as the training procedure progressed. After 40,000 iterations of training, the loss and accuracy became stable.
Table 3

Change of learning rate as the training procedures went

EpochsLearning rate
0–190.007
20–390.0035
40–590.0021
60–790.00105
80–990.0007
100–1190.00035
120–1390.00007
140–1590.000035
160–1790.000007
180–1890.0000007
190–1990.00000007

Performance merit

We used the accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) to assess the performance of the DCNN. Additionally, we applied a receiver operating characteristic (ROC) curve and computed the AUC. Notably, the detection of RH was a multiclass classification problem (normal, grade 1 RH, grade 2 RH and grade 3 RH). We transformed it into four binary classification problems for assessment and comparison of the performance levels: RH (yes vs. no), grade 1 RH (yes vs. no), grade 2 RH (yes vs. no) and grade 3 RH (yes vs. no). All statistical analyses were performed using SPSS version 19 (IBM Corp., Armonk, NY, U.S.A.).

Results

We used 14,910 fundus images of 1131 newborns from the validation dataset to assess the performance of the trained DCNN. The performance levels of each classification are listed in Table 4, and the ROC curve of each classification is shown in Figure 3.
Table 4

Performance levels for machines experiments using fundus images dataset

ClassificationSensitivitySpecificityPPVNPVAccuracyAUC (95%CI)
RH99.5799.2099.8298.1499.501 (0)
Grade 1 RH97.8597.8593.6299.3097.850.995 (0.001)
Grade 2 RH99.9299.9899.9699.9699.961 (0)
Grade 3 RH93.1199.2697.3198.0497.880.989 (0.001)

Abbreviations: 95%CI, 95% confidence interval; AUC, area under ROC curve; NPV, negative predictive value; PPV, positive predictive value; RH, retinal hemorrhage. All values indicate percentages except for AUC.

Figure 3

Receiver operating characteristic curve for performance of the neural network in the binary classification problem of retinal hemorrhage.

(A) hemorrhage versus no hemorrhage, (B) grade 1 hemorrhage versus the others, (C) grade 2 hemorrhage versus the others and (D) grade 3 versus the others

Receiver operating characteristic curve for performance of the neural network in the binary classification problem of retinal hemorrhage.

(A) hemorrhage versus no hemorrhage, (B) grade 1 hemorrhage versus the others, (C) grade 2 hemorrhage versus the others and (D) grade 3 versus the others Abbreviations: 95%CI, 95% confidence interval; AUC, area under ROC curve; NPV, negative predictive value; PPV, positive predictive value; RH, retinal hemorrhage. All values indicate percentages except for AUC. The trained DCNN had high performance for prediction of RH regardless of severity, with a sensitivity of 99.57%, specificity of 99.20%, PPV predictive value of 99.82%, NPV of 98.14% and accuracy of 99.50%. The AUC was 1, indicating excellent performance of the DCNN for the diagnosis of RH. The DCNN showed varying performance for classifying the severity of RH. The accuracy for grade 1, 2 and 3 RH was 97.85%, 99.96% and 97.88%, respectively. The AUC for each classification was 0.995, 1.000 and 0.989, respectively, implying that our DCNN has the best performance for classification of grade 2 RH. Overall, our model showed high performance levels in the detection and grading of RH using fundus images. We also assessed the performance level of the DCNN on the multiclass classification problem (normal and grade 1, 2 or 3 RH). When the image classification obtained by the DCNN was in accordance with the original label, it was regarded as a correct classification. The overall accuracy was then calculated as the number of correct classifications divided by the total number of images. Finally, we obtained 14,528 correctly classified images, yielding an overall accuracy of 97.44%.

Discussion

With the increasing understanding of the importance of neonatal ocular healthcare among parents, pediatric experts and ophthalmologists, there is a growing demand for early ocular examination worldwide, especially in developing countries. However, the numbers of ophthalmologists and individuals involved in ocular healthcare of newborns and children are still limited and are not increasing quickly enough to meet the rising demand. Thus, the development of automated and assistive systems for ocular screening is expected. The present study has demonstrated that artificial intelligence based on application of a DCNN can successfully classify and grade RH with high accuracy of 97.85 to 99.96% and an overall accuracy of 97.44%. Compared with ophthalmoscopy or the red reflex test, wide-field digital imaging is the neonatal ocular screening method most widely recommended by the American Association of Pediatrics. This technique has some advantages in the comprehensive screening and documentation of neonatal ocular disorders. The system can be operated by a trained doctor or nurse with minimal clinical experience requirements. It also causes less distress in infants, increasing parents’ and ophthalmologists’ acceptance of this method for detection of RH and other disorders. The system can detect some fundus disorders other than RH, including retinopathy of prematurity [21-23], ocular birth defects [9], persistent hyperplastic primary vitreous [24], retinoblastoma [25] and others. Moreover, the digital system can generate high-resolution images of a wide range of the fundus that provide instant documentation of neonatal ocular disorders. The present study is the first to demonstrate that fundus images can also be used for automated detection of RH by combining these images with a DCNN. RH is the most frequently encountered abnormality during neonatal universal ocular screening by the RetCam shuttle, accounting for more than 85% of all findings [9]. Most birth-related RH will spontaneously and quickly resolve because of its small size and number [8]. However, RH of great size, long duration or special location should be given closer attention because it can be a clinical sign of cerebral injury [4] or may affect the future development of vision. In the present study, we developed a DCNN that can successfully classify the severity of RH, which will support the long-term management of fundus hemorrhage. In conclusion, this was a pilot exploration of the combination of artificial intelligence with digital fundus images for the detection of neonatal ocular abnormalities. We applied a DCNN in the classification and grading of RH, the most frequently encountered fundus abnormality in newborns, at high performance levels. In our next phase of development, we will further explore more DCNN models related to other ocular diseases or develop a comprehensive model for all the ocular diseases.
  21 in total

1.  Screening for retinopathy of prematurity in China: a neonatal units-based prospective study.

Authors:  Yu Xu; Xiaohong Zhou; Qi Zhang; Xunda Ji; Qin Zhang; Jianxing Zhu; Chao Chen; Peiquan Zhao
Journal:  Invest Ophthalmol Vis Sci       Date:  2013-12-19       Impact factor: 4.799

2.  Incidence and rate of disappearance of retinal hemorrhage in newborns.

Authors:  M V Emerson; D J Pieramici; K M Stoessel; J P Berreen; R F Gariano
Journal:  Ophthalmology       Date:  2001-01       Impact factor: 12.079

3.  Outcomes of 957 preterm neonatal fundus examinations in a Guangzhou NICU through 2008 to 2011.

Authors:  Rui-Hong Ju; Xiao-Yun Ke; Jia-Qing Zhang; Min Fu
Journal:  Int J Ophthalmol       Date:  2012-08-18       Impact factor: 1.779

4.  Incidence, distribution, and duration of birth-related retinal hemorrhages: a prospective study.

Authors:  Lindsey A Hughes; Kristina May; John F Talbot; Michael Andrew Parsons
Journal:  J AAPOS       Date:  2006-04       Impact factor: 1.220

5.  Birth-related retinal hemorrhages in healthy full-term newborns and their relationship to maternal, obstetric, and neonatal risk factors.

Authors:  Qi Zhao; Yanli Zhang; Yu Yang; Zijing Li; Yu Lin; Ran Liu; Chunyi Wei; Xiaoyan Ding
Journal:  Graefes Arch Clin Exp Ophthalmol       Date:  2015-05-17       Impact factor: 3.117

6.  Screening for retinopathy of prematurity employing the retcam 120: sensitivity and specificity.

Authors:  D B Roth; D Morales; W J Feuer; D Hess; R A Johnson; J T Flynn
Journal:  Arch Ophthalmol       Date:  2001-02

7.  Inflicted childhood neurotrauma (shaken baby syndrome): ophthalmic findings.

Authors:  Brian J Forbes; Cindy W Christian; Alexander R Judkins; Kasia Kryston
Journal:  J Pediatr Ophthalmol Strabismus       Date:  2004 Mar-Apr       Impact factor: 1.402

Review 8.  Newborn retinal hemorrhages: a systematic review.

Authors:  Patrick Watts; Sabine Maguire; Thomas Kwok; Bnar Talabani; Mala Mann; Jarmila Wiener; Zoe Lawson; Alison Kemp
Journal:  J AAPOS       Date:  2013-01-28       Impact factor: 1.220

Review 9.  Fundus hemorrhages in infancy.

Authors:  B Kaur; D Taylor
Journal:  Surv Ophthalmol       Date:  1992 Jul-Aug       Impact factor: 6.048

10.  Targeting the HA2 subunit of influenza A virus hemagglutinin via CD40L provides universal protection against diverse subtypes.

Authors:  X Fan; A M Hashem; Z Chen; C Li; T Doyle; Y Zhang; Y Yi; A Farnsworth; K Xu; Z Li; R He; X Li; J Wang
Journal:  Mucosal Immunol       Date:  2014-07-23       Impact factor: 7.313

View more
  2 in total

Review 1.  Accuracy of Deep Learning Algorithms for the Diagnosis of Retinopathy of Prematurity by Fundus Images: A Systematic Review and Meta-Analysis.

Authors:  Jingjing Zhang; Yangyang Liu; Toshiharu Mitsuhashi; Toshihiko Matsuo
Journal:  J Ophthalmol       Date:  2021-08-06       Impact factor: 1.909

2.  Analysis of non-retinopathy of prematurity (ROP)-related fundus hemorrhage in preterm infants in China.

Authors:  Sun Yaoyao; Deng Kaixin; Li Xiaoxin; Zhao Min; Jiang Yanrong; Yin Hong; Qi Huijun; Qian Tong; Linghu Dandan; Yu Wenzhen; Zhao Mingwei; Liang Jianhong
Journal:  Front Pediatr       Date:  2022-09-26       Impact factor: 3.569

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.