| Literature DB >> 32369035 |
Jaehyeong Chun1, Youngjun Kim2, Kyoung Yoon Shin2, Sun Hyup Han2, Sei Yeul Oh2, Tae-Young Chung2, Kyung-Ah Park2, Dong Hui Lim2,3.
Abstract
BACKGROUND: Accurately predicting refractive error in children is crucial for detecting amblyopia, which can lead to permanent visual impairment, but is potentially curable if detected early. Various tools have been adopted to more easily screen a large number of patients for amblyopia risk.Entities:
Keywords: amblyopia; cycloplegic refraction; deep convolutional neural network; deep learning; mobile phone; photorefraction; refractive error; screening
Year: 2020 PMID: 32369035 PMCID: PMC7238094 DOI: 10.2196/16225
Source DB: PubMed Journal: JMIR Med Inform
Figure 1Overview of the proposed deep convolutional neural network architecture. The photorefraction image inputs pass through 17 convolutional layers and one fully connected layer, and the outputs of the network assign the probabilities for each refractive error class given the image. We also generate the localization map highlighting the important regions from the final convolutional feature maps of the layer i (i=1, 2, 3, or 4).
Figure 2Structure of the basic block and the shortcut connection. The basic block consists of two 3×3 convolutional layers, two Batch Normalization layers, and a Rectified Linear Unit (ReLU) activation function. The shortcut connection adds the input vector of the basic block to the output of the basic block.
Configuration of the deep convolutional network.
| Layer type, feature map | Filters | Kernel | Stride | Padding | Learning rate | |||
|
| ||||||||
|
| 224×224×3 | —a | — | — | — | 0.0 (freeze) | ||
|
| ||||||||
|
| 112×112×64 | 64 | 7×7×3 | 2 | 3 | 0.0 (freeze) | ||
|
| ||||||||
|
| 112×112×64 | — | — | — | — | 0.0 (freeze) | ||
|
| ||||||||
|
| 56×56×64 | 1 | 3×3 | 2 | 1 | 0.0 (freeze) | ||
|
| ||||||||
|
|
| |||||||
|
|
| 56×56×64 | 64 | 3×3×64 | 1 | 1 | 0.0 (freeze) | |
| 56×56×64 | 64 | 3×3×64 | 1 | 1 | 0.0 (freeze) | |||
|
|
| |||||||
|
| 56×56×64 | 64 | 3×3×64 | 1 | 1 | 0.0 (freeze) | ||
| 56×56×64 | 64 | 3×3×64 | 1 | 1 | 0.0 (freeze) | |||
|
| ||||||||
|
|
| |||||||
|
|
| 28×28×128 | 128 | 3×3×64 | 2 | 1 | 1e-10 | |
| 28×28×128 | 128 | 3×3×128 | 1 | 1 | 1e-10 | |||
| 28×28×128 | 128 | 1×1×64 | 2 | 0 | 1e-10 | |||
|
|
| |||||||
|
| 28×28×128 | 128 | 3×3×128 | 1 | 1 | 1e-10 | ||
| 28×28×128 | 128 | 3×3×128 | 1 | 1 | 1e-10 | |||
|
| ||||||||
|
|
| |||||||
|
|
| 14×14×256 | 256 | 3×3×128 | 2 | 1 | 1e-8 | |
| 14×14×256 | 256 | 3×3×256 | 1 | 1 | 1e-8 | |||
| 14×14×256 | 256 | 1×1×128 | 2 | 0 | 1e-8 | |||
|
| ||||||||
|
| 14×14×256 | 256 | 3×3×256 | 1 | 1 | 1e-8 | ||
| 14×14×256 | 256 | 3×3×256 | 1 | 1 | 1e-8 | |||
|
| ||||||||
|
|
| |||||||
|
|
| 7×7×512 | 512 | 3×3×256 | 2 | 1 | 1e-6 | |
| 7×7×512 | 512 | 3×3×512 | 1 | 1 | 1e-6 | |||
| 7×7×512 | 512 | 1×1×64 | 2 | 0 | 1e-6 | |||
|
|
| |||||||
|
| 7×7×512 | 512 | 3×3×512 | 1 | 1 | 1e-6 | ||
| 7×7×512 | 512 | 3×3×512 | 1 | 1 | 1e-6 | |||
|
| ||||||||
|
| 1×1×512 | 1 | 7×7 | 7 | 0 | — | ||
|
| ||||||||
|
| 1×7 | — | — | — | — | 1e-5 | ||
|
| ||||||||
|
| 1×7 | — | — | — | — | — | ||
aNot applicable.
Dataset participant demographics.
| Characteristic | Value | |
| Total images, n | 305 | |
|
|
| |
|
| ≤−5.0 Da | 25 |
|
| >−5.0 D and ≤−3.0 D | 18 |
|
| >−3.0 D and ≤−0.5 D | 50 |
|
| >−0.5 D and <+0.5 D | 84 |
|
| ≥+0.5 D and <+3.0 D | 87 |
|
| ≥+3.0 D and <+5.0 D | 29 |
|
| ≥+5.0 D | 12 |
| Girls, n (%) | 191 (62.6) | |
| Age, mean (SD) | 4.32 (1.87) | |
aD: diopters.
Figure 3Examples of photorefraction images from the seven different refractor error classes. A bright crescent appears in the pupillary reflex, and its size and shape indicate the diopter (D) value.
Results for five-fold cross-validation.
| Iterationa | Validation accuracy (%) (N=31) | Test accuracy (%) (N=61) |
| First iteration | 87.1 | 83.6 |
| Second iteration | 80.6 | 80.3 |
| Third iteration | 80.6 | 82.0 |
| Fourth iteration | 83.9 | 78.7 |
| Fifth iteration | 83.9 | 83.6 |
| Average | 83.2 | 81.6 |
aIn each iteration, our network was trained using the rest of the validation and test dataset (213 training images).
Performance of our deep convolutional neural network with the overall test dataset.
| Class | Number | Accuracy (%) |
| ≤−5.0 Da | 25 | 80.0 |
| >−5.0 D and ≤−3.0 D | 18 | 77.8 |
| >−3.0 D and ≤−0.5 D | 50 | 82.0 |
| >−0.5 D and <+0.5 D | 84 | 83.3 |
| ≥+0.5 D and <+3.0 D | 87 | 82.8 |
| ≥+3.0 D and <+5.0 D | 29 | 79.3 |
| ≥+5.0 D | 12 | 75.0 |
| Total | 305 | 81.6 |
aD: diopter.
Confusion matrix for refractive error classification of our deep convolutional neural network.
| Ground truth | Predictive value | Accuracy (%) | |||||||
| ≤−5.0 Da | >−5.0 D and ≤−3.0 D | >−3.0 D and ≤−0.5 D | >−0.5 D and <+0.5 D | ≥+0.5 D and <+3.0 D | ≥+3.0 D and <+5.0 D | ≥+5.0 D | |||
| ≤−5.0 D | 20b | 3 | 2 | 0 | 0 | 0 | 0 | 80.0 | |
| >−5.0 D and ≤−3.0 D | 1 | 14b | 2 | 0 | 1 | 0 | 0 | 77.8 | |
| >−3.0 D and ≤−0.5 D | 1 | 4 | 41b | 4 | 0 | 0 | 0 | 82.0 | |
| >−0.5 D and <+0.5 D | 0 | 0 | 5 | 70b | 8 | 1 | 0 | 83.3 | |
| ≥+0.5 D and <+3.0 D | 0 | 0 | 1 | 10 | 72b | 4 | 0 | 82.8 | |
| ≥+3.0 D and <+5.0 D | 0 | 0 | 0 | 1 | 4 | 23b | 1 | 79.3 | |
| ≥+5.0 D | 0 | 0 | 0 | 0 | 1 | 2 | 9b | 75.0 | |
| Overall accuracy (%) | —c | — | — | — | — | — | — | 81.6 | |
aD: diopter.
bNumber of correct predictions of our deep convolutional neural network.
cNot applicable.
Performance comparison of the proposed model and baseline models.
| Model | Accuracy (%) |
| The proposed deep convolutional neural network | 81.6 |
| Pretrained VGG-11 | 70.8 |
| Pretrained SqueezeNet | 77.4 |
| Support Vector Machine | 65.2 |
| Random Forest | 62.9 |
| Simple convolutional neural network | 70.8 |
Figure 4Examples of photorefraction images correctly classified by deep neural networks. (A), (B), (C) were identified as ≥+0.5 D and <+3.0 D, ≥+3.0 D and <+5.0 D, and ≥+5.0 D, respectively. The first layers captured low-level features, such as edge and color. With deeper layers, the network focused on high-level features that were regarded as important aspects for classification.