| Literature DB >> 31695786 |
Jie Xu1, Kanmin Xue2, Kang Zhang3.
Abstract
With the recent developments in deep learning technologies, artificial intelligence (AI) has gradually been transformed from cutting-edge technology into practical applications. AI plays an important role in disease diagnosis and treatment, health management, drug research and development, and precision medicine. Interdisciplinary collaborations will be crucial to develop new AI algorithms for medical applications. In this paper, we review the basic workflow for building an AI model, identify publicly available databases of ocular fundus images, and summarize over 60 papers contributing to the field of AI development. © The author(s).Entities:
Keywords: artificial intelligence; deep learning; machine learning; ophthalmology
Year: 2019 PMID: 31695786 PMCID: PMC6831476 DOI: 10.7150/thno.38065
Source DB: PubMed Journal: Theranostics ISSN: 1838-7640 Impact factor: 11.556
Figure 1A typical deep learning neural network with multiple deep layers between input and output layers
Figure 2Workflow diagram of developing a deep learning-based medical diagnostic algorithm.
Summary of publicly available databases of ocular retinal images
| Database | Number of images | Camera Model | Image Resolution (pixels) | Field of View | Application |
|---|---|---|---|---|---|
| DRIVE | 40 | Canon CR5 | 768×584 | 45° | Blood vessel segmentation |
| STARE | 400 | Topcon trv-50 | 605×700 | 35° | Blood vessel segmentation; Optic disk detection |
| Image-Ret | |||||
| DIARETDB0 | 130 | unknown | 1500×1152 | 50° | Diabetic retinopathy detection |
| DIARETDB1 | 89 | unknown | 1500×1152 | 50° | Diabetic retinopathy detection |
| e-ophtha | |||||
| e-ophtha_EX | 82 | OPHDIAT Tele-medical network | 2048×1360; 1440×960 | - | Diabetic retinopathy detection |
| e-ophtha_MA | 381 | 1440×960; 2544×1696 | Diabetic retinopathy detection | ||
| HEI-MED | 169 | Zeiss Visucam PRO fundus camera | 2196×1958 | 45° | Hard exudate detection; Diabetic macular edema assessment |
| Retinopathy Online Challenge | 100 | Canon CR5-45-NM | 768×576; 1058×061; 1389×1383 | 45° | Microaneurysms detection |
| Messidor | 1200 | TopCon TRC NW6 | 1440×960; 2240×1488; 2304×1536 | 45° | Diabetic retinopathy detection |
| RIM-ONE | 169 | Nidek AFC-210 with a body of a Canon EOS 5D | 2144 × 1424 | - | Optic nerve head segmentation |
| DRIONS-DB | 110 | Color analogue fundus camera digitized by HP-PhotoSmart-S20 scanner | 600×400 | - | Optic nerve head segmentation |
Summary of influential papers on ophthalmic image analysis
| Disease | Procedures / examinations | Data sets | Deep learning techniques | Performance | Reference | |
|---|---|---|---|---|---|---|
| Keratoconus | Pentacam | 194 normal, 454 keratoconus, 67 forme fruste, 28 astigmatic, 117 after refractive surgery | SVM | Acc: 98.9%; Sen: 99.1%; Spe: 98.5%; AUC:0.99 | Hidalgo et al. | |
| Dual Scheimpflug Analyzer | 177 normal, 148 keratoconus | Decision Tree | Sen: 100%; Spe: 99.5% | Smadja et al. | ||
| Pentacam HR | 30 normal, 15 unilateral keratoconus, 30 bilateral keratoconus | FNN | Bilateral keratoconus versus normal; AUC: 0.99; Sen: 100%; Spe: 95% | Kovacs et al. | ||
| Pterygium | Anterior segment photographed images | 2,692 non-pterygium, 325 pterygium | Shape features + SVM/ANN | Acc: 91.27%; AUC: 0.956; Sen: 88.7%; Spe: 88.3% | Zaki et al. | |
| Cataract | Slit-lamp image | 5,378 images with decimal grading scores ranging from 0.1 to 5.0. | CRNN | 70.7% exact integral agreement ratio (R0); 88.4% decimal grading error ≤ 0.5 (Re0.5); 99.0% decimal grading error ≤ 1.0 (Re1.0 ). | Gao et al. | |
| Slit-lamp image | 476 normal, 410 cataract | DCNN | Cataract vs Normal: | Long et al. | ||
| Fundus image | 767 normal, 472 cataract (246 mild cataract,128 moderate cataract, and 98 severe cataract) | SVM, BPNN | Acc: 93.2% for detection,84.5% for grading; Sen:94.2% for detection,74.6-89.3% for grading; | Yang et al. | ||
| Slit-lamp image | 476 normal, 410 pediatric cataract | CNN, SVM | Acc, Sen, and Spe: classification (97.07%, 97.28%, and 96.83%,) three-degree grading area (89.02%, 86.63%, and 90.75%) density (92.68%, 91.05%, and 93.94%) location (89.28%, 82.70%, and 93.08%) | Liu et al. | ||
| POAG | Fundus image | Training set:125,189; Validation set: 71,896 | DLS | AUC: 0.942; Sen: 96.4%; Spe: 87.2% | Ting et al. | |
| Fundus image | Training set: 31,745; Validation set: 8,000 | DCNN | AUC: 0.98; Acc: 92.9%; Sen: 95.6%; Spe: 92.0%; AUC: 0.986 | Li et al. | ||
| Fundus image | 589 normal, 837 glaucoma | CNN | Acc: 98.13%; Sen: 98%; Spe: 98.3% | Raghavendra et al. | ||
| Fundus image | 30 normal, 30 open-angle glaucoma | SVW | Acc:91.67; Sen:90%;Spe:93.33% | Krishnan et al. | ||
| Visual field | Training set:257; Test set: 129 | ANN | AUC: 0.890; Sen: 78.3%; Spe: 89.5% | Oh et al. | ||
| Fundus image | 266 normal, 72 mild, 86 moderate, 86 severe glaucoma | SVM | Acc: 93.1%; Sen: 89.75%; Spe: 96.2% | Acharya et al. | ||
| Fundus image and SLO image | Normal/glaucoma Fundus images:85/39; Normal/glaucoma SLO images: 46/19 | RIFM | Acc for Fundus images: 94.4%,SLO images: 93.9%;Sen for Fundus images: 92.3%,SLO images: 89.5%;Spe for Fundus images: 95.3%,SLO images: 93.5% | Haleem et al. | ||
| Visual fields | 53 glaucoma eyes, 108 normal eyes | FNN | AUC: 92.6%, The sensitivity was 77.8%,54.6%, and 50.0%, respectively, at the specificity of 90%, 95%,and 99% | Asaoka et al. | ||
| DR | Fundus image | Training set:76,370; Validation set: 112,648 | DLS | For referable DR: AUC: 0.936; Sen: 90.5; Spe: 91.6%; For vision-threatening DR: AUC: 0.958; Sen: 100%; Spe: 91.1% | Ting et al. | |
| Fundus image | Development Data Set (EyePACS in the United States and 3 eye hospitals in India): 128,175 Validation Data Set (EyePACS-1: 9,963; Messidor-2: 1,748) | DCNN | AUC: 0.991 for EyePACS,0.990 for Messidor;Sen: 90.3% for EyePACS,87% for Messidor;Spe: 98.1% for EyePACS,98.5% for Messidor | Gulshan et al. | ||
| Fundus image | 170 DR, 170 normal | PNN-GA, SVM quadratic kernels | PNN-GA: Acc:99.41%,Sen:99.41% | Ganesan et al. | ||
| Fundus image | EyePACS: 75,137 DR images; External validation: MESSIDOR 2 and E-Ophtha | DCNN | AUC 0.94 for Messidor 2, 0.95 for E-Ophtha;Sen 93% for Messidor 2,87% for E-Ophtha;Spe 90% for Messidor 2,94% for E-Ophtha | Gargeya et al. | ||
| Fundus image | Training set: 327 diabetic patients; Validation set: 725 diabetic patients | LASSO | Acc: 89.2%; AUC: 0.90; Sen: 75%; Spe: 89.6% | Oh et al. | ||
| Fundus image | Training set: 400; Testing set: 9,954 | Ensemble of classifiers with hidden Markov chain for context information, trained by genetic algorithm | Sen: 92.2%; Spe: 90.4% | Tang et al. | ||
| Fundus image | Messidor-2 dataset: 1,748 | CNN | Referable DR: AUC: 0.980; Sen: 96.8%; Spe: 87%; Vision threatening DR: AUC: 0.989; Sen: 100%; Spe: 90.8% | Abramoff et al. | ||
| Fundus image | 4,445 DR; 5,494 normal | DCNN | Acc: 0.81 | Takahashi et al. | ||
| Fundus image | DIARETDB1, FAZ, MESSIDOR, Private dataset: 750 (Normal: 150, mild NPDR: 150, moderate NPDR: 150, severe NPDR: 150, PDR: 150) | DNN | AUC: 0.924; Sen: 92.18%; Spe: 94.50% | Abbas et al. | ||
| DME | SD-OCT | Training set: 11,349; DEM; 51,140 normal; Validation set: 250 DME, 250 normal | CNN | Acc: 98.2%; Sen: 96.8%; Spe: 99.6% | Kermany et al. | |
| DME | Fundus image | 283 DR; 1,086 normal | Ensemble of Gaussian mixture model and SVM with RBF kernel | Acc: 96.8%; Sen: 97.3%; Spe: 95.9% | Akram et al. | |
| AMD | Fundus image | Training set 72,610; Validation set: 35,948 | DLS | AUC: 0.931; Sen: 93.2%; Spe: 88.7% | Ting et al. | |
| Fundus image | AREDS dataset: >130,000 | DCNN | AUC: 0.94∼0.96 Acc: 88.4%∼91.6% Sen: 71%∼88.4% Spe: 91.4%∼94.1% | Burlina et al. | ||
| Fundus image | AREDS dataset: 5,664 | DCNN | Acc 79.4% (4-class) 81.5% (3-class); 93.4% (2-class) | Burlina et al. | ||
| SD-OCT | Training and validation sets: 1,012 (AMD: 701; normal: 311); Test:100 (AMD: 50, normal: 50) | DCNN | Acc: 96%; Sen: 100%; Spe: 92% | Treder et al. | ||
| Fundus image | 135 AMD subjects, 135 normal subjects | Feature extracted by Discrete wavelet transform and others for feature selection and classification | Average Acc: 93.7%; Sen: 91.11%; Spe: 96.3% | Mookiah et al. | ||
| OCT | 48,312 AMD; 52,690 normal | DCNN | AUC: 0.975; Sen: 92.6%; Spe: 93.7% | Lee et al. | ||
| SD-OCT | 1,289 | CNN | The mean Dice coefficient for human interrater reliability and deep learning were 0.750 and 0.729, respectively. | Lee et al. | ||
| CNV | SD-OCT | Training set: 37,206 CNV, 51,140 normal; Validation set: 250 CNV, 250 normal | CNN | Acc: 100%; Sen:100%; Spe:100% | Kermany et al. | |
| CNV | Fluorescein angiography | 33 | AdaBoost | Accuracy: 83.26% | Tsai et al. |
RBFNN, radial basis function neural network; SVM, support vector machine; MLP, multi-layer perceptron; CRNN, convolutional-recursive neural networks; DCNN, Deep-learning convolutional neural network; BPNN, Back propagation neural network; DLS, deep learning system; CNN-FE, convolutional neural networks feature-exaggerated; MLP-BP, Multilayer Perceptron with Back Propagation; RIFM, Regional Image Features Model; FNN, feed-forward neural network; PNN-GA, probabilistic neural network-genetic algorithms; LASSO, least absolute shrinkage and selection operator; NB, naive Bayes; PNN, probabilistic neural network; RBF, Radial basis function; SD-OCT, spectral domain optical coherence tomography; SLO, Scanning Laser Ophthalmoscopy. Acc, accuracy; Sen, sensitivity; Spe, specificity; Vs, versus; AUC, area under the curve; POAG, primary open-angle glaucoma; AMD, age-related macular degeneration; OCT, optical coherence tomography; DR, diabetic retinopathy.
Figure 3Illustrations of transfer learning: a neural network is pretrained on ImageNet and subsequently trained on retinal, OCT, X-ray images, B-scans for different disease classifications