| Literature DB >> 35978087 |
Ramin Almasi1, Abbas Vafaei2, Elahe Kazeminasab1, Hossein Rabbani3.
Abstract
Microaneurysms (MAs) are pathognomonic signs that help clinicians to detect diabetic retinopathy (DR) in the early stages. Automatic detection of MA in retinal images is an active area of research due to its application in screening processes for DR which is one of the main reasons of blindness amongst the working-age population. The focus of these works is on the automatic detection of MAs in en face retinal images like fundus color and Fluorescein Angiography (FA). On the other hand, detection of MAs from Optical Coherence Tomography (OCT) images has 2 main advantages: first, OCT is a non-invasive imaging technique that does not require injection, therefore is safer. Secondly, because of the proven application of OCT in detection of Age-Related Macular Degeneration, Diabetic Macular Edema, and normal cases, thanks to detecting MAs in OCT, extensive information is obtained by using this imaging technique. In this research, the concentration is on the diagnosis of MAs using deep learning in the OCT images which represent in-depth structure of retinal layers. To this end, OCT B-scans should be divided into strips and MA patterns should be searched in the resulted strips. Since we need a dataset comprising OCT image strips with suitable labels and such large labelled datasets are not yet available, we have created it. For this purpose, an exact registration method is utilized to align OCT images with FA photographs. Then, with the help of corresponding FA images, OCT image strips are created from OCT B-scans in four labels, namely MA, normal, abnormal, and vessel. Once the dataset of image strips is prepared, a stacked generalization (stacking) ensemble of four fine-tuned, pre-trained convolutional neural networks is trained to classify the strips of OCT images into the mentioned classes. FA images are used once to create OCT strips for training process and they are no longer needed for subsequent steps. Once the stacking ensemble model is obtained, it will be used to classify the OCT strips in the test process. The results demonstrate that the proposed framework classifies overall OCT image strips and OCT strips containing MAs with accuracy scores of 0.982 and 0.987, respectively.Entities:
Mesh:
Year: 2022 PMID: 35978087 PMCID: PMC9385621 DOI: 10.1038/s41598-022-18206-8
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Recent works on MA detection from fundus photographs (CPM = Competition measure, CV = cross validation, BV = Blood vessel, ROC dataset = retinopathy online challenge datas et AUC = area under curve, ROC = receiver operating characteristic, SSAE = stacked sparse auto encoder, NB = Naive Bayesian, KNN = K-nearest neighbor, SVM = support vector machine, MMMF = multi-scale and multi-orientation sum of matched filter, and LLDA = local linear discrimination analysis).
| Author | Dataset | Image type | Initial candidates method | Classifier for false positive reduction | Results |
|---|---|---|---|---|---|
| Habib et al. [ | New dataset based on the MESSIDOR dataset | Color fundus | Gaussian matched filter | Tree Ensemble classifier | ROC score = 41.5% |
| Hatanaka et al. [ | DIARETDB1 and ROC datasets | Color fundus | DCNN (GoogLeNet) | DCNN and three-layer perceptron with 48 features | Sensitivity = 84% |
| Shan et al. [ | DIARETDB dataset | Color fundus | SSAE with two hidden layers plus a Softmax classifier | AUC = 96.2% | |
| With tenfold CV | |||||
| f-score = 91.3% | |||||
| Deepa et al. [ | DIARETDB1 dataset | Color fundus | CNN | Sensitivity = 97.62% Specificity = 100% Accuracy = 97.75% | |
| Zhang et al. [ | IDRiD_VOC dataset | Color fundus | Deep neural network with a multilayer attention mechanism | Refining initial candidates using spatial relationships between MAs and BVs | Sensitivity = 86.8% |
| Eftekhari et al.[ | ROC and E-Ophtha-MA datasets | Color fundus | Basic CNN | Thresholded probability map and final CNN | CPM = 0.461 |
| Long et al.[ | E-Ophtha-MA and DIARETDB1 datasets | Color fundus | BV removal, shape characteristics and connected components analysis | Classification using NB, KNN and SVM | AUC -ROC = 87% on E-Ophtha-MA |
| AUC of ROC = 86% on DIARETDB1 | |||||
| Wu et al.[ | ROC dataset | Color fundus | MMMF | Extracting 37 dimensional features and using modified KNN, LLDA, and SVM | Averaged number of false positives per image = 0.286 |
Some recent works on OCT image classification considering diabetic retinopathy (MCME = multi-scale convolutional mixture of experts, LBP = local binary patterns, BoF = bag of features, SURF = speeded up robust features, and MLP = multilayer perceptron).
| Author | Dataset | Image type | Method | Results |
|---|---|---|---|---|
| Lemaître et al.[ | SERI Dataset of normal, DME-cyst, DME-Exudate | SD-OCT images | LBP features, different mapping strategies and using linear and nonlinear classifiers | Sensitivity = 81.2% Specificity = 93.7% |
| Rasti et al.[ | Duke university dataset and private dataset of AMD, DME, and normal | Macular OCT | MCME model containing CNN experts and gating network are fed by specific scales of the input pattern | Precision = 98.86% |
| AUC-ROC = 99.85% | ||||
| Shih et al.[ | Dataset of DME, CNV, Drusen, and normal[ | OCT images | Pre-trained VGG16 | Accuracy = 99.48% |
| Tsuji et al.[ | Dataset of DME, CNV, Drusen, and normal[ | OCT images | Capsule network | Accuracy = 99.6% |
| Kazeminasab et al.[ | Dataset of MA and normal OCT strips | OCT images | BoF and SURF, MLP classifier | Accuracy = 94.5% |
Figure 1Overall block diagram of proposed method (SLO = Scanning Laser Ophthalmoscopy).
Figure 2The process of creating OCT strips for MA, normal, abnormal, and vessel with the help of corresponding FA. (a) Red circle shows MA in FA. (b) B-scan Corresponds to green line in (a). (c) Cropped ROI from (b). (d–f) Creating strip for normal class. (g–i) Creating strip for abnormal class. (j–l) Creating strip for vessel class (in color).
Figure 3Overall structure of the stacked generalization ensemble.
Figure 4Basic structure of CNNs used in stacking ensemble[39].
Per class and weighted average measures for method in[29], stand-alone CNNs and stacking ensemble of our method.
| Model per class parameter | Inception V3 | VGG16 | VGG19 | Xception | Method in[ | MLP ensemble |
|---|---|---|---|---|---|---|
| Abnormal | 0.948 | 0.792 | 0.922 | 0.883 | 0.931 | 0.974 |
| MA | 0.883 | 0.883 | 0.987 | 0.922 | 0.962 | 0.987 |
| Normal | 0.935 | 0.948 | 0.922 | 0.857 | 0.943 | 0.961 |
| Vessel | 0.896 | 0.961 | 0.987 | 0.974 | 0.946 | 1 |
| 0.913 | 0.91 | 0.958 | 0.916 | 0.945 | 0.982 | |
| Abnormal | 0.813 | 0.464 | 1 | 0.619 | 0.952 | 0.928 |
| MA | 0.682 | 1 | 0.945 | 0.739 | 0.964 | 0.944 |
| Normal | 0.895 | 0.944 | 0.797 | 1 | 0.97 | 0.947 |
| Vessel | 0.95 | 1 | 0.963 | 1 | 0.964 | 1 |
| 0.851 | 0.888 | 0.921 | 0.873 | 0.963 | 0.961 | |
| Abnormal | 0.929 | 0.929 | 0.571 | 0.929 | 0.86 | 0.928 |
| MA | 0.882 | 0.471 | 1 | 1 | 0.913 | 1 |
| Normal | 0.85 | 0.85 | 0.95 | 0.45 | 0.907 | 0.9 |
| Vessel | 0.731 | 0.885 | 1 | 0.923 | 0.883 | 1 |
| 0.831 | 0.792 | 0.909 | 0.818 | 0.890 | 0.961 | |
| Abnormal | 0.952 | 0.762 | 1 | 0.873 | 0.866 | 0.984 |
| MA | 0.883 | 1 | 0.983 | 0.9 | 0.941 | 0.983 |
| Normal | 0.964 | 0.982 | 0.912 | 1 | 0.86 | 0.982 |
| Vessel | 0.980 | 1 | 0.980 | 1 | 0.887 | 1 |
| 0.95 | 0.952 | 0.967 | 0.818 | 0.888 | 0.989 | |
| Abnormal | 0.867 | 0.619 | 0.727 | 0.743 | – | 0.928 |
| MA | 0.769 | 0.64 | 0.971 | 0.85 | – | 0.971 |
| Normal | 0.872 | 0.895 | 0.863 | 0.621 | – | 0.923 |
| Vessel | 0.826 | 0.939 | 0.981 | 0.96 | – | 1 |
| 0.833 | 0.803 | 0.902 | 0.808 | – | 0.961 | |
| 0.981 | 0.97 | 0.99 | 0.989 | – | 0.998 | |
Accuracy = (TP + TN)/(TP + TN + FP + FN), Precision = TP/(TP + FP), Recall(sensitivity) = TP/(TP + FN), Specificity = TN/(TN + FP), F1score = (2 × Recall × Precision)/(Recall + Precision) = (2TP)/(2TP + FP + FN), ROC-AUC score: the closer this criterion is to one, the greater the number of correctly predicted cases. On the other hand, the closer this criterion is to zero, the greater the number of incorrectly predicted cases.
Figure 5Confusion matrices. (a) The relationship between confusion matrix and TP, TN, FP, and FN for MA class. (b–f) Confusion matrices of different CNNs and stacking ensemble (rows: True labels and cols: Predicted labels) (in color).