| Literature DB >> 26221613 |
Zsolt Torok1, Tunde Peto2, Eva Csosz3, Edit Tukacs4, Agnes M Molnar5, Andras Berta6, Jozsef Tozser7, Andras Hajdu8, Valeria Nagy9, Balint Domokos10, Adrienne Csutak6.
Abstract
Background. It is estimated that 347 million people suffer from diabetes mellitus (DM), and almost 5 million are blind due to diabetic retinopathy (DR). The progression of DR can be slowed down with early diagnosis and treatment. Therefore our aim was to develop a novel automated method for DR screening. Methods. 52 patients with diabetes mellitus were enrolled into the project. Of all patients, 39 had signs of DR. Digital retina images and tear fluid samples were taken from each eye. The results from the tear fluid proteomics analysis and from digital microaneurysm (MA) detection on fundus images were used as the input of a machine learning system. Results. MA detection method alone resulted in 0.84 sensitivity and 0.81 specificity. Using the proteomics data for analysis 0.87 sensitivity and 0.68 specificity values were achieved. The combined data analysis integrated the features of the proteomics data along with the number of detected MAs in the associated image and achieved sensitivity/specificity values of 0.93/0.78. Conclusions. As the two different types of data represent independent and complementary information on the outcome, the combined model resulted in a reliable screening method that is comparable to the requirements of DR screening programs applied in clinical routine.Entities:
Mesh:
Substances:
Year: 2015 PMID: 26221613 PMCID: PMC4499636 DOI: 10.1155/2015/623619
Source DB: PubMed Journal: J Diabetes Res Impact factor: 4.011
Characteristics of the participants.
| Total number of participants enrolled | Eye examination | ||||
|---|---|---|---|---|---|
| 52 | 104 | ||||
|
| |||||
| Non-DR | DR | Eyes included | Eyes excluded | ||
| Proliferative | Nonproliferative | Tear fluid sampling | Retina photography | ||
|
| |||||
| 13 | 15 | 24 | 74 | 9 | 21 |
Figure 1Microaneurysm detection. (a) Original retina image; (b) CLAHE contrast enhancement; (c) median filtering; (d) top-hat transform; (e) raw MA candidates.
Performance measures of the screening methods.
| Screening method | SENS | SPC | ACC | PREC | NPV |
| LRP | LRN | |
|---|---|---|---|---|---|---|---|---|---|
| Image processing | Mean | 0.84 | 0.81 | 0.84 | 0.94 | 0.63 | 0.89 | 4.42 | 0.20 |
| SD | 0.11 | 0.04 | 0.10 | 0.13 | 0.11 | 0.11 | 2.36 | 0.11 | |
|
| |||||||||
| Proteomics | Mean | 0.87 | 0.68 | 0.82 | 0.89 | 0.63 | 0.88 | 2.72 | 0.19 |
| SD | 0.17 | 0.12 | 0.11 | 0.21 | 0.15 | 0.16 | 1.24 | 0.12 | |
|
| |||||||||
| Combined method | Mean | 0.93 | 0.78 | 0.89 | 0.93 | 0.78 | 0.93 | 4.23 | 0.09 |
| SD | 0.18 | 0.19 | 0.15 | 0.23 | 0.22 | 0.18 | 1.32 | 0.07 | |
Performance measures of the image processing based, the tear proteomics based, and the combined screening methods. SENS: sensitivity, SPC: specificity, ACC: accuracy, PREC: precision (positive predictive value), NPV: negative predictive value, F1: F-measure, LRP: likelihood ratio positive, and LRN: likelihood ratio negative.
Figure 2Application of machine learning algorithm in the combined model. Learning phase (above): we randomly select a subpopulation of the total patient group, called the training group, and then use the known clinical diagnosis to split the training group into a DR group and a non-DR group. The clinical diagnosis, the number of MAs on the retina images, and the protein concentration values are the inputs of the machine learning algorithm. The algorithms are able to tell which data patterns are the most characteristic for the DR and non-DR groups. Assessment phase (below): in the following steps, we use the data from the validation group. The number of MAs and the protein concentration values constitute the input of the algorithm, but we do not use the information from clinical diagnosis. The learning algorithm compares the new data to the characteristic patterns that are known from the learning phase and will make its own decision (normal/DR) for each patient as the output of the model. For the assessment of the performance of the model, we compare the output with the known clinical diagnosis.