Literature DB >> 35888063

Which Color Channel Is Better for Diagnosing Retinal Diseases Automatically in Color Fundus Photographs?

Sangeeta Biswas1, Md Iqbal Aziz Khan1, Md Tanvir Hossain1, Angkan Biswas2, Takayoshi Nakai3, Johan Rohdin4.   

Abstract

Color fundus photographs are the most common type of image used for automatic diagnosis of retinal diseases and abnormalities. As all color photographs, these images contain information about three primary colors, i.e., red, green, and blue, in three separate color channels. This work aims to understand the impact of each channel in the automatic diagnosis of retinal diseases and abnormalities. To this end, the existing works are surveyed extensively to explore which color channel is used most commonly for automatically detecting four leading causes of blindness and one retinal abnormality along with segmenting three retinal landmarks. From this survey, it is clear that all channels together are typically used for neural network-based systems, whereas for non-neural network-based systems, the green channel is most commonly used. However, from the previous works, no conclusion can be drawn regarding the importance of the different channels. Therefore, systematic experiments are conducted to analyse this. A well-known U-shaped deep neural network (U-Net) is used to investigate which color channel is best for segmenting one retinal abnormality and three retinal landmarks.

Entities:  

Keywords:  color fundus photographs; deep neural network; detection of retinal diseases; segmentation of retinal landmarks

Year:  2022        PMID: 35888063      PMCID: PMC9321111          DOI: 10.3390/life12070973

Source DB:  PubMed          Journal:  Life (Basel)        ISSN: 2075-1729


1. Introduction

Diagnosing retinal diseases at their earliest stage can save a patient’s vision since, at an early stage, the diseases are more likely to be treatable. However, ensuring regular retina checkups for each citizen by ophthalmologists is infeasible not only in developing countries with huge populations but also in developed countries with small populations. The main reason is that the number of ophthalmologists compared to citizens is very small. It is particularly true for low-income and low-middle-income countries with huge populations, such as Bangladesh and India. For example, according to a survey conducted by the International Council of Ophthalmology (ICO) in 2010 [1], there were only four ophthalmologists per million people in Bangladesh. For India, the number was 11. Even for high-income countries with a small population, such as Switzerland and Norway, the numbers of ophthalmologists per million were not very high (91 and 68, respectively). More than a decade later, in 2021, these numbers remain roughly the same. Moreover, 60+ people (who are generally at high risk of retinal diseases) are increasing in most countries. The shortage of ophthalmologists and the necessity of regular retina checkups at low cost inspired researchers to develop computer-aided systems to detect retinal diseases automatically. Different kinds of imaging technologies (e.g., color fundus photography, monochromatic retinal photography, wide-field imaging, autofluorescence imaging, indocyanine green angiography, scanning laser ophthalmoscopy, Heidelberg retinal tomography and optical coherence tomography) have been developed for the clinical care and management of patients with retinal diseases [2]. Among them, color fundus photography is available and affordable in most parts of the world. A color fundus photograph can be captured using a non-mydriatic fundus camera, handled by non-professional personnel, and delivered online to major ophthalmic institutions for follow-up in the case a disease is suspected. Moreover, there are many publicly available data sets of color fundus photographs such as CHASE_DB1 [3,4], DRIVE [5], HRF [6], IDRiD [7], Kaggle EyePACS data set [8], Messidor [9], STARE [10,11] and UoA_DR [12] to help researchers compare the performances of their proposed approaches. Therefore, color fundus photography is used more widely than other retinal imaging techniques for automatically diagnosing retinal diseases. In color fundus photographs, the intensity of colors reflected from the retina are recorded in three color channels, red, green, and blue. In this paper, we investigate which color channel is better for the automatic detection of retinal diseases as well as the segmentation of retinal landmarks. Although the detection of retinal diseases is the main objective of computer-aided diagnostic (CAD) systems, segmentation is also an important part of many CAD systems. For example, structural changes in the central retinal blood vessels (CRBVs) may indicate diabetic retinopathy (DR). Therefore, a technique for segmenting CRBVs is often an important step in DR detection systems. Similarly, optic disc (OD) segmentation is important for some glaucoma detection algorithms. In this work, we first extensively survey the usage of the different color channels in previous works. Specifically, we investigate works on four retinal diseases (i.e., glaucoma, age-related macular degeneration (AMD), and DR, diabetic macular edema (DME)) which are the major causes of blindness [13,14,15,16] as well as works on the segmentation of retinal landmarks, such as OD, macula/fovea and CRBVs, and retinal atrophy. We notice that the focus of the previous works was not to investigate which of the different channels (or combination of channels) is the best for the automatic analysis of fundus photographs. At the same time, there does not seem to be complete consensus on this since different studies used different channels (or combinations of channels). Therefore, to better understand the importance of the different color channels, we develop color channel-specific U-shaped deep neural networks (i.e., U-Nets [17]) for segmenting OD, macula, and CRBVs. We also develop U-Nets for segmenting retinal atrophy. The U-Net is well-known for its excellent performance in medical image segmentation tasks. The U-Net can segment images in great detail, even using very few images in the training phase. It is shown in [17] that a U-Net trained using only 30 images outperformed a sliding window convolutional neural network for the ISBI neuronal structures in the EM stacks challenge 2012. To the best of our knowledge, a systematic exploration of the importance of different color channels for the automatic processing of color fundus photographs has not been undertaken before. Naturally, a better understanding of the effectiveness of different color channels can reduce the amount of development time of future algorithms. In the long term, it may also affect the design of new fundus cameras and the procedures for capturing fundus photographs, e.g., the appropriate light conditions. The organization of this paper is as follows: in Section 2, we describe briefly the different color channels of a color fundus photograph, in Section 3, we survey which color channels were used in previous works for the automatic detection of retinal diseases and segmentation. In Section 4, we describe our setup for U-Nets based experiments. In Section 5, we show the performance of color channel-specific U-Nets. At last, in Section 6, we draw conclusions about our findings. Some steps of image pre-processing in more detail and additional experiments are described in the Appendices.

2. Fundus Photography

Our retina does not have any illumination power. Moreover, it is a minimally reflective surface. Therefore, a fundus camera which is a complex optical system, needs to illuminate and capture the low reflected light of the retina simultaneously while imaging [18]. A single image sensor coated with a color filter array (CFA) is used more commonly to capture the reflected light in a fundus camera. In a CFA, in general, color filters are arranged following the Bayer pattern [19], developed by the Eastman Kodak company, as shown in Figure 1a. Instead of using three filters for capturing three primary colors (i.e., red, green and blue) reflected from the retina, only one filter is used per pixel to capture one primary color in the Bayer pattern. In this pattern, the number of green filters is twice the number of blue and red filters. Different kinds of demosaicing techniques are applied to get full color fundus photographs [20,21,22]. Some sophisticated and expensive fundus cameras do not use a CFA with a Bayer pattern to distinguish color, rather they use a direct imaging sensor with three layers of photosensitive elements as shown in Figure 1b. No demosaicing technique is necessary for getting full color fundus photographs from such fundus cameras.
Figure 1

Sensors used in fundus cameras: (a) commonly used a single layered sensor coated with a color filter array having a Bayer pattern and (b) less commonly used three-layered direct imaging sensor. R: Red, G: Green, B: Blue.

As shown in Figure 2, in a color fundus photograph, we can see the major retinal landmarks, such as the optic disc (OD), macula, and central retinal blood vessels (CRBVs), on the colored foreground surrounded by the dark background. As can be seen in Figure 3, different color channels highlight different things in color fundus photographs. We can see the boundary of the OD more clearly and the choroid in more detail in the red channel. The red channel helps us segment the OD more accurately and see the choroidal blood vessels and choroidal lesions such as nevi or tumors more clearly than the other two color channels. The CRBVs and hemorrhages can be seen in the green channel with excellent contrast. The blue channel allows us to see the retinal nerve fiber layer (RNFL) defects and epiretinal membranes more clearly than the other two color channels.
Figure 2

A color fundus photograph. We can see the retinal landmarks, i.e., optic disc, macula, and central retinal blood vessels, on the circular and colored foreground, surrounded by a dark background. Source of image: publicly available DRIVE data set and image file: 21_training.tif.

Figure 3

Pros and cons of different color channels. 1st column i.e., (a,e,i,m): RGB fundus photographs, 2nd column i.e., (b,f,j,n): red channel images, 3rd column i.e., (c,g,k,o): green channel images, and 4th column i.e., (d,h,l,p): blue channel images. Choroidal blood vessels are clearly visible in the red channel, as shown inside the red box in (b). Lens flares are more visible in the blue channel, as shown inside the blue box in (d). Atrophy and diabetic retinopathy affected areas are more clearly visible in the green channel as shown inside the green boxes in (g,k). As shown inside the blue box in (l), the blue channel is prone to underexposure. The red channel is prone to overexposure, as shown inside the red box in (m). Source of fundus photographs: (a) PALM/PALM-Training400/H0025.jpg, (e) PALM/PALM-Training400/P0010.jpg, (i) UoA_DR/94/94.jpg, and (m) CHASE_DB1/images/Image_11L.jpg.

3. Previous Works on Diagnosing Retinal Disease Automatically

Many diseases can be the cause of retinal damage, such as glaucoma, age-related macular degeneration (AMD), diabetic retinopathy (DR), diabetic macular edema (DME), retinal artery occlusion, retinal vein occlusion, hypertensive retinopathy, macular hole, epiretinal membrane, retinal hemorrhage, lattice degeneration, retinal tear, retinal detachment, intraocular tumors, penetrating ocular trauma, pediatric and neonatal retinal disorders, cytomegalovirus retinal infection, uveitis, infectious retinitis, central serous retinopathy, retinoblastoma, endophthalmitis, and retinitis pigmentosa. Among them, glaucoma, AMD, DR, and DME drew the main focus of researchers for color fundus photograph-based automation. One reason could be that for many cases, these causes lead to irreversible complete vision loss, i.e., blindness if they are left undiagnosed and untreated. According to the information reported in [23,24], glaucoma, AMD, and DR are among the five most common causes of vision impairment in adults. Among billion people living in 2020, million people experienced moderate or severe vision impairment (MSVI) and million people were blind. Glaucoma was the cause of MSVI for million people, whereas AMD for million and DR for million people. Glaucoma was the cause of blindness for million people, whereas AMD for million and DR for million people [24]. Therefore, in our literature survey, we investigate the color channels used in previously published studies for automatically diagnosing glaucoma, DR, AMD, and DME. We also survey works on segmentation of retinal landmarks, such as OD, macula/fovea and CRBVs, and retinal atrophy. We consider both original studies and reviews as the source of information. However, our survey includes only original studies written in English and published in SJR ranked Q1 and Q2 journals. Note that SJR (SCImago Journal Rank) is an indicator developed by SCImago from the widely known algorithm Google PageRank [25]. This indicator shows the visibility of the journals contained in the Scopus database from 1996. We used different keywords such as ‘automatic retinal disease detection’, ‘automatic diabetic retinopathy detection’, ‘automatic glaucoma detection’, ‘detect retinal disease by deep learning’, ‘segment macula’, ‘segment optic disc’, and ‘segment central retinal blood vessels’ in the Google search engine to find previous studies. After finding a paper, we checked the SJR rank of the journal. We used the reference list of papers published in Q1/Q2 journals; we especially benefited from the review papers related to our area of interest. In this paper, we include our findings based on information reported in 199 journal papers. As shown in Table 1, the green channel dominates non-neural network-based previous works, whereas RGB images (i.e., red, green, and blue channels together) dominate neural network-based previous works. Few works were based on the red and blue channels and they were mainly for atrophy segmentation. See Table 2, Table 3, Table 4, Table 5 and Table 6 for the color channel distribution in our studied previous works.
Table 1

Color distribution in previous works for the automatic detection of retinal diseases and segmentation of retinal landmarks and atrophy. NN: Neural network-based approaches, Non-NN: Non-neural network-based approaches.

ColorNumber of Papers
Disease DetectionSegmentation
Non-NNNNNon-NNNN
TotalQ1Q2TotalQ1Q2TotalQ1Q2TotalQ1Q2
(42)(30)(12)(35)(28)(7)(77)(56)(21)(37)(28)(9)
RGB1899292451410428226
R7522111596000
G2211114225943161082
B330110871000
Gr633541752303
Table 2

Color channel used in non-neural Network (Non-NN) based previous works for automatically detecting diseases in retina. DR: Diabetic Retinopathy, AMD: Age-related Macular Degeneration, DME: Diabetic Macular Edema, R: Red, G: Green, B: Blue, Gr: Grayscale weighted summation of Red, Green and Blue.

YearGlaucomaAMD & DMEDR
ReferenceColorReferenceColorReferenceColor
2000 Hipwell [26]G, B
2002 Walter [27]G
2004 Klein [28]RGB
2007 Scott [29]RGB
2008 Kose [30]RGBAbramoff [31]RGB
Gangnon [32]RGB
2010Bock [33]GKose [34]Gr
Muramatsu [35]R, G
2011Joshi [36]RAgurto [37]GFadzil [38]RGB
2012Mookiah [39]GrHijazi [40]RGB
Deepak [41]RGB, G
2013 Akram [42]RGB
Oh [43]RGB
2014Fuente-Arriaga [44]R, G Akram [45]RGB
Noronha [46]RGBMookiah [47]GCasanova [48]RGB
2015Issac [49]R, GMookiah [50]R, GJaya [51]RGB
Oh [52]G, Gr
2016Singh [53]G, GrAcharya [54]GBhaskaranand [55]RGB
Phan [56]G
Wang [57]RGB
2017Acharya [58]GrAcharya [59]GLeontidis [60]RGB
Maheshwari [61]R, G, B, Gr
Maheshwari [62]G
2018 Saha [63]G, RGB
2020 Colomer [64]G
Table 3

Color channel used in neural network (NN) based previous works for automatically detecting diseases in retina. DR: Diabetic Retinopathy, AMD: Age-related Macular Degeneration, DME: Diabetic Macular Edema, Gr: Grayscale weighted summation of Red, Green and Blue, R: Red, G: Green, B: Blue.

YearGlaucomaAMD & DMEDR
ReferenceColorReferenceColorReferenceColor
1996 Gardner [65]RGB
2009Nayak [66]R, G
2014 Ganesan [67]Gr
2015 Mookiah [68]G
2016Asoka [69]Gr Abramoff [70]RGB
Gulshan [71]RGB
2017Zilly [72]G, GrBurlina [73]RGBAbbas [74]RGB
Ting [75]RGBBurlina [76]RGBGargeya [77]RGB
Quellec [78]RGB
2018Ferreira [79]RGB, GrGrassmann [80]RGBKhojasteh [81]RGB
Raghavendra [82]RGBBurlina [83]RGBLam [84]RGB
Li [85]RGB
Fu [86]RGB
Liu [87]RGB
2019Liu [88]R, G, B, GrKeel [89]RGBLi [90]RGB
Diaz-Pinto [91]RGBPeng [92]RGBZeng [93]RGB
Matsuba [94]RGBRaman [95]RGB
2020 Singh [96]RGB
Gonzalez-Gonzalo [97]RGB
2021Gheisari [98]RGB
Table 4

Color channel used in non-neural network (Non-NN) based previous works for segmenting retinal landmarks. OD: Optic Disc, CRBVs: Central Retinal Blood Vessels, Gr: Grayscale weighted summation of Red, Green and Blue, R: Red, G: Green, B: Blue.

YearODMacula/FoveaCRBVs
ReferenceColorReferenceColorReferenceColor
1989 Chaudhuri [99]G
1999 Sinthanayothin [100]RGB
2000 Hoover [10]RGB
2004Lowell [101]GrLi [102]RGB
2006 Soares [103]G
2007Xu [104]RGBNiemeijer [105]GRicci [106]G
Abramoff [107]R, G, BTobin [108]G
2008Youssif [109]RGB
2009 Niemeijer [110]GCinsdikici [111]G
2010Welfer [112]G
Aquino [113]R, G
Zhu [114]RGB
2011Lu [115]R, GWelfer [116]GCheung [117]RGB
Kose [118]RGB
You [119]G
2012 Bankhead [120]G
Qureshi [121]GFraz [4]G
Fraz [122]G
Li [123]RGB
Lin [124]G
Moghimirad [125]G
2013Morales [126]GrChin [127]RGBAkram [128]G
Gegundez [129]GBadsha [130]Gr
Budai [6]G
Fathi [131]G
Fraz [132]G
Nayebifar [133]G, B
Nguyen [134]G
Wang [135]G
2014Giachetti [136]G, GrKao [137]GBekkers [138]G
Aquino [139]R, GCheng [140]G
2015Miri [141]R, G, B Dai [142]G
Mary [143]R Hassanien [144]G
Harangi [145]RGB, G Imani [146]G
Lazar [147]G
Roychowdhury [148]G
2016Mittapalli [149]RGBMedhi [150]RAslani [151]G
Roychowdhury [152]GOnal [153]GrBahadarkhan [154]G
Sarathi [155]R, G Christodoulidis [156]G
Orlando [157]G
2018 Ramani [158]GKhan [159]G
Chalakkal [160]RGBXia [161]G
2019Thakur [162]Gr Khawaja [163]G
Naqvi [164]R, G Wang [165]RGB
2020Dharmawan [166]R, G, BCarmona [167]GSaroj [168]Gr
Guo [169]GZhang [170]G
Zhou [171]G
2021 Kim [172]G
Table 5

Color channel used in neural network (NN) based previous works for segmenting retinal landmarks. OD: Optic Disc, CRBVs: Central Retinal Blood Vessels, Gr: Grayscale weighted summation of Red, Green and Blue, R: Red, G: Green, B: Blue.

YearODMacula/FoveaCRBVs
ReferenceColorReferenceColorReferenceColor
2011 Marin [173]G
2015 Wang [174]G
2016 Liskowski [175]G
2017 Barkana [176]G
Mo [177]RGB
2018Fu [178]RGBAl-Bander [179]GrGuo [180]G
Guo [181]RGB
Hu [182]RGB
Jiang [183]RGB
Oliveira [184]G
Sangeethaa [185]G
2019Wang [186]RGB, Gr Jebaseeli [187]G
Chakravarty [188]RGB Lian [189]RGB
Gu [190]RGB Noh [191]RGB
Tan [192]RGB Wang [193]Gr
Jiang [194]RGB
2020Gao [195]RGB Feng [196]G
Jin [197]RGB Tamim [198]G
Sreng [199]RGB
Bian [200]RGB
Almubarak [201]RGB
Tian [202]RGB
Zhang [203]RGB
Xie [204]RGB
2021Bengani [205]RGBHasan [206]RGBGegundez-Arias [207]RGB
Veena [208]RGB
Wang [209]RGB
Table 6

Color channel used for automatically detecting atrophy in retina. R: Red, G: Green, B: Blue.

YearNon-NNNN
ReferenceColorReferenceColor
2011Lu [210]R, B
2012Cheng [211]R, G, B
Lu [212]R, B
2018Septiarini [213]R, G
2020Li [214]R, G, BChai [215]RGB
Son [216]RGB
2021 Sharma [217]RGB

4. Experimental Setup

4.1. Hardware & Software Tools

We performed all experiments using TensorFlow’s Keras API 2.0.0, OpenCV 4.2.0, and Python 3.6.9. We used a standard PC with 32 GB memory, Intel 10th Gen Core i5-10400 Processor with six cores per socket, and Intel UHD Graphics 630 (CML GT2).

4.2. Data Sets

We used RGB color fundus photographs from seven publicly available data sets: (1) Child Heart Health Study in England (CHASE) data set [3,4], (2) Digital Retinal Images for Vessel Extraction (DRIVE) data set [5], (3) High-Resolution Fundus (HRF) data set [6], (4) Indian Diabetic Retinopathy Image Dataset (IDRiD) [7], (5) Pathologic Myopica Challenge (PALM) data set [218], (6) STructured Analysis of the Retina (STARE) data set [10,11], and (7) University of Auckland Diabetic Retinopathy (UoA-DR) data set [12]. Images in these data sets were captured by different fundus cameras for different kinds of research objectives, as shown in Table 7.
Table 7

Data sets used in our experiments.

Data SetHeight × WidthField-of-ViewFundus CameraNumber of Images
CHASE_DB1 960×999 30 Nidek NM-200-D28
DRIVE 584×565 45 Canon CR5-NM 3CCD40
HRF 3264×4928 45 Canon CR-145
IDRiD 2848×4288 50 Kowa VX-10α81
PALM 1444×1444 2056×2124 45 Zeiss VISUCAM 500 NM400
STARE 605×700 35 TopCon TRV-5020
UoA-DR 2056×2124 45 Zeiss VISUCAM 500200
Since all of the seven data sets do not have manually segmented images for all retinal landmarks and atrophy, we cannot use all of them for all kinds of segmentation tasks. Therefore, instead of seven data sets we used five data sets for the experiments of segmenting CRBVs, three data sets for OD, and two data sets for macula, while only one data set for the experiments of segmenting retinal atrophy. We emphasize to have reliable results. For that we used the majority of the data (i.e., 55% of the data) as the test data. We prepared one training and one validation set. By combining 25% of the data from each data set, we prepared the training set, whereas we prepared the validation set by combining 20% of the data from each data set. By taking the rest of the 55% of the data from each data set, we prepared individual test sets for each type of segmentation. See Table 8 for the number of images in the training, validation, and test sets. Note that the training set is used to tune the parameters of the U-Net (i.e., weights and biases), the validation set is used to tune the hyperparameters (such number of epochs, learning rate, and activation function), and the test set is used to evaluate the performance of the U-Net.
Table 8

Training, validation and test sets used in our experiments.

Segmentation ofData SetNumber of Images in
Training SetValidation SetTest Set
CRBVsCHASE_DB17516
DRIVE10822
HRF11925
STARE5411
UoA-DR5040110
Optic DiscIDRiD201645
PALM10080220
UoA-DR5040110
MaculaPALM10080220
UoA-DR5040110
AtrophyPALM10080220

4.3. Image Pre-Processing

We prepared four types of 2D fundus photographs: , , , and . By splitting 3D color fundus photographs into three color channels (i.e., red, green and blue), we prepared , , . Moreover, by performing a weighted summation of , , , we prepared the grayscale image, . By a grayscale image, we generally mean an image whose pixels have only one value representing the amount of light. It can be visualized as different shades of gray. An 8-bit grayscale image has pixel values in the range 0–255. There are many ways to convert a color image into a grayscale image. In this paper, we use a function from the OpenCV library where each grey pixel is generated according to the following scheme: . This conversion scheme is frequently used in computer vision and implemented in different toolboxes, e.g., GIMP and MATLAB [219] including OpenCV. The background of a fundus photograph does not contain any information about the retina, which can be helpful for manual or automatic retina-related tasks. Sometimes background noise can be misleading. In order to avoid the interference of the background noise in any decision, we need to use a binary background mask, which has zero for the pixels of the background and for the pixels of the foreground, where n is the number of bits used for the intensity of each pixel. For an 8-bit image, . Except the DRIVE and HRF data sets, background masks are not provided for the other five data sets. Therefore, we followed the steps described in Appendix A to generate the background masks for all data sets. We generated binary background masks for DRIVE and HRF data sets in order to keep the same set up for all data sets. Overall, has a higher intensity than and in all data sets, whereas has a lower intensity compared to and . Moreover, in , the foreground is less likely to overlap with the background noise than and . In , the foreground intensity has the highest possibility to be overlapped with the intensity of the background noise, as shown in Figure 4. Therefore, we use (i.e., the red channel image) for generating the binary background masks.
Figure 4

There is a noticeable overlap in the histograms of the foreground and the background in the blue channel. Histograms are slightly overlapped in the green channel. In the red channel, histograms are not overlapped and easily separable. Therefore, by setting 0 to the pixels lower than the threshold value, and setting 255 to the pixels higher than the , we can easily generate the background mask from the red channeled image. Source of fundus photograph: STARE data set and image file: im0139.ppm.

We used the generated background mask and followed the steps described in Appendix B for cropping out the background as much as possible and removing background noise outside the field-of-view (FOV). Since cropped fundus photographs of different data sets have different resolutions as shown in Table 7, we re-sized all masked and cropped fundus photographs to by bicubic interpolation so that we could use one U-Net. After resizing fundus photographs, we applied contrast limited adaptive histogram equalization (CLAHE) [220] to improve the contrast of each single colored image. Then we re-scaled pixel values to . Note that, re-scaling pixel values to is not necessary for fundus photographs. However, we did it to keep the input and output in the same range. We did not apply any other pre-processing techniques to the images. Similar to the fundus photographs, reference masks provided by the data sets for segmenting OD, CRBVs and retinal atrophy can have an unnecessary and noisy background. We, therefore, cropped out the unnecessary background of the provided reference masks and removed noise outside the field-of-view area by following the steps described in Appendix B. Since some provided masks are not binary masks, we turned them into 2D binary masks by following the steps described in Appendix C. No data set provides binary masks for segmenting the macula. Instead the center of the macula are provided by the PALM and UoA-DR. We generated binary masks for segmenting macula using the center values of the macula and the OD masks of the PALM and UoA-DR by following the steps described in Appendix D. We re-sized all kinds of binary masks to by bicubic interpolation. We then re-scaled pixel values to , since we used the sigmoid function as the activation function in the output layer of the U-Net and the range of this function is .

4.4. Setup for U-Net

We trained color-specific U-Nets with an architecture as shown in Table A3 of Appendix E. To train our U-Nets, we set Jaccard co-efficient loss (JCL) as the loss function; RMSProp with a learning rate of as the optimizer and . We reduced the learning rate if there was no change in the for more than 30 consecutive epochs. We stopped the training if the did not change in 100 consecutive epochs. We trained all color-specific U-Nets five times to avoid the effect of randomness caused by different factors, including weight initialization and dropout, on the U-Net’s performance. That means, in total, we trained 100 U-Nets, among which 25 U-Nets for OD segmentation (i.e., five models for each RGB, gray, red, green, and blue), 25 U-Nets for macula segmentation, 25 U-Nets for CRBVs segmentation, and 25 U-Nets for atrophy segmentation. We estimate the performance of each model separately and then report of the performance for each category.
Table A3

Architecture of our U-Net. #Params: Number of parameters.

LayerOutput Shape# Params
Input(256, 256, 1)0
Convolution (strides = (1, 1), filters = 16, kernel = (3, 3), activation = ELU)(256, 256, 16)160
Dropout (0.1)(256, 256, 16)0
Convolution (strides = (1, 1), filters = 16, kernel = (3, 3), activation = ELU, name = C1)(256, 256, 16)2320
Convolution (strides = (2, 2), filters = 16, kernel = (3, 3), activation = ELU)(128, 128, 16)2320
Convolution (strides = (1, 1), filters = 32, kernel = (3, 3), activation = ELU)(128, 128, 32)4640
Dropout (0.1)(128, 128, 32)0
Convolution (strides = (1, 1), filters = 32, kernel = (3, 3), activation = ELU, name = C2)(128, 128, 32)9248
Convolution (strides = (2, 2), filters = 32, kernel = (3, 3), activation = ELU)(64, 64, 32)9248
Convolution (strides = (1, 1), filters = 64, kernel = (3, 3), activation = ELU)(64, 64, 64)18,496
Dropout (0.2)(64, 64, 64)0
Convolution (strides = (1, 1), filters = 64, kernel = (3, 3), activation = ELU, name = C3)(64, 64, 64)36,928
Convolution (strides = (2, 2), filters = 64, kernel = (3, 3), activation = ELU)(32, 32, 64)36,928
Convolution (strides = (1, 1), filters = 128, kernel = (3, 3), activation = ELU)(32, 32, 128)73,856
Dropout (0.2)(32, 32, 128)0
Convolution (strides = (1, 1), filters = 128, kernel = (3, 3), activation = ELU, name = C4)(32, 32, 128)147,584
Convolution (strides = (2, 2), filters = 128, kernel = (3, 3), activation = ELU)(16, 16, 128)147,584
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU)(16, 16, 256)295,168
Dropout (0.3)(16, 16, 256)0
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU, name = C5)(16, 16, 256)590,080
Convolution (strides = (2, 2), filters = 256, kernel = (3, 3), activation = ELU)(8, 8, 256)590,080
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU)(8, 8, 256)590,080
Dropout (0.3)(8, 8, 256)0
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU)(8, 8, 256)590,080
Transposed Convolution (strides = (2, 2), filters = 256, kernel = (2, 2), activation = ELU, name = U1)(16, 16, 256)262,400
Concatenation (C5, U1)(16, 16, 512)0
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU)(16, 16, 256)1,179,904
Dropout (0.3)(16, 16, 256)0
Convolution (strides = (1, 1), filters = 256, kernel = (3, 3), activation = ELU)(16, 16, 256)590,080
Transposed Convolution (strides = (2, 2), filters = 128, kernel = (2, 2), activation = ELU, name = U2)(32, 32, 128)131,200
Concatenation (C4, U2)(32, 32, 256)0
Convolution (strides = (1, 1), filters = 128, kernel = (3, 3), activation = ELU)(32, 32, 128)295,040
Dropout (0.2)(32, 32, 128)0
Convolution (strides = (1, 1), filters = 128, kernel = (3, 3), activation = ELU)(32, 32, 128)147,584
Transposed Convolution (strides = (2, 2), filters = 64, kernel = (2, 2), activation = ELU, name = U3)(64, 64, 64)32,832
Concatenation (C3, U3)(64, 64, 128)0
Convolution (strides = (1, 1), filters = 64, kernel = (3, 3), activation = ELU)(64, 64, 64)73,792
Dropout (0.2)(64, 64, 64)0
Convolution (strides = (1, 1), filters = 64, kernel = (3, 3), activation = ELU)(64, 64, 64)36,928
Transposed Convolution (strides = (2, 2), filters = 32, kernel = (2, 2), activation = ELU, name = U4)(128, 128, 32)8224
Concatenation (C2, U4)(128, 128, 64)0
Convolution (strides = (1, 1), filters = 32, kernel = (3, 3), activation = ELU)(128, 128, 32)18,464
Dropout (0.1)(128, 128, 32)0
Convolution (strides = (1, 1), filters = 32, kernel = (3, 3), activation = ELU)(128, 128, 32)9248
Transposed Convolution (strides = (2, 2), filters = 16, kernel = (2, 2), activation = ELU, name = U5)(256, 256, 16)2064
Concatenation (C1, U5)(256, 256, 16)0
Convolution (strides = (1, 1), filters = 16, kernel = (3, 3), activation = ELU)(256, 256, 16)4624
Dropout (0.1)(256, 256, 16)0
Convolution (strides = (1, 1), filters = 16, kernel = (3, 3), activation = ELU)(256, 256, 16)2320
Convolution (strides = (1, 1), filters = 1, kernel = (1, 1), activation = Sigmoid, name = Output)(256, 256, 1)17

4.5. Evaluation Metrics

In segmentation, the U-Net shall predict whether a pixel is part of the object in question (e.g., OD) or not. Ideally, it should therefore output: However, instead of 0/1, the output of the U-Net is in the range [0, 1] for each pixel since we use sigmoid as the activation function in the last layer. The output can be interpreted as the probability that the pixel is part of the mask. To obtain a hard prediction (0/1), we use a threshold of 0.5. By comparing the hard prediction to the reference, it is decided whether the prediction is a true positive (TP), true negative (TN), false-positive (FP), or false negative (FN). Using those results for each pixel in the test set, we estimated the performance of the U-Net using four metrics. We used three metrics that are commonly used in classification tasks (i.e., precision, recall, and area-under-curve (AUC)) and one metric which is commonly used in image segmentation tasks (i.e., mean intersection-over-union (MIoU), also known as Jaccard index or Jaccard similarity coefficient). We computed precision = TP / (TP + FP) and recall = TP / (TP + FN) for both semantic classes together. On the other hand, we computed IoU = TP / (TP + FP + FN) for each semantic class (i.e., 0/1) and then averaged over the classes to estimate MIoU. We estimated the AUC for the receiver operating characteristic (ROC) curve using a linearly spaced set of thresholds. Note that AUC is a threshold-independent metric, unlike precision, recall, and MIoU, which are threshold-dependent metrics.

5. Performance of Color Channel Specific U-Net

Comparing the results as shown in Table 9, Table 10, Table 11 and Table 12, we can say that the U-Net is more successful at segmenting the OD and less successful at segmenting CRBVs for all channels. The U-Net performs better when all three color channels (i.e., RGB images) are used together than when the color channels are used individually. For segmenting the OD, the red and gray channels are better than the green and blue channels (see Table 9). For segmenting CRBVs the green channel performs better than other single channels, whereas both the red and blue channels perform poorly (see Table 10). For macula segmentation, there is no clear winner among gray and green channels. Although, the blue channel is a bad choice for segmenting the CRBVs, it is reasonably good at segmenting macula (see Table 11). For segmenting retinal atrophy, the green channel is better than other single channel and the blue channel is also a good choice (see Table 12).
Table 9

Performance (mean ± standard deviation) of U-Nets using different color channels for segmenting optic disc.

ColorDatasetPrecisionRecallAUCMIoU
RGBIDRiD0.897 ± 0.0180.877 ± 0.0100.940 ± 0.0050.896 ± 0.003
PALM0.859 ± 0.0090.862 ± 0.0130.933 ± 0.0060.873 ± 0.003
UoA_DR0.914 ± 0.0120.868 ± 0.0060.936 ± 0.0030.895 ± 0.004
GrayIDRiD0.868 ± 0.0200.902 ± 0.0160.952 ± 0.0070.892 ± 0.004
PALM0.758 ± 0.0200.737 ± 0.0250.870 ± 0.0110.788 ± 0.009
UoA_DR0.907 ± 0.0070.840 ± 0.0050.923 ± 0.0020.876 ± 0.008
RedIDRiD0.892 ± 0.0060.872 ± 0.0080.936 ± 0.0040.892 ± 0.004
PALM0.798 ± 0.0040.824 ± 0.0120.912 ± 0.0060.837 ± 0.003
UoA_DR0.900 ± 0.0070.854 ± 0.0060.928 ± 0.0030.885 ± 0.003
GreenIDRiD0.837 ± 0.0230.906 ± 0.0090.953 ± 0.0040.882 ± 0.008
PALM0.708 ± 0.0120.718 ± 0.0130.859 ± 0.0060.771 ± 0.004
UoA_DR0.895 ± 0.0090.821 ± 0.0100.912 ± 0.0050.869 ± 0.006
BlueIDRiD0.810 ± 0.0380.715 ± 0.0110.858 ± 0.0050.799 ± 0.010
PALM0.662 ± 0.0320.692 ± 0.0190.845 ± 0.0090.748 ± 0.008
UoA_DR0.873 ± 0.0120.800 ± 0.0090.901 ± 0.0040.851 ± 0.002
Table 10

Performance (mean ± standard deviation) of U-Nets using different color channels for segmenting CRBVs.

ColorDatasetPrecisionRecallAUCMIoU
RGBCHASE_DB10.795 ± 0.0050.638 ± 0.0040.840 ± 0.0020.696 ± 0.018
DRIVE0.851 ± 0.0070.519 ± 0.0090.781 ± 0.0040.696 ± 0.013
HRF0.730 ± 0.0170.633 ± 0.0070.838 ± 0.0050.651 ± 0.021
STARE0.822 ± 0.0090.488 ± 0.0100.766 ± 0.0060.654 ± 0.011
UoA_DR0.373 ± 0.0030.341 ± 0.0080.669 ± 0.0050.556 ± 0.004
GrayCHASE_DB10.757 ± 0.0190.635 ± 0.0160.834 ± 0.0090.648 ± 0.040
DRIVE0.864 ± 0.0140.529 ± 0.0140.786 ± 0.0080.673 ± 0.032
HRF0.721 ± 0.0320.617 ± 0.0080.825 ± 0.0050.605 ± 0.038
STARE0.810 ± 0.0210.522 ± 0.0220.784 ± 0.0110.619 ± 0.031
UoA_DR0.373 ± 0.0070.298 ± 0.0220.648 ± 0.0120.540 ± 0.009
RedCHASE_DB10.507 ± 0.0180.412 ± 0.0070.703 ± 0.0050.602 ± 0.001
DRIVE0.713 ± 0.0260.391 ± 0.0160.705 ± 0.0100.637 ± 0.005
HRF0.535 ± 0.0270.349 ± 0.0140.680 ± 0.0080.581 ± 0.004
STARE0.646 ± 0.0400.271 ± 0.0110.649 ± 0.0080.563 ± 0.005
UoA_DR0.304 ± 0.0110.254 ± 0.0120.621 ± 0.0060.539 ± 0.002
GreenCHASE_DB10.781 ± 0.0170.676 ± 0.0210.858 ± 0.0070.691 ± 0.059
DRIVE0.862 ± 0.0110.541 ± 0.0260.794 ± 0.0120.703 ± 0.047
HRF0.754 ± 0.0180.662 ± 0.0200.856 ± 0.0080.647 ± 0.077
STARE0.829 ± 0.0180.558 ± 0.0280.806 ± 0.0110.662 ± 0.052
UoA_DR0.384 ± 0.0070.326 ± 0.0230.662 ± 0.0120.552 ± 0.011
BlueCHASE_DB10.581 ± 0.0240.504 ± 0.0230.751 ± 0.0100.638 ± 0.004
DRIVE0.771 ± 0.0160.449 ± 0.0150.736 ± 0.0080.657 ± 0.007
HRF0.473 ± 0.0160.279 ± 0.0160.633 ± 0.0070.558 ± 0.004
STARE0.446 ± 0.0140.242 ± 0.0180.608 ± 0.0070.535 ± 0.003
UoA_DR0.316 ± 0.0100.271 ± 0.0150.630 ± 0.0070.540 ± 0.002
Table 11

Performance (mean ± standard deviation) of U-Nets using different color channels for segmenting macula.

ColorDatasetPrecisionRecallAUCMIoU
RGBPALM0.732 ± 0.0160.649 ± 0.0290.825 ± 0.0140.753 ± 0.009
UoA_DR0.804 ± 0.0270.713 ± 0.0430.858 ± 0.0210.794 ± 0.012
GrayPALM0.712 ± 0.0240.638 ± 0.0160.819 ± 0.0070.744 ± 0.003
UoA_DR0.811 ± 0.0170.712 ± 0.0180.858 ± 0.0080.796 ± 0.005
RedPALM0.719 ± 0.0130.648 ± 0.0150.823 ± 0.0070.749 ± 0.005
UoA_DR0.768 ± 0.0060.726 ± 0.0130.863 ± 0.0060.790 ± 0.003
GreenPALM0.685 ± 0.0200.641 ± 0.0040.820 ± 0.0020.739 ± 0.005
UoA_DR0.791 ± 0.0130.693 ± 0.0110.848 ± 0.0050.783 ± 0.005
BluePALM0.676 ± 0.0200.637 ± 0.0190.817 ± 0.0090.734 ± 0.002
UoA_DR0.801 ± 0.0350.649 ± 0.0130.826 ± 0.0060.769 ± 0.012
Table 12

Performance (mean ± standard deviation) of U-Nets using different color channels for segmenting atrophy.

ColorDatasetPrecisionRecallAUCMIoU
RGBPALM0.719 ± 0.0330.638 ± 0.0300.814 ± 0.0140.707 ± 0.019
GrayPALM0.630 ± 0.0210.571 ± 0.0250.777 ± 0.0120.658 ± 0.039
RedPALM0.514 ± 0.0100.430 ± 0.0290.705 ± 0.0130.596 ± 0.015
GreenPALM0.695 ± 0.0090.627 ± 0.0320.808 ± 0.0150.714 ± 0.011
BluePALM0.711 ± 0.0150.578 ± 0.0160.785 ± 0.0080.687 ± 0.018
To better understand the performance of U-Nets, we manually inspect all images together with their reference and predicted masks. As shown in Table 13, we see that for the majority number of cases, all color-specific U-Nets can generate at least partially accurate masks for segmenting OD and macula. When the retinal atrophy severely affects any retina, no channel-specific U-Net can generate accurate masks for segmenting OD and macula, as shown in Figure 5 and Figure 6. For many cases multiple areas in the generated masks are pointed as OD (see Figure 5d–f) and macula (see Figure 6d). As shown in Table 14, it happens more in the gray channel for the macula and in the green channel for the OD.
Table 13

Number of cases where a U-Net marks OD and macula correctly in the masks. N: Total number of fundus photographs in the test set.

Segmentation forNNumber of Cases in
RGBGrayRedGreenBlue
Optic Disc (OD)375329324316303297
Macula330270265271265267
Figure 5

Failure case of OD segmentation. (a) RGB image overlaid by reference mask for OD segmentation, (b) RGB image overlaid by inaccurately predicted OD mask, (c) Grayscale image overlaid by inaccurately predicted mask for OD segmentation, (d) Red channel image overlaid by inaccurately predicted mask for OD segmentation, (e) Green channel image overlaid by inaccurately predicted mask for OD segmentation, and (f) Blue channel image overlaid by inaccurately predicted mask for OD segmentation. Source of image: PALM/P0159.jpg.

Figure 6

Failure case of macula segmentation. (a) RGB image overlaid by reference mask for macula segmentation, (b) RGB image overlaid by inaccurate predicted macula mask, (c) Grayscale image overlaid by inaccurately predicted mask for macula segmentation, (d) Red channel image overlaid by inaccurately predicted mask for macula segmentation, (e) Green channel image overlaid by inaccurately predicted mask for macula segmentation, and (f) Blue channel image overlaid by inaccurately predicted mask for macula segmentation. Source of image: PALM/P0159.jpg.

Table 14

Number of cases where a U-Net marks multiple places as OD and macula in the masks. N: Total number of fundus photographs in the test set.

Segmentation forNNumber of Cases in
RGBGrayRedGreenBlue
Optic Disc (OD)3752926434643
Macula3301725141714
We find that our U-Nets trained for the RGB, gray, and green channel images can segment thick vessels quite well, whereas they are in general not good at segmenting thin blood vessels. As shown in Figure 7b,e, Figure 7c,f, and Figure 7h,k, discontinuity occurs in the thin vessels segmented by our U-Nets.
Figure 7

Examples of generated masks by the color-specific U-Nets for segmenting the CRBVs. The reference mask and the generated masks are shown in the first and third rows, whereas different color channels overlaid by masks are shown in the second and fourth rows. (a) the reference mask & (d) RGB fundus photograph overlaid by the reference mask, (b) generated mask by the U-Net trained by the RGB fundus photographs & (e) RGB image overlaid by the mask in (b), (c) generated mask by the U-Net trained by the grayscale fundus photographs & (f) Grayscaled image overlaid by the mask in (c), (g) generated mask by the U-Net trained by the red channel fundus photographs & (j) Red channeled fundus photograph overlaid by the mask in (g), (h) generated mask by the U-Net trained by the green channel fundus photographs & (k) Green channel image overlaid by the mask in (h), and (i) generated mask by the U-Net trained by the blue channel fundus photographs & (l) Blue channel image overlaid by the mask in (i). Source of image: CHASE_DB1/Image_14R.jpg.

The performance of U-Nets also depends to some extent on how accurately CRBVs are marked in the reference masks. Among the five data sets, the reference masks of the DRIVE data set are very accurate for both thick and thin vessels. That could be one reason we get the best performance for this data set. On the contrary, we get the worst performance for the UoA-DR data set because of the inaccurate reference masks (see Appendix F for more details). If the reference masks have inaccurate information, then the estimated performance of the U-Nets will be lower than what it should be. Two things can happen when reference masks are inaccurate. The first thing is that inaccurate reference masks in the training set may deteriorate the performance of the U-Net. However, if most reference masks are accurate enough, the deterioration may be small. The second thing is that inaccurate reference masks in the test set can generate inaccurate values for the estimated metrics. These two cases happen for the UoA-DR data set. Our U-Nets can tackle the negative effect of inaccurate reference masks in the training set of the UoA-DR. Our U-Nets learn to predict the majority of the thick vessels and some parts of thin vessels quite accurately for the UoA-DR data set. However, because of the inaccurate reference masks of the test data, the precision and recall are extremely low for all channels for the UoA-DR data set. We also notice that quite often, the red channel is affected by the overexposure, whereas the blue channel is affected by the underexposure (see Table 15). Both kinds of inappropriate exposure wash out retinal information that causes low entropy. Therefore, the generated masks for segmenting CRBVs do not have lines in the inappropriately exposed parts of a fundus photograph (see the overexposed part of the red channel in Figure 7j and the underexposed part of the blue channel in Figure 7l). Note that histograms of inappropriately exposed images are highly skewed and have low entropy (as shown in Figure 8).
Table 15

Number of inappropriately exposed fundus photographs. N: Total number RGB fundus photographs in the test set of a specific data set.

Data SetNNumber of Cases in Each Color Channel
Where |skewness|>6.0Where entropy<3.0
GrayRedGreenBlueGrayRedGreenBlue
CHASE_DB1280100130403
DRIVE40012000103
HRF4500000002
IDRiD81020600023
PALM40000140002121
STARE20020100004
UoA-DR2000002200088
Figure 8

Example of overexposed red channel and underexposed blue channel of a retinal image. First row shows different channels of a fundus photograph and second row shows their corresponding histograms. Histograms of inappropriately exposed images are highly skewed and have low entropy. Source of image: CHASE_DB1/Image_11R.jpg.

It is not surprising that using all three color channels (i.e., RGB images) as input to the U-Net performs the best since the convolutional layers of the U-Net are flexible enough to use all information from the three color channels appropriately. By using multiple filters in each convolutional layer, U-Nets can extract multiple features from the retinal images, many of which are appropriate for segmentation. As discussed in Section 3, previous works based on non-neural network-based models usually used one color channel, most likely because these models could not be benefited from the information contained in three channels. The fact that the individual color channel performs well in certain situations raises two questions regarding the camera design: Would it be worth it to develop cameras with only one color channel rather than red, green, and blue, possibly customized for retina analysis? Could a more detailed representation of the spectrum than RGB improve the automatic analysis of retinas? The RGB representation captures the information from the spectrum that the human eye can recognize. Perhaps this is not all information from the spectrum that an automatic system could have used. To fully answer those questions, many hardware developments would be needed. However, an initial analysis to address the first question could be to tune the weights used to produce the grayscale image from the RGB images.

6. Conclusions

We conduct an extensive survey to investigate which color channel in color fundus photographs is most commonly preferred for automatically diagnosing retinal diseases. We find that the green channel images dominate previous non-neural network-based works while all three color channels together, i.e., RGB images, dominate neural network-based works. In non-neural network-based works, researchers almost ignored the red and blue channels, reasoning that these channels are prone to poor contrast, noise, and inappropriate exposure. However, no works provided a conclusive experimental comparison of the performance of different color channels. In order to fill up that gap we conduct systematic experiments. We use a well-known U-shaped deep neural network (U-Net) to investigate which color channel is best for segmenting retinal atrophy and three retinal landmarks (i.e., central retinal blood vessels, optic disc, and macula). In our U-Net based segmentation approach, we see that segmentation of retinal landmarks and retinal atrophy can be conducted more accurately when RGB images are used than when a single channel is used. We also notice that as a single channel, the red channel is bad for segmenting the central retinal blood vessels, but better than other single channels for the optic disc segmentation. Although, the blue channel is a bad choice for segmenting the central retinal blood vessels, it is reasonably good for segmenting macula and very good for segmenting retinal atrophy. For all cases, RGB images perform the best which reveals the fact that the red and blue channels can provide supplementary information to the green channel. Therefore, we can conclude that all color channels are important in color fundus photographs.
Table A1

Performance of our approach of generating background masks.

Data SetPrecisionRecallAUCMIoU
DRIVE0.9970.9970.9960.995
HRF1.0001.0001.0001.000
Table A2

Distribution of provided binary and non-binary masks for segmenting CRBVs, optic discs, macula and retinal atrophy. n: total number of provided masks, m: number of provided binary masks.

Segmentation TypeCHASE_DB1DRIVEHRFIDRiDPALMSTAREUoA-DR
nmnmnmnmnmnmnm
CRBVs2804004500000400200200
Optic Disc000000810400000200200
Macula00000000000000
Retinal Atrophy0000000031100000
Table A4

Effect of different amounts of training data on the performance (mean ± standard deviation) of U-Nets trained using different color channels for segmenting CRBVs. Note that the CLAHE is applied in the pre-processing stage.

CHASE_DB1
Data Split Color Precision Recall AUC MIoU
25% Training, 20% Validation, 55% TestRGB0.569 ± 0.2030.448 ± 0.0410.729 ± 0.0590.537 ± 0.046
GRAY0.615 ± 0.0810.412 ± 0.0410.735 ± 0.0240.503 ± 0.051
RED0.230 ± 0.0300.332 ± 0.0530.613 ± 0.0100.474 ± 0.006
GREEN0.782 ± 0.0260.526 ± 0.0200.792 ± 0.0070.606 ± 0.045
BLUE0.451 ± 0.1140.370 ± 0.0180.683 ± 0.0320.485 ± 0.008
25% Training, 20% Validation, 25% TestRGB0.571 ± 0.2070.441 ± 0.0450.724 ± 0.0620.538 ± 0.048
GRAY0.624 ± 0.0800.407 ± 0.0360.731 ± 0.0230.502 ± 0.050
RED0.244 ± 0.0370.342 ± 0.0460.619 ± 0.0090.474 ± 0.007
GREEN0.791 ± 0.0260.515 ± 0.0220.787 ± 0.0080.602 ± 0.044
BLUE0.449 ± 0.1160.362 ± 0.0170.677 ± 0.0330.484 ± 0.008
55% Training, 20% Validation, 25% TestRGB0.816 ± 0.0120.541 ± 0.0240.784 ± 0.0120.684 ± 0.018
GRAY0.803 ± 0.0020.515 ± 0.0260.775 ± 0.0100.671 ± 0.016
RED0.389 ± 0.0390.363 ± 0.0270.680 ± 0.0210.504 ± 0.028
GREEN0.838 ± 0.0050.583 ± 0.0170.806 ± 0.0090.687 ± 0.038
BLUE0.648 ± 0.0190.383 ± 0.0120.698 ± 0.0060.601 ± 0.010
DRIVE
Data Split Color Precision Recall AUC MIoU
25% Training, 20% Validation, 55% TestRGB0.796 ± 0.0360.443 ± 0.0650.749 ± 0.0280.622 ± 0.072
GRAY0.835 ± 0.0160.419 ± 0.0220.739 ± 0.0090.590 ± 0.066
RED0.362 ± 0.0980.342 ± 0.0720.628 ± 0.0150.476 ± 0.007
GREEN0.846 ± 0.0100.463 ± 0.0250.758 ± 0.0090.671 ± 0.027
BLUE0.537 ± 0.0780.297 ± 0.0280.660 ± 0.0220.512 ± 0.026
25% Training, 20% Validation, 25% TestRGB0.839 ± 0.0350.442 ± 0.0680.749 ± 0.0300.626 ± 0.073
GRAY0.874 ± 0.0180.413 ± 0.0230.737 ± 0.0090.592 ± 0.068
RED0.400 ± 0.1080.352 ± 0.0730.637 ± 0.0140.476 ± 0.009
GREEN0.896 ± 0.0090.462 ± 0.0250.760 ± 0.0090.676 ± 0.028
BLUE0.575 ± 0.0800.300 ± 0.0240.663 ± 0.0200.512 ± 0.027
55% Training, 20% Validation, 25% TestRGB0.896 ± 0.0050.539 ± 0.0100.787 ± 0.0060.732 ± 0.014
GRAY0.895 ± 0.0040.528 ± 0.0120.781 ± 0.0050.731 ± 0.006
RED0.660 ± 0.0850.316 ± 0.0370.674 ± 0.0170.520 ± 0.038
GREEN0.904 ± 0.0030.533 ± 0.0080.786 ± 0.0030.718 ± 0.024
BLUE0.783 ± 0.0420.386 ± 0.0440.705 ± 0.0210.645 ± 0.037
HRF
Data Split Color Precision Recall AUC MIoU
25% Training, 20% Validation, 55% TestRGB0.792 ± 0.0060.537 ± 0.0210.799 ± 0.0130.597 ± 0.024
GRAY0.776 ± 0.0040.497 ± 0.0170.781 ± 0.0110.579 ± 0.025
RED0.204 ± 0.0240.258 ± 0.0170.591 ± 0.0140.467 ± 0.002
GREEN0.821 ± 0.0130.578 ± 0.0120.824 ± 0.0060.624 ± 0.037
BLUE0.155 ± 0.0020.361 ± 0.0100.580 ± 0.0010.482 ± 0.008
25% Training, 20% Validation, 25% TestRGB0.759 ± 0.0060.535 ± 0.0230.797 ± 0.0140.593 ± 0.023
GRAY0.741 ± 0.0050.503 ± 0.0170.782 ± 0.0110.576 ± 0.025
RED0.197 ± 0.0210.245 ± 0.0170.586 ± 0.0130.467 ± 0.002
GREEN0.794 ± 0.0160.581 ± 0.0130.824 ± 0.0060.619 ± 0.036
BLUE0.149 ± 0.0040.368 ± 0.0130.578 ± 0.0020.480 ± 0.007
55% Training, 20% Validation, 25% TestRGB0.781 ± 0.0080.608 ± 0.0050.824 ± 0.0040.693 ± 0.013
GRAY0.768 ± 0.0100.573 ± 0.0170.807 ± 0.0090.677 ± 0.022
RED0.512 ± 0.0090.271 ± 0.0210.641 ± 0.0130.536 ± 0.011
GREEN0.788 ± 0.0060.647 ± 0.0090.846 ± 0.0030.674 ± 0.060
BLUE0.274 ± 0.1100.341 ± 0.0470.620 ± 0.0320.500 ± 0.019
STARE
Data Split Color Precision Recall AUC MIoU
25% Training, 20% Validation, 55% TestRGB0.556 ± 0.2040.300 ± 0.0730.659 ± 0.0730.478 ± 0.008
GRAY0.619 ± 0.0500.283 ± 0.0580.680 ± 0.0330.478 ± 0.017
RED0.148 ± 0.0030.222 ± 0.0330.516 ± 0.0090.468 ± 0.000
GREEN0.600 ± 0.2420.351 ± 0.0300.680 ± 0.0820.483 ± 0.019
BLUE0.167 ± 0.0360.145 ± 0.0340.518 ± 0.0210.469 ± 0.001
25% Training, 20% Validation, 25% TestRGB0.531 ± 0.1950.334 ± 0.0820.672 ± 0.0820.482 ± 0.009
GRAY0.607 ± 0.0550.314 ± 0.0660.691 ± 0.0390.483 ± 0.020
RED0.143 ± 0.0030.231 ± 0.0380.512 ± 0.0110.471 ± 0.000
GREEN0.587 ± 0.2430.376 ± 0.0480.688 ± 0.0920.488 ± 0.024
BLUE0.164 ± 0.0320.142 ± 0.0390.517 ± 0.0200.472 ± 0.001
55% Training, 20% Validation, 25% TestRGB0.756 ± 0.0140.448 ± 0.0310.749 ± 0.0150.610 ± 0.038
GRAY0.748 ± 0.0100.504 ± 0.0260.770 ± 0.0100.656 ± 0.017
RED0.181 ± 0.0200.293 ± 0.0690.558 ± 0.0080.474 ± 0.006
GREEN0.749 ± 0.0130.550 ± 0.0250.795 ± 0.0120.659 ± 0.038
BLUE0.163 ± 0.0070.324 ± 0.0590.547 ± 0.0060.469 ± 0.004
UoA_DR
Data Split Color Precision Recall AUC MIoU
25% Training, 20% Validation, 55% TestRGB0.320 ± 0.0110.398 ± 0.0080.699 ± 0.0060.541 ± 0.015
GRAY0.315 ± 0.0110.353 ± 0.0160.675 ± 0.0070.526 ± 0.017
RED0.203 ± 0.0130.260 ± 0.0160.614 ± 0.0060.516 ± 0.005
GREEN0.332 ± 0.0070.415 ± 0.0180.705 ± 0.0090.534 ± 0.014
BLUE0.237 ± 0.0120.260 ± 0.0080.620 ± 0.0070.526 ± 0.006
25% Training, 20% Validation, 25% TestRGB0.313 ± 0.0110.395 ± 0.0080.697 ± 0.0050.540 ± 0.015
GRAY0.306 ± 0.0110.350 ± 0.0160.673 ± 0.0080.524 ± 0.017
RED0.201 ± 0.0130.259 ± 0.0150.614 ± 0.0060.516 ± 0.005
GREEN0.326 ± 0.0070.412 ± 0.0170.704 ± 0.0090.532 ± 0.014
BLUE0.232 ± 0.0110.257 ± 0.0070.618 ± 0.0060.524 ± 0.005
55% Training, 20% Validation, 25% TestRGB0.333 ± 0.0050.445 ± 0.0120.717 ± 0.0040.557 ± 0.007
GRAY0.330 ± 0.0030.413 ± 0.0140.700 ± 0.0060.559 ± 0.004
RED0.289 ± 0.0110.299 ± 0.0070.641 ± 0.0040.543 ± 0.003
GREEN0.335 ± 0.0020.470 ± 0.0100.728 ± 0.0040.564 ± 0.004
BLUE0.281 ± 0.0120.280 ± 0.0130.630 ± 0.0060.540 ± 0.004
Table A5

Performance (mean ± standard deviation) of U-Nets trained using different color channels for segmenting CRBVs when CLAHE is not applied on the retinal images in the pre-processing stage. Note that 55% data was used for training, whereas, 25% data is for validation and 25% data for testing.

DatabaseColorPrecisionRecallAUCMIoU
CHASEDB1RGB0.676 ± 0.0570.419 ± 0.0370.727 ± 0.0200.576 ± 0.051
GRAY0.629 ± 0.0780.406 ± 0.0520.714 ± 0.0250.570 ± 0.060
RED0.217 ± 0.0120.353 ± 0.0260.611 ± 0.0060.476 ± 0.009
GREEN0.802 ± 0.0170.530 ± 0.0190.781 ± 0.0090.672 ± 0.023
BLUE0.589 ± 0.0230.373 ± 0.0160.690 ± 0.0060.556 ± 0.050
DRIVERGB0.856 ± 0.0240.470 ± 0.0170.750 ± 0.0100.693 ± 0.011
GRAY0.855 ± 0.0210.464 ± 0.0300.746 ± 0.0150.693 ± 0.024
RED0.297 ± 0.0090.376 ± 0.0170.619 ± 0.0030.472 ± 0.010
GREEN0.886 ± 0.0060.509 ± 0.0100.771 ± 0.0050.722 ± 0.004
BLUE0.504 ± 0.1710.331 ± 0.0430.642 ± 0.0310.551 ± 0.071
HRFRGB0.757 ± 0.0140.533 ± 0.0230.784 ± 0.0100.664 ± 0.026
GRAY0.730 ± 0.0100.520 ± 0.0110.776 ± 0.0060.655 ± 0.011
RED0.164 ± 0.0020.311 ± 0.0100.577 ± 0.0010.483 ± 0.005
GREEN0.791 ± 0.0070.603 ± 0.0080.820 ± 0.0030.705 ± 0.008
BLUE0.153 ± 0.0040.347 ± 0.0220.576 ± 0.0030.476 ± 0.006
STARERGB0.579 ± 0.0770.348 ± 0.0300.696 ± 0.0200.497 ± 0.023
GRAY0.379 ± 0.1460.312 ± 0.0670.624 ± 0.0410.487 ± 0.032
RED0.157 ± 0.0040.444 ± 0.0550.558 ± 0.0100.456 ± 0.017
GREEN0.592 ± 0.0850.442 ± 0.0210.742 ± 0.0100.517 ± 0.033
BLUE0.164 ± 0.0050.327 ± 0.0560.546 ± 0.0130.474 ± 0.003
UoADRRGB0.323 ± 0.0030.411 ± 0.0040.699 ± 0.0020.555 ± 0.004
GRAY0.319 ± 0.0030.372 ± 0.0190.679 ± 0.0090.556 ± 0.005
RED0.238 ± 0.0170.220 ± 0.0140.598 ± 0.0080.522 ± 0.005
GREEN0.328 ± 0.0090.438 ± 0.0190.713 ± 0.0080.563 ± 0.004
BLUE0.262 ± 0.0120.261 ± 0.0080.619 ± 0.0040.535 ± 0.002
  161 in total

1.  Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response.

Authors:  A Hoover; V Kouznetsova; M Goldbaum
Journal:  IEEE Trans Med Imaging       Date:  2000-03       Impact factor: 10.048

2.  Automatic assessment of macular edema from color retinal images.

Authors:  K Sai Deepak; Jayanthi Sivaswamy
Journal:  IEEE Trans Med Imaging       Date:  2011-12-08       Impact factor: 10.048

3.  Measuring retinal vessel tortuosity in 10-year-old children: validation of the Computer-Assisted Image Analysis of the Retina (CAIAR) program.

Authors:  Christopher G Owen; Alicja R Rudnicka; Robert Mullen; Sarah A Barman; Dorothy Monekosso; Peter H Whincup; Jeffrey Ng; Carl Paterson
Journal:  Invest Ophthalmol Vis Sci       Date:  2009-03-25       Impact factor: 4.799

4.  Graph convolutional network based optic disc and cup segmentation on fundus images.

Authors:  Zhiqiang Tian; Yaoyue Zheng; Xiaojian Li; Shaoyi Du; Xiayu Xu
Journal:  Biomed Opt Express       Date:  2020-05-13       Impact factor: 3.732

5.  Retinal vascular tree reconstruction with anatomical realism.

Authors:  Kai-Shun Lin; Chia-Ling Tsai; Chih-Hsiangng Tsai; Michal Sofka; Shih-Jen Chen; Wei-Yang Lin
Journal:  IEEE Trans Biomed Eng       Date:  2012-08-23       Impact factor: 4.538

6.  Automated detection of fovea in fundus images based on vessel-free zone and adaptive Gaussian template.

Authors:  E-Fong Kao; Pi-Chen Lin; Ming-Chung Chou; Twei-Shiun Jaw; Gin-Chung Liu
Journal:  Comput Methods Programs Biomed       Date:  2014-08-14       Impact factor: 5.428

7.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

8.  An adaptive threshold based image processing technique for improved glaucoma detection and classification.

Authors:  Ashish Issac; M Partha Sarathi; Malay Kishore Dutta
Journal:  Comput Methods Programs Biomed       Date:  2015-08-10       Impact factor: 5.428

9.  Age-related macular degeneration and blindness due to neovascular maculopathy.

Authors:  F L Ferris; S L Fine; L Hyman
Journal:  Arch Ophthalmol       Date:  1984-11

10.  Dense U-net Based on Patch-Based Learning for Retinal Vessel Segmentation.

Authors:  Chang Wang; Zongya Zhao; Qiongqiong Ren; Yongtao Xu; Yi Yu
Journal:  Entropy (Basel)       Date:  2019-02-12       Impact factor: 2.524

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.