Literature DB >> 35547359

Segmentation of Optic Disc and Cup Using Modified Recurrent Neural Network.

J Surendiran1, S Theetchenya2, P M Benson Mansingh3, G Sekar3, M Dhipa4, N Yuvaraj5, V J Arulkarthick6, C Suresh7, Arram Sriram8, K Srihari9, Assefa Alene10.   

Abstract

Glaucoma is one of the leading factors of vision loss, where the people tends to lose their vision quickly. The examination of cup-to-disc ratio is considered essential in diagnosing glaucoma. It is hence regarded that the segmentation of optic disc and cup is useful in finding the ratio. In this paper, we develop an extraction and segmentation of optic disc and cup from an input eye image using modified recurrent neural networks (mRNN). The mRNN use the combination of recurrent neural network (RNN) with fully convolutional network (FCN) that exploits the intra- and interslice contexts. The FCN extracts the contents from an input image by constructing a feature map for the intra- and interslice contexts. This is carried out to extract the relevant information, where RNN concentrates more on interslice context. The simulation is conducted to test the efficacy of the model that integrates the contextual information for optimal segmentation of optical cup and disc. The results of simulation show that the proposed method mRNN is efficient in improving the rate of segmentation than the other deep learning models like Drive, STARE, MESSIDOR, ORIGA, and DIARETDB.
Copyright © 2022 J. Surendiran et al.

Entities:  

Mesh:

Year:  2022        PMID: 35547359      PMCID: PMC9085314          DOI: 10.1155/2022/6799184

Source DB:  PubMed          Journal:  Biomed Res Int            Impact factor:   3.246


1. Introduction

Glaucoma is a long-term eye condition in which the optic nerve is gradually weakened due to increased pressure within the eye. A condition known as glaucoma affects around 60 million people globally and is the second biggest cause of blindness after cataracts. Glaucoma will affect an estimated 80 million individuals worldwide by 2020, according to current estimates [1]. Glaucoma can permanently damage the visual nerve if it is not detected early enough. Early detection of glaucoma is critical to ensuring that patients receive effective first-line medical treatment [2-4]. Finding the cup-to-disc ratio is a difficult, time-consuming, and expensive job that currently done solely by experts. Because of this, automated methods of glaucoma image detection and assessment are critical. To automatically detect the optic nerve head image, there are two methods [5]. Using image feature extraction for binary categorization of normal and abnormal circumstances is the first strategy. While clinical indications like the cup-to-disc ratio and inferior, superior, nasal, and temporal (ISNT) zones control the optic disc area in the second approach, it is less prevalent. More than a million axons exit the retina and exit the eye through the scleral canal to carry visual information to the brain, all of which are contained in one little disc called the optic disc. In glaucoma, looking at the optic disc clarifies the link between optic nerve cupping and visual field loss [6]. Disc-to-cup ratio (DCR) is a measurement that compares the disc vertical diameter to the cup vertical diameter. Different techniques have been used for optic disc (OD), optic cup (OC) (Figure 1), or optic disc with optic cup segmentation.
Figure 1

Retinal fundus image—OD and OC.

In this paper, we develop an extraction and segmentation of optic disc and cup from an input eye image using modified recurrent neural networks (mRNN). The mRNN use the combination of recurrent neural network (RNN) with fully convolutional network (FCN) that exploits the intra- and interslice contexts. The major contributions of the work involve the following: In preprocessing, we eliminate the duplication, similar data, null set, etc., and rest of other datas will be forward to next stage The proposed method FCN extracts the contents from an input image by constructing a feature map for the intra- and interslice contexts. This is carried out to extract the relevant information, where RNN concentrates more on interslice context The simulation is conducted to test the efficacy of the model that integrates the contextual information for optimal segmentation of optical cup and disc

2. Related Works

Using the Hough transform, Zhu and Rangayyan [7] devised an automatic segmentation approach for detecting the centre and radius of a circle that is close to the OD edge. This method was also utilised by Gonzalez and Woods [8] to find the OD edge. After normalising colour image components and converting them to luminance components, we thresholded the image effective region to figure out the reference intensity needed for circle selection. For the OD automated segmentation, Welfer et al. [9] have developed a new adaptive technique that relies on anatomical models of vascular organisation combined with mathematical morphology. A two-stage methodology was used: first, the OD location was determined by analysing data from the main vessel arcade, where vessels were detected to determine the foreground and background images of the green channel. The background region was then identified by employing the RMIN operator, and the optic disc edge was then detected. There is an algorithm proposed by Aquino et al. [10] that uses the localization methods to obtain an optic disc-like pixel from the OD. It utilises three distinct detecting strategies. For each approach, there is a candidate OD pixel, and a voting system is used to determine which pixel is the final classified one. The green channel was chosen because of its high level of contrast. As an initial level set for the active contours used in optic disc segmentation, Tjandrasa et al. [11] utilised the Hough transform [12]. Prior to doing the image enhancement for OD segmentation, the image is converted to grayscale. In the next stage, the blood vessels are surgically removed to make the segmentation process go more smoothly. When low pixel values are detected, a threshold is used to blur the blood vessels. The median filter is then applied to soften the edges. A Hough transform is used to find a circle that matches the OD position in the image. It was recently hypothesised that an edge detection method with a circular Hough transform and a statistically deformable model may be used to measure OD. The disc shape was modelled with the point distribution model using a number of landmarks. The image was analysed and the influence of blood vessels was reduced using a preprocessing procedure. Using a heuristic voting mechanism, the best channel was also found. OD segmentation based on peripapillary atrophy removal was proposed by Cheng et al. [12]. There were three components to the algorithm: edge filtering, constraint elliptical Hough transformation, and the final step. Starting with a region of interest and then looking for the outer boundaries of that zone, this algorithm was built from there. The noise was removed using a low pass filter, and the first derivative from each row of the ROI was then calculated as described above. The detected disc boundary was used to determine a ring region, which was then divided into four equal sections to avoid spurious segmentation between the PPA and the OD. According to Zhang et al. [13], a new OD localization method based on 1D projection has been developed. The horizontal position of OD was found using the vascular scatter degree. The brightness and edge gradients around OD were used to determine its vertical placement. Using a binary mask created by a morphological erosion technique, the retinal image region of interest could be located without resorting to further postprocessing. As a result of the preprocessing steps, a vertical window was created and moved over the vessels map to determine the degree of arterial scatter and find the OD horizontal placement using a 1D horizontal projection signal. According to Fraga et al. [14], an approach for OD segmentation was developed that included many phases. The retinal image was normalised using the retinex algorithm to reduce contrast variability and improve process dependability. Both vessel convergence analysis and the fuzzy Hough transform are employed to find the optic disc location. The optical disc was found using both methods.

3. Joint OD Segmentation

The optic disc is a critical component of the retinal fundus image (RFI) [15] as in Figure 2. The study includes OD detection as a preprocessing component and then segmentation, which is a common step in most retinopathy screening procedures.
Figure 2

Flowchart for proposed algorithm.

RNN stands for recurrent neural network, which is a deep representation of a neural network. It learns from the input sequence through nonlinear mapping and then transforms to a hidden state sequence to get the output. In RNN, the recurrent operation is accomplished by the use of an LSTM or recurrent unit. Using a recursive approach, hidden states are discovered by iterating over a recurrent module, which keeps the data in memory for a longer period of time. The process of obtaining important information from the provided input image makes it effective. After that, the recurrent unit is trained in the hidden states with LSTM to get the desired output. Both LSTM and a linear fully connected layer are found in an RNN. The RNN extracts the ROI points from a normal eye image using the ROI points as an input. As a result, the RNN receives numerous images containing ROI points as input. On the input test image, the output obtains similar ROI points, where the ROIs are mapped based on ground truth data. The ROI feature vector has 1024 components, while the RNN generates 512 items as an output. The ROI vector on an input test image is output by the LFCL layer on LSTM cells placed on top of it. Images having ROI patterns on the eyes are fed into the RNN to train it. Pupil, iris, and eyelid information are used to generate random ROI sites at the beginning of the process. After that, the patterns are repositioned to their original locations using ground truth values that provide enough background information. In the end, the ROI points are placed at the correction positions, i.e., on the upper and lower eyelids, corners, and the iris, after a number of iterations have taken place. The RNN's stability is critical because it causes the hidden layers to freeze, allowing the ROI points to be modified first and then the points in the neighbourhood. Training losses can be calculated by comparing the rebuilt ROI to the original ROI. This is calculated by adding up all of the squared errors in the eye region across all of the images. where N is the eye images, U finds the shape of eye, β is the parameter vectors (estimated), α is the ground truth vector, and λ is the factor that helps in controlling the reconstruction of subset pairs. The process flow of the proposed mRNN is given below. After the detection of the optic disc, the optical cup is segmented using the segmentation process, which is given in the following section.

4. OC Segmentation Process

A section of the OD that is white and cup-shaped is the OC region. The optic disc connects the optic nerve to the retina at the back of the eye interior, in a circular area. From the retina in the eye, the optic nerve transmits sight-related impulses to the brain via the optic nerve. The optic disc at the back of the eye contains millions of retinal nerve fibres that bundle and travel to the brain through the optic nerve. When compared to the complete optic disc, the cup at the centre of the disc is fairly small. An algorithm for segmenting the OC region (Figure 3) from the test image maps all ROI points or subset points. Iteration 2 of the algorithm includes instructions for separating out the regions of the eyes for segmentation. To exclude all noncontributing regions, we employ a background subtraction technique.
Figure 3

OC region (the layer while in the centre of eye ball—white region).

4.1. Sørensen-Dice Coefficients

Using a Dice coefficient, you may determine how similar two test images are by comparing their subset pairings. The Hausdorff distance is used to compare the distances between the similar subset pairs in the left OD and right OD image subsets A and B. OC is then segmented, when the maximum distance between pairs of subsets exceeds a threshold value. It is possible to calculate the threshold distance by comparing the distance between two identical ROI spots on two distinct healthy OC images used as a ground truth. The Hausdorff distance method measures the distance between the initial eye subset pairings in real-world images. When measuring Hausdorff distance, this distance is used to find amblyopia in a particular test image. The Sørensen-Dice coefficient is computed between the ground truth (A) and segmented OC (B), which is given below: where C is the mean of B. With regard to binary values, a and b refer to the two different groups of image pixels. This can be written as follows: The OD pairs of left eye A and right eye B is regarded as a positive one and it is then calculated as below: The sign (x) function is represented as below: ∑asign(b) ∈ 0, OC subset fails to overlap, and then C is assigned as unity upon a ∈ 0, 1. Here, the values are considered binary if C and ∑ab = 0.

5. Results and Discussions

In this section, we present the comparative results between the proposed method and existing methods in terms of various performance metrics including accuracy, F-measure, specificity, sensitivity, and mean average percentage error. The comparison is conducted in terms of various datasets that includes Drive [16], STARE [17], MESSIDOR [18], ORIGA [19], and DIARETDB0 [20]. The simulation is conducted on a high-end computing system running on a Ryzen 5 powered 4500 CPU with 16 GB RAM and 4 GB GPU. The results of training the segmentation algorithm are given in Figures 4 and 5, and the validation of segmentation is given in Figure 6.
Figure 4

The results of training the segmentation algorithm, where red denotes the proposed method, blue denotes the level set method, black denotes the FCM, and green denotes the ground truth.

Figure 5

After training the segmentation of OC is conducted with original images (a–c) and then on preprocessing (a1–c1), feature extraction (a2–c2), detection of OD (a3–c3), and segmentation of OC (a4–c4).

Figure 6

The results of testing the segmentation algorithm, where red denotes the proposed method, blue denotes the level set method, black denotes the FCM, and green denotes the ground truth.

Table 1 shows the training and testing time of segmenting the OC after the detection OD. The results show that the testing of classifier is effective since it offers a marginal difference with the testing results.
Table 1

Training and testing time of segmenting the OC.

ImagesTrainingTesting
Image 19953 s10121 s
Image 210251 s11299 s
Image 312418 s12658 s
Figure 7 shows the results of proposed mRNN segmentation algorithm against various existing methods including level set method, FCM, and ANN. The results of simulation on various datasets show that the proposed method achieves higher rate of segmentation accuracy than other methods.
Figure 7

Segmentation accuracy of OC with OD detection.

Figure 8 shows the results of sensitivity between the proposed mRNN segmentation algorithm against various existing methods including level set method, k-means, FCM, and ANN. The results of simulation on various datasets show that the proposed method achieves higher rate of sensitivity than other methods.
Figure 8

Sensitivity of OC with OD detection.

Figure 9 shows the results of specificity between the proposed mRNN segmentation algorithm against various existing methods including level set method, FCM, and ANN. The results of simulation on various datasets show that the proposed method achieves higher rate of specificity than other methods.
Figure 9

Specificity of OC with OD detection.

Figure 10 shows the results of F-measure between the proposed mRNN segmentation algorithm against various existing methods including level set method, FCM, and ANN. The results of simulation on various datasets show that the proposed method achieves higher rate of F-measure than other methods.
Figure 10

F-measure of OC with OD detection.

Figure 11 shows the results of F-measure between the proposed mRNN segmentation algorithm against various existing methods including level set method, FCM, and ANN. The results of simulation on various datasets show that the proposed method achieves higher rate of F-measure than other methods.
Figure 11

MAPE of OC with OD detection.

6. Conclusions

In this paper, we develop an extraction and segmentation of OD and OC from an input eye image using mRNN. The mRNN use the combination of RNN with FCN that exploits the intra- and interslice contexts. The FCN extracts the contents from an input image by constructing a feature map for the intra- and interslice contexts. This is carried out to extract the relevant information, where RNN concentrates more on interslice context. The simulation is conducted to test the efficacy of the model that integrates the contextual information for optimal segmentation of optical cup and disc. The results of simulation show that the proposed method is efficient in improving the rate of segmentation than the other deep learning models. In future, the proposed modelling can improvised with the several utilization of machine learning or deep learning methods.
  10 in total

1.  Automated localisation of the optic disc, fovea, and retinal blood vessels from digital colour fundus images.

Authors:  C Sinthanayothin; J F Boyce; H L Cook; T H Williamson
Journal:  Br J Ophthalmol       Date:  1999-08       Impact factor: 4.638

2.  Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response.

Authors:  A Hoover; V Kouznetsova; M Goldbaum
Journal:  IEEE Trans Med Imaging       Date:  2000-03       Impact factor: 10.048

3.  Automatic optic disc segmentation with peripapillary atrophy elimination.

Authors:  Jun Cheng; Jiang Liu; Damon Wing Kee Wong; Fengshou Yin; Carol Cheung; Mani Baskaran; Tin Aung; Tien Yin Wong
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2011

4.  Detecting the optic disc boundary in digital fundus images using morphological, edge detection, and feature extraction techniques.

Authors:  Arturo Aquino; Manuel Emilio Gegundez-Arias; Diego Marin
Journal:  IEEE Trans Med Imaging       Date:  2010-06-17       Impact factor: 10.048

5.  ORIGA(-light): an online retinal fundus image database for glaucoma analysis and research.

Authors:  Zhuo Zhang; Feng Shou Yin; Jiang Liu; Wing Kee Wong; Ngan Meng Tan; Beng Hai Lee; Jun Cheng; Tien Yin Wong
Journal:  Annu Int Conf IEEE Eng Med Biol Soc       Date:  2010

6.  Detection of the optic nerve head in fundus images of the retina using the Hough transform for circles.

Authors:  Xiaolu Zhu; Rangaraj M Rangayyan; Anna L Ells
Journal:  J Digit Imaging       Date:  2010-06       Impact factor: 4.056

7.  Segmentation of the optic disk in color eye fundus images using an adaptive morphological approach.

Authors:  Daniel Welfer; Jacob Scharcanski; Cleyson M Kitamura; Melissa M Dal Pizzol; Laura W B Ludwig; Diane Ruschel Marinho
Journal:  Comput Biol Med       Date:  2009-12-31       Impact factor: 4.589

Review 8.  Pharmacotherapy of intraocular pressure - part II. Carbonic anhydrase inhibitors, prostaglandin analogues and prostamides.

Authors:  Ciro Costagliola; Roberto dell'Omo; Mario R Romano; Michele Rinaldi; Lucia Zeppa; Francesco Parmeggiani
Journal:  Expert Opin Pharmacother       Date:  2009-12       Impact factor: 3.889

9.  The number of people with glaucoma worldwide in 2010 and 2020.

Authors:  H A Quigley; A T Broman
Journal:  Br J Ophthalmol       Date:  2006-03       Impact factor: 4.638

Review 10.  Pharmacotherapy of intraocular pressure: part I. Parasympathomimetic, sympathomimetic and sympatholytics.

Authors:  Ciro Costagliola; Roberto dell'Omo; Mario R Romano; Michele Rinaldi; Lucia Zeppa; Francesco Parmeggiani
Journal:  Expert Opin Pharmacother       Date:  2009-11       Impact factor: 3.889

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.