Literature DB >> 34110385

Classifying Retinal Degeneration in Histological Sections Using Deep Learning.

Daniel Al Mouiee1,2,3, Erik Meijering1,2, Michael Kalloniatis4, Lisa Nivison-Smith4, Richard A Williams5, David A X Nayagam5,6, Thomas C Spencer6,7, Chi D Luu8,9, Ceara McGowan6, Stephanie B Epp6, Mohit N Shivdasani1,6.   

Abstract

Purpose: Artificial intelligence (AI) techniques are increasingly being used to classify retinal diseases. In this study we investigated the ability of a convolutional neural network (CNN) in categorizing histological images into different classes of retinal degeneration.
Methods: Images were obtained from a chemically induced feline model of monocular retinal dystrophy and split into training and testing sets. The training set was graded for the level of retinal degeneration and used to train various CNN architectures. The testing set was evaluated through the best architecture and graded by six observers. Comparisons between model and observer classifications, and interobserver variability were measured. Finally, the effects of using less training images or images containing half the presentable context were investigated.
Results: The best model gave weighted-F1 scores in the range 85% to 90%. Cohen kappa scores reached up to 0.86, indicating high agreement between the model and observers. Interobserver variability was consistent with the model-observer variability in the model's ability to match predictions with the observers. Image context restriction resulted in model performance reduction by up to 6% and at least one training set size resulted in a model performance reduction of 10% compared to the original size. Conclusions: Detecting the presence and severity of up to three classes of retinal degeneration in histological data can be reliably achieved with a deep learning classifier. Translational Relevance: This work lays the foundations for future AI models which could aid in the evaluation of more intricate changes occurring in retinal degeneration, particularly in other types of clinically derived image data.

Entities:  

Mesh:

Year:  2021        PMID: 34110385      PMCID: PMC8196406          DOI: 10.1167/tvst.10.7.9

Source DB:  PubMed          Journal:  Transl Vis Sci Technol        ISSN: 2164-2591            Impact factor:   3.283


Introduction

The methodological approaches to characterizing the anatomy and physiology of the normal and diseased retina are heavily centered around techniques and technologies which can produce and analyze images of this tissue and its components with a high resolution. At a cellular level, histological techniques are widely used to visualize and characterize anatomical, neurochemical, and metabolic characteristics of the retina. The cellular complexity of the retina, however, means analyzing histological data is often very resource intensive, requiring significant time and expertise to accurately interpret images. This issue is particularly prevalent in the assessment of retinal degeneration where histology can reveal a myriad of disease changes including cell loss, functional synaptic changes, corrupted retinal circuitry, cell migration, formation of scar tissue, and even cellular reprogramming.– Detailed analyses of such changes can play an important role in assessing the preclinical therapeutic or protective effects or even the safety aspects of various treatments such as retinal prosthetics, gene therapy, or optogenetic techniques on retinal neurons, or aid in further understanding treatment mechanisms and predicting the efficacy of specific treatments. Analysis of retinal histology may also be useful in a clinical setting such as for intraocular tumors or uveitis, which requires examination of the underlying pathology to aid in accurate diagnosis., For such investigations, through vitreoretinal surgery, retinal biopsies may be performed and histological analysis may be conducted by a retinal pathologist. Thus, there is a need for more efficient ways to assess complex retinal histology images. Artificial intelligence (AI), and more specifically deep learning, may offer a solution to time consuming, subjective analysis of retinal histology. AI has already been successfully implemented in several areas of ophthalmology, such as analysis of clinical retinal images, including grading of retinal fundus photographs for diabetic retinopathy,, assessing retinal optical coherence tomography (OCT) images for age-related macular degeneration prognosis,– and detecting macular holes. With regards to histological data, AI has been extensively used to study histological changes occurring in other organs– and has also been used in one study to perform a connectomic reconstruction in the inner retina using electron microscopy images. However, to the best of our knowledge, no study has used AI to categorize features of standard hematoxylin and eosin (H&E) based retinal histology images comprising all layers. Thus, as a first step towards assessing the viability of using AI techniques for analyzing retinal histology, we aimed to develop an automated neural network to perform classification of histological retinal images into predefined stages of retinal degeneration. As the stages of retinal degeneration are based off a set of rules that are not necessarily distinct, but instead represent a continuum of neuronal changes across the degeneration stages, we first attempted to automate the classification of retinal degeneration based on predefined specialist-approved criteria, specific to retinal features clearly observable in H&E images. This included the development and optimization of a convolutional neural network (CNN) capable of classifying H&E-stained retinal sections into different retinal degeneration stages and comparing its performance to classifications made by several observers with varying degrees of expertise in retinal network identification.

Methods

Data Preparation and Preprocessing

The images used for this study were obtained from animals that were part of another study in our group, which assessed the response of visual cortex neurons to electrical stimulation of healthy versus degenerate retina., All animal procedures were approved by the Bionics Institute Animal Research Ethics Committee (Project #14 304AB), complied with the Association for Research in Vision and Ophthalmology statement for use of animals in ophthalmic research, and were in accordance with the Australian Code for the Care and Use of Animals for Scientific Purposes and with the National Institutes of Health, USA, guidelines regarding the care and use of animals for experimental procedures. Adenosine triphosphate (ATP) was used to induce photoreceptor loss followed by retinal degeneration and remodeling in one eye of four healthy adult cats., After up to 23 weeks postinjection, the animal was euthanized, transcardially perfused, and the eyes enucleated and the retina dissected. Representative retinal strips (∼2 mm wide) were dehydrated and embedded in paraffin, sectioned at 5 µm, and stained with H&E. Full details regarding the histological processing techniques have been previously published.– The length of the full retinal sections ranged between 8500 to 9500 µm. These full sections (Fig. 1) were imaged using a high-resolution slide scanner (Aperio Scanscope XT, Leica Biosystems), which used 20×/0.75 Plan Apo objective lenses, 40× scanning magnification, 0.25 µm/pixel resolution, and produced the final files in SVS (TIFF) format.
Figure 1.

An example of an H&E-stained retinal section. The region enclosed by the black dashed boundary represents the pocket in which electrode arrays were implanted in the suprachoroidal space for another study. The flanking regions enclosed by the blue dashed boundaries represent the retinal segments from which the images were sampled.

An example of an H&E-stained retinal section. The region enclosed by the black dashed boundary represents the pocket in which electrode arrays were implanted in the suprachoroidal space for another study. The flanking regions enclosed by the blue dashed boundaries represent the retinal segments from which the images were sampled. The data preprocessing procedure involved cropping the sections into images 250 µm in length and only including the retinal portion (i.e., from the retinal pigment epithelium to the nerve fiber layer). We also chose to focus only on the areas of the retinal sections that were not above the suprachoroidally implanted electrode arrays for retinal stimulation, so that any possible confounding damage to the retina caused by acute array implantation was minimized (Fig. 1). We also only selected retinal sections that were largely devoid of major histological artifacts as identified by an expert in retinal networks (MK). After applying the exclusion criteria, 70 sections were analyzed in total; 32 control sections (eight sections from each of the four control noninjected eyes) and 38 degenerated sections (8–11 sections from each of the four ATP-injected eyes). From these 70 sections, 454 cropped images were obtained (Fig. 2). The QuPath open-source software was used for the image cropping and file generation, in conjunction with ImageJ.,
Figure 2.

Examples of each degeneration stage, A) healthy, B) mild damage, C) moderate damage, D)–E) severe damage. The following annotations represent the retinal cellular layers in the segment; GCL: ganglion cell layer, INL: inner nuclear layer, ONL: outer nuclear layer, RPE: retinal pigment epithelium.

Examples of each degeneration stage, A) healthy, B) mild damage, C) moderate damage, D)–E) severe damage. The following annotations represent the retinal cellular layers in the segment; GCL: ganglion cell layer, INL: inner nuclear layer, ONL: outer nuclear layer, RPE: retinal pigment epithelium.

Degeneration Criteria

Prior to cropping sections, a few samples were first examined by the retinal networks expert (MK) who inferred that it was possible to observe damage gradation along a retinal section. The expert also asserted that distinct biological features could be derived for four different stages of degeneration in the ATP-induced eyes (Table 1). These features were determined using a combination of predefined rules and numerical measurements of some retinal components, as the expert depended on his experience and retinal knowledge to establish such criteria. A discriminative boundary was thus drawn between healthy and diseased retinae, where the healthy instances consistently showed no biological abnormalities (with the exception of some minor remaining histological artifacts such as focal retinal detachments) in any of the cellular layers and the outer segments (OS) of the photoreceptors displayed a rigid formation along the retinal section (Fig. 2a). The inner and outer nuclear layers (INL, ONL) had approximately 3 to 6 and 8 to 15 rows of cell nuclei, respectively. These layers were mostly in parallel alignment and the layers of nuclei within them were closely spaced together making a clearly distinct layer. The retinal pigment epithelium (RPE) was also intact. Using these characteristics, the first stage was labeled as “Healthy.” Instances of this Healthy stage were observed in images obtained from both the ATP-induced eyes and the control eyes, as ATP-induced retinal degeneration is known to be patchy and nonuniform across the retina.–
Table 1.

The Degeneration Criteria Used to Implement the 4-Class Classification System

Stage NumberStage NameBiological Features
1 Healthy Typical retinal layers observed (RPE, OS, ONL, OPL, INL, IPL, GCL), normal retinal lamination evident and the outer photoreceptors were organized. No sign of retinal damage (i.e., absence of a layer, low cell body density in nuclear layers, cell migration in plexiform layers). Some histological artifacts may be present but minimal. Some sections may be cut slightly obliquely.
2 Mild damageReduction in ONL thickness (compared to normal ONL thickness of 8–15 rows of nuclei) and/or nuclei density. Outer photoreceptor (outer and inner segment) and RPE disorganization as the previous class. Retinal lamination still observed with distinction between nuclear and plexiform layers evident.
3 Moderate damageLarge reduction in ONL thickness and/or nuclear density (over half that of the normal retina). Further degenerative alterations to the outer segments and RPE. Retinal lamination no longer preserved; signs of discontinuity between nuclear and plexiform layers.
4 Severe damageComplete loss of outer retinal layers including ONL, outer segments and RPE. No clear lamination of nuclear and plexiform layers. Evidence of cell migration.
The Degeneration Criteria Used to Implement the 4-Class Classification System The second stage was characterized to contain more disrupted features. The OS lost its straight extension feature from the ONL, in comparison to the Healthy stage. We note that the outer part of the photoreceptors points to the nodal point of the eye, and as such, their orientation will vary due to the section's orientation but also the retinal location. The direction these features extended to was not considered an essential part of grading as this was dependent on the plane of section and retinal location. The ONL also showed signs of damage, evident as a decrease in thickness and/or density of nuclei. The lamination of the retinal layers was still well preserved and the RPE showed slight disruption such as cell migration to OS and ONL, all of which constituted the second stage named “Mild Damage” (Fig. 2b). A third degeneration stage was defined when nearly all of the inner and outer photoreceptor segments were absent. The thickness of the ONL was considerably reduced compared to normal healthy tissue and the lamination of the layers was less distinct, for example, with nuclei present in the plexiform layers. This stage was labeled as “Moderate Damage” (Fig. 2c). The final stage of degeneration was characterized as a complete loss of the ONL, which led to an absence of clear lamination of the cellular layers. Mass cell migration and cell death were also visible in many cases in this stage. This last stage, labeled “Severe Damage,” included retinal areas where the INL and IPL were still observed (Fig. 2d) to retinal areas where the layering of the retina was unrecognizable (Fig. 2e).

Data Classification

Initially, a sample of 20 cropped images (5 from control eyes and 15 from ATP eyes) were given to two experts; one in retinal networks (MK) and another in retinal pathology (RW). The experts were asked to identify best sample images for the four different classes based on the criteria. This was considered an important step as we intended that all the remaining images for training and testing the AI model would be classified based on the same criteria and using the ideal example images as a guide. The 454 cropped images were split into two different sets; training and testing, which constituted 81% and 19% (369 and 85 images) of the preprocessed dataset, respectively. Once ideal examples were determined, the training set images were then independently classified by two trained graders (DA and MS) into the four degeneration classes. Both graders were given the ideal example images to use as a guide and consulted the written criteria during classification. A retinal expert would adjudicate if the graders did not reach the same classification for any given image. Out of the 369 training set images, 67 required expert adjudication. The testing set was provided to six different observers; two retinal experts, the two trained graders who classified the training set, and two ophthalmology researchers (Table 2). Each observer classified the testing images independently and were asked to base their decision for each image on the ideal examples set and degeneration criteria.
Table 2.

Observer Description

Observer NumberObserver AliasObserver Description
1DATrained grader
2MSTrained grader and ophthalmology researcher
3DNOphthalmology researcher
4RWRetinal pathology expert
5LN-SOphthalmology researcher
6MKRetinal networks expert
Observer Description

Deep Learning Approach

A custom convolutional neural network (CNN) was developed and trained for the task of classifying the images into the four degeneration classes (Fig. 3). The model takes a preprocessed retinal image as an input and performs a series of mathematical operations on the image's pixels to finally predict its degeneration class. The model is made up of several interconnected convolutional layers, where two successive layers constitute a “block,” a flattening layer that converts the convoluted image sequence into a vector that can be passed into two fully connected layers, before producing the maximum class probability indicating the model's prediction (Fig. 3).
Figure 3.

A general schematic of the convolutional neural network's architecture and the classification's end-to-end workflow. Each layer's input was equal to the previous layer's output, while its output size was equal to its assigned block's convolutional filter size (Table 3).

A general schematic of the convolutional neural network's architecture and the classification's end-to-end workflow. Each layer's input was equal to the previous layer's output, while its output size was equal to its assigned block's convolutional filter size (Table 3).
Table 3.

The 12 Different Architectures Used in the Architecture Search Experiments.

Architecture LabelNumber of Convolutional BlocksOutput Filter Size of Each Convolutional Block
Arch_012[64, 128]
Arch_023[64, 96, 128]
Arch_034[64, 96, 128, 256]
Arch_042[128, 256]
Arch_053[128, 256, 320]
Arch_064[128, 256, 320, 512]
Arch_072[256, 512]
Arch_083[256, 320, 512]
Arch_094[256, 320, 512, 1024]
Arch_102[512, 1024]
Arch_113[512, 720, 1024]
Arch_124[256, 512, 720, 1024]
A variation of hold-out was implemented to train the CNN on the training set, from which a 20% subset, namely a validation set, was chosen randomly for each training epoch. The following hyperparameters were maintained for all final experimentations: 2000 epochs for training, batch size of 128, learning rate of 1e-3, and a stochastic gradient descent optimizer. The number of convolutional blocks and the size of the filters in each layer were tested in a series of architecture search (AS) experiments (Table 3). Our AS tested different variations of the two architectural parameters, while keeping the remaining hyperparameters fixed. A total of 12 different architectures were trained and then evaluated on the unobserved testing set. Interobserver variability was also evaluated to compare with the best model's performance on the six observer testing sets. The code can be found online on GitHub (see Supplementary code). The 12 Different Architectures Used in the Architecture Search Experiments.

Model Evaluation and Data Analysis

The model's training and testing performance was evaluated using a weighted-F1 score used to measure the accuracy of our model by compensating for the imbalanced distribution of class data instances, while evaluating the balance between precision and recall (Eq. 1). To calculate an F1 score, either the observer classification or the model's prediction has to be designated as the true label. A total of five training rounds per configuration were run and a mean weighted-F1 score was calculated. The model with the highest mean weighted-F1 score across all observers was selected as the best model. True accuracy, defined as the number of true predictions (true positives and true negatives) out of all predictions, and cross entropy loss (log loss), a metric used to evaluate the model's divergence from the actual true classification, as a function of the number of epochs were then visualized for the best model. For the training set F1 score, only the training (and validation) grader labels were defined as the true labels for the classification, hence requiring one weighted-F1 score to be measured. For the six observer testing sets, two weighted-F1 scores were calculated, once when the model's prediction labels were designated to be the true labels, and another when the observer's labels were designated as such. where: : True positive instances of class i, : False positive instances of class i, : False negative instances of class i, N: Total number of samples Equation 1: Weighted-F1 equations Interobserver variability was visualized using confusion matrices, which capture the raw agreement between any two observers (including the model). The variability was quantified using Cohen's kappa coefficient, as it is standardized to produce a score in the range [0-1] and can directly evaluate interobserver agreement (Eq. 2). A substantially high level of agreement between the observers can be inferred from kappa scores in the range [0.8–1] and lower scores of [0.6–0.79] and [0–0.59] indicate a moderate and low degree of agreement, respectively. Weighted-F1 scores between the observers were also measured to reinforce the kappa measurements and further validate the degree of observer agreement. Heatmaps of the testing set image classifications were generated to demonstrate the raw predictions made by all observers and were compared to the model's prediction. where: : True positive instances of class i, : False positive instances of class i, : False negative instances of class i Equation 2: Cohen's Kappa Score equations

Results

Architecture Search

Of the 12 different architectures tested, all architectures performed consistently well and Arch_04 was found to have the highest mean weighted-F1 score of 84.4% across all observers (Fig. 4). This configuration was selected as the best model for subsequent analyses and was renamed RDP-Net-Ret (Retinal Degeneration Prediction Network—Retina). For simplicity, this best model will be referred to as RDP-Net in this study. Its training and validation true accuracies converged to 1 and 0.928, respectively (Fig. 5a), and losses converged to 0.04 and 0.2, respectively (Fig. 5b). It is also interesting to note that the second and third best architectures were Arch_05 and Arch_01, both of which were among the less complex architectures considered.
Figure 4.

The mean weighted-F1 scores on the testing set for all architectures across the six observers.

Figure 5.

Training and validation A) true accuracy and B) cross entropy log loss curve for RDP-Net.

The mean weighted-F1 scores on the testing set for all architectures across the six observers. Training and validation A) true accuracy and B) cross entropy log loss curve for RDP-Net.

Interobserver Variability

Confusion matrices between RPD-Net and two of the six observers (DA and MS) who were also the training set graders, demonstrated high agreement (high diagonal values in the matrices) regardless of whether RDP-Net's grading or the observers’ grading were designated as the true labels and across all four classes of retinal degeneration. Confusion matrices for all other observers were similar (Supplementary Figures S1 and S2) and there were no systematic confusions between the observers and RDP-Net. When viewing the raw agreement between the observers and the model, we found that the model as well as nearly all the observers labeled most images the same but interobserver variability was clearly present, especially for images showing some level of damage (Fig. 7a). Observer 5 was an exception, who tended to be more conservative when classifying images into the three damage classes, as several images were considered by this observer to be that of healthy retina (Fig. 7a), but they were classified as mild damage by other observers and the model. Another interesting observation is that all predicted misclassifications by the model (except for one image) were consistently one class above or below the observer's true label (Fig. 6, Fig. 7a). The overall agreements between the model and observers, and between pairs of observers were quantified with the kappa score (Fig. 7b). When comparing RDP-Net's level of agreement with other observers, kappa scores varied between 0.59–0.86 (M = 0.7633, SD = 0.0961) while the interobserver kappa scores varied between 0.56–0.91 (M = 0.771, SD = 0.109). Both groups of kappa scores were not significantly different from each other (t-test, t(10) = −0.15, P = 0.88) (Fig. 7b). Similarly, the weighted-F1 scores between RDP-Net and the individual observers varied between 76% to 91%, and 70% to 94%, depending on which values were taken as the true labels, once again demonstrating equivalence between model-observer variability and interobserver variability (Fig. 7c).
Figure 7.

Interobserver variability plots: A) The agreement heatmap between all observers and RDP-Net for the 85 testing set images, where each column represents a single image and a row represents an observer's classification for all 85 images. B) The Cohen kappa scores between all observers and RDP-Net. C) The weighted-F1 scores between the observers and RDP-Net, with the columns representing the true labels.

Figure 6.

The confusion matrices which measured the agreement between; A) Observer 1's (DA) predictions and RDP-Net's labels, B) Observer 2's (MS) predictions and RDP-Net's labels, C) RDP-Net's predictions and Observer 1's true labels, D) RDP-Net's predictions and Observer 2's true labels.

The confusion matrices which measured the agreement between; A) Observer 1's (DA) predictions and RDP-Net's labels, B) Observer 2's (MS) predictions and RDP-Net's labels, C) RDP-Net's predictions and Observer 1's true labels, D) RDP-Net's predictions and Observer 2's true labels. Interobserver variability plots: A) The agreement heatmap between all observers and RDP-Net for the 85 testing set images, where each column represents a single image and a row represents an observer's classification for all 85 images. B) The Cohen kappa scores between all observers and RDP-Net. C) The weighted-F1 scores between the observers and RDP-Net, with the columns representing the true labels.

Data Experimentation

To observe changes in RDP-Net's performance as the input data were changed, we undertook a series of experiments in which several data parameters were varied, and the model performance was re-evaluated. The first of these experiments involved retraining the model on the same training images when they were halved to 125 µm in width, with the number of these images doubling as a result, to evaluate RDP-Net's robustness when dealing with less retinal context. Once retrained, the model agreement against the original observer testing set classifications was recalculated. RDP-Net showed no significant difference in kappa scores when trained on the original training set or the halved training set (t(15) = −1.3, P = 0.21) (Fig. 7b vs Fig. 8a). A heatmap and weighted-F1 scores for this experiment (similar to Figs. 7a and 7c) displayed similar results to the overall kappa scores (Supplementary Fig. S3).
Figure 8.

(A) The Cohen kappa scores between all observers and RDP-Net when trained on a new training set with half the width of the original images. B) The mean weighted-F1 scores when RDP-Net was trained on different sizes of the original training set. An asterisk (*) indicates significant difference between the two associated groups (Tukey test, P < 0.05). The error bars represent standard errors.

(A) The Cohen kappa scores between all observers and RDP-Net when trained on a new training set with half the width of the original images. B) The mean weighted-F1 scores when RDP-Net was trained on different sizes of the original training set. An asterisk (*) indicates significant difference between the two associated groups (Tukey test, P < 0.05). The error bars represent standard errors. The second experiment involved keeping the cropped images for training as full width but reducing the training set size by randomly eliminating images from the training set, to determine how much training played a role in RDP-Net's performance. This reduction was done in 10% steps down to a reduction of 30% (i.e., the model was trained on sizes of 90%, 80%, and 70% of the original training set). The reduction was limited to 30% as beyond this value, further reduction would require modification of certain hyperparameters, which would change the model architecture. Once retrained, the model agreement against the original observer classifications was recalculated. A gradual decline in agreement performance was observed across all observers as the training set size was reduced. A one-way ANOVA comparing the mean performance across the six observers for the four training set sizes, showed that performance was significantly different for at least one pair of different training set sizes (F3,20 = 7.28, P = 0.002, ηp2 = 0.52). Tukey post hoc tests revealed that there were two training size pairs (100%, 70%) and (90%, 70%) which resulted in no significant performance difference (Fig. 8b), while other pairs showed no significance.

Discussion

The present study showed that a convolutional neural network can accurately and reliably classify histologically stained retinal images into separate predefined degeneration classes. This, to our knowledge, is the first attempt to use AI for such a task. RDP-Net's training and validation metrics indicated that it reached its maximum potential using the chosen architectural parameters, with no signs of under- or overfitting. RDP-Net was highly accurate with no significant differences in interobserver versus RDP-Net and observers’ kappa scores. RDP-Net also demonstrated robustness with similar levels of agreement to observers when trained on images with less context. However, these results were not replicated when the number of training images was reduced, suggesting RDP-Net's performance is more dependent on the quantity, rather than the context of the training set.

Model Complexity Versus Performance

Certain architectures, like CNNs, are known to be effective in image classification tasks, as their highly tunable complexity offers unprecedented access for machines to learn without the hassle of feature engineering. Although this is highly advantageous for medical imaging tasks, which require vast expert resources to process the relevant features, its current multitude indicates that the search space for an optimal model capable of performing expert-level classification is exponentially large. Additionally, many architectural parameters, which control how a neural network may perform, further exacerbate this search space. Our three best performing configurations (Arch_04, Arch_05, and Arch_01) showed that (1) the top performing models were not the most complex architectures tested, (2) the best performing model, Arch_04 or RDP-Net, was the second simplest architecture out of the top three performing models and (3) the search space's local maximum, which was derived from the 12 tested architectures, could potentially be a global maximum for this specific classification task (Fig. 4). When compared to more prominent and well-known architectures such as ResNET and Inception used in image classification with nonretinal histological– and nonhistological images,,, RDP-Net was far less complex (two layers as opposed to 18–27 layer architectures), hinting at the notion that not only are the visual features within histological data easily categorized by an algorithm, but the well-segmented structure of the retina and its easily distinguishable features when stained with H&E, could have played a large role in reducing the complexity of RDP-Net. This prompts the need for further investigation into deep learning parameter optimization for models such as RDP-Net coupled with similar data to prove these hypotheses.

Model and Observer Classifications Reflect a Continuum of Retinal Degeneration

Another interesting observation from our results is that all predicted misclassifications by the model (except for one image) were consistently one class above or below the observer's true label (Fig. 6, Fig. 7b). In addition, certain images for which the observers classified them differently, when asked about their classification attempt, responded that these images contained one or more features from two adjacent classes but they in the end had to choose one class. The reasons for these apparent nonagreements suggest that retinal degeneration stages are not necessarily discrete. One can think of these contentious images to be representatives of degenerative gradients in which the retina may be transitioning into a higher degenerative stage and therefore could be considered as belonging to hidden stages that exist between the present defined stages. It is possible that further degeneration stages could be defined to fine grain the boundaries between the existing stages, and guide future studies to solidify universally accepted degeneration stages, but ultimately there may always be a continuum of degeneration rather than precisely discrete and independent stages. Another possible way to get around this issue is by tweaking the surrounding context of input and thus directly influencing training of the model, resulting in a more well-defined and sensible representation of the different retinal degeneration stages. Indeed, reducing the number of neighboring features in a single data instance does not significantly affect the model's performance, and thus, this can then be extended to an optimization problem where the search space of these different data parameters can be fine-tuned collectively and produce an even more powerful automated tool.

Further Applicability of RDP-Net

Our results demonstrate that with the current architecture of RDP-Net, a scenario in which a robust digital copy of an observer can be developed and implemented in practice is highly probable since the model performed best when compared against the two observers responsible for grading the training set. Future iterations of the model could be generalized enough so it conforms to another observer's expectations, regardless of that observer's involvement in the degeneration criteria design and training set classification. This would not only result in significant time and resource savings when analyzing large datasets but could also aid in standardization of criteria used to assess retinal degeneration across individual researchers. Secondly, while the model in this study was applied to H&E images, the current version of the model could be adapted to other histological data or adapted to quantify features such as cell density, retinal thickness, or degree of antibody staining. Its portability will need to be tested before applying it on other histological data, but it certainly has the potential to be used as a transfer learning model. Other data may illustrate different visual features more prominently, in which case it may require the model to undergo a new architecture search experiment to evaluate the performance of different configurations for the set task, using the appropriate training and testing data. Thirdly, it is possible, and indeed likely, that the personal and professional interactions between pairs of observers over more than a decade of collaboration influenced the manner in which degeneration criteria were weighted, further influencing their classifications. This is perhaps best exemplified by the relatively lower interobserver variability between observers 3 and 4 (who have worked together for over a decade), as well as 3 and 1, and 1 and 2 compared to other pairs of observers. This suggests that a tacit diffusion of shared understanding may underpin the transferal of information with the human network and future versions of the model could take advantage of this phenomenon by having the model trained on degeneration criteria that are standardized across several like-minded experts.

Model Interpretability

While RDP-Net demonstrates that it is capable of performing this classification task with reasonably high accuracy comparable to that of the observers, it is important to understand the level of explainability that can be extracted from its performance. This is currently an active area of research, due to the notorious black box nature of CNNs., Despite this, our drafted criteria and ideal example images serve as a stepping stone into our model's interpretability, as its accuracy is preserved when attempting to classify different classes of degeneration. The feature space on which RDP-Net makes its predictions intersects largely, if not completely, with the proposed criteria features, as our results demonstrate. This is important as previous attempts to incorporate AI into medicine lacked this level of transparency when certain models would completely misinterpret data. However, the experiments to exactly determine what these features are, are out of this study's scope. Additionally, mechanically quantifying these particular visual features, for example, the nuclear density in different retinal layers or the thickness of these layers, may further solidify the exact feature space on which the model is searching. Rather than experts giving subjective opinions when evaluating these visual features, fully quantifiable and more informative metrics may assist them, and by extension the model, to make more informed decisions about the retina's cellular condition. Hence, identifying precisely, and perhaps quantifying, such feature spaces will be highly important, as clinicians and researchers will be able to establish new ways to incorporate AI into pipelines safely with high confidence.

Limitations

This study only used one animal model of retinal degeneration to train and test RDP-Net and therefore the performance of RDP-Net for other retinal degenerations is not clear. In particular, our dataset was derived from the feline retina and so species-specific differences in retinal anatomy could affect its performance for other datasets. A large amount of evidence, however, suggests anatomical changes such as cell migration, synaptic remodeling, and cell loss, that are most likely to have been used as features by this model to perform classifications, are common to other models of photoreceptor degeneration, including in humans. In addition, the inherent patchy nature of degeneration present in our chosen degeneration model provided the advantage of being able to encounter multiple degeneration stages within the same retinal sample. Future studies, however, should utilize datasets from different models of retinal degeneration preferably with several different etiologies to confirm the accuracy and robustness of the model. Computationally, the experimental search space could also be increased to explore more architectures and increase the accuracy of our model.

Conclusions

This study demonstrates proof of concept of a novel attempt to classify retinal degeneration on a cellular level using AI. Our results indicate that the development of digital experts, that can be personalized to specific experts, while maintaining a certain degree of generalizability across other expert opinions, is realistic. Our study lays the foundation for implementing powerful automated frameworks that can alleviate resource heavy tasks such as interpretation of histological data and provide meaningful and objective insights into retinal disease.

Supplementary Data Code

Code used for architecture training/testing can be found here: https://github.com/dalmouiee/RDP-Net-Ret. Raw data (image training and testing datasets) can be made available on request.
  39 in total

1.  Functional remodeling of glutamate receptors by inner retinal neurons occurs from an early stage of retinal degeneration.

Authors:  Jacqueline Chua; Erica L Fletcher; Michael Kalloniatis
Journal:  J Comp Neurol       Date:  2009-06-10       Impact factor: 3.215

2.  Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs.

Authors:  Varun Gulshan; Lily Peng; Marc Coram; Martin C Stumpe; Derek Wu; Arunachalam Narayanaswamy; Subhashini Venugopalan; Kasumi Widner; Tom Madams; Jorge Cuadros; Ramasamy Kim; Rajiv Raman; Philip C Nelson; Jessica L Mega; Dale R Webster
Journal:  JAMA       Date:  2016-12-13       Impact factor: 56.272

3.  The role of choroidal and retinal biopsies in the diagnosis and management of atypical presentations of uveitis.

Authors:  Lyndell L Lim; Eric B Suhler; James T Rosenbaum; David J Wilson
Journal:  Trans Am Ophthalmol Soc       Date:  2005

4.  Deep learning can predict microsatellite instability directly from histology in gastrointestinal cancer.

Authors:  Jakob Nikolas Kather; Alexander T Pearson; Niels Halama; Dirk Jäger; Jeremias Krause; Sven H Loosen; Alexander Marx; Peter Boor; Frank Tacke; Ulf Peter Neumann; Heike I Grabsch; Takaki Yoshikawa; Hermann Brenner; Jenny Chang-Claude; Michael Hoffmeister; Christian Trautwein; Tom Luedde
Journal:  Nat Med       Date:  2019-06-03       Impact factor: 53.440

5.  Artificial Intelligence-Based Breast Cancer Nodal Metastasis Detection: Insights Into the Black Box for Pathologists.

Authors:  Yun Liu; Timo Kohlberger; Mohammad Norouzi; George E Dahl; Jenny L Smith; Arash Mohtashamian; Niels Olson; Lily H Peng; Jason D Hipp; Martin C Stumpe
Journal:  Arch Pathol Lab Med       Date:  2018-10-08       Impact factor: 5.534

6.  Deep Learning-Based Automated Classification of Multi-Categorical Abnormalities From Optical Coherence Tomography Images.

Authors:  Wei Lu; Yan Tong; Yue Yu; Yiqiao Xing; Changzheng Chen; Yin Shen
Journal:  Transl Vis Sci Technol       Date:  2018-12-28       Impact factor: 3.283

7.  Feedback from artificial intelligence improved the learning of junior endoscopists on histology prediction of gastric lesions.

Authors:  Thomas K L Lui; Kenneth K Y Wong; Loey L Y Mak; Elvis W P To; Vivien W M Tsui; Zijie Deng; Jiaqi Guo; Li Ni; Michael K S Cheung; Wai K Leung
Journal:  Endosc Int Open       Date:  2020-01-22

Review 8.  Neural remodeling in retinal degeneration.

Authors:  Robert E Marc; Bryan W Jones; Carl B Watt; Enrica Strettoi
Journal:  Prog Retin Eye Res       Date:  2003-09       Impact factor: 21.198

9.  Techniques for processing eyes implanted with a retinal prosthesis for localized histopathological analysis.

Authors:  David A X Nayagam; Ceara McGowan; Joel Villalobos; Richard A Williams; Cesar Salinas-LaRosa; Penny McKelvie; Irene Lo; Meri Basa; Justin Tan; Chris E Williams
Journal:  J Vis Exp       Date:  2013-08-02       Impact factor: 1.355

Review 10.  Causability and explainability of artificial intelligence in medicine.

Authors:  Andreas Holzinger; Georg Langs; Helmut Denk; Kurt Zatloukal; Heimo Müller
Journal:  Wiley Interdiscip Rev Data Min Knowl Discov       Date:  2019-04-02
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.