Literature DB >> 35953633

Stage-independent biomarkers for Alzheimer's disease from the living retina: an animal study.

Hugo Ferreira1, Pedro Serranho1,2, Pedro Guimarães1, Rita Trindade1, João Martins1,3,4,5, Paula I Moreira4,5,6,7, António Francisco Ambrósio3,4,5, Miguel Castelo-Branco1,5, Rui Bernardes8,9.   

Abstract

The early diagnosis of neurodegenerative disorders is still an open issue despite the many efforts to address this problem. In particular, Alzheimer's disease (AD) remains undiagnosed for over a decade before the first symptoms. Optical coherence tomography (OCT) is now common and widely available and has been used to image the retina of AD patients and healthy controls to search for biomarkers of neurodegeneration. However, early diagnosis tools would need to rely on images of patients in early AD stages, which are not available due to late diagnosis. To shed light on how to overcome this obstacle, we resort to 57 wild-type mice and 57 triple-transgenic mouse model of AD to train a network with mice aged 3, 4, and 8 months and classify mice at the ages of 1, 2, and 12 months. To this end, we computed fundus images from OCT data and trained a convolution neural network (CNN) to classify those into the wild-type or transgenic group. CNN performance accuracy ranged from 80 to 88% for mice out of the training group's age, raising the possibility of diagnosing AD before the first symptoms through the non-invasive imaging of the retina.
© 2022. The Author(s).

Entities:  

Mesh:

Substances:

Year:  2022        PMID: 35953633      PMCID: PMC9372147          DOI: 10.1038/s41598-022-18113-y

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

Despite the abundant number of studies and distinct diagnostic approaches followed, it is still widely accepted that even with the current state-of-the-art techniques, Alzheimer’s disease (AD) remains broadly undiagnosed for over a decade before first suspicions[1,2] because screening the entire asymptomatic population is not feasible[1]. The striking impact of this delayed diagnosis is the loss of the time window where intervention could have a substantial effect by halting or delaying the disease progression. Therefore, a fundamentally distinct approach is of paramount importance; due to the general increase in life expectancy, AD affects an increasing number of people worldwide, with substantial social and economic impact and a significant burden to patients, family members, and caregivers[2]. In the last decade, the use of artificial intelligence/machine learning (ML) systems for disease diagnosis exploded because of the advances in computing power and artificial neural networks (NN), especially on the development and use of convolution neural networks (CNN)[3,4]. These, when given the proper dataset and adequate training, can analyse and classify images. Indeed, the use of CNNs is now widespread in many fields, medicine included. Its use has been proven efficient, achieving impressive performance in disease diagnosis and identifying lesions within a wealth of medical image modalities[4-14]. Despite the many and clear advantages of NNs, typically, NNs are trained to discriminate between the examples provided but are not forced to learn features that can be applied out of the conditions of the training set. Consequently, it can only be expectable that these systems allow overcoming the lack of experts in a particular field by processing/analyzing significant amounts of data, make the diagnosis faster or cheaper, or all the above. However, a NN is typically limited by the training data (ground truth); concerning medical data, early diagnosis is out of reach. The use of the retina and animal models of disease to study AD is well-grounded because of the hurdles in studying neurodegenerative disorders in humans based on brain imaging. The retina is the only part of the central nervous system (CNS) readily available through non-invasive optical means[15] and shares its embryonic origin with the brain[16]. Thus, it has been used as a source of information on the changes unfolding in the CNS and as a surrogate for changes unfolding in the brain. Another obstacle in studying neurodegenerative disorders is the time scale in which these changes unfold. As mentioned above, AD is considered to develop for many years before possible diagnosis, with all associated consequences; unknown changes at the onset of the disease and the impossibility to intervene when changes are still minute and would have a more significant impact on disease progression. The above reasons led us to design and perform the present study in which we aim to answer two scientific questions in the context of discrimination between the retinas of wild-type and the triple-transgenic (3xTg-AD) mouse model of AD: (1) can a neural network learn features to discriminate the two groups consistently in a longitudinal period, and (2) can a neural network be trained to discriminate in time points outside the trained/learned period?

Material and methods

Ethics

This study was approved by the Animal Welfare Committee of the Coimbra Institute for Clinical and Biomedical Research (iCBR), Faculty of Medicine, University of Coimbra, and by Direção-Geral de Alimentação e Veterinária (DGAV). All procedures involving mice were conducted following the ARRIVE guidelines[17] and also as per the Association for Research in Vision and Ophthalmology statement for animal use, and in agreement with the European Community Directive Guidelines for the care and use of nonhuman animals for scientific purposes (2010/63/EU), transposed into the Portuguese law in 2013 (DL113/2013). All animals were housed in a certified facility, with a temperature-controlled environment under a 12:12 h light–dark cycle, and food was provided ad libitum. All acquisitions were performed during the light phase. OCT acquisitions were performed under light anaesthesia with 80 mg/kg ketamine and 5 mg/kg xylazine (IP), the pupils dilated with 0.5% tropicamide and 2.5% phenylephrine, and ocular analgesia was achieved with oxibuprocaine.

Data source and selection

The retinas of both eyes of wild type (WT) (C57BL6/129S) and 3xTg-AD mice were imaged using a Micron IV OCT System (Phoenix Technology Group, Pleasanton, CA, USA) at the ages of 11 (weight: 14.9 ± 2.6 g), 2 (weight: 22.4 ± 2 g), 3 (weight: 25.4 ± 2.1 g), 4 (weight: 27.2 ± 2.1 g), 8 (weight: 31.6 ± 3.4 g), and 12-months-old (weight: 33.5 ± 3.8 g). Volumes with insufficient image quality, low contrast, or leading to segmentation errors were excluded from the dataset. Also, from the initial 57 WT and 57 3xTg-AD mice, respectively, 7 and 13 deceased over the time window of the study (mice died between the ages of 1–2 (4 WT and 1 3xTg-A), 2–3 (1 3xTg-A), 3–4 (2 3xTg-A), 4–8 (2 WT and 4 3xTg-A) and 8–12 (1 WT and 5 3xTg-A)). For the remaining volumes, OCT data was segmented as disclosed in[16] to compute five mean value fundus (MVF) images (Fig. 1) corresponding to retinal layers/layer-aggregate of the neuroretina: the retinal nerve fibre layer and ganglion cell layer complex (RNFL-GCL), the inner plexiform layer (IPL), the inner nuclear layer (INL), the outer plexiform layer (OPL), and the outer nuclear layer (ONL). In MVF images[18], each pixel’s value is the average of the respective A-scan values within the boundaries of the layer of interest, therefore mimicking a fundus photograph if taken in the absence of all remaining layers.
Figure 1

MVF (mean-value fundus) images examples from two left eyes of mice two months old (top: WT, bottom: 3xTg-AD). From left to right: RNFL-GCL (retinal nerve fibre layer-ganglion cell layer) complex, IPL (inner plexiform layer), INL (inner nuclear layer), OPL (outer plexiform layer), and ONL (outer nuclear layer).

MVF (mean-value fundus) images examples from two left eyes of mice two months old (top: WT, bottom: 3xTg-AD). From left to right: RNFL-GCL (retinal nerve fibre layer-ganglion cell layer) complex, IPL (inner plexiform layer), INL (inner nuclear layer), OPL (outer plexiform layer), and ONL (outer nuclear layer). The number of acquisitions per group, eye, and age used in the present study is presented in Table 1. Altogether, these yielded a total of 1144 MVF images per layer/layer-aggregate.
Table 1

Number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age.

Mouse groupEyeAge (months)
OneTwoThreeFourEightTwelve
WTOD535048534438
OS505141524736
3xTg-ADOD445152494640
OS525254524643
Total199204195206183157
Number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age.

Pre-processing and data augmentation

Because subtle differences may exist within the dataset between the two mouse groups, e.g. average intensity, intensity distribution, etc., all MVF images were normalised to ensure the same intensity distribution for all images, thus, limiting the possibility that discrimination is based on simple intensity-based metrics. First, we applied contrast-limited adaptive histogram equalisation (CLAHE) to adjust pixel intensities towards a uniform grayscale distribution and enhance local contrast. The resulting image was then low-pass filtered (LPF) (Gaussian filter) to find broader regional differences in intensity (filter parameters: 121 × 121 pixels’ kernel and standard deviation of 25). A new image resulting from the element-wise division of CLAHE by LPF was then computed and normalized to have zero mean and unit variance. Finally, this image’s variance was adjusted by multiplying it by 8 (ensuring all images are kept within the original range) and adding 128 (8-bit images). Toward data augmentation, for each pre-processed image used for the training of the NN, two new images were created: one as the vertical flip (top–bottom flip), and another as a rotation of the original image of either 90° or 270°, being the angle chosen at random.

Transfer learning from the inception-v3 model

While CNN models require substantial datasets for effective training, our database (per layer/layer-aggregate) consists of 1144 images only, yet to be split into training, validation, and test sets. The standard approach to overcome this obstacle is through the application of transfer learning, which was followed in this study by choosing the Inception-v3 network trained on the ImageNet dataset[19] proposed by Szegedy[20]. Transfer learning was applied both via feature extraction and fine-tuning. For the present application, the Inception-v3’s classification layer was replaced by a global average pooling layer, followed by a fully-connected layer with a sigmoid activation function to classify images as belonging to the WT or 3xTg-AD mouse group. Furthermore, only the weights of the last layer were updated during the training, and the fine-tuning was extended only to the last inception module. The above procedure for transfer learning was independently repeated five times as an independent CNN was created for the RNFL-GCL complex, the IPL, INL, OPL, and ONL layers.

Implementation

Data pre-processing was implemented in MATLAB Release 2021b (The Math-Works, Inc., Natick, Massachusetts, United States). All neural network models were built, trained and verified in Python 3.7.9 using the Keras 2.7[21] framework with TensorFlow 2.7[22] as the backend. The network training was performed on a desktop, equipped with an AMD Ryzen 9 3900 × CPU @3.8 GHz with 12 cores, 64 GB RAM and an Nvidia RTX 3060 with 12 GB of memory, using CUDA version 11.5.

Evaluation metrics

Performance metrics need to be chosen considering the problem being addressed and the distribution of the labels in the dataset. In binary classification tasks, like the present one, accuracy, sensitivity, specificity, and F1-score are commonly used to evaluate a model’s performance and its ability as a class predictor. Consequently, these were chosen for the present study.

Training and testing

To address both questions put forward in the present study, data was split per animal to ensure no data from one animal can be present in more than one of the training, validation, and test sets. Also, and of the utmost importance, not all time points were considered for the training and validation sets. Specifically, only data from mice at the ages of 3, 4, and 8 months were considered for training and validation, leaving out data from 1, 2, and 12-month-old mice for testing purposes only. This division allows for testing the classification ability of the neural network on mice either younger or older than those used for training. Data was split into train/validation and test sets, with 80% of mice being used in the train/validation set (46 WT and 46 3xTg-AD mice) and 20% used in the test set (11 WT and 11 3xTg-AD mice). Furthermore, a similar scheme was applied to split mice into the train (75%) and validation (25%) sets. The number of eyes in the train/validation set is presented in Table 2; the test set is presented in Table 3.
Table 2

Train (T) and validation set (V): number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age.

Mouse groupEyeAge (months)
ThreeFourEight
TVTVTV
WTOD29103211269
OS2583210289
3xTg-ADOD3210299289
OS321131102910
Total118391244011137
Table 3

Test set: number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age.

Mouse groupEyeAge (months)
OneTwoThreeFourEightTwelve
WTOD10991098
OS99810107
3xTg-ADOD811101198
OS1110111179
Total383938423532
Train (T) and validation set (V): number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age. Test set: number of acquisitions (OCT volume scans) per group (WT–wild-type mice; 3xTg-AD–transgenic mice), eye (OD–right eye; OS–left eye), and age. Two independent tests were conducted, one testing only within the period considered for the training to answer the first scientific question (determine the feasibility of classification between WT and AD within the same age range), and one testing exclusively for ages outside of the training period to answer the second scientific question (determine the classification ability of the neural model outside its training age range).

Results

Performance metrics were evaluated for the two tests detailed in the previous section. Results are shown in Table 4 (mice of 3, 4 and 8 months old in the training/validation/test sets) and Table 5 (mice of 3, 4 and 8 months old in the training and validation sets and mice of 1, 2 and 12 months old in the test set).
Table 4

Performance metrics for the classification into wild-type and the triple-transgenic mouse model of Alzheimer’s disease mouse groups, using 3, 4, and 8-month-old mice for the train, validation, and test sets; for the layer/layer-aggregates: the retinal nerve fibre layer and ganglion cell layer complex (RNFL-GCL), the inner plexiform layer (IPL), the inner nuclear layer (INL), the outer plexiform layer (OPL), and the outer nuclear layer (ONL).

Retinal layerAccuracySensitivitySpecificityF1-score
RNFL-GCL0.8850.9660.8210.905
IPL0.9131.0000.8210.922
INL0.8520.9830.7140.872
OPL0.8260.8810.7680.839
ONL0.8000.9490.6430.829
Table 5

Performance metrics for the classification into wild-type and the triple-transgenic mouse model of Alzheimer’s disease mouse groups, using 3, 4, and 8-month-old mice for the train and validation sets, and younger (1 and 2-months-old), and older (12-months-old) mice for the test set; for the layer/layer-aggregates: the retinal nerve fibre layer and ganglion cell layer complex (RNFL-GCL), the inner plexiform layer (IPL), the inner nuclear layer (INL), the outer plexiform layer (OPL), and the outer nuclear layer (ONL).

Retinal layerAccuracySensitivitySpecificityF1-score
RNFL-GCL0.8810.8770.8850.885
IPL0.8350.8250.8460.839
INL0.8440.8250.8650.847
OPL0.8070.7020.9230.792
ONL0.7980.7890.8080.804
Performance metrics for the classification into wild-type and the triple-transgenic mouse model of Alzheimer’s disease mouse groups, using 3, 4, and 8-month-old mice for the train, validation, and test sets; for the layer/layer-aggregates: the retinal nerve fibre layer and ganglion cell layer complex (RNFL-GCL), the inner plexiform layer (IPL), the inner nuclear layer (INL), the outer plexiform layer (OPL), and the outer nuclear layer (ONL). Performance metrics for the classification into wild-type and the triple-transgenic mouse model of Alzheimer’s disease mouse groups, using 3, 4, and 8-month-old mice for the train and validation sets, and younger (1 and 2-months-old), and older (12-months-old) mice for the test set; for the layer/layer-aggregates: the retinal nerve fibre layer and ganglion cell layer complex (RNFL-GCL), the inner plexiform layer (IPL), the inner nuclear layer (INL), the outer plexiform layer (OPL), and the outer nuclear layer (ONL). Overall, the accuracy is high for both test sets, showing the discrimination power of these neural networks to distinguish between WT and transgenic mice. Even more so as the training group incorporates retinas of distinct ages, hence forcing the network for each layer/aggregate to learn common features across this time window. Expectedly, all metrics show a decrease in performance for the second scenario, in which the test set is outside the age range of the training set. Surprisingly, the decrease in performance can be considered modest despite the large differences in the age of the older animals in the training group (8-months-old) and testing group (12-months-old). The same applies to the younger group, despite the lower age difference. As classification errors distribute evenly between left and right eyes, in Table 6 we present the distribution of classification errors by group and age.
Table 6

Classification errors (number of errors/number of cases) per time point and mouse group (WT–wild-type mice; 3xTg-AD–transgenic mice).

Retinal layerMouse groupAge (months)SubtotalTotal
OneTwoThreeFourEightTwelve
RNFL-GCLWT

1/19

5.3%

3/18

16.7%

3/17

17.6%

4/20

20.0%

3/19

15.8%

2/15

13.3%

16/108

14.8%

27/224 (12.1%)
3xTg-AD

3/19

15.8%

4/21

19.0%

2/21

9.5%

0/22

0.0%

0/16

0.0%

2/17

11.8%

11/116

9.5%

IPLWT

2/19

10.5%

3/18

16.7%

5/17

29.4%

3/20

15.0%

2/19

10.5%

3/15

20.0%

18/108

16.7%

33/224 (14.7%)
3xTg-AD

11/19

57.9%

3/21

14.3%

0/21

0.0%

0/22

0.0%

0/16

0.0%

1/17

5.9%

15/116

12.9%

INLWT

4/19

21.1%

2/18

11.1%

4/17

23.5%

7/20

35.0%

5/19

26.3%

3/15

20.0%

25/108

12.0%

41/224 (16.7%)
3xTg-AD

9/19

47.4%

3/21

14.3%

0/21

0.0%

0/22

0.0%

1/16

5.9%

3/17

17.6%

16/116

13.8%

OPLWT

1/19

5.3%

2/18

11.1%

4/17

23.5

6/20

30.0%

3/19

15.8%

1/15

6.7%

17/108

15.7%

47/224 (21.0%)
3xTg-AD

15/19

78.9%

4/21

19.0%

1/21

4.8%

1/22

4.5%

5/16

31.3%

4/17

23.5%

30/116

25.9%

ONLWT

2/19

10.5%

6/18

33.3%

3/17

17.6%

10/20

50.0%

7/19

36.8%

3/15

20.0%

31/108

28.7%

51/224 (22.8%)
3xTg-AD

12/19

63.2%

3/21

14.3%

1/21

4.8%

0/22

0.0%

2/16

12.5%

2/17

11.8%

20/116

17.2%

Total

60/190

31.6%

33/195

16.9%

23/190

12.1%

31/210

14.8%

28/175

16.0%

24/160

15.0%

199/112 (17.8%)
Classification errors (number of errors/number of cases) per time point and mouse group (WT–wild-type mice; 3xTg-AD–transgenic mice). 1/19 5.3% 3/18 16.7% 3/17 17.6% 4/20 20.0% 3/19 15.8% 2/15 13.3% 16/108 14.8% 3/19 15.8% 4/21 19.0% 2/21 9.5% 0/22 0.0% 0/16 0.0% 2/17 11.8% 11/116 9.5% 2/19 10.5% 3/18 16.7% 5/17 29.4% 3/20 15.0% 2/19 10.5% 3/15 20.0% 18/108 16.7% 11/19 57.9% 3/21 14.3% 0/21 0.0% 0/22 0.0% 0/16 0.0% 1/17 5.9% 15/116 12.9% 4/19 21.1% 2/18 11.1% 4/17 23.5% 7/20 35.0% 5/19 26.3% 3/15 20.0% 25/108 12.0% 9/19 47.4% 3/21 14.3% 0/21 0.0% 0/22 0.0% 1/16 5.9% 3/17 17.6% 16/116 13.8% 1/19 5.3% 2/18 11.1% 4/17 23.5 6/20 30.0% 3/19 15.8% 1/15 6.7% 17/108 15.7% 15/19 78.9% 4/21 19.0% 1/21 4.8% 1/22 4.5% 5/16 31.3% 4/17 23.5% 30/116 25.9% 2/19 10.5% 6/18 33.3% 3/17 17.6% 10/20 50.0% 7/19 36.8% 3/15 20.0% 31/108 28.7% 12/19 63.2% 3/21 14.3% 1/21 4.8% 0/22 0.0% 2/16 12.5% 2/17 11.8% 20/116 17.2% 60/190 31.6% 33/195 16.9% 23/190 12.1% 31/210 14.8% 28/175 16.0% 24/160 15.0% Also, except for the RNFL-GCL complex, the classifications for the remaining layers present a significant portion of their errors at the age of 1-month-old. Indeed, this time point alone is responsible for 30% of the classification errors. Overall, the innermost layers of the retina present fewer errors, with these steadily increasing from the RNFL-GCL complex to the ONL. Unsurprisingly, percentage-wise, fewer errors were found for the time points on which NNs were trained. On the other hand, most errors were found for the first time point, at the age of 1-month-old, while the error for the last time point is within the range of those for the three to 8-months-old. Indeed, the error at the 2-month-old time point is higher than that of 12-months-old despite the latter being further away from the training ages.

Discussion

In this work, we used neural networks to classify computed fundus images from the retina of WT and 3xTg-AD mice from OCT data. This study did not intend to make associations with cell and molecular biology markers. Indeed, this mouse model of AD is well studied and documented in the literature from multiple viewpoints and assessed by numerous techniques analysing brain and retina functional, biochemical, biological and structural parameters[16,23-25]. We kept all animals for the entire project timeframe in the present study instead of sacrificing some along the study period. This approach allowed us to perform the analysis herein that, otherwise, would require a much larger number of animals. A single model of the AD was considered as adding more would have rendered this study not feasible because of the number of animals involved. Nevertheless, this proof-of-concept can now be tested in other animal models of disease. Instead of using the traditional approach of training and classifying cases of the same age or disease stage (a cross-sectional study), we aimed to perform this study across multiple ages to learn if a neural network can learn consistent features from cases at different stages of the natural progression, both on the ageing and disease progression. This is of crucial importance for AD since, in clinical practice, disease staging and duration is often unknown. Thus, understanding if a network can learn consistent features across multiple stages and ages is of relevance. From the achieved performance in the first of the two scenarios (classification of test cases within the same age range of the training and validation tests), it is clear the capacity of all five networks to learn consistent features over time and use those correctly. Despite the high sensitivity in identifying 3xTg-AD mice across all retinal layers/layer-aggregates, the decrease in performance from the inner to the outer layers is evident from both the accuracy and F1-score, but above all from the specificity, which presents a significant decrease from the INL downwards. In consequence, these results suggest that the innermost layers offer more consistent characteristics over time and that these are specific for each of the considered mouse groups in the present study. Also, these results demonstrate the validity of the approach followed in assessing retinal changes due to the presence of human genes introduced in mice. However, prior studies of our research group using this animal model of Alzheimer’s disease have shown a reduction of significant differences in key metrics considered hallmarks of AD in the brain, especially in males, e.g., neural cell death, neuroinflammation, glial activity, etc.[23,24]. Still, the results of the present study show that substantial differences in the retina are present and detectable by our classification method. Of paramount importance, though, is the demonstration of the possibility of achieving interesting classification results when classifying cases outside the training range, that is, younger and older mice than those in the training and validation sets. Indeed, layer-wise, similar results to the former scenario were found with the classification performance decreasing, in general, from the inner to the outer retinal layers. The first time point stands out from this analysis as it is responsible for 30% of the classification errors. This may be linked to the yet developing CNV (mice’s CNS increase in mass 4–15 weeks post-natal[26]), thus presenting more similar retinas, between groups, at this age, which would explain the much better performance achieved for the latter time point even though there is a significant gap in mice’ age comparing the oldest mice in both groups, 8 and 12-months-old, respectively in the training and validation groups. Interestingly, this would explain the better performance achieved in classifying mice of 12-months-old compared to the performance in classifying 2-month-old mice, despite the lower age difference to the training and validation sets. However, another possible explanation is that changes in the retina precede those in the brain, opening two possibilities. It suggests that introduced genes play a role in the neurodevelopment of the mouse model of AD, which would require revisiting prior studies and their conclusions or that the retina would be the right location to detect the early changes in the CNV associated with AD.

Conclusions

In conclusion, the present work demonstrates the possibility of training a neural network to learn common features, across all ages, of the central nervous system and using those to identify cases outside the training range. This is of paramount importance as it puts forward the hypothesis of training a neural network based on existing data from patients and control groups and applying it to identify emerging cases or shed light on the changes unfolding in the human central nervous system. Furthermore, while our present study focuses on AD, this methodology might be extendable to other neurodegenerative diseases, following the same rationale that changes in the retina can be linked to changes in the brain, as both are part of the CNS.
  17 in total

1.  Automatic glaucoma classification using color fundus images based on convolutional neural networks and transfer learning.

Authors:  Juan J Gómez-Valverde; Alfonso Antón; Gianluca Fatti; Bart Liefers; Alejandra Herranz; Andrés Santos; Clara I Sánchez; María J Ledesma-Carbayo
Journal:  Biomed Opt Express       Date:  2019-01-25       Impact factor: 3.732

2.  Ocular fundus reference images from optical coherence tomography.

Authors:  Pedro Guimarães; Pedro Rodrigues; Conceição Lobo; Sérgio Leal; João Figueira; Pedro Serranho; Rui Bernardes
Journal:  Comput Med Imaging Graph       Date:  2014-02-22       Impact factor: 4.790

Review 3.  The Retina as a Window or Mirror of the Brain Changes Detected in Alzheimer's Disease: Critical Aspects to Unravel.

Authors:  Samuel Chiquita; Ana C Rodrigues-Neves; Filipa I Baptista; Rafael Carecho; Paula I Moreira; Miguel Castelo-Branco; António F Ambrósio
Journal:  Mol Neurobiol       Date:  2019-01-05       Impact factor: 5.590

4.  Convolutional neural networks for multi-class brain disease detection using MRI images.

Authors:  Muhammed Talo; Ozal Yildirim; Ulas Baran Baloglu; Galip Aydin; U Rajendra Acharya
Journal:  Comput Med Imaging Graph       Date:  2019-10-10       Impact factor: 4.790

5.  Automatic Detection and Classification of Colorectal Polyps by Transferring Low-Level CNN Features From Nonmedical Domain.

Authors:  Ruikai Zhang; Yali Zheng; Tony Wing Chung Mak; Ruoxi Yu; Sunny H Wong; James Y W Lau; Carmen C Y Poon
Journal:  IEEE J Biomed Health Inform       Date:  2016-12-05       Impact factor: 5.772

6.  Application of convolutional neural network in the diagnosis of the invasion depth of gastric cancer based on conventional endoscopy.

Authors:  Yan Zhu; Qiu-Cheng Wang; Mei-Dong Xu; Zhen Zhang; Jing Cheng; Yun-Shi Zhong; Yi-Qun Zhang; Wei-Feng Chen; Li-Qing Yao; Ping-Hong Zhou; Quan-Lin Li
Journal:  Gastrointest Endosc       Date:  2018-11-16       Impact factor: 9.427

Review 7.  Improving bioscience research reporting: the ARRIVE guidelines for reporting animal research.

Authors:  Carol Kilkenny; William J Browne; Innes C Cuthill; Michael Emerson; Douglas G Altman
Journal:  PLoS Biol       Date:  2010-06-29       Impact factor: 8.029

8.  Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning.

Authors:  Hoo-Chang Shin; Holger R Roth; Mingchen Gao; Le Lu; Ziyue Xu; Isabella Nogues; Jianhua Yao; Daniel Mollura; Ronald M Summers
Journal:  IEEE Trans Med Imaging       Date:  2016-02-11       Impact factor: 10.048

9.  Transfer learning with deep convolutional neural network for liver steatosis assessment in ultrasound images.

Authors:  Michał Byra; Grzegorz Styczynski; Cezary Szmigielski; Piotr Kalinowski; Łukasz Michałowski; Rafał Paluszkiewicz; Bogna Ziarkiewicz-Wróblewska; Krzysztof Zieniewicz; Piotr Sobieraj; Andrzej Nowicki
Journal:  Int J Comput Assist Radiol Surg       Date:  2018-08-09       Impact factor: 2.924

10.  A Transfer Learning Approach for Malignant Prostate Lesion Detection on Multiparametric MRI.

Authors:  Quan Chen; Shiliang Hu; Peiran Long; Fang Lu; Yujie Shi; Yunpeng Li
Journal:  Technol Cancer Res Treat       Date:  2019-01-01
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.