Literature DB >> 34027015

Transfer learning-trained convolutional neural networks identify novel MRI biomarkers of Alzheimer's disease progression.

Yi Li1, Annat Haber1, Christoph Preuss2, Cai John1, Asli Uyar1, Hongtian Stanley Yang2, Benjamin A Logsdon3, Vivek Philip2, R Krishna Murthy Karuturi1, Gregory W Carter1,2.   

Abstract

INTRODUCTION: Genome-wide association studies (GWAS) for late onset Alzheimer's disease (AD) may miss genetic variants relevant for delineating disease stages when using clinically defined case/control as a phenotype due to its loose definition and heterogeneity.
METHODS: We use a transfer learning technique to train three-dimensional convolutional neural network (CNN) models based on structural magnetic resonance imaging (MRI) from the screening stage in the Alzheimer's Disease Neuroimaging Initiative consortium to derive image features that reflect AD progression.
RESULTS: CNN-derived image phenotypes are significantly associated with fasting metabolites related to early lipid metabolic changes as well as insulin resistance and with genetic variants mapped to candidate genes enriched for amyloid beta degradation, tau phosphorylation, calcium ion binding-dependent synaptic loss, APP-regulated inflammation response, and insulin resistance. DISCUSSION: This is the first attempt to show that non-invasive MRI biomarkers are linked to AD progression characteristics, reinforcing their use in early AD diagnosis and monitoring.
© 2020 The Authors. Alzheimer's & Dementia: Diagnosis, Assessment & Disease Monitoring published by Wiley Periodicals, LLC on behalf of Alzheimer's Association.

Entities:  

Keywords:  Alzheimer's disease; convolutional neural networks; deep learning; disease progression; imaging phenotypes; machine learning; magnetic resonance imaging; transfer learning

Year:  2021        PMID: 34027015      PMCID: PMC8120261          DOI: 10.1002/dad2.12140

Source DB:  PubMed          Journal:  Alzheimers Dement (Amst)        ISSN: 2352-8729


BACKGROUND

Alzheimer's disease (AD) is a progressive neurodegenerative disorder that slowly degrades memory and cognitive functions. It is neuropathologically defined by intracellular neurofibrillary tangles and aggregated amyloid beta (Aβ) plaques, both of which can currently be estimated accurately only post mortem. The phenotype in current case/control genome‐wide association studies (GWAS) for late onset AD (LOAD) , , are based largely on clinical assessments, in which mild cognitive impairment (MCI) and AD are determined by designed memory and cognitive tests and clinical observations. These criteria fail to reflect early AD hallmark characteristics such as Aβ plaques and neurofibrillary tangles and highlight advanced AD, leaving the MCI category widely heterogeneous and poorly understood. Consequently, current GWAS for LOAD usually exclude MCI and therefore may miss critical genetic variants associated with early AD characteristics and progression. Non‐invasive brain imaging modalities such as magnetic resonance imaging (MRI) and positron emission tomography (PET) are promising tools for monitoring AD progression and its diagnosis. Imaging provides precise quantitative phenotypes, and numerous methods have been proposed for analyzing neuropathology with MRI. , , , , However, the high dimensionality of these phenotypes makes it challenging to extract concise and interpretable information. Summary measures for pre‐defined regions of interest (ROI) are suboptimal for predicting the onset and progression of AD because they are derived independently of AD status.

HIGHLIGHTS

Novel non‐invasive magnetic resonance imaging (MRI) biomarkers of Alzheimer's disease progression. Transfer learning. Deep convolutional neural networks. In this article, we make use of deep convolutional neural networks (CNN) , , , to simultaneously extract relevant features and classify patients using (structural) MRI data from the Alzheimer's Disease Neuroimaging Initiative (ADNI) consortium. Deep CNN have become the state‐of‐the‐art methods for image classification due to their ability to form translation invariant hierarchical image features. To mitigate the scarcity of images in the ADNI dataset, and accommodate the high number of model parameters that need to be learned, we adopt a transfer learning technique. , This technique uses an independent data‐trained 3D CNN model that is then fine‐tuned using our dataset of 1381 images. This greatly augments our image dataset and ensures the learned CNN model is more robust to overfitting. Because our baseline MRI images were taken 3 years prior to clinical labeling and provide holistic snapshots of brain states, our CNN‐derived image features reflect earlier and more specific AD characteristics than the memory and cognitive performance features used for assessing AD and MCI. This is supported by the significant associations we find between the CNN‐derived phenotypes and early AD‐related metabolites and genes (Figure 1). To our knowledge, this is the first attempt to link non‐invasive MRI biomarkers with AD progression characteristics.
FIGURE 1

Graphic summary of the analytical approach. AMP‐AD, Accelerating Medicines Partnership‐Alzheimer's Disease; APOE, apolipoprotein E; CNN, convolutional neural network; GWAS, genome‐wide association studies; MRI, magnetic resonance imaging

Graphic summary of the analytical approach. AMP‐AD, Accelerating Medicines Partnership‐Alzheimer's Disease; APOE, apolipoprotein E; CNN, convolutional neural network; GWAS, genome‐wide association studies; MRI, magnetic resonance imaging

METHODS

MRI and clinically labeled data from the ADNI consortium

Data for this study were obtained from the ADNI database (adni.loni.usc.edu). ADNI is a longitudinal study in which initial imaging is followed by annual reimaging. MRI images taken at the initial stage are later categorized into four major classes based on their follow‐up status: control, AD, stable MCI subjects who maintain the same disease status throughout the follow‐up period, and progressive MCI subjects who convert from MCI to AD sometime during the follow‐up period. MCI were counted as stable MCI only if they were followed‐up for at least 3 years in this study. The conversion and follow‐up timelines for the 526 MCI patients are shown in Table S1B in supporting information.

RESEARCH IN CONTEXT

Systematic review: Case/control genome‐wide association studies (GWAS) for late onset Alzheimer's disease (AD) may miss genetic variants relevant for delineating disease stages because the cases highlight advanced AD and widely heterogeneous mild cognitive impairment patients are usually excluded. More precise phenotypes for AD are in demand. Interpretation: Convolutional neural networks (CNN) trained on structural magnetic resonance imaging (MRI) and clinical labels integrated AD classification and image feature extraction in one step; transfer learning–trained CNN were more robust to overfitting, yielding more accurate image features that predict AD progression. CNN‐derived image phenotypes were significantly associated with metabolites related to early lipid metabolic changes and insulin resistance, and with genetic variants mapped to candidate genes enriched for amyloid beta degradation, tau phosphorylation, calcium ion binding‐dependent synaptic loss, APP‐regulated inflammation response, and insulin resistance. Future direction: Relating the MRI biomarkers to specific regions in original MRI images that drive the AD classification. We downloaded 817 screening images from the ADNI‐1 cohort, 104 ADNI‐GO new participants, and 624 ADNI‐2 new participants. Because AD patients rarely convert back, we included 162 Year 1 and 95 Year 2 images from ADNI‐1 patients who were diagnosed as AD at screening, 155 of which also had MRIs at screening, totaling 1802 images (Table S1A). The adding of Year 1/2 AD images to screening images was expected to help CNN more accurately recognize progression‐related image features. However, there were no duplicate subjects in the downstream metabolite and GWAS analysis. Some subjects in ADNI‐1 have two MRI scans from the same session; we kept the one in the “Scaled_2″ directory as recommended by ADNI MRI core team. We filtered out Year 1 to 2 AD MRIs with the rank of 4 or ‐1 (based on downloaded MRIMPRANK.csv) and ADNI‐GO/2 MRIs with the quality of 4 or none (based on downloaded MAYOADIRL_MRI_IMAGEQC_12_08_15.csv). When there was more than one MRI scan from the same session after filtering, we kept the one with the highest quality or the latest timepoint when qualities were equal. FreeSurfer , software (‐autorecon1 option) was applied to correct motions, normalize image intensities, and strip bone tissue, followed by manual checking of sagittal slice 101 of each MRI image, ensuring that the mean intensity of the white matter was around 110 and skull was stripped correctly (FreeSurfer suggestions). Inappropriate skull stripping was rescued by running mri_watershed with different watershed thresholds. MRIs with incorrect mean intensity of the white matter or inappropriate skull stripping after rescuing were excluded. To investigate whether co‐registration is necessary when applying CNN to brain image analysis, we used the Talairach transformation calculated in FreeSurfer (‐autorecon1) to obtain MNI305 atlas‐registered MRIs. Age, four cognitive scores (Mini‐Mental State Examination [MMSE], Clinical Dementia Rating‐Sum of Boxes [CDR‐SOB], Functional Activities Questionnaire [FAQ], Alzheimer's Disease Assessment Scale [ADAS]), and various image summary measures at screening as well as sex, education level, apolipoprotein E (APOE) genotype, and ethnicity were downloaded. The subjects used for training CNN had MRIs and all selected covariates, with only one subject missing cognitive score examination date at screening. Most of the MRIs had examination date within 2 months from that of cognitive scores (Table S1C). The distributions of age, sex, MMSE, and APOE genotype are shown in Table 1.
TABLE 1

Demographic assessment and APOE ε4 genotype distribution in ADNI and AIBL data

ADNI dataAIBL data
No. of APOE ε4 copiesNo. of APOE ε4 copies
No. of subjectsAgeMale/female012No. of subjectsAgeMale/female012MMSE
Control37374.3 ± 6182/19127491910770.8 ± 751/567630129.1 ± 1.1
AD25174.8 ± 8134/11784114537473.2 ± 829/4523371420.3 ± 5.6
sMCI42473.1 ± 8255/169246141371077.2 ± 78/254128.0 ± 1.5
pMCI23073.9 ± 7134/9677114391174.9 ± 67/416426.3 ± 1.7
P‐Value 1 0.0470.265.72 × 10–30 0.280.264.51 × 10–10 2.93 × 10–16
P‐Value 2 0.720.643.21 × 10–09 0.250.430.0330.18

Notes: Age is presented in a mean ± standard deviation format.

ADNI, sMCI, and pMCI were estimated until 3 years from screening (for CNN training); AIBL, sMCI, and pMCI were estimated until 6 years from baseline (for CNN evaluation).

P‐value1: P value of comparing AD and controls.

P‐value2: P value of comparing sMCI and pMCI.

Abbreviations: AD, Alzheimer's disease; ADNI, Alzheimer's Disease Neuroimaging Initiative; AIBL, Australian Imaging, Biomarker & Lifestyle Flagship Study of Ageing; APOE, apolipoprotein E; MMSE, Mini‐Mental State Examination; pMCI, progressive MCI; sMCI, stable MCI.

Demographic assessment and APOE ε4 genotype distribution in ADNI and AIBL data Notes: Age is presented in a mean ± standard deviation format. ADNI, sMCI, and pMCI were estimated until 3 years from screening (for CNN training); AIBL, sMCI, and pMCI were estimated until 6 years from baseline (for CNN evaluation). P‐value1: P value of comparing AD and controls. P‐value2: P value of comparing sMCI and pMCI. Abbreviations: AD, Alzheimer's disease; ADNI, Alzheimer's Disease Neuroimaging Initiative; AIBL, Australian Imaging, Biomarker & Lifestyle Flagship Study of Ageing; APOE, apolipoprotein E; MMSE, Mini‐Mental State Examination; pMCI, progressive MCI; sMCI, stable MCI.

CNN and feature formation

CNN is a type of supervised multiple‐layer neural network that adopts learnable convolutional kernels to detect hierarchical image features. Because the same kernel slides over the whole image, the detected image features are translation invariant. To reinforce this, input images are often augmented during CNN training via transformations such as multiple scaling and cropping. A loss value at the last layer of a CNN is computed in the forward pass and iteratively minimized by back‐propagating the loss to all hidden layers to update their weights based on the stochastic gradient descent rule. The pre‐trained 3D CNN model used for the transfer learning adopted ResNext101 network structure, which consists of 101 layers, and was trained using 300,000 Kinetics video clips. Only the parameters in the last few layers of ResNext101 were fine‐tuned during the training stage of our dataset. We added nodes to the second‐to‐last layer in ResNext101 structure to accommodate covariates (Figure S1 in supporting information). Our preliminary classification results showed that progressive MCI was frequently predicted as AD by the CNN model, suggesting that the CNN classification possibly reclassified patients with pending diagnosis. We therefore trained our CNN models with the target classes of controls, stable MCI, and broad AD (AD and progressive MCI). To maximize the chance of obtaining an accurate CNN model, we generated 10‐fold sample splits. The three classes of subjects were evenly divided into 10 folds in a class‐wise fashion; for each sample split, one fold was used as an independent test set, the remaining nine folds were randomly split into training and validation sets with the ratio of 9:1. One CNN model was learned on the training and validation sets in each of the 10 sample splits, and the one with the highest classification accuracy on the (independent) test set was selected as the best model for downstream analyses. The second‐to‐last layer of our CNN was the only layer that provided input for the class probabilities at the last layer, and therefore contained the features that are the most predictive of the classification. This layer yielded 2048 image features in the adopted ResNext101 structure (Figure S1). Covariates entered CNN at this layer and their effects were passed forward to compute the loss at the last layer, which was propagated backward to all the hidden layers including the second‐to‐last layer. Hence, when there are covariates in our CNN model, the extracted image features are covariate adjusted. To reduce the number of image features for downstream analyses, we applied principal component analysis (PCA) to the 2048 image features, used the broken stick model to estimate the number of PCs needed, followed by a L1‐norm regularized regression model (Lasso ) to select the most informative PCs for distinguishing stable and progressive MCIs. These PCs are hereafter referred to as CNN‐derived image phenotypes. After CNN was trained, covariates were not needed to obtain the image phenotypes, but needed for disease status predictions. We generated two sets of CNN‐derived phenotypes. For the first, we trained a CNN model with age at screening, sex, education level, MRI field strength indicator (1.5T or 3T), and ethnicity as covariates (hereafter Image CNN). For the second set, we included APOE genotype as an additional covariate, along with four APOE‐correlated cognitive scores at screening (hereafter Augmented CNN model). To evaluate the performance of our CNN‐derived image phenotypes, we correlated them with metabolites and genetic variants. We also compared them with conventional image summary measures, cognitive scores, and clinical labels. See supporting information for additional details.

AIBL MRIs as a validation dataset

We used MR images from the Australian Imaging, Biomarker & Lifestyle Flagship Study of Ageing (AIBL) to evaluate the performance of the trained CNNs. AIBL, designed similarly to ADNI, is a longitudinal study that follows‐up participants every 18 months until 6 years from screening. All subjects were assumed to be White, a subset of whose MRIs were provided at the ADNI website. We selected 207 subjects who had one MRI scan at baseline and either remained as controls/MCI/AD for at least 3 years or converted to AD during the 6‐year follow‐up. There were only 10 stable MCI and 11 progressive MCI among 207 subjects. The MRIs, 60 of which were of 1.5T magnetic strength and 147 of 3T, went through the same pre‐processing step as ADNI MRIs. Participants’ characteristics were described in Dang et al. and Ellis et al. Age, sex, APOE ε4 genotype, and MMSE are summarized in Table 1 and used here as covariates. Education level and cognitive scores of ADAS and FAQ, not available to our access, were assigned values of zero to remove their effects in our CNNs. CDR was available and used here instead of CDR‐SOB. Youden's J statistics was applied to the class probabilities of CNN to determine the predicted class category.

RESULTS

Deep 3D CNN models classify transition from MCI to AD accurately

The confusion matrices for the best CNN models on unregistered MRIs are shown in Table 2. The Augmented CNN yielded prediction accuracy of 0.992 for broad AD, 0.986 for controls, and between 0.911 (1‐year follow‐up) to 0.801 (the final visit) for progressive MCI. In comparison, the Image CNN achieved prediction accuracy of 0.913 for broad AD, 0.906 for controls, and between 0.822 (1‐year follow‐up) to 0.69 (the final visit) for progressive MCI. The Image CNN had lower power to differentiate stable MCI from healthy controls. Both models had lower error rate of predicting stable MCI as broad AD with longer follow‐up period (0.409 to 0.192 for Augmented CNN, 0.376 to 0.2 for Image CNN), implying that some of the stable MCI that were predicted by CNN as broad AD converted to AD when tracked for longer than 3 years. We also trained the two CNN models using co‐registered MRIs. As shown in Figure 2A and B, CNN performance on non‐registered and registered MRIs was not significantly different, verifying that CNN is able to learn translation‐invariant image features. However, Image CNN with the same structure had a bigger performance difference between training and test samples on registered MRI than on non‐registered MRI (Figure 2C), implying that Image CNN might be somewhat overfitting for registered MRI. This is probably because the transferred pre‐trained CNN was trained on non‐registered images. Hence, all the downstream analyses are based on non‐registered MRIs.
TABLE 2

Confusion matrix for CNN predictions

Image CNN (trained without APOE genotype and cognitive score as covariates) Augmented CNN (trained with APOE genotype and cognitive score as covariates)
Clinical labelControlStable MCIBroad ADControlStable MCIBroad AD
3Y_ctrl (373)0.895 (334)0.0107 (4)0.0938 (35)0.96 (358)0.0268 (10)0.0134 (5)
AD (482)0.0851 (41)0.00207 (1)0.913 (440)0 (0)0.0083 (4)0.992 (478)
1Y_sMCI (425)0.442 (188)0.181 (77)0.376 (160)0.231 (98)0.36 (153)0.409 (174)
2Y_sMCI (336)0.518 (174)0.226 (76)0.256 (86)0.277 (93)0.435 (146)0.289 (97)
3Y_sMCI (296)0.551 (163)0.24 (71)0.209 (62)0.297 (88)0.476 (141)0.226 (67)
4Y_sMCI (278)0.558 (155)0.241 (67)0.201 (56)0.317 (88)0.478 (133)0.205 (57)
5Y_sMCI (271)0.561 (152)0.232 (63)0.207 (56)0.325 (88)0.48 (130)0.196 (53)
final_sMCI (255)0.557 (142)0.243 (62)0.2 (51)0.318 (81)0.49 (125)0.192 (49)
1Y_pMCI (101)0.178 (18)0 (0)0.822 (83)0.0198 (2)0.0693 (7)0.911 (92)
2Y_pMCI (190)0.195 (37)0.00526 (1)0.8 (152)0.0368 (7)0.0737 (14)0.889 (169)
3Y_pMCI (230)0.209 (48)0.0087 (2)0.783 (180)0.0522 (12)0.0826 (19)0.865 (199)
4Y_pMCI (248)0.226 (56)0.0403 (10)0.734 (182)0.0484 (12)0.109 (27)0.843 (209)
5Y_pMCI (255)0.231 (59)0.0549 (14)0.714 (182)0.0471 (12)0.118 (30)0.835 (213)
final_pMCI (271)0.255 (69)0.0554 (15)0.69 (187)0.0701 (19)0.129 (35)0.801 (217)

Notes: Number of samples is given in parentheses. The fraction at each entry stands for the ratio of the number of CNN predictions belonging to the column category to the number of samples belonging to the row (clinical) category. That pMCI or AD were predicted as broad AD, and control/sMCI were predicted as non broad AD can be viewed as correct predictions in a broad sense.

Abbreviations: AD, Alzheimer's disease; ADNI, Alzheimer's Disease Neuroimaging Initiative; APOE, apolipoprotein E; CNN, convolutional neural network; pMCI, progressive MCI; sMCI, stable MCI.

FIGURE 2

Average AUC (area under the receiver operating characteristic curve) of predicting stable and progressive mild cognitive impairment (MCI) among 10 sample splits for six follow‐up periods, comparing registered and non‐registered images. A, Image CNN model on test samples (with error bar). B, Augmented CNN model on test samples (with error bar). C, Comparison of average AUC between training and test samples. CNN, convolutional neural network; pMCI, progressive mild cognitive impairment; sMCI, stable mild cognitive impairment

Confusion matrix for CNN predictions Notes: Number of samples is given in parentheses. The fraction at each entry stands for the ratio of the number of CNN predictions belonging to the column category to the number of samples belonging to the row (clinical) category. That pMCI or AD were predicted as broad AD, and control/sMCI were predicted as non broad AD can be viewed as correct predictions in a broad sense. Abbreviations: AD, Alzheimer's disease; ADNI, Alzheimer's Disease Neuroimaging Initiative; APOE, apolipoprotein E; CNN, convolutional neural network; pMCI, progressive MCI; sMCI, stable MCI. Average AUC (area under the receiver operating characteristic curve) of predicting stable and progressive mild cognitive impairment (MCI) among 10 sample splits for six follow‐up periods, comparing registered and non‐registered images. A, Image CNN model on test samples (with error bar). B, Augmented CNN model on test samples (with error bar). C, Comparison of average AUC between training and test samples. CNN, convolutional neural network; pMCI, progressive mild cognitive impairment; sMCI, stable mild cognitive impairment

Trained CNNs applied to AIBL MRIs

CNN analysis was also informative in our AIBL validation dataset, especially for Image CNN as the relevant cognitive data were not available for Augmented CNN. Area under the curve (AUC) of distinguishing controls from AD was 0.78 for Image CNN and 0.76 for Augmented CNN. AUC of differentiating stable and progressive MCI were lower—0.61 and 0.6, respectively (Table S3 in supporting information)—probably due to the smaller sample size (21 subjects total). The Augmented CNN performed worse than the Image CNN, probably because ADAS, FAQ and CDR‐SOB were not available. Potential reasons for the reduced AUC compared to ADNI analysis include (1) MRI acquisitions in ADNI and AIBL used slightly different protocols and (2) some overfitting in the trained CNN in spite of the transfer learning, as the trained CNN was fitted to a North American population whereas AIBL data were drawn from Australia.

Image phenotypes derived from 3D CNNs

Using Lasso, we selected principal components 1, 4, and 9 as informative phenotypes from Image CNN model (hereafter imageCNN.PC1, imageCNN.PC4, imageCNN.PC9), explaining 0.257, 0.035, 0.019 of the variance of the 2048 CNN‐derived image features, respectively. Only PC 2 (hereafter augmentedCNN.PC2) was selected from Augmented CNN model, explaining 0.064 of the variance of the 2048 image features. Only imageCNN.PC4 and augmentedCNN.PC2 had a high correlation coefficient of 0.72; all other pairwise correlations of PCs were below 0.3.

Image phenotypes are associated with early AD‐related metabolites

Seven metabolites were found to be significantly (P < 0.05/55 = 0.0009) associated with the four CNN‐derived phenotypes in the ADNI‐1 participants (Table S4 in supporting information), including two phosphatidylcholines (PC) metabolites (PC ae C44:4 associated with imageCNN.PC1, PC aa C32:3 with augmentedCNN.PC2), and three sphingomyelin (SM) metabolites (SM C16:1, SM C18:0, and SM C20:2, with augmentedCNN.PC2). These PC or SM metabolites were previously found to be significantly associated with cerebrospinal fluid (CSF) Aβ 1‐42 and/or CSF tau in ADNI‐1 cohort, either directly or indirectly. Significant branched‐chain amino acids included histidine (with imageCNN.PC9) and isoleucine (with augmentedCNN.PC2), both of which have been previously implicated in insulin resistance. , We need to caution the reader that the interpretation may be overstated as the metabolite association was performed only in ADNI‐1 participants while image‐derived phenotypes were obtained from ADNI‐1/GO/2 participants.

GWAS using CNN‐derived image phenotypes

The results for meta GWAS on imputed SNPs (see supporting information) using the four CNN‐derived image phenotypes are shown in Table 3 and Table S5 in supporting information. QQ plots for all phenotypes show no obvious inflation of large P‐values (> 1 × 10–4), with λGC between 0.993 and 1.01, indicating that our association analyses have accounted for population substructure well (Figure S2 in supporting information).
TABLE 3

Genome‐wide association results based on the two sets of CNN‐derived phenotypes

Lead SNPImage phenotypeChrPosition (hg19)A1A2AF 1 (A2)GMMAT score (A1)GMMAT standard error P‐valueSNP typeeQTL genes in AMP‐ADNearby genes (±15 KB)AMP‐AD logFC 2 Other associated phenotypes [43]
Image CNN model
rs11558606PC91230814668AG0.0725.2431.1535.41 × 10–6 missense COG2* 0.096edu
rs6672949 # PC4137985911CT0.361‐12.8602.6731.51 × 10–6 intergenic ZC3H12A, MEAF6*, SNIP1 ‐0.222NA
rs34707417PC9271708810TG0.2368.6051.9066.38 × 10–6 intronic DYSF NANA
rs12361440PC91174396631GA0.151‐12.192.5431.63 × 10–6 intergenic POLD3, CHRDL2 NANA
rs11062078PC4122135487CT0.26‐11.172.3973.14 × 10–6 intergenic CACNA1C, DCP1B 3 ‐0.089edu, high
rs35047PC91231163186GT0.13611.5162.5315.36 × 10–6 intergenic DDX11 TSPAN11* 0.466NA
rs12588868PC91492909309TC0.49916.1143.6319.07 × 10–6 intronic SLC24A4 NAedu
rs429498PC41956455746GA0.441‐12.8302.6641.46 × 10–6 intergenic BRSK1 NLRP13, NLRP8 NANA
rs8115712PC4204746960GA0.32811.4382.5356.43 × 10–6 intergenic RASSF2 NANA
rs8116731PC92016718784AG0.114‐11.8202.2711.96 × 10–7 intronic SNRPB2*, OTOR ‐0.132NA
rs117100735PC42018041336GT0.0634.4611.0058.97 × 10–6 intergenic OVOL2 NANA
rs35278766PC42122815262CT0.076.2901.3744.68 × 10–6 intronic NCAM2* 0.299edu, high, cog
Augmented CNN model
rs112175941PC21150558293TA0.1596.5941.4002.48 × 10–6 intergenic RPRD2, CTSS ENSA, MCL1 3 0.407edu, math, high
rs6698178 # PC2137964765AT0.3518.9541.9012.47 × 10–6 intronic EPHA10, GNL2, DNALI, MEAF6 ZC3H12A, MEAF6* ‐0.222NA
rs144916872PC2225125902GT0.15.7731.1718.28 × 10–7 intronic ADCY3 ADCY3* 0.306NA
rs8654PC2596498783AG0.407‐8.7401.9356.30 × 10–6 synonymous RIOK2 RIOK2 NANA
rs7821522PC2851650500AC0.2067.1361.5172.56 × 10–6 intronic SNTG1* ‐0.352NA
rs1426205PC21549117232CA0.1376.3051.3332.24 × 10–6 UTR CEP152, SHC4 3 ‐0.327NA
rs67805160PC21683734755GT0.1676.7301.4644.26 × 10–6 intronic CDH13* ‐0.533edu, math, high, cog
rs4984939PC216893181GA0.445‐5.3781.2139.33 × 10–6 intergenic LMF1* ‐0.347edu, math, high, cog

Allele frequency.

Log fold change of transcript abundance for AD cases versus controls in AMP‐AD RNA‐Seq data.

cog, cognitive test performance; edu, educational attainment; high, highest math achievement; math, self‐reported math ability; PC, principal component.

These two SNPs are in high LD and have P‐values < 1 × 10–5 in GWAS from the two CNN models.

Gene with differential expression in AMP‐AD RNA‐Seq data.

Abbreviations: AMP‐AD, Accelerating Medicines Partnership‐Alzheimer's Disease; CNN, convolutional neural network; SNP, single nucleotide polymorphism.

Genome‐wide association results based on the two sets of CNN‐derived phenotypes Allele frequency. Log fold change of transcript abundance for AD cases versus controls in AMP‐AD RNA‐Seq data. cog, cognitive test performance; edu, educational attainment; high, highest math achievement; math, self‐reported math ability; PC, principal component. These two SNPs are in high LD and have P‐values < 1 × 10–5 in GWAS from the two CNN models. Gene with differential expression in AMP‐AD RNA‐Seq data. Abbreviations: AMP‐AD, Accelerating Medicines Partnership‐Alzheimer's Disease; CNN, convolutional neural network; SNP, single nucleotide polymorphism. Using imageCNN.PC1 as a phenotype, we obtained genome‐wide significant P‐values (<5 × 10–8) for genetic variants at the APOE/TOMM40 locus (Figure S2A). However, its QQ plot shows no upward deviation from the diagonal line when single nucleotide polymorphisms (SNPs) at the APOE/TOMM40 locus are excluded (Figure S2B), suggesting that imageCNN.PC1 is not significantly associated with any genetic variants outside APOE given the current sample size. In contrast, using imageCNN.PC4, imageCNN.PC9, and augmentedCNN.PC2 as phenotypes revealed no significant variants at the APOE/TOMM40 locus (P < 1 × 10–5). Their QQ plots show moderate excess of low P‐values even when SNPs at the APOE/TOMM40 locus are excluded (Figure S2C‐2E), suggesting that these phenotypes are significantly associated with variants outside APOE given the current sample size. GWAS based on imageCNN.PC4 and imageCNN.PC9 identified 116 and 41 significant (P < 1 × 10–5) SNPs, respectively (Figure 3A), which were mapped to 17 protein‐coding genes (within ±15 kb) according to FUMA. Identified SLC24A4 (rs12588868, P = 9.07 × 10–6) is a known AD gene; two genes, CACNA1C (rs11062078, P = 3.14 × 10–6) and DYSF (rs34707417, P = 6.38 × 10–6), were significantly enriched in the Accelerating Medicines Partnership‐Alzheimer's Disease (AMP‐AD) gene co‐expression submodules (Figure S3 in supporting information). The Gene Ontology (GO) annotations for the enriched submodules include regulation of action potential and calcium‐mediated signaling for CACNA1C and regulation of endocytosis for DYSF (Table S7 in supporting information). DYSF has been reported to be significantly associated (P < 1 × 10–4) with AD in an exome array association study.
FIGURE 3

Manhattan plots for (A) principal components (PCs) 1, 4, and 9 of the Image CNN‐derived image features, and (B) PC 2 of the Augmented CNN‐derived image features. Gene names in red text are for imageCNN.PC1, those in blue text are for imageCNN.PC4, imageCNN.PC9, and augmentedCNN.PC2

Manhattan plots for (A) principal components (PCs) 1, 4, and 9 of the Image CNN‐derived image features, and (B) PC 2 of the Augmented CNN‐derived image features. Gene names in red text are for imageCNN.PC1, those in blue text are for imageCNN.PC4, imageCNN.PC9, and augmentedCNN.PC2 AugmentedCNN.PC2 identified 130 SNPs with P‐values < 1 × 10–5 (Figure 3B), which were mapped to 12 protein‐coding genes (within ±15 kb) according to FUMA. Two genes, CDH13 (rs67805160, P = 4.26 × 10–6) and ENSA (rs112175941, P = 2.48 × 10–6) were significantly enriched in AMP‐AD gene co‐expression submodules (Figure S3), whose GO annotations include calcium‐dependent cell–cell adhesion for CDH13, negative regulation of dephosphorylation, and regulation of hormone/insulin secretion for ENSA (Table S8 in supporting information).

Candidate genes link image phenotypes to AD‐related functions

For Image CNN, three candidate genes (SLC24A4, CACNA1C, DYSF) are related to the GO term of calcium ion binding, which may in turn play a role in synaptic plasticity. NCAM2 mediates synaptic adhesion, and Aβ‐dependent disruption of NCAM2 functions in the AD hippocampus contributes to synapse loss. BRSK1 is the eQTL target gene of significant rs429498 in the AMP‐AD RNA‐Seq data, and mediates phosphorylation of tau. Therefore, we believe that Image CNN detected image patterns that are related to calcium ion binding, Aβ‐mediated synaptic loss, and tau phosphorylation. For Augmented CNN, CDH13 negatively regulates axon growth and LMF1 is required for maturation and transport of active lipoprotein lipase (LPL). Previous studies have established that LPL is a novel Aβ‐binding protein promoting cellular uptake and subsequent degradation of Aβ. ENSA is an inhibitor of protein phosphatase 2A (PP2A) that regulates tau phosphorylation directly. ADCY3 loss‐of‐function variants increase the risk of obesity and type 2 diabetes. ZC3H12A (a.k.a. MCPIP1), detected by both CNN models, is an APP‐regulated inflammation respondent in NT2 cells. All together, we believe that augmentedCNN.PC2 represented both early (Aβ and tau related) and late (insulin resistance/diabetes and inflammation response) AD characteristics.

DISCUSSION

Augmented CNN model achieves higher prediction accuracy than Image CNN model in the ADNI cohort. The high accuracy achieved by both models, as well as the four cognitive scores and APOE genotype, for distinguishing stable and progressive MCI (Table S2 in supporting information) implies that image, cognitive performance, and genetics have complementary roles in disease status prediction. ImageCNN.PC1 is the only CNN‐derived phenotype to identify genome‐wide significant SNPs at the APOE locus because principal components are uncorrelated with one another by definition and Augmented CNN‐derived image features were APOE‐adjusted. This phenotype also had the highest power (AUC = 0.784) to predict the clinical conversion of MCI to AD among all the phenotypes we considered (Table S2), and the highest Pearson correlation with cognitive scores (0.55 with ADAS and 0.52 with CDR‐SOB). However, its QQ plot shows no excess of low P‐values outside of the APOE locus. We therefore believe that imageCNN.PC1 represents image features that are mainly redundant with APOE genotype and cognitive performance. The other three CNN‐derived phenotypes show low correlations with cognitive scores (<0.1) and have relatively low power to predict the clinical conversion of MCI to AD (Table S2). This is probably due to the conversion assessment being largely based on cognitive performance in the first place. However, these phenotypes are associated with early‐stage markers of disease. For example, augmentedCNN.PC2 is significantly associated with the largest number of lipid metabolites (three sphingomyelin and one phosphatidylcholines; Table S4), which have been previously found significantly associated with CSF Aβ 1‐42 and/or CSF tau. Furthermore, imageCNN.PC9 is significantly associated with a metabolite of histidine, which has been implicated in insulin resistance and p‐tau. , Their most significant SNPs map to protein‐coding genes that are enriched for diverse AD stages, ranging from early Aβ, tau phosphorylation, and calcium ion binding‐related synaptic loss to late energy hypo‐utilization and inflammation response. Moreover, our CNN‐derived phenotypes compare favorably to other AD‐related phenotypes (cognitive scores, image summary measures, and clinical labels) in terms of metabolite association and GWAS findings (Table S4 and S6 in supporting information). Although the CNN‐derived image phenotypes could be explained to some degree by a linear combination of image summary measures from ROIs—with the highest explaining R2 of 0.358 for imageCNN.PC1 (Figure S4A in supporting information), followed by the explaining R2 of 0.133 for augmentedCNN.PC2 (Figure S4B) the majority of the image phenotypes were unexplained by the ROIs, showing that the CNN‐derived image phenotypes provide novel MRI biomarkers. These findings suggest that our CNN‐derived image phenotypes reflect AD progression better than other common phenotypes and refine the genetic associations to key subprocesses for LOAD. Three reasons may explain this. First, the transfer learning technique greatly augments the ADNI image data, making the learned CNN models more robust to overfitting. Second, unlike our image phenotypes, case/control LOAD GWAS often exclude MCI due to their uncertain disease status and could misdiagnose healthy controls that develop AD later, yielding a less precise and less specific phenotype. Third, the categorical clinical labels are more prone to errors due to the use of thresholds than the continuous CNN‐derived image phenotypes. Although we have applied FreeSurfer to the downloaded pre‐processed MRI images to correct motion and normalize image intensities, we acknowledge that some confounding effects may not have been accounted for, due to the different MRI acquisition parameters adopted at different sites. This could also explain the lower prediction power in the independent AIBL MRIs. One direction for future studies is to explore CNN training strategies that can better tolerate inaccurate target labels; another one is to identify the regions in the original MRI images that drive the CNN classification.

AUTHOR CONTRIBUTIONS

YL and GWC designed the project. YL and CJ modified the 3D‐CNN structure, YL performed the CNN training, image‐derived phenotype construction, and downstream analysis. CP annotated the associated variants and genes, AH carried out the AMP‐AD gene expression submodule enrichment analysis, AU and HSY performed RNA‐Seq data analysis, KK and VP advised on data analysis. YL and GWC drafted the manuscript. YL, AH, and GWC revised the manuscript. GWC obtained the funding and was responsible for the study supervision. All authors read and approved the final manuscript.

DATA SHARING

Summary statistics from the GWAS based on the CNN‐derived image phenotypes will be made available for download upon publication from https://www.synapse.org/#!Synapse:syn21069604. The Python codes for CNN fine‐tuning and prediction are also available there. However, the ADNI MRI images, which are controlled data, are not allowed to accompany the Python codes. URLs ADNI: http://www.adni‐info.org ADNI database: http://adni.loni.usc.edu/ AIBL: https://aibl.csiro.au/adni/index.html AIBL hosted at ADNI: http://adni.loni.usc.edu/study‐design/collaborative‐%20studies/aibl/ AMP‐AD: https://adknowledgeportal.synapse.org FUMA: http://fuma.ctglab.nl/

CONFLICTS OF INTEREST

The authors have no conflicts of interest to disclose. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file. Supporting Information. Click here for additional data file.
  32 in total

1.  Reducing the dimensionality of data with neural networks.

Authors:  G E Hinton; R R Salakhutdinov
Journal:  Science       Date:  2006-07-28       Impact factor: 47.728

2.  Greatwall phosphorylates an inhibitor of protein phosphatase 2A that is essential for mitosis.

Authors:  Satoru Mochida; Sarah L Maslen; Mark Skehel; Tim Hunt
Journal:  Science       Date:  2010-12-17       Impact factor: 47.728

3.  Deep learning based imaging data completion for improved brain disease diagnosis.

Authors:  Rongjian Li; Wenlu Zhang; Heung-Il Suk; Li Wang; Jiang Li; Dinggang Shen; Shuiwang Ji
Journal:  Med Image Comput Comput Assist Interv       Date:  2014

4.  Histidine supplementation improves insulin resistance through suppressed inflammation in obese women with the metabolic syndrome: a randomised controlled trial.

Authors:  R N Feng; Y C Niu; X W Sun; Q Li; C Zhao; C Wang; F C Guo; C H Sun; Y Li
Journal:  Diabetologia       Date:  2013-01-30       Impact factor: 10.122

5.  Whole genome association study of brain-wide imaging phenotypes for identifying quantitative trait loci in MCI and AD: A study of the ADNI cohort.

Authors:  Li Shen; Sungeun Kim; Shannon L Risacher; Kwangsik Nho; Shanker Swaminathan; John D West; Tatiana Foroud; Nathan Pankratz; Jason H Moore; Chantel D Sloan; Matthew J Huentelman; David W Craig; Bryan M Dechairo; Steven G Potkin; Clifford R Jack; Michael W Weiner; Andrew J Saykin
Journal:  Neuroimage       Date:  2010-01-25       Impact factor: 6.556

Review 6.  Genetic analysis of quantitative phenotypes in AD and MCI: imaging, cognition and biomarkers.

Authors:  Li Shen; Paul M Thompson; Steven G Potkin; Lars Bertram; Lindsay A Farrer; Tatiana M Foroud; Robert C Green; Xiaolan Hu; Matthew J Huentelman; Sungeun Kim; John S K Kauwe; Qingqin Li; Enchi Liu; Fabio Macciardi; Jason H Moore; Leanne Munsie; Kwangsik Nho; Vijay K Ramanan; Shannon L Risacher; David J Stone; Shanker Swaminathan; Arthur W Toga; Michael W Weiner; Andrew J Saykin
Journal:  Brain Imaging Behav       Date:  2014-06       Impact factor: 3.978

7.  Cadherin-13, a risk gene for ADHD and comorbid disorders, impacts GABAergic function in hippocampus and cognition.

Authors:  O Rivero; M M Selten; S Sich; S Popp; L Bacmeister; E Amendola; M Negwer; D Schubert; F Proft; D Kiser; A G Schmitt; C Gross; S M Kolk; T Strekalova; D van den Hove; T J Resink; N Nadif Kasri; K P Lesch
Journal:  Transl Psychiatry       Date:  2015-10-13       Impact factor: 6.222

8.  Deep Learning in Alzheimer's Disease: Diagnostic Classification and Prognostic Prediction Using Neuroimaging Data.

Authors:  Taeho Jo; Kwangsik Nho; Andrew J Saykin
Journal:  Front Aging Neurosci       Date:  2019-08-20       Impact factor: 5.750

9.  Gene discovery and polygenic prediction from a genome-wide association study of educational attainment in 1.1 million individuals.

Authors:  James J Lee; Robbee Wedow; Aysu Okbay; Edward Kong; Omeed Maghzian; Meghan Zacher; Tuan Anh Nguyen-Viet; Peter Bowers; Julia Sidorenko; Richard Karlsson Linnér; Mark Alan Fontana; Tushar Kundu; Chanwook Lee; Hui Li; Ruoxi Li; Rebecca Royer; Pascal N Timshel; Raymond K Walters; Emily A Willoughby; Loïc Yengo; Maris Alver; Yanchun Bao; David W Clark; Felix R Day; Nicholas A Furlotte; Peter K Joshi; Kathryn E Kemper; Aaron Kleinman; Claudia Langenberg; Reedik Mägi; Joey W Trampush; Shefali Setia Verma; Yang Wu; Max Lam; Jing Hua Zhao; Zhili Zheng; Jason D Boardman; Harry Campbell; Jeremy Freese; Kathleen Mullan Harris; Caroline Hayward; Pamela Herd; Meena Kumari; Todd Lencz; Jian'an Luan; Anil K Malhotra; Andres Metspalu; Lili Milani; Ken K Ong; John R B Perry; David J Porteous; Marylyn D Ritchie; Melissa C Smart; Blair H Smith; Joyce Y Tung; Nicholas J Wareham; James F Wilson; Jonathan P Beauchamp; Dalton C Conley; Tõnu Esko; Steven F Lehrer; Patrik K E Magnusson; Sven Oskarsson; Tune H Pers; Matthew R Robinson; Kevin Thom; Chelsea Watson; Christopher F Chabris; Michelle N Meyer; David I Laibson; Jian Yang; Magnus Johannesson; Philipp D Koellinger; Patrick Turley; Peter M Visscher; Daniel J Benjamin; David Cesarini
Journal:  Nat Genet       Date:  2018-07-23       Impact factor: 38.330

10.  Aβ-dependent reduction of NCAM2-mediated synaptic adhesion contributes to synapse loss in Alzheimer's disease.

Authors:  Iryna Leshchyns'ka; Heng Tai Liew; Claire Shepherd; Glenda M Halliday; Claire H Stevens; Yazi D Ke; Lars M Ittner; Vladimir Sytnyk
Journal:  Nat Commun       Date:  2015-11-27       Impact factor: 14.919

View more
  5 in total

1.  Feature Fusion and Detection in Alzheimer's Disease Using a Novel Genetic Multi-Kernel SVM Based on MRI Imaging and Gene Data.

Authors:  Xianglian Meng; Qingpeng Wei; Li Meng; Junlong Liu; Yue Wu; Wenjie Liu
Journal:  Genes (Basel)       Date:  2022-05-07       Impact factor: 4.141

2.  Research on Pathogenic Hippocampal Voxel Detection in Alzheimer's Disease Using Clustering Genetic Random Forest.

Authors:  Wenjie Liu; Luolong Cao; Haoran Luo; Ying Wang
Journal:  Front Psychiatry       Date:  2022-04-07       Impact factor: 5.435

3.  Prediction of Early Alzheimer Disease by Hippocampal Volume Changes under Machine Learning Algorithm.

Authors:  Qun Shang; Qi Zhang; Xiao Liu; Lingchen Zhu
Journal:  Comput Math Methods Med       Date:  2022-05-06       Impact factor: 2.809

Review 4.  Transfer Learning Approaches for Neuroimaging Analysis: A Scoping Review.

Authors:  Zaniar Ardalan; Vignesh Subbian
Journal:  Front Artif Intell       Date:  2022-02-21

5.  Portable Vertical Graphene@Au-Based Electrochemical Aptasensing Platform for Point-of-Care Testing of Tau Protein in the Blood.

Authors:  Yibiao Liu; Xingyun Liu; Mifang Li; Qiong Liu; Tailin Xu
Journal:  Biosensors (Basel)       Date:  2022-07-25
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.