| Literature DB >> 36253382 |
Sheng Liu1, Arjun V Masurkar2,3, Henry Rusinek4,5, Jingyun Chen2,4, Ben Zhang4, Weicheng Zhu1, Carlos Fernandez-Granda6,7, Narges Razavian8,9,10,11.
Abstract
Early diagnosis of Alzheimer's disease plays a pivotal role in patient care and clinical trials. In this study, we have developed a new approach based on 3D deep convolutional neural networks to accurately differentiate mild Alzheimer's disease dementia from mild cognitive impairment and cognitively normal individuals using structural MRIs. For comparison, we have built a reference model based on the volumes and thickness of previously reported brain regions that are known to be implicated in disease progression. We validate both models on an internal held-out cohort from The Alzheimer's Disease Neuroimaging Initiative (ADNI) and on an external independent cohort from The National Alzheimer's Coordinating Center (NACC). The deep-learning model is accurate, achieved an area-under-the-curve (AUC) of 85.12 when distinguishing between cognitive normal subjects and subjects with either MCI or mild Alzheimer's dementia. In the more challenging task of detecting MCI, it achieves an AUC of 62.45. It is also significantly faster than the volume/thickness model in which the volumes and thickness need to be extracted beforehand. The model can also be used to forecast progression: subjects with mild cognitive impairment misclassified as having mild Alzheimer's disease dementia by the model were faster to progress to dementia over time. An analysis of the features learned by the proposed model shows that it relies on a wide range of regions associated with Alzheimer's disease. These findings suggest that deep neural networks can automatically learn to identify imaging biomarkers that are predictive of Alzheimer's disease, and leverage them to achieve accurate early detection of the disease.Entities:
Mesh:
Substances:
Year: 2022 PMID: 36253382 PMCID: PMC9576679 DOI: 10.1038/s41598-022-20674-x
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.996
Demographic characteristics of the ADNI and NACC cohorts.
| Dataset | ADNI | NACC | ||||
|---|---|---|---|---|---|---|
| Subject characteristics | CN | MCI | AD | CN | MCI | AD |
| Age, mean (sd) | 77.3 (5.6) | 76.5 (7.3) | 76.5 (7.3) | 69.1 (9.4)* (p-val < 0.01) | 74.4 (8.5)* (p-val < 0.01) | 73.9 (8.8)* (p-val: < 0.01) |
| Male | 394 (50.4%) | 659 (60.5%) | 406 (54.3%) | 489 (38.2%)* (p-val < 0.01) | 128 (39.8%)* (p-val < 0.01) | 219 (49.5%) (p-val:0.433) |
| Female | 388 (49.6%) | 430 (39.5%) | 342 (45.7%) | 792 (61.8%)* (p-val < 0.01) | 194 (60.2%)* (p-val < 0.01) | 223 (50.5%)* (p-val:0.02) |
| Education, average years (sd) | 17.2 (3.1) | 16.7 (3.2) | 16.1 (3.5) | 16.3 (2.6)* (p-val < 0.01) | 15.7 (2.8)* (p-val < 0.01) | 15.1 (3.3)* (p-val < 0.01) |
| APOE4, n (%) | 224 (28.6%) | 567 (52.1%) | 496 (66.3%) | 479 (37.4%)* (p-val < 0.01) | 146 (45.3%)* (p-val:0.03) | 202 (45.7%)* (p-val < 0.01) |
Significance of differences between NACC and ADNI for each cognitive category are reported. Statistical significance (at p-value < 0.05) is indicated by *.
Figure 1Overview of the deep learning framework and performance for Alzheimer’s automatic diagnosis. (a) Deep learning framework used for automatic diagnosis. (b) Receiver operating characteristic (ROC) curves for classification of cognitively normal (CN), mild cognitive impairment (MCI) and Alzheimer’s disease (AD), computed on the ADNI held-out test set. (c) ROC curves for classification of cognitively normal (CN), mild cognitive impairment (MCI) and Alzheimer’s disease (AD) on the NACC test set. (d) Visualization using t-SNE projections of the features computed by the proposed deep-learning model. Each point represents a scan. Green, blue, red colors indicate predicted cognitive groups. CN and AD scans are clearly clustered. (e) Visualization using t-SNE projections of the 138 volumes and thickness in the ROI-volume/thickness model. Compared to (d) the separation between CN and AD scans is less marked. The t-SNE approach is described in details in the methods section.
Classification performance in ADNI held-out set and an external validation set.
| ADNI held-out | NACC external validation | |||
|---|---|---|---|---|
| Deep learning model | ROI-volume/thickness | Deep learning model | ROI-volume/thickness | |
| Cognitively Normal (CN) | 87.59 (95% CI: 87.13–88.05) | 84.45 (95% CI: 84.19–84.71) | 85.12 (95% CI: 85.26–84.98) | 80.77 (95% CI: 80.55–80.99) |
| Mild Cognitive Impairment (MCI) | 62.59 (95% CI: 62.01–63.17) | 56.95 (95% CI: 56.27–57.63) | 62.45 (95% CI: 62.82–62.08) | 57.88 (95% CI: 57.53–58.23) |
| Alzheimer’s Disease Dementia (AD) | 89.21 (95% CI: 88.88–89.54) | 85.57 (95% CI: 85.16–85.98) | 89.21 (95% CI: 88.99–89.43) | 81.03 (95% CI: 80.84–81.21) |
Area under ROC curve for classification performance based on the deep learning model vs the ROI-volume/thickness model, for ADNI held-out set and NACC external validation set. Deep learning model outperforms ROI-volume/thickness-based model in all classes.
Figure 2Progression analysis for MCI subjects. (a) Progression analysis based on the deep learning model. (b) Progression analysis based on the ROI-volume/thickness model. The subjects in the ADNI test set are divided into two groups based on the classification results of the deep learning model from their first scan diagnosed as MCI: group A if the prediction is AD, and group B if it is not. The graph shows the fraction of subjects that progressed to AD at different months following the first scan diagnosed as MCI for both groups. Subjects in group A progress to AD at a significantly faster rate, suggesting that the features extracted by the deep-learning model may be predictive of the transition.
Figure 3Performance across different subgroups. Performance of the deep learning model (in blue) and of the ROI-volume/thickness model (in red) for different subpopulations of individuals, separated according to sex, education, and ApoE4 status.
Figure 4Datasize Impact. Performance of the baseline ROI-volume/thickness model (left) and the proposed deep learning model (right) when trained on datasets with different sizes (obtained by randomly subsampling the training set). The performance of the ROI-volume/thickness model improves when the training data increases from 50 to 70%, but remains essentially stagnant after further increases. In contrast, the performance of the deep learning model consistently improves as the size of the training set increases. Given that the deep learning model is trained on a very small dataset compared to standard computer-vision tasks for natural images, this suggests that building larger training sets is a promising avenue to further improving performance.
Figure 5(a–c) Visualization of the aggregated importance of each voxel (in yellow) in the deep learning model when classifying subjects into CN/MCI/AD. For each subject, the importance map was computed using the gradient of the deep-learning model with respect to its input (Details in “Methods” section). The computed gradients are visualized over the MNI T1-weighted template. (d–f) Top 30 regions of interest, sorted by their normalized gradient count, which quantifies their importance (see “Methods” section), for each of the classes.