| Literature DB >> 30584016 |
Silvia Basaia1, Federica Agosta1, Luca Wagner2, Elisa Canu1, Giuseppe Magnani3, Roberto Santangelo3, Massimo Filippi4.
Abstract
We built and validated a deep learning algorithm predicting the individual diagnosis of Alzheimer's disease (AD) and mild cognitive impairment who will convert to AD (c-MCI) based on a single cross-sectional brain structural MRI scan. Convolutional neural networks (CNNs) were applied on 3D T1-weighted images from ADNI and subjects recruited at our Institute (407 healthy controls [HC], 418 AD, 280 c-MCI, 533 stable MCI [s-MCI]). CNN performance was tested in distinguishing AD, c-MCI and s-MCI. High levels of accuracy were achieved in all the classifications, with the highest rates achieved in the AD vs HC classification tests using both the ADNI dataset only (99%) and the combined ADNI + non-ADNI dataset (98%). CNNs discriminated c-MCI from s-MCI patients with an accuracy up to 75% and no difference between ADNI and non-ADNI images. CNNs provide a powerful tool for the automatic individual patient diagnosis along the AD continuum. Our method performed well without any prior feature engineering and regardless the variability of imaging protocols and scanners, demonstrating that it is exploitable by not-trained operators and likely to be generalizable to unseen patient data. CNNs may accelerate the adoption of structural MRI in routine practice to help assessment and management of patients.Entities:
Keywords: Alzheimer's disease; Convolutional neural networks; Deep learning; Diagnosis; Mild cognitive impairment; Prediction
Mesh:
Year: 2018 PMID: 30584016 PMCID: PMC6413333 DOI: 10.1016/j.nicl.2018.101645
Source DB: PubMed Journal: Neuroimage Clin ISSN: 2213-1582 Impact factor: 4.881
Literature review on structural MRI (T1-weighted images) and deep learning in AD and MCI patient classification.
| Studies (chronological order) | Dataset | Sample size | MCI conversion to AD? | Deep learning architecture | Input | Data augmentation | Validation | Transfer learning | Classifications & Accuracy (%) |
|---|---|---|---|---|---|---|---|---|---|
| Natural Image Bases to Represent Neuroimaging Data ( | ADNI | 200 AD, 232 HC, 411 MCI | NO | Sparse Auto-encoder with Convolutional Neural Network | Normalized 3D T1-weighted images | YES (serial scans from each subject) | Independent sample | NO | AD |
| Predicting Alzheimer's disease: a neuroimaging study with 3D convolutional neural networks ( | ADNI | 755 AD, 755 HC, 755 MCI | NO | Sparse Auto-encoder with 3D Convolutional Neural Network | Normalized 3D T1-weighted images | YES (serial scans from each subject) | Independent sample | NO | AD |
| Alzheimer's disease diagnostics by a deeply supervised adaptable 3D convolutional network ( | ADNI | 70 from each class (AD, HC, MCI) | NO | 3D convolutional autoencoder | Normalized 3D T1-weighted images | NO | Not specified | YES | AD |
| DeepAD: Alzheimer's Disease Classification | ADNI | 91 AD, 211 HC | NO | Convolutional Neural Network | Normalized 3D T1-weighted images | YES (3D to 2D conversion) | Not specified | NO | AD |
| Deep ensemble learning of sparse regression models for brain disease diagnosis ( | ADNI | 186 AD, 226 HC, 167 converters MCI, 226 stable MCI | YES | Multiple sparse regression models with deep Convolutional Neural Network | GM volumes | NO | 10-fold cross validation | NO | AD |
Abbreviations: AD = Alzheimer's disease; ADNI = Alzheimer's Disease Neuroimaging Initiative; GM = GRAY matter; HC = healthy controls; MCI = Mild Cognitive Impairment.
Demographic and clinical features of AD and MCI patients and healthy controls from the ADNI dataset.
| HC | AD | c-MCI | s-MCI | P AD | P c-MCI | P s-MCI | P AD | P AD | P c-MCI | |
|---|---|---|---|---|---|---|---|---|---|---|
| N | 352 | 294 | 253 | 510 | ||||||
| Age [years] | 74.53 ± 6.16 (56.20–89.60) | 75.13 ± 7.75 (55.10–90.90) | 73.80 ± 7.35 (55.00–88.30) | 72.33 ± 7.68 (54.40–91.40) | 1.00 | 1.00 | <0.001 | 0.20 | <0.001 | 0.05 |
| Gender [women/men] | 185/167 | 136/158 | 102/151 | 223/287 | 0.12 | 0.003 | 0.01 | 0.17 | 0.51 | 0.35 |
| Education [years] | 16.30 ± 2.76 (6–20) | 15.14 ± 3.02 (4–20) | 15.76 ± 2.84 (6–20) | 16.02 ± 2.82 (4–20) | <0.001 | 0.13 | 0.93 | 0.07 | <0.001 | 1.00 |
| CDR sum of boxes | 0.03 ± 0.12 (0–1) | 4.46 ± 1.61 (1−10) | 1.95 ± 0.97 (0.5–5.5) | 1.29 ± 0.76 (0.5–4) | <0.001 | <0.001 | <0.001 | <0.001 | <0.001 | <0.001 |
| MMSE | 29.07 ± 1.16 (24–30) | 23.12 ± 2.1 (18–27) | 26.91 ± 1.78 (23−30) | 27.99 ± 1.71 (23–30) | <0.001 | <0.001 | <0.001 | <0.001 | <0.001 | <0.001 |
Values are numbers or means ± standard deviations (range). P values refer to ANOVA models, followed by post-hoc pairwise comparisons (Bonferroni-corrected for multiple comparisons), or Chi-squared test. Abbreviations: AD = Alzheimer's Disease; CDR = Clinical Dementia Rating Scale; HC = healthy controls; MCI = Mild Cognitive Impairment (c = converters; s = stable); MMSE = Mini Mental State Examination; N = Number.
Demographic, clinical and neuropsychological features of AD and MCI patients and healthy controls from the Milan dataset.
| HC | AD | c-MCI | s-MCI | P AD | P c-MCI | P s-MCI | P AD | P AD | P c-MCI | |
|---|---|---|---|---|---|---|---|---|---|---|
| N | 55 | 124 | 27 | 23 | ||||||
| Age [years] | 67.1 ± 6.8 (56.1–81.8) | 68.3 ± 8.1 (48.5–85.6) | 71.6 ± 7.5 (55.3–85.7) | 68.2 ± 6.4 (52.1–80.7) | 1.00 | 0.08 | 1.00 | 0.25 | 1.00 | 0.71 |
| Gender [women/men] | 29/26 | 69/55 | 13/14 | 10/13 | 0.42 | 0.44 | 0.31 | 0.31 | 0.20 | 0.48 |
| Education [years] | 12.2 ± 4.7 (5–24) | 9.5 ± 4.5 (1–18) | 10.4 ± 4.4 (4–18) | 11.3 ± 3.7 (5–18) | 0.001 | 0.56 | 1.00 | 1.00 | 0.38 | 1.00 |
| Disease duration [years] | – | 3.5 ± 2.0 (0.0–10.2) | 3.0 ± 1.5 (1.0–6.1) | 3.0 ± 1.7 (0.6–6.0) | – | – | – | 0.65 | 0.64 | 1.00 |
| CDR | – | 1.2 ± 0.6 (0.5–3) | 0.5 ± 0.2 (0.5–1) | 0.5 ± 0.1 (0.5–1) | – | – | – | <0.001 | <0.001 | 1.00 |
| CDR sum of boxes | – | 5.1 ± 2.2 (2−12) | 2.2 ± 1.0 (1–4.5) | 2.4 ± 1.1 (1–4.5) | – | – | – | <0.001 | <0.001 | 1.00 |
| MMSE | 29.1 ± 1.0 (26–30) | 19.8 ± 4.5 (9–27) | 26.8 ± 1.7 (24–30) | 27.4 ± 2.0 (23–30) | <0.001 | 0.17 | 0.64 | <0.001 | <0.001 | 1.00 |
| Verbal memory | ||||||||||
| RAVLT, immediate recall | 43.4 ± 9.0 (25–60) | 15.0 ± 7.1 (0–40) | 19.6 ± 6.0 (8–32) | 23.7 ± 7.8 (10–34) | <0.001 | <0.001 | <0.001 | 0.11 | 0.001 | 0.62 |
| RAVLT, delayed recall | 8.9 ± 3.3 (4–15) | 0.4 ± 0.9 (0–3) | 1.2 ± 1.9 (0–6) | 1.9 ± 2.1 (0–7) | <0.001 | <0.001 | <0.001 | 0.53 | 0.04 | 1.00 |
| Digit span, forward | 5.9 ± 1.1 (4–9) | 4.6 ± 1.1 (0–7) | 4.9 ± 0.8 (3–6) | 5.4 ± 1.2 (3–8) | <0.001 | 0.01 | 0.47 | 0.99 | 0.02 | 1.00 |
| Memory prose | 9.7 ± 7.0 (3–17) | 2.2 ± 2.5 (0–14) | 5.5 ± 3.1 (0−11) | 7.3 ± 3.6 (2–15) | <0.001 | 0.14 | 1.00 | <0.001 | <0.001 | 0.30 |
| Visuospatial memory | ||||||||||
| Spatial span, forward | 5.1 ± 1.0 (4–7) | 3.0 ± 1.2 (0–6) | 4.2 ± 0.7 (3–6) | 4.5 ± 0.5 (4–5) | <0.001 | 0.02 | 0.46 | <0.001 | <0.001 | 1.00 |
| Rey's figure, recall | 17.7 ± 5.9 (9–33) | 2.8 ± 3.8 (0–26) | 5.5 ± 3.0 (0−13) | 10.0 ± 5.5 (2−21) | <0.001 | <0.001 | <0.001 | 0.03 | <0.001 | 0.002 |
| Visuospatial abilities | ||||||||||
| Rey's figure, copy | 33.2 ± 2.4 (27–36) | 15.3 ± 10.0 (0–35) | 24.3 ± 9.1 (0–36) | 27.7 ± 5.3 (15–36.0) | <0.001 | 0.001 | 0.15 | <0.001 | <0.001 | 0.99 |
| Clock Drawing Test | 8.9 ± 0.9 (8–10) | 3.5 ± 3.9 (0−10) | 7.1 ± 3.2 (0–10) | 8.0 ± 3.1 (0–10) | <0.001 | 1.00 | 1.00 | 0.002 | <0.001 | 1.00 |
| Attention and executive functions | ||||||||||
| Attentive matrices | 48.8 ± 7.6 (32–57) | 31.0 ± 12.7 (2–56) | 43.3 ± 8.2 (30–56) | 46.5 ± 7.9 (33–58) | <0.001 | 0.43 | 1.00 | <0.001 | <0.001 | 1.00 |
| Raven coloured progressive matrices | 29.9 ± 3.8 (22–35) | 16.8 ± 8.2 (2–35) | 23.3 ± 5.5 (10−31) | 26.1 ± 5.9 (9–33) | <0.001 | 0.01 | 0.37 | <0.001 | <0.001 | 1.00 |
| Semantic fluency | 42.4 ± 8.9 (27–60) | 19.0 ± 9.0 (3–55) | 29.5 ± 7.0 (16–42) | 32.7 ± 10.3 (12–55) | <0.001 | <0.001 | 0.001 | <0.001 | <0.001 | 1.00 |
| Phonemic fluency | 36.7 ± 10.4 (18–55) | 16.6 ± 10.3 (0–43) | 24.0 ± 10.3 (11–48) | 30.6 ± 13.8 (10–66) | <0.001 | <0.001 | 0.32 | 0.01 | <0.001 | 0.22 |
| Language | ||||||||||
| Token test | 33.2 ± 2.1 (29–36) | 25.5 ± 5.8 (5–36) | 30.5 ± 3.1 (24–35) | 31.6 ± 2.3 (25–34) | <0.001 | 0.26 | 1.00 | <0.001 | <0.001 | 1.00 |
Values are numbers or means ± standard deviations (range). Disease duration was defined as years from onset to date of MRI scan. P values refer to ANOVA models, followed by post-hoc pairwise comparisons (Bonferroni-corrected for multiple comparisons), or Chi-squared test. Abbreviations: AD = Alzheimer's Disease; CDR = Clinical Dementia Rating Scale; HC = healthy controls; MCI = Mild Cognitive Impairment (c = converters; s = stable); MMSE = Mini Mental State Examination; N = Number; RAVLT = Rey Auditory Verbal Learning Test.
Fig. 1Architecture of a typical convolutional neural network. a) Input layer: the data is given to the network. b) Convolutional layer: neurons identify the main features that characterize the images, storing the information into a ‘feature map’ (e.g., red, blue and yellow blocks). c) Pooling layer: the size of each feature map is reduced with a downsampling operation along the spatial dimension (e.g., red, blue and yellow blocks). d) Fully-connected layer: the neurons are connected to all neurons from the previous layer. e) Output layer: the step that returns the probability of the input data to belong to each class.
Fig. 2Flowchart of the main steps of the experiments performed. MRI data of each classification dataset (AD vs HC, c-MCI vs HC, s-MCI vs HC, AD vs c-MCI, AD vs s-MCI, c-MCI vs s-MCI) were randomly split into a large training and validation set (90% of images) and a testing set (10% of images). Data augmentation was applied on images selected for training and validation. See text for further details.
Fig. 3Examples of images after data augmentation, i.e., deformation, cropping, rotation, flipping, and scaling. Axial and coronal images are shown. A = anterior; L = left; P = posterior; R = right.
Binary classification results on testing datasets.
| Accuracy | Sensitivity | Specificity | ||
|---|---|---|---|---|
| AD | ADNI dataset | 99.2% | 98.9% | 99.5% |
| ADNI + Milan dataset | 98.2% | 98.1% | 98.3% | |
| c-MCI | ADNI dataset | 87.1% | 87.8% | 86.5% |
| ADNI + Milan dataset | 87.7% | 87.3% | 88.1% | |
| s-MCI | ADNI dataset | 76.1% | 75.1% | 77.1% |
| ADNI + Milan dataset | 76.4% | 75.1% | 77.8% | |
| AD | ADNI dataset | 75.4% | 74.5% | 76.4% |
| ADNI + Milan dataset | 75.8% | 74.8% | 77.1% | |
| AD | ADNI dataset | 85.9% | 83.6% | 88.3% |
| ADNI + Milan dataset | 86.3% | 84.0% | 88.7% | |
| c-MCI | ADNI dataset | 75.1% | 74.8% | 75.3% |
| ADNI + Milan dataset | 74.9% | 75.8% | 74.1% |
Abbreviations: AD = Alzheimer's disease; ADNI = Alzheimer's Disease Neuroimaging Initiative; HC = healthy controls; MCI = Mild Cognitive Impairment (c = converters; s = stable).