| Literature DB >> 32976486 |
Obioma Pelka1,2, Christoph M Friedrich1,3, Felix Nensa2, Christoph Mönninghoff4, Louise Bloch1,3, Karl-Heinz Jöckel3, Sara Schramm3, Sarah Sanchez Hoffmann5, Angela Winkler5, Christian Weimar5, Martha Jokisch5.
Abstract
Detection and diagnosis of early and subclinical stages of Alzheimer's Disease (AD) play an essential role in the implementation of intervention and prevention strategies. Neuroimaging techniques predominantly provide insight into anatomic structure changes associated with AD. Deep learning methods have been extensively applied towards creating and evaluating models capable of differentiating between cognitively unimpaired, patients with Mild Cognitive Impairment (MCI) and AD dementia. Several published approaches apply information fusion techniques, providing ways of combining several input sources in the medical domain, which contributes to knowledge of broader and enriched quality. The aim of this paper is to fuse sociodemographic data such as age, marital status, education and gender, and genetic data (presence of an apolipoprotein E (APOE)-ε4 allele) with Magnetic Resonance Imaging (MRI) scans. This enables enriched multi-modal features, that adequately represent the MRI scan visually and is adopted for creating and modeling classification systems capable of detecting amnestic MCI (aMCI). To fully utilize the potential of deep convolutional neural networks, two extra color layers denoting contrast intensified and blurred image adaptations are virtually augmented to each MRI scan, completing the Red-Green-Blue (RGB) color channels. Deep convolutional activation features (DeCAF) are extracted from the average pooling layer of the deep learning system Inception_v3. These features from the fused MRI scans are used as visual representation for the Long Short-Term Memory (LSTM) based Recurrent Neural Network (RNN) classification model. The proposed approach is evaluated on a sub-study containing 120 participants (aMCI = 61 and cognitively unimpaired = 59) of the Heinz Nixdorf Recall (HNR) Study with a baseline model accuracy of 76%. Further evaluation was conducted on the ADNI Phase 1 dataset with 624 participants (aMCI = 397 and cognitively unimpaired = 227) with a baseline model accuracy of 66.27%. Experimental results show that the proposed approach achieves 90% accuracy and 0.90 F1-Score at classification of aMCI vs. cognitively unimpaired participants on the HNR Study dataset, and 77% accuracy and 0.83 F1-Score on the ADNI dataset.Entities:
Mesh:
Substances:
Year: 2020 PMID: 32976486 PMCID: PMC7518632 DOI: 10.1371/journal.pone.0236868
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
HNR Study explorative analysis.
Summary statistics computed on the sub-study of the HNR Study adopted for the proposed fusion approach. Participant Data denotes the sociodemographic data (age, marital status, education, gender) and genetic data (APOE-ε4). The total number of particpants is n = 120.
| Participant Data | aMCI | Controls | Sum | |
|---|---|---|---|---|
| ageyr | 46–55 | 1 (50.00%) | 1 (50.00%) | 2 (1.67%) |
| 56–65 | 15 (60.00%) | 10 (40.00%) | 25 (20.83%) | |
| 66–75 | 31 (48.44%) | 33 (51.56%) | 64 (53.33%) | |
| 76–85 | 14 (48.28%) | 15 (51.72%) | 29 (21.17%) | |
| gender | Female | 24 (50.00%) | 24 (50.00%) | 48 (40.00%) |
| Male | 37 (51.39%) | 35 (48.61%) | 72 (60.00%) | |
| educationyr | <= 10 | 15 (55.56%) | 12 (44.44%) | 27 (22.50%) |
| 11–13 | 37 (53.62%) | 32 (46.38%) | 69 (57.50%) | |
| >= 14 | 9 (37.50%) | 15 (62.50%) | 24 (20.00%) | |
| marital status | Married | 49 (51.04%) | 47 (48.96%) | 96 (80.00%) |
| Widowed | 8 (57.14%) | 7 (42.86%) | 14 (11.67%) | |
| Divorced | 4 (57.14%) | 3 (42.86%) | 7 (5.83%) | |
| Single | 0 (0%) | 2 (100%) | 2 (1.67%) | |
| APOE- | Positive | 21 (63.64%) | 12 (36.36%) | 33 (27.50%) |
| Negative | 40 (45.98%) | 47 (54.02%) | 87 (72.50%) | |
| Sum | 61 (50.83%) | 59 (49.17%) | 120 (100.00%) | |
aMCI = Amnestic Mild cognitive impairment
Controls = Cognitively unimpaired
ADNI Phase 1 dataset explorative analysis.
Summary statistics computed on ADNI Phase 1 dataset adopted for the proposed fusion approach. Participant Data denotes the sociodemographic data (age, marital status, education, gender) and genetic data (APOE-ε4). The total number of particpants is n = 624.
| Participant Data | aMCI | Controls | Sum | |
|---|---|---|---|---|
| ageyr | 46–55 | 3 (100.00%) | 0 (00.00%) | 2 (00.48%) |
| 56–65 | 52 (89.66%) | 6 (10.34%) | 58 (09.29%) | |
| 66–75 | 158 (58.52%) | 112 (41.48%) | 270 (43.27%) | |
| 76–90 | 184 (62.80%) | 109 (37.20%) | 293 (46.96%) | |
| gender | Female | 141 (56.40%) | 109 (43.60%) | 250 (40.01%) |
| Male | 256 (68.45%) | 118 (31.55%) | 374 (59.94%) | |
| educationyr | <= 10 | 20 (66.67%) | 10 (33.33%) | 30 (04.80%) |
| 11–13 | 79 (72.48%) | 30 (27.52%) | 109 (17.46%) | |
| >= 14 | 298 (61.44%) | 187 (38.56%) | 485 (77.72%) | |
| marital status | Married | 318 (67.23%) | 155 (32.77%) | 473 (75.80%) |
| Widowed | 48 (55.17%) | 39 (44.83%) | 87 (13.94%) | |
| Divorced | 25 (59.52%) | 17 (40.48%) | 42 (06.73%) | |
| Single | 6 (27.27%) | 16 (72.72%) | 22 (03.53%) | |
| APOE- | Positive | 185 (52.56%) | 167 (47.44%) | 352 (56.41%) |
| Negative | 212 (77.94%) | 60 (22.06%) | 272 (43.59%) | |
| Sum | 397 (63.62%) | 227 (36.38%) | 624 (100.00%) | |
aMCI = Amnestic Mild cognitive impairment
Controls = Cognitively unimpaired
Fig 1Marker for branding.
Generated markers applied for fusing sociodemographic data and APOE-ε4 data with 2D slices of MRI scans. Each marker denotes the different values for clinical data variables. Participant Data denote the sociodemographic data variables (age, marital status, education, gender) and genetic data variable (APOE-ε4). The markers were randomly distributed amongst values per variable.
Fig 2Branding approach.
Proposed branding approach of fusing sociodemographic data (age, education, marital status and gender) and genetic data (APOE-ε4) with 2D slices of an MRI scan. The marker positions and sizes of each clinical data variable branded are displayed. The 2D slice was randomly selected from an MRI scan of the sub-study from the HNR Study.
Fig 3CLAHE image preprocessing.
2D slice from a MRI scan before and after applying the Contrast Limited Adaptive Histogram Equation (CLAHE) preprocessing method. The 2D slice was randomly selected from an MRI scan of the sub-study from the HNR Study.
Fig 4NL-MEANS image preprocessing.
2D slice from a MRI scan before and after applying the Non-Local Means (NL-MEANS) preprocessing method. The 2D slice was randomly selected from an MRI scan of the sub-study from the HNR Study.
Fig 5Complete proposed approach.
Complete workflow of the proposed approach. Sociodemographic data and APOE-ε4 are fused with MRI scans 2D slice-wise and further enhanced by augmenting contrast intensified and blurred image adaptions as two extra layer completing the RGB channels. DeCAF representations are extracted and used as visual representations for training the aMCI vs control classification model.
Cross-validation prediction on HNR Study.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the k = 5-fold cross validation sets from the training set with n = 99 participants of the sub-study from the HNR Study. The values are the average and standard deviation rates across all k = 5 cross validation sets. Visual representation were extracted using the ImageNet database [64].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.64 (± 0.26) | 0.64 (± 0.21) | |
| Sensitivity | 0.70 (± 0.12) | 0.74 (± 0.11) | |
| 0.69 (± 0.09) | 0.71 (± 0.08)) | ||
| Accuracy | 0.70 (± 0.16) | 0.70 (± 0.07) |
Cross-validation prediction on HNR Study.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the k = 5-fold cross validation sets from the training set with n = 99 participants of the sub-study from the HNR Study. The values are the average and standard deviation rates across all k = 5-fold cross validation sets. Visual representation were extracted using the ChestX-Ray8 database [16].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.68 (± 0.13) | 0.74 (± 0.27) | |
| Sensitivity | 0.70 (± 0.07) | 0.68 (± 0.23) | |
| 0.70 (± 0.03) | 0.70 (± 0.16) | ||
| Accuracy | 0.69 (± 0.04) | 0.71 (± 0.13) |
Prediction accuracy on HNR Study test set.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the independent test set with n = 21 participants of the sub-study from the HNR Study. Visual representation were extracted using the ImageNet database [64].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.82 | 0.64 | |
| Sensitivity | 0.70 | ||
| 0.74 | 0.78 | ||
| Accuracy | 0.76 | 0.76 |
Prediction accuracy on HNR Study test set.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the independent test set with n = 21 participants of the sub-study from the HNR Study. Visual representation were extracted using the ChestX-Ray8 database [16].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.73 | 0.64 | |
| Sensitivity | 0.90 | ||
| 0.82 | 0.83 | ||
| Accuracy | 0.81 | 0.81 |
Cross-validation prediction on ADNI Phase 1 dataset.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the k = 5-fold cross validation sets from the training set with n = 561 participants of the ADNI Phase 1 dataset. The values are the average and standard deviation rates across all k = 5-fold cross validation sets. Visual representation were extracted using the ImageNet database [64].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.44 (± 0.08) | 0.47 (± 0.09) | |
| Sensitivity | 0.82 (± 0.06) | 0.81 (± 0.03) | |
| 0.79 (± 0.04) | 0.80 (± 0.03)) | ||
| Accuracy | 0.69 (± 0.06) | 0.71 (± 0.04) |
Cross-validation prediction on ADNI Phase 1 dataset.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the k = 5-fold cross validation sets from the training set with n = 561 participants of the ADNI Phase 1 dataset. The values are the average and standard deviation rates across all k = 5-fold cross validation sets. Visual representation were extracted using the ChestX-Ray8 database [16].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.41 (± 0.08) | 0.39 (± 0.05) | |
| Sensitivity | 0.67 (± 0.04) | 0.70 (± 0.03) | |
| 0.67 (± 0.02) | 0.68 (± 0.00)) | ||
| Accuracy | 0.58 (± 0.03) | 0.59 (± 0.01) |
Prediction accuracy on ADNI Phase 1 test set.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the independent test set with n = 63 participants of the ADNI Phase 1 dataset. Visual representation were extracted using the ImageNet database [64].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.48 | 0.52 | |
| Sensitivity | 0.77 | ||
| 0.78 | 0.79 | ||
| Accuracy | 0.66 | 0.71 |
Prediction accuracy on ADNI Phase 1 test set.
Prediction performance of the LSTM classification model using various image input types. The highlighted values are the best per evaluation metric. Evaluation was calculated on the independent test set with n = 63 participants of the ADNI Phase 1 dataset. Visual representation were extracted using the ChestX-Ray8 database [16].
| Original | Branded | Wide and Deep | |
|---|---|---|---|
| Specificity | 0.43 | 0.48 | |
| Sensitivity | 0.70 | ||
| 0.61 | 0.69 | ||
| Accuracy | 0.53 | 0.61 |
Fig 6Classification activation mapping.
Gradient-weighted Class Activation Mapping (Grad-CAM) image, highlighting important image regions used for distinguishing between aMCI and controls by the classification models. The 2D slice was randomly chosen from the sub-study of the HNR Study.
Ablation study on HNR Study test set.
Prediction performance of the LSTM classification model on the ablation study. Each sociodemographic data variable, as well as the genetic data APOE-ε4 was subsequently omitted, prior to the MRI branding. Evaluation was calculated on the independent test set with n = 21 participants of the sub-study from the HNR Study. Visual representation were extracted using the ImageNet database [64].
| Specificity | Sensitivity | Accuracy | ||
|---|---|---|---|---|
| All data variables | ||||
| Without age | 0.82 | 0.84 | 0.86 | |
| Without APOE- | 0.91 | 0.70 | 0.78 | 0.81 |
| Without gender | 0.80 | 0.86 | 0.86 | |
| Without education | 0.82 | 0.80 | 0.80 | 0.81 |
| Without marital status | 0.80 | 0.84 | 0.86 |
Ablation study on HNR Study test set.
Prediction performance of the LSTM classification model on the ablation study. Each sociodemographic data variable, as well as the genetic data APOE-ε4 was subsequently omitted, prior to the MRI branding. Evaluation was calculated on the independent test set with n = 21 participants of the sub-study from the HNR Study. Visual representation were extracted using the ChestX-Ray8 database [16].
| Specificity | Sensitivity | Accuracy | ||
|---|---|---|---|---|
| All data variables | ||||
| Without age | 0.74 | 0.78 | 0.76 | |
| Without APOE- | 0.82 | 0.80 | 0.86 | 0.86 |
| Without gender | 0.70 | 0.82 | 0.81 | |
| Without education | 0.74 | 0.70 | 0.78 | 0.76 |
| Without marital status | 0.80 | 0.78 | 0.76 |
Ablation study on ADNI Phase 1 test set.
Prediction performance of the LSTM classification model on the ablation study. Each sociodemographic data variable, as well as the genetic data APOE-ε4 was subsequently omitted, prior to the MRI branding. Evaluation was calculated on the independent test set with n = 63 participants of the ADNI Phase 1 dataset. Visual representation were extracted using the ImageNet database [64].
| Specificity | Sensitivity | Accuracy | ||
|---|---|---|---|---|
| All data variables | ||||
| Without age | 0.57 | 0.72 | 0.73 | 0.63 |
| Without APOE- | 0.39 | 0.80 | 0.74 | 0.65 |
| Without gender | 0.61 | 0.72 | 0.74 | 0.68 |
| Without education | 0.52 | 0.72 | 0.73 | 0.65 |
| Without marital status | 0.57 | 0.75 | 0.75 | 0.68 |
Ablation study on ADNI Phase 1 test set.
Prediction performance of the LSTM classification model on the ablation study. Each sociodemographic data variable, as well as the genetic data APOE-ε4 was subsequently omitted, prior to the MRI branding. Evaluation was calculated on the independent test set with n = 63 participants of the ADNI Phase 1 dataset. Visual representation were extracted using the ChestX-Ray8 database [16].
| Specificity | Sensitivity | Accuracy | ||
|---|---|---|---|---|
| All data variables | ||||
| Without age | 0.50 | 0.66 | 0.71 | 0.60 |
| Without APOE- | 0.58 | 0.69 | 0.68 | 0.62 |
| Without gender | 0.52 | 0.70 | 0.66 | 0.61 |
| Without education | 0.49 | 0.71 | 0.70 | 0.64 |
| Without marital status | 0.54 | 0.72 | 0.72 | 0.65 |