| Literature DB >> 35601611 |
Solale Tabarestani1, Mohammad Eslami2, Mercedes Cabrerizo1, Rosie E Curiel3,4, Armando Barreto1, Naphtali Rishe1, David Vaillancourt4,5,6, Steven T DeKosky4,5, David A Loewenstein3,4,7, Ranjan Duara4,7, Malek Adjouadi1,4.
Abstract
With the advances in machine learning for the diagnosis of Alzheimer's disease (AD), most studies have focused on either identifying the subject's status through classification algorithms or on predicting their cognitive scores through regression methods, neglecting the potential association between these two tasks. Motivated by the need to enhance the prospects for early diagnosis along with the ability to predict future disease states, this study proposes a deep neural network based on modality fusion, kernelization, and tensorization that perform multiclass classification and longitudinal regression simultaneously within a unified multitask framework. This relationship between multiclass classification and longitudinal regression is found to boost the efficacy of the final model in dealing with both tasks. Different multimodality scenarios are investigated, and complementary aspects of the multimodal features are exploited to simultaneously delineate the subject's label and predict related cognitive scores at future timepoints using baseline data. The main intent in this multitask framework is to consolidate the highest accuracy possible in terms of precision, sensitivity, F1 score, and area under the curve (AUC) in the multiclass classification task while maintaining the highest similarity in the MMSE score as measured through the correlation coefficient and the RMSE for all time points under the prediction task, with both tasks, run simultaneously under the same set of hyperparameters. The overall accuracy for multiclass classification of the proposed KTMnet method is 66.85 ± 3.77. The prediction results show an average RMSE of 2.32 ± 0.52 and a correlation of 0.71 ± 5.98 for predicting MMSE throughout the time points. These results are compared to state-of-the-art techniques reported in the literature. A discovery from the multitasking of this consolidated machine learning framework is that a set of hyperparameters that optimize the prediction results may not necessarily be the same as those that would optimize the multiclass classification. In other words, there is a breakpoint beyond which enhancing further the results of one process could lead to the downgrading in accuracy for the other.Entities:
Keywords: Alzheimer’s disease; longitudinal regression; multitask learning; neural network; prediction; progression
Year: 2022 PMID: 35601611 PMCID: PMC9120529 DOI: 10.3389/fnagi.2022.810873
Source DB: PubMed Journal: Front Aging Neurosci ISSN: 1663-4365 Impact factor: 5.702
Demographic characteristics of subjects used in this study.
| Parameter | Value | Total | Alzheimer | MCI-C | MCI-NC | Control |
| Subjects | Number | 1,117 | 157 | 191 | 441 | 328 |
| Gender | f/m | 485/632 | 73/84 | 75/116 | 184/257 | 153/175 |
| Age | Year (mean ± std) | 73.84 ± 7.07 | 76.77 ± 6.99 | 73.86 ± 7.47 | 70.85 ± 7.19 | 75.01 ± 5.71 |
| Education | Year (mean ± std) | 16.04 ± 2.78 | 14.63 ± 3.15 | 16.09 ± 2.74 | 16.09 ± 2.63 | 16.36 ± 2.68 |
| MMSE | Number (mean ± std) | 27.43 ± 2.46 | 23.24 ± 1.96 | 27.23 ± 1.75 | 28.30 ± 1.59 | 29.15 ± 1.01 |
| CDR | Number (mean ± std) | 1.25 ± 1.36 | 3.98 ± 1.51 | 1.62 ± 0.92 | 1.24 ± 0.74 | 0.03 ± 0.13 |
Label f/m stands for the number of females in comparison to males. Age, years of education, MMSE, and CDR of subjects in each category are presented by mean ± standard variation of that variable.
FIGURE 1Number of subjects in each of the four subgroups of AD at different time points.
FIGURE 2The average trajectories of (A) RAVLT, (B) MMSE, and (C) ADAS11 score for subjects for four different classes of AD.
FIGURE 3Design architecture of the proposed KTM network.
Summary of multimodal features used for training and testing the KTMnet dataset.
| Source | Features |
| MRI | Ventricular volume, Hippocampus volume, Whole Brain volume, Entorhinal Cortical thickness, Fusiform, Middle temporal gyrus, and intracranial volume (ICV) |
| PET | FDG, Pittsburgh Compound-B (PIB), AV45 |
| Cognitive test | Rey Auditory Verbal Learning Test (RAVLT Immediate, RAVLT Learning, RAVLT Forgetting, RAVLT Perc Forgetting), Functional Activities Questionnaires (FAQ), Everyday Cognition (Ecog) scales: (EcogPtMem, EcogPtLang, EcogPtVisspat, EcogPtPlan, EcogPtOrgan, EcogPtDivatt, EcogPtTotal, EcogSPMem, EcogSPLang, EcogSPVisspat, EcogSPPlan, EcogSPOrgan, EcogSPDivatt, and EcogSPTotal) |
| CSF | Amyloid Beta (ABETA), Phosphorylated Tau Protein (PTAU), and Total Tau Protein (TAU) |
| Risk factors | Age, gender, years of education, and APOE4 |
Comparison of longitudinal regression performance of the proposed network in contrast to other methods reported in the literature.
| T0 | T06 | T12 | T24 | |||||||
| Study | Data | Subjects | RMSE | Corr | RMSE | Corr | RMSE | Corr | RMSE | Corr |
|
| MRI + PET | 202 | 1.80 ± 0.13 | 0.57 ± 0.23 | − | − | − | − | − | − |
|
| MRI + DEM | 1,984 | 2.37 | 0.57 | − | − | − | − | − | − |
|
| MRI | 755 | 2.37 ± 0.19 | 0.57 ± 0.05 | − | − | − | − | − | − |
|
| MRI | 445 | 1.75 ± 0.20 | 0.75 ± 0.08 | 2.31 ± 0.29 | 0.79 ± 0.10 | 2.48 ± 0.40 | 0.79 ± 0.12 | 3.00 ± 0.38 | 0.83 ± 0.06 |
|
| MRI + PET + CSF | 186 | 2.11 ± 0.35 | 0.65 ± 0.27 | − | − | − | − | − | − |
| Elastic net | Multimodal*b | 1,117 | 1.84 ± 0.35 | 0.71 | 2.58 ± 0.34 | 0.54 | 2.91 ± 0.53 | 0.51 | 3.64 ± 0.56 | 0.50 |
| SVR | Multimodal*b | 1,117 | 1.75 ± 0.44 | 0.42 | 2.02 ± 0.53 | 0.54 | 2.52 ± 31 | 0.54 | 3.12 ± 0.41 | 0.51 |
| Random forest | Multimodal*b | 1,117 | 1.74 ± | 0.78 | 1.98 ± 0.45 | 0.67 | 2.36 ± 0.36 | 0.73 | 3.15 ± 0.28 | 0.70 |
|
| Multimodal*b | 1,620 | 1.62 ± 0.24 | 0.82 | 1.78 ± 0.22 | 0.86 | 2.24 ± 0.24 | 0.80 | 2.38 ± 0.21 | 0.81 |
| KTMnet | Multimodal*b | 1,117 | 1.79 ± 0.12 | 0.66 ± 0.81 | 2.10 ± 0.15 | 0.71 ± 0.92 | 2.42 ± 0.28 | 0.71 ± 0.41 | 2.97 ± 0.45 | 0.75 ± 3.10 |
*Multimodal here refers to using MRI, PET, DEM, CSF, and cognitive measurements without the inclusion of ADAS11, ADAS13, and CDRSB.
FIGURE 4Scatter plots of predicted MMSE values.
Comparison of 4-way multiclass classification performance of methodologies reported in the literature using ADNI dataset.
| Study | Data | Subjects | Validation method | Accuracy |
|
| MRI | 758 | 10-fold | 46.30 ± 4.24 |
|
| MRI + PET | 331 | 10-fold | 53.79 ± 4.76 |
|
| MRI + PET | 202 | 10-fold | 0.619 ± 1.54 |
|
| MRI + PET + DEM | 202 | Independent test | 51.80 |
|
| MRI + PET | 202 | 10-fold | 61.06 ± 1.40 |
|
| MRI + PET + CSF | 805 | 10-fold | 53.72 (max) |
| SVM | MRI + PET + CSF + COG + DEM | 1,117 | 10-fold | 58.49 ± 4.01 |
| Random forest | MRI + PET + CSF + COG + DEM | 1,117 | 10-fold | 60.28 ± 2.83 |
| KTMnet | MRI + PET + CSF + COG + DEM | 1,117 | 10-fold | 66.85 ± 3.77 |
FIGURE 5Comparison of ROC curves of the KTMnet for AD vs. MCI-C vs. MCI-NC vs. CN.
FIGURE 6Confusion matrix of the KTMnet model.
Comparison of different configurations of the proposed model discussed as design exploration study.
| T0 | T06 | T12 | T24 | Acc | |||||
| Experiment | RMSE | Corr | RMSE | Corr | RMSE | Corr | RMSE | Corr | |
| Design exploration 1 | 5.93 ± 1.29 | 0.52 ± 0.32 | 6.02 ± 1.17 | 0.50 ± 0.43 | 5.855 ± 1.30 | 0.51 ± 0.21 | 6.45 ± 1.08 | 0.52 ± 041 | 60.98 ± 3.07 |
| Design exploration 2 | 1.84 ± 0.15 | 0.62 ± 0.27 | 2.46 ± 0.22 | 0.61 ± 0.18 | 2.50 ± 0.25 | 0.58 ± 0.25 | 3.17 ± 0.32 | 0.69 ± 0.38 | 64.42 ± 4.37 |
| Design exploration 3 | 2.19 ± 0.20 | 0.56 ± 0.76 | 2.39 ± 0.35 | 0.62 ± 0.31 | 2.63 ± 0.29 | 0.62 ± 0.43 | 3.25 ± 0.32 | 0.70 ± 035 | 63.16 ± 5.13 |
| KTMnet | 1.79 ± 0.12 | 0.66 ± 0.81 | 2.10 ± 0.15 | 0.71 ± 0.92 | 2.42 ± 0.28 | 0.71 ± 0.41 | 2.97 ± 0.45 | 0.75 ± 0.31 | 66.85 ± 3.77 |
Summary of prediction tasks accomplished in the literature.
| Method | Multitask | Classification type | Class name | Regression type | Modality | Subjects |
|
| No | Multiclass | AD-MCI-CN | − | MRI | 397 |
| RELM ( | No | Multiclass | AD-MCI-CN | − | MRI | 214 |
|
| No | Multiclass | AD/MCI/CN and AD/MCI-C/MCI-NC/CN) | − | MRI—PET | 202 |
| JRMI ( | Yes | Multiclass | AD/MCI/CN and AD/MCI-C/MCI-NC/CN | Single time point | MRI—PET | 202 |
| DM2L ( | Yes | Binary and multiclass | AD/MCI/CN and AD/pMCI/sMCI/CN | Single time point | MRI—Demographic | 1,984 |
| DW-S2MTL ( | No | Binary and multiclass | AD/MCI/CN and AD/pMCI/sMCI/CN | − | MRI—PET—CSF | 805 |
| SMKMTL ( | No | Binary | AD/MCI-C/MCI-NC/CN | Multiple cognitive scores | MRI | 788 |
| SAE ( | No | Multiclass | AD/MCI-C/MCI-NC/CN | − | MRI and (MRI + PET) | 758–331 |
| SMTL ( | No | − | AD/MCI/CN | 4 time points | MRI | 445 |
| MSMT ( | No | − | CN/MCI/AD | 4 time points | Multimodal | 818 |
| CNN ( | No | Binary | AD/pMCI/sMCI/CN | − | MRI + PET | 397 |
| M3T ( | Yes | Binary | MCI-C/MCI-NC and AD/CN and MCI/CN | 2y changes of MMSE | MRI + PET + CSF | 186 |
| MSJL ( | No | Binary | AD/CN, MCI/CN, MCI-C/MCI-NC | Single time point | MRI + PET + CSF | 202 |
FIGURE 7Boxplot for RMSE of mixture category of subjects using different combinations of modalities. Here C1 stands for MRI + PET + RF, C2 stands for MRI + PET + RF + COG, C3 stands for MRI + PET + RF + CSF, C4 stands for MRI + PET + RF + COG + CSF.
FIGURE 8Boxplot for accuracy of multiclass classification achieved through the proposed network based on a different combination of modalities. Here C1 stands for MRI + PET + RF, C2 stands for MRI + PET + RF + COG, C3 stands for MRI + PET + RF + CSF, C4 stands for MRI + PET.