| Literature DB >> 32415917 |
Jessica Dafflon1, Walter H L Pinaya2,3, Federico Turkheimer1, James H Cole1, Robert Leech1, Mathew A Harris4, Simon R Cox5,6, Heather C Whalley4, Andrew M McIntosh4, Peter J Hellyer1.
Abstract
The use of machine learning (ML) algorithms has significantly increased in neuroscience. However, from the vast extent of possible ML algorithms, which one is the optimal model to predict the target variable? What are the hyperparameters for such a model? Given the plethora of possible answers to these questions, in the last years, automated ML (autoML) has been gaining attention. Here, we apply an autoML library called Tree-based Pipeline Optimisation Tool (TPOT) which uses a tree-based representation of ML pipelines and conducts a genetic programming-based approach to find the model and its hyperparameters that more closely predicts the subject's true age. To explore autoML and evaluate its efficacy within neuroimaging data sets, we chose a problem that has been the focus of previous extensive study: brain age prediction. Without any prior knowledge, TPOT was able to scan through the model space and create pipelines that outperformed the state-of-the-art accuracy for Freesurfer-based models using only thickness and volume information for anatomical structure. In particular, we compared the performance of TPOT (mean absolute error [MAE]: 4.612 ± .124 years) and a relevance vector regression (MAE 5.474 ± .140 years). TPOT also suggested interesting combinations of models that do not match the current most used models for brain prediction but generalise well to unseen data. AutoML showed promising results as a data-driven approach to find optimal models for neuroimaging applications.Entities:
Keywords: age prediction; automated machine learning; cortical features; neuroimaging; predictive modelling; structural imaging
Mesh:
Year: 2020 PMID: 32415917 PMCID: PMC7416036 DOI: 10.1002/hbm.25028
Source DB: PubMed Journal: Hum Brain Mapp ISSN: 1065-9471 Impact factor: 5.399
FIGURE 1Overview of experimental design. The subject's structural MRI is used to create a parcellation of cortical and subcortical regions. The data set was split into two independent sets: TPOT training set and evaluation set. The TPOT training set was passed to TPOT, which depending on the specified configuration performed feature selection, feature transformation, feature generation or a combination of those and evaluated the model's performance. For each generation, a 10‐fold cross‐validation was performed and the best models for that specific generation were identified, crossed‐over/mutated and passed to the next generation. At the last generation, the pipeline with the lowest mean absolute error (MAE) was identified and returned by TPOT. We then retrained the optimised pipeline on the independent evaluation set and tested its performance using a 10‐fold cross‐validation. Finally, we compared the MAEs between different TPOT configurations and between TPOT and RVR
Overview of the demographics and imaging parameters for the different datasets
| Cohort | N | Age mean ( | Age range | Sex male/female | Repository details | Scanner (field strength) | Scan | Voxel dimensions |
|---|---|---|---|---|---|---|---|---|
| ABIDE (Autism Brain Imaging Data Exchange) | 147 | 24.43 (4.89) | 18–40 | 130/17 | INDI | Various (all 3T) | MPRAGE | Various |
| Beijing Normal University | 151 | 21.36 (1.95) | 18–28 | 63/88 | INDI | Siemens (3T) | MPRAGE | 1.33 × 1.0 × 1.0 |
| Berlin School of Brain and Mind | 49 | 30.99 (7.08) | 20–60 | 24/25 | INDI | Siemens Tim Trio (3T) | MPRAGE | 1.0 × 1.0 × 1.0 |
| CADDementia | 12 | 62.33 (6.26) | 58–79 | 9/3 | http://caddementia.grand‐challenge.org | GE Signa (3T) | 3D IRFSPGR | 0.9 × 0.9 × 1.0 |
| Cleveland Clinic | 31 | 43.55 (11.14) | 24–60 | 11/20 | INDI | Siemens Tim Trio (3T) | MPRAGE | 2.0 × 1.0 × 1.2 |
| ICBM (International Consortium for Brain Mapping) | 42 | 27.71 (5.75) | 24–60 | 14/28 | LONI IDA | Siemens Magnetom (1.5T) | MPRAGE | 1.0 × 1.0 × 1.0 |
| IXI (Information eXtraction from Images) | 394 | 46.21 (16.11) | 20–86 | 159/235 | http://biomedic.doc.ic.ac.uk/brain‐development | Philips Intera (3T); Philips Gyroscan Intera (1.5T); GE Signa (1.5T) | T1‐FFE; MPRAGE | 0.9375 × 0.93751 × 1.2 |
| MCIC (MIND Clinical Imaging Consortium) | 92 | 32.33 (11.92) | 18–60 | 63/29 | COINS | Siemens Sonata/Trio (1.5/3T); GE Signa (1.5T) | MPRAGE; SPGR | 0.625 × 0.625 × 1.5 |
| MIRIAD (Minimal Interval Resonance Imaging in Alzheimer's Disease) | 23 | 69.66 (7.18) | 58–85 | 12/11 | https://www.ucl.ac.uk/drc/research/miriad‐scan‐database | GE Signa (1.5T) | 3D IRFSPGR | 0.9375 × 0.93751 × 1.5 |
| NEO2012 (Adelstein, 2011) | 39 | 29.59 (8.38) | 20–49 | 18/21 | INDI | Siemens Allegra (3T) | MPRAGE | 1.0 × 1.0 × 1.0 |
| Nathan Kline Institute (NKI)/Rockland | 151 | 41.92 (18.24) | 18–85 | 94/57 | INDI | Siemens Tim Trio (3T) | MPRAGE | 1.0 × 1.0 × 1.0 |
| OASIS (Open Access Series of Imaging Studies) | 61 | 42.82 (20.42) | 18–89 | 20/41 | http://www.oasis‐brains.org/ | Siemens Vision (1.5T) | MPRAGE | 1.0 × 1.0 × 1.25 |
| TRAIN‐39 | 35 | 22.77 (2.52) | 18–28 | 10/25 | INDI | Siemens Allegra (3T) | MPRAGE | 1.33 × 1.33 × 1.3 |
| UK BIOBANK | 9080 | 62.45 (7.48) | 45–79 | 4334/4746 | https://biobank.ctsu.ox.ac.uk/crystal/crystal/docs/brain_mri.pdf | Siemens Skyra (3T) | MPRAGE | 1.0 × 1.0 × 1.0 |
|
|
|
|
|
|
|
|
|
|
Note. ABIDE consortiums comprising data from various sites with different scanners/parameters.
Abbreviations: COINS, Collaborative Informatics and Neuroimaging Suite (http://coins.mrn.org); INDI, International Neuroimaging Data‐sharing Initiative (http://fcon_1000.projects.nitrc.org); LONI, Laboratory of Neuro Imaging Image & Data Archive (https://ida.loni.usc.edu).
OASIS scans were acquired four times and then averaged to increase signal‐to‐noise ratio.
FIGURE 2Overview of the models count for each generation from one repetition for the different configurations experiments. Models with a darker colour were more popular then models with lighter colour. Across the four experiments, random forest, K‐nearest neighbours, linear regression and extra trees regressors are the models with the highest count per generation. To make sure that all models were represented, we had 1,000 models in the first generation and 100 models were passed on for the following generations
FIGURE 3Overview of the ensembles for the different analysis configurations at each repetition and their performance. (a) Schematic overview of the models composing the ‘best’ ensembles yielded by TPOT at each repetition. A darker colour represents models with higher counts. Random forest regression, extra trees regressors, lasso lars and linear regression were the most frequently represented. Despite the different models combinations among the different preprocessing analysis (b), initial population size (c) and mutation/cross‐over rate (d), there was no difference in the yielded performance
FIGURE 4Comparison of model's performance between TPOT and RVR. We compared the MAE (top panel left) and Pearson's correlation (top panel right) between true and predicted age of the optimised model suggested by TPOT with and RVR on the test set. The lower panels show the predicted versus the true age for one of the optimal pipelines suggested by TPOT (left) and RVR (right). Note that although both models use the same subject to make prediction, the scales of the TPOT and RVR predictions are different, and the RVR model predicts young subject to be younger and old as older. Asterisks show differences that are statistically significant at p < .01 (t‐test corrected); error bars indicate ±1SD
Comparison between TPOT and RVR. Although TPOT has a significant higher accuracy and Pearson's correlation when using the original data distribution, when using the uniformly distributed data set both models had a similar performance (the values represent ±SD)
| MAE |
|
| Pearson's correlation |
|
| |
|---|---|---|---|---|---|---|
| TPOT | 4.612 ± .124 | < | −6.441 | .874 ± .012 | < | 3.745 |
| RVR | 5.474 ± 0.140 | .813 ± .0102 | ||||
| TPOT (uniform distribution) | 5.594 ± .0706 | >.5 | −0.616 | .917 ± .027 | >.5 | 0.007 |
| RVR (uniform distribution) | 5.975 ± .525 | .919 ± .013 |
Note. The bold values correspond to analysis with a significant p‐value (p < 0.05).
Comparison between TPOT and RVR time‐complexity
| RVR | TPOT | |
|---|---|---|
| TPOT training | ‐ | About 6 hr (6.61 ± 0.39 hr) |
| Training | About 8 min (519.37 ± 2.62 s) | About 5 min (276.02 ± 2.51 s) |
| Inference | 0.001 ± 3.85 | 0.174 ± 0.06 s |