| Literature DB >> 30919255 |
Vanessa Gómez-Verdejo1, Emilio Parrado-Hernández1, Jussi Tohka2.
Abstract
An important problem that hinders the use of supervised classification algorithms for brain imaging is that the number of variables per single subject far exceeds the number of training subjects available. Deriving multivariate measures of variable importance becomes a challenge in such scenarios. This paper proposes a new measure of variable importance termed sign-consistency bagging (SCB). The SCB captures variable importance by analyzing the sign consistency of the corresponding weights in an ensemble of linear support vector machine (SVM) classifiers. Further, the SCB variable importances are enhanced by means of transductive conformal analysis. This extra step is important when the data can be assumed to be heterogeneous. Finally, the proposal of these SCB variable importance measures is completed with the derivation of a parametric hypothesis test of variable importance. The new importance measures were compared with a t-test based univariate and an SVM-based multivariate variable importances using anatomical and functional magnetic resonance imaging data. The obtained results demonstrated that the new SCB based importance measures were superior to the compared methods in terms of reproducibility and classification accuracy.Entities:
Keywords: Alzheimer’s Disease; Bagging; MRI; Schizophrenia; Support Vector Machines; Variable importance
Mesh:
Year: 2019 PMID: 30919255 PMCID: PMC6841656 DOI: 10.1007/s12021-019-9415-3
Source DB: PubMed Journal: Neuroinformatics ISSN: 1539-2791
Quantitative results with synthetic data
| Method | ACC | SEN | SPE | MAE |
|---|---|---|---|---|
| SCB | 0.889 ± 0.002 | 0.392 ± 0.004 | ||
| SCBconf | 0.879 ± 0.008 | 0.208 ± 0.011 | 0.957 ± 0.002 | |
| SVM+perm | 0.797 ± 0.005 | 0.076 ± 0.004 | 0.411 ± 0.004 | |
| SVMmar+perm | 0.891 ± 0.007 | 0.233 ± 0.011 | 0.945 ± 0.001 | 0.411 ± 0.004 |
| t-test+NGB | 0.818 ± 0.010 | 0.259 ± 0.013 | 0.949 ± 0.002 | 0.396 ± 0.004 |
The values shown are averages and standard deviations over 10 different training sets. ACC is the classification accuracy evaluated using a large test set, SEN is the sensitivity of the variable selection, SPE is the specificity of the variable selection, and MAE is the mean absolute error. See the text for details. Variables are selected using the α-threshold of 0.05
The result of the best performing method is highlighted in bold-face
Fig. 1Variable importance Z-scores (absolute values) on a plane cutting through Thalami and Superior Frontal Gyri with synthetic data. The voxels in the areas surrounded by red color were important in the ground-truth and the voxels outside those areas were not
Quantitative results with the ADNI split-half experiment
| SCBconf | SCB | SVM+perm | SVMmar+perm | T-test+NGB | |
|---|---|---|---|---|---|
| ACC | 0.766 | 0.713 | 0.745 | 0.704 | |
| ΔACC | 0.030 | 0.047 | 0.041 | 0.045 | |
| Nsel | 2067 ± 255 | 4420 ± 420 | 1884 ± 2286 | 14686 ± 8838 | 10253 ± 2278 |
| mHD | 1.536 ± 0.105 | 1.174 ± 0.049 | 2.952 ± 0.843 | 1.648 ± 2.736 | |
| mHDsta | 1.546 ± 0.111 | 2.938 ± 3.590 | 2.719 ± 2.774 | 1.707 ± 0.705 | |
| MAE | 0.278 ± 0.007 | 0.267 ± 0.064 | 0.445 ± 0.051 | 0.197 ± 0.020 |
The values listed are the averaged values over 100 resampling runs followed, where reasonable, by their standard deviations. mHD and mHDsta are computed in voxels. ACC is the classification accuracy, ΔACC is the variability of the ACC (13), Nsel is the number of selected voxels, mHD is the modified Hausdorff distance (14), mHDsta is the modified Hausdorff distance when all methods are forced to select the same number of variables, and MAE is the mean absolute error between the variable importance p-values obtained using independent training sets
The result of the best performing method is highlighted in bold-face
Fig. 2The difference in the numbers of selected voxels between two independent training sets within each split-half resampling run. The SCB methods were more stable with respect to the number of selected voxels than the other methods. Especially, SVM+perm and SVMmar+perm suffered from an excess variability
Fig. 3Variable importance Z-scores from a randomly selected example run of the ADNI split-half experiment. The Z-scores are thresholded at |Z| > 1.96, corresponding to two-sided alpha threshold of 0.05. Positive Z values indicate positive weights. Axial slices at the z-coordinate of the MNI stereotactic space of 0mm, -10mm -20mm, and -30mm are shown
Average accuracy and number of selected voxels with the 10-fold CV with the COBRE experiment
| SCBconf | SCB | SVM+perm | SVMmar+perm | T-test+NGB | |
|---|---|---|---|---|---|
| ACC | 0.695 ± 0.154 | 0.731 ± 0.136 | 0.692 ± 0.129 | 0.709 ± 0.170 | |
| Nsel | 4251 ± 598 | 11085 ± 588 | 2433 ± 216 | 6975 ± 442 | 26757 ± 3397 |
The values after ± refer to the standard deviations over 10 CV-folds
Fig. 4Median magnitudes of variable importance Z scores among 10 CV runs with COBRE data. The Z-scores are thresholded at |Z| > 1.96. Note that if a variable lights up then it was selected during at least half of the CV runs. ’Pos’ and ’Neg’ quantifiers refer to the strength of the positive and negative connectedness that were separated in the analysis. We do not visualize whether the classifier weights are negative or positive to avoid clutter. Axial slices at the z-coordinate of the MNI stereotactic space of 15mm, 0mm -15mm, and -30mm are shown. Complete maps are available in the NeuroVault service http://neurovault.org/collections/MOYIOPDI/
Fig. 5The classification accuracy and the number of selected variables across 10 CV folds with COBRE data with and without FDR based multiple comparisons correction. Whether FDR correction is included or not made no difference to the classification performance of the methods