| Literature DB >> 35650452 |
Henricus G Ruhe1,2, Christian F Beckmann1,3,4, Andre F Marquand1,3,5, Saige Rutherford6,7,8, Seyed Mostafa Kia1,3,9, Thomas Wolfers10,11, Charlotte Fraza1,3, Mariam Zabihi1,3, Richard Dinga1,3, Pierre Berthet10,11, Amanda Worker5, Serena Verdi12,13.
Abstract
Normative modeling is an emerging and innovative framework for mapping individual differences at the level of a single subject or observation in relation to a reference model. It involves charting centiles of variation across a population in terms of mappings between biology and behavior, which can then be used to make statistical inferences at the level of the individual. The fields of computational psychiatry and clinical neuroscience have been slow to transition away from patient versus 'healthy' control analytic approaches, probably owing to a lack of tools designed to properly model biological heterogeneity of mental disorders. Normative modeling provides a solution to address this issue and moves analysis away from case-control comparisons that rely on potentially noisy clinical labels. Here we define a standardized protocol to guide users through, from start to finish, normative modeling analysis using the Predictive Clinical Neuroscience toolkit (PCNtoolkit). We describe the input data selection process, provide intuition behind the various modeling choices and conclude by demonstrating several examples of downstream analyses that the normative model may facilitate, such as stratification of high-risk individuals, subtyping and behavioral predictive modeling. The protocol takes ~1-3 h to complete.Entities:
Mesh:
Year: 2022 PMID: 35650452 PMCID: PMC7613648 DOI: 10.1038/s41596-022-00696-5
Source DB: PubMed Journal: Nat Protoc ISSN: 1750-2799 Impact factor: 17.021
Figure 1Conceptual Overview of Normative Modeling.
A) Classical example of normative modeling: the use of height and weight growth charting in pediatrics. B) Case-control models (left) theoretically make assumptions that there is a boundary that can separate groups and that there is within-group homogeneity. In reality (right), there is nested variation across controls and patient groups and within-group heterogeneity, resulting in unclear separation boundaries. Normative modeling is well equipped to handle this reality. C) An example application of normative modeling in computational psychiatry using neuroimaging data. Mean cortical thickness (y-axis) is predicted from age (x-axis) using a training set consisting of multi-site structural MRI from neurotypical controls and a test set consisting of neurotypical controls and patient groups. Every dot indicates the deviation score for a single individual from normal development. D) Regression model equation and design matrix setup for the model shown in panel C.
Figure 2Practical Overview of Normative Modeling Framework.
The workflow consists of four stages: data selection, data preparation, algorithm & modeling, and evaluation & interpretation, which are visualized by the numbered shaded blue boxes. The steps involved at each of these stages are summarized in the box below and highlighted in the images above.
Figure 3Overview of Resources for Running a Normative Modeling Analysis.
A) Detailed documentation, including installation instructions, input/output descriptions of all classes and functions implemented in the python package, tutorials for all algorithms, frequently asked questions, a glossary explaining acronyms and other jargon, references to existing normative modeling literature, and a citation guide, is available online. B) Example of the documentation showing the required input, expected output of the main function used in the pcntoolkit software, the estimate function. C) All of the code and data used in this protocol is available to run in the cloud via Google Colab. Additional tutorials (shown under the tutorials header in panel A) are also available to run in Google Colab.
PCNtoolkit Normative Modeling Algorithm Overview*.
| Algorithm | Implemented in PCNtoolkit? | Transfer to new sites? | Fast compute time with large sample sizes? | Model non-Gaussianity? | Federated learning framework? |
|---|---|---|---|---|---|
| Gaussian Process Regression (GPR) | yes | no | no | no | no |
| Hierarchical Bayesian Regression (HBR) | yes | yes | yes | yes | yes |
| Bayesian Linear Regression (BLR) | yes | yes | yes | yes | no |
| Generalized additive models of location, scale, and shape (GAMLSS) | no | yes | yes | yes | no |
Random feature approximation and neural processes algorithms are not well documented in the PCNtoolkit and do not have tutorials available, thus these algorithms are not included in the table and are only recommended for advanced users who can implement the code on their own.
The vanilla GPR algorithm implemented in the PCNtoolkit cannot model non-Gaussianity and does not scale well to large datasets. However, this is a question of implementation, and there are versions of GPs algorithms that satisfy these criteria82,83.
Implemented in R, see this GitHub repository.
Normative Model Metrics.
The ‘Individual or summary?’ column refers to whether there is a value for every subject or if the metric is summarized across all subjects. For summary metrics, there is one value per brain region (model), and for individual metrics there are n_subjects x n_brain_regions values.
| Variable name | Full name | Definition | Interpretation | Individual or summary? |
|---|---|---|---|---|
|
| True data | individual | ||
|
| Predictive mean | individual | ||
|
| Predictive noise variance | Represents uncertainty in the data. | individual | |
|
| Predictive modeling variance | Represents uncertainty in model estimation. | individual | |
| Z | Deviation score | A statistical estimate (Z-score) of how much each subject deviates from the normative range. | Z > 2 ‘extreme’ positive deviation | individual |
| Rho | Pearson correlation between true and predicted responses | A measure of linear correlation between true and predicted responses. It is the ratio between the covariance of true and predicted values and the product of their standard deviations. | Ranges between -1 and 1. Closer to 1 = better model performance. | summary |
| pRho | Parametric p-value for the Pearson correlation | The probability of obtaining test results at least as extreme as the results actually observed, under the assumption that the null hypothesis is true. | Ranges between 0 and 1. Closer to 0 = more statistically significant. | summary |
| SMSE | Standardized mean squared error | The square root of the squared residual between the mean prediction and the target at each test point, averaged over samples in the test set, normalized by the variance of the targets in the test set. | Closer to 0 = better (more accurate) model performance. | summary |
| EV | Explained variance | The proportion to which the predicted value accounts for the variance of the true value. Sensitive to the mean fit, dependent on flexibility of the model. | Closer to 1 = better model performance. | summary |
| MSLL | Mean standardized log-loss | The log loss minus the loss that would be obtained under the trivial model which predicts using a Gaussian with the mean and variance of the training data, averaged over the test set. Sensitive to the variance, penalizes the flexibility of the model. | More negative = better model performance. | summary |
Figure 4Visualization of Normative Model Evaluation Metrics.
A) A ridge plot showing the distribution across all brain regions of the standardized mean squared error (SMSE), an evaluation metric that represents accuracy, visualized for each site in the test set. Visualizing for each test site can help identify if there are sites where the model is performing poorly. Ideally, the distribution will be Gaussian and should look similar across all sites. Small shifts in the mean across sites is to be expected and is acceptable. B) Explained variance is shown for cortical thickness of every brain region in the Destrieux parcellation) and volume of subcortical regions. Visualizing the evaluation metrics in brain space helps to identify patterns and see the big picture. C) The number of extreme deviations (both positive and negative) are counted for each individual in the test set, group ID is used to plot the distribution of the extreme deviation count for each group. A statistical test can be done on the count to determine if there is a significant difference between groups. Testing group differences in the count of deviations does not require there to be spatial overlap of the deviations within the group (i.e., this test can account for within-group heterogeneity of deviations). D) The normative trajectory for an example brain region (lateral ventricle) showing age (x-axis) versus the predicted volume (y-axis). The centiles of variation are shown by the lines and shaded confidence intervals. Each subject in the test set is plotted as a single point. E-F) Extreme deviations, separated into positive (E) and negative (F), are summarized for each group. For each brain region, the number of subjects with an extreme deviation in that region is counted, then divided by the group sample size, to show the percent of subjects with an extreme deviation. These visualizations demonstrate the benefit of normative modeling as there is within group heterogeneity that other methods (i.e., case-control group difference testing) are not equipped to handle. Abbreviations: HC = Controls, MDD=Major Depressive Disorder, SZ=Schizophrenia, SAD=Social Anxiety Disorder, EP=Early Psychosis.