Literature DB >> 35496170

Predicting Verbal Learning and Memory Assessments of Older Adults Using Bayesian Hierarchical Models.

Endris Assen Ebrahim1,2, Mehmet Ali Cengiz1.   

Abstract

Verbal learning and memory summaries of older adults have usually been used to describe neuropsychiatric complaints. Bayesian hierarchical models are modern and appropriate approaches for predicting repeated measures data where information exchangeability is considered and a violation of the independence assumption in classical statistics. Such models are complex models for clustered data that account for distributions of hyper-parameters for fixed-term parameters in Bayesian computations. Repeated measures are inherently clustered and typically occur in clinical trials, education, cognitive psychology, and treatment follow-up. The Hopkins Verbal Learning Test (HVLT) is a general verbal knowledge and memory assessment administered repeatedly as part of a neurophysiological experiment to examine an individual's performance outcomes at different time points. Multiple trial-based scores of verbal learning and memory tests were considered as an outcome measurement. In this article, we attempted to evaluate the predicting effect of individual characteristics in considering within and between-group variations by fitting various Bayesian hierarchical models via the hybrid Hamiltonian Monte Carlo (HMC) under the Bayesian Regression Models using 'Stan' (BRMS) package of R. Comparisons of the fitted models were done using leave-one-out information criteria (LOO-CV), Widely applicable information criterion (WAIC), and K-fold cross-validation methods. The full hierarchical model with varying intercepts and slopes had the best predictive performance for verbal learning tests [from the Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) study dataset] using the hybrid Hamiltonian-Markov Chain Monte Carlo approach.
Copyright © 2022 Ebrahim and Cengiz.

Entities:  

Keywords:  Hamiltonian Monte Carlo; Verbal Learning Test; hierarchical; model; predicting

Year:  2022        PMID: 35496170      PMCID: PMC9046850          DOI: 10.3389/fpsyg.2022.855379

Source DB:  PubMed          Journal:  Front Psychol        ISSN: 1664-1078


1. Introduction

Verbal learning and memory tests are highly varied among older-aged adults due to various influences. Early cognitive intervention in older adults is a critical program to reduce the future risk of dementia (Thomas et al., 2019). The efficacy of the Chinese form Hopkins Verbal Learning Test (HVLT) for screening dementia and mild cognitive impairment in a Chinese population showed that HVLT scores were affected by age, education, and sex (Shi et al., 2012). The dataset of Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) study consists of two hierarchies in which four different repeated measures are nested within each participant (Luo and Wang, 2014). The outcome measures of the cognitive training interventions were the total HVLT from three learning trials and the baseline measure (Gross, 2011). Bayesian logistic and hierarchical probit models of accuracy data that allow two levels of mixed-effects in repeated-measures designs have been implemented. The Bayes factor through the Bayesian information criterion estimate and the Widely applicable information criterion (WAIC) model selection techniques were used (Song et al., 2017). Duff (2016) used stepwise regression model to scrutinize the effect of age, education, and gender on HVLT scores in 290 cognitively intact older adults. The study revealed that age was negatively correlated with the HVLT score, while education status was positively correlated. Moreover, there were fewer gender differences among four repeatedly measured verbal learning tests (Lekeu et al., 2009). Another study showed that besides capabilities through training, personal characteristics like age, unmarried status, and lower occupational cognitive requirements increased the likelihood of cognitive risk (Silva et al., 2012). Higher educational levels and active engagement in exercise may contribute to cognitive reserve and have a protective effect on cognitive decline in late life (Shen et al., 2021). Gender effects on neuropsychological performance were negligible when the age and educational status of elderly people were controlled (Welsh-Bohmer et al., 2009). Recently, the Markov chain Monte Carlo (MCMC) methods have been widely used to generate samples from complicated and high-dimensional distributions (Hadfield, 2017). Among all Bayesian computational methods, the Hamiltonian Monte Carlo (HMC) (Almond, 2014) approach is the most efficient for approximating complex data structure models and converges faster than the traditional Metropolis-Hastings and Gibbs methods (Kruschke and Vanpaemel, 2015). The common MCMC approaches show poor performance and tremendously slow convergence in complex parameter structures (Yao and Stephan, 2021). The HVLT is the ultimate in situations calling for multiple neuropsychological assessments (Benedict et al., 1998). Classical statistical inferences and single-level models have limitations for predicting naturally nest data. Bayesian hierarchical models (Congdon, 2020) were able to predict verbal learning test and memory scores from baseline personal characteristics, such as age, gender, cognitive status [mini-mental state exam (MMSE) score], years of education, and participants’ booster training and reasoning ability measured by training progress (Kuslansky et al., 2004). In Bayesian inference, the WAIC, the leave one out information criterion (LOO-IC), and K-fold cross-validation (K-fold-CV) are recently developed measures of complexity penalized fitting models (Almond, 2014; Sivula et al., 2020). In this article, model comparisons and model selections were performed using these three methods under the Bayesian Regression Models using ‘Stan’ (BRMS) package of R (Bürkner, 2018). In most cases, WAIC and LOO-IC showed a slight preference for the random slope model over other models (Bürkner, 2018). However, the general model selection principle shows to choose the null model when diffuse priors are used in the parameters to be included or rejected by the algorithms (Liu, 2000). Therefore, in this article, we used the HMC approach to fit the three different Bayesian hierarchical models and select the best predictive model.

2. Materials and Methods

2.1 Data and Variables

The ACTIVE study was a randomized controlled trial conducted in 1999–2001 at six diverse research centers in the United States and organized by the New England Research Institutes (NERI). A total of 1,575 purposively selected older adults were included in this study (Willis et al., 2015), in which 26% of the participants were African American. The ACTIVE dataset accessed from the study of Willis et al. (2015) has 13 variables. However, this modeling paper used six explanatory variables, and the dependent variable HVLT is used as repeated measures of learning tests and memory ability. In this dataset, HVLT has four different repeated measurement scores doi: 10.3886/ICPSR04248.v3.

2.2 Bayesian Hierarchical Model for Repeated Measures Data

Suppose X is the matrix of explanatory variables, and Y is the outcome variable that is the Total Hopkins Verbal Learning Test Score (). Besides the classical statistics, a more flexible Bayesian model is required that can accommodate the varying correlation between covariates and independent variables that occur in repeated measures-type longitudinal data. The general form of the Bayesian hierarchical model for repeated measures data can be expressed as: Where Y denotes the vector of outcome variable; β denotes a vector of fixed effects parameters; U denotes a vector of associated random effects (specifictoeachsubject); X is a matrix of covariates (explanatory variables); Z denotes a block diagonal matrix of covariates for the random effects as a complement of X embraced of m blocks that each block has n × q dimension matrix and ε denote a column vector of residuals. We assumed that the random effects U∼N(0, Ω) and the residuals . Where U and ε are independently distributed. Based on the unknown vector of φ and φ, the unknown random effects in Ω and R can be written as Σ = (φΩ, φR) (Laird and Ware, 1982). Where X is divided into two columns corresponding to fixed effects and a corresponding random effects design matrix denoted as and , respectively. And the parameters are divided into fixed effects β( and random effects β( = U. Cov(u,u) = Var(u) = Ω and It can be assumed that the hyperparameters of both the intercept and the coefficient/slope model have uniform hyper-prior distributions with appropriate assumptions for the parameters μ, μβ, σ, σβ ve ρ. Then, the mathematical form of the three possible Bayesian hierarchical models (Nalborczyk and Vilain, 2019) for predicting the verbal learning and memory test with two (group/subject and time) random effects (Hilbe, 2009) can be written as follows:

Model 1: Null Model

Here, the model is fitted by varying the intercept without including any predictor variable. Thus, this model shows the overall within and between-subject variations of the outcome variable (Goldstein et al., 2009).

Model 2: Varying Intercept Model

Here, the BRMS command is fitted in R with varying intercepts for both clusters (i.e., participating subjects) and repeated measures (i.e., measurement time point) by including all predictor variables in the model. Thus, this model can be called a random intercept and fixed slope model (McGlothlin and Viele, 2018).

Model 3: Varying Slopes

Here, we can focus on examining the dependence between the random intercepts and the random coefficients (Bafumi and Gelman, 2011). In this case, we are interested in whether the effects of age and reasoning skill have correlations with variations in verbal and memory test skills measured by trail scores. Where S is the covariance matrix, is the corresponding correlation matrix, and ρ is the association between intercepts and coefficients used in the calculation of S. The prior matrix R is the LKJ-correlation (Lewandowski et al., 2009) with a parameter ζ(zeta) which regulates the strength of the association. As shown in Figure 1 above, each component of the mixed effect model appears in the graph as a node. The dotted arrows represent deterministic (fixed) dependencies between the parameters (e.g., from β to μ), whereas the solid arrows represent probabilistic (random) dependencies (e.g., from to Y) (Bürkner, 2018). The hyper-parameters of the varying both intercept and slope model (μα, μβ, σα, σβ, and ρ) can be assumed to have hyper-prior distributions with appropriate assumptions for the parameters (Liu, 2016; Congdon, 2020).
FIGURE 1

A varying intercept and slope model (Bayesian Framework).

A varying intercept and slope model (Bayesian Framework).

2.3 Bayesian Information Criterion for Model Comparison and Selection

Watanabe’s Widely Applicable (WAIC)

WAIC (Watanabe, 2010) could be achieved as an improvement over the divergence-based information criterion (DIC) for Bayesian models. The deviation term used in the calculation of the WAIC is Log-Point Based -Requires Predictive-Density (LNTTY). LNTTY is calculated as: The whole p(θ) is the posterior distribution used in the calculation of LNTTY. Similar to LNTTY, WAIC’s penalty term is purely Bayesian and is computed as: Where p is the penalty term which is the variance of the log-predictive-density terms aggregated over N data points. Thus, the WAIC can be calculated as:

Leave-One-Out Information Criteria (LOO-CV)

Bayesian leave-one-out cross-validation (LOO-CV) is different from the WAIC. Because there is no penalty term in its calculation. LOO-CV can be computed as: Where p(θ) is the posterior distribution based on a sub-set of the data at point i from the dataset. LNTTY used i data points to calculate both the posterior distribution and the parameter estimation. Here, in contrast, the log-pointwise predictive density (LPPD) is used the same for prediction only. Therefore, there is no need for a penalty term to correct potential bias by using the data twice (Vehtari et al., 2017).

K-Fold Cross-Validation

Sometimes, multiple Pareto Corrected Significance Sampling (PSIS-LOO) fails, and it takes too long to remodel in the iteration. Therefore, we can estimate LOO-CV using K-fold-CV by separating the data into completely random multiples, which leads to looking at each cross-validation estimate distinctly (Vehtari et al., 2018). The Bayesian K-fold-CV partitions the dataset into k subsets y(k = 1, 2, …, K). The Bayesian hierarchical model (BHM) generates each training dataset y separately, which returns a p(θ) = p(θ|y() posterior distribution (Vehtari and Gelman, 2014). To preserve reliability with WAIC and LOO-IC, defining the predictive accuracy of every point in the dataset is essential. Therefore, the log-predictive distribution function is Using “S” simulations corresponding to a subset of k (usually K = 10) containing the i data point and the posterior distribution P(θ|y(). The overall estimate of the expected log point predictive density for a new dataset is determined as follows: Therefore, a point estimate of the k-fold value is the sum of the iterative folds from the data points.

2.4 The Hamiltonian Monte Carlo Algorithm in Bayesian Regression Models Using ‘Stan’ Package of R

Similar to Gibbs sampling, HMC practices a proposal distribution that changes subject to the recent location in the parameter space (Liu, 2000). However, unlike the Gibbs algorithm, HMC does not rely on computing the conditional posterior distribution of parameters and sampling from it (Mai and Zhang, 2018). HMC has two advantages over other MCMC methods: little or no autocorrelation of the samples and fast mixing, i.e., the chain converges to the distribution immediately (Nalborczyk and Vilain, 2019). Therefore, it is the best approach for continuous distributions with low (auto) correlation and low rejection of samples. When the model parameters are continuous rather than discrete, HMC, also known as Hybrid Monte Carlo, can overpower such random walk behavior using a clever scheme of supplementary variables that converts the tricky of sampling from the targeted function into the simulating Hamiltonian dynamics (Britten et al., 2021). HMC is an MCMC algorithm that avoids the random walk behavior and sensitivity to correlated parameters that outbreak other MCMC approaches by performing a series of steps informed by first-order gradient information (Hilbe, 2009). The HMC algorithm is based on the Hamiltonian (total energy) calculating the trajectory for a time t = 0, …, T and then taking the final position X(T) = X. The steps of the algorithm are as follows: HMC algorithm Choose a starting point and a velocity distribution θ0 = X0q(v) for n = 0, … Set the initial position as X(t = 0) = X Draw a random initial velocity, v(t = 0)∼q(v); Integrate the orbit numerically with the total energy for some time (use the Leapfrog method): Calculate the probability of acceptance: Set X = X(t = T) Increment

3. Results

In practice, the three basic Bayesian hierarchical models have been fitted in BRMS default settings, and population-level (fixed) effects and subject-level (random) effects were obtained (Luo et al., 2021). All three models (Models 1, 2, and 3) had both fixed and random (mixed) parts but with different estimated parameter types. In the result, the estimate shows the posterior mean and Est. Error is the SD for each parameter. Model convergence was achieved well enough both the bulk effective sample size (Bulk_ESS) and the tail effective sample size (Tail_ESS) for the 95% CIs were adequate (Vehtari et al., 2017). In general, every parameter is summarized using the posterior distribution’s mean (“Estimate”) and SD (“Est. Error”), as well as two-sided 95% credible intervals as lower and upper bounds based on quintiles. Table 1 of the fixed effects shows that the posterior mean verbal testing score was estimated to be 26.33 with an SD of 0.73. The 95% credible interval shows that the posterior distribution mean (intercept) was significant. On the other hand, the random effect showed significant verbal score test variation between groups (participant subjects) and within-subjects (between different measurements of different time points). Thus, according to the null model, the HVLT score showed more between-group/subject variation than within-group (between repeated measurements) variation.
TABLE 1

Results from the fitted null model: Model 1.

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS R^ 95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept26.33120.7331137118751.0124.850127.7214
Random EffectsLowerUpper
σintercept,subject4.31050.085281014501.004.15244.4751
σintercept,time1.30350.6456204724291.000.57543.0562
σe(sigma)3.11340.0256331532961.013.04623.1662
Results from the fitted null model: Model 1. Table 2 showed that the coefficient of booster training was positive with a zero overlapping 95% CI. This indicates that, on average, there is little evidence that taking booster training increases elderly adults’ verbal learning and memory test scores by 0.1865, but the evidence-based on the data and random intercept model. On the other hand, adults’ years of education (edu) estimate was negative with a zero overlapping 95% CI. This negative estimate indicates that, on average, in the random intercept model, there is little evidence that increasing the years of education decreases elderly adults’ verbal learning and memory test scores by 0.0034 units.
TABLE 2

Results from the fitted varying intercept model: Model 2.

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS R^ 95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept9.23141.941112601891.005.471212.9510
Age−0.12110.021292617021.01−0.1611−0.0854
Edu−0.00340.0011413928381.01−0.01010.0042
Booster0.18650.175464518381.00−0.15110.5432
Gender2.65640.201591016071.002.27523.0654
Reason0.14640.011298016731.000.13100.4232
MMSE0.60120.0462103221281.000.50130.7012
Random effectsLowerUpper
σintercept,subject3.03120.0654114622711.002.88453.1645
σintercept,time1.26540.6572185221211.000.58323.0812
σe(sigma)3.11020.0312426430291.003.04623.1761
Results from the fitted varying intercept model: Model 2. According to the predictive effects of each explanatory variable shown in Figure 2 and Table 3, taking booster training, age, and gender were the most influential factors affecting participants’ cognitive verbal test and memory ability. Table 3 reveals that there is also an adverse association between the intercepts and coefficients for reasoning ability, which implies reasoning ability has a large average score value showing additional variability by poor reasoning ability than by good reasoning ability. Nevertheless, it can be seen that the slope estimate of such a model is even further unreliable than that of the preceding models, as it can be clearly understood from the associated standard error and the size of the 95% CIs. Table 3 also showed that booster training had a significant positive predictive effect on elderly adults’ verbal learning and memory test scores. In contracts, adults’ years of education had a significant negative impact on elderly adults’ verbal learning and memory test scores.
FIGURE 2

Bayesian hierarchical varying slope convergence diagnosis.

TABLE 3

Results from the fitted varying slope mode: Model 3.

Outcome variableCovariates
EstimateEst. ErrorBulk_ESSTail_ESS R^ 95% CI
Fixed effectsLowerUpper
Total hopkins verbal learning test score (THVLTS)Intercept9.84122.0602315729181.005.852313.9344
Age−0.12110.0213284627201.01−0.1513−0.0823
Edu (education)−0.00330.0012552327701.00−0.01220.0012
Booster0.14120.1703336228761.00−0.21450.4831
Gender2.55050.2004323628661.012.17122.9331
Reason0.14440.0113308728671.000.13130.4402
MMSE0.58030.0512325630421.000.48220.6840
Random effectsLowerUpper
σintercept,subject1.92221.28331114881.000.07244.3111
σintercept,time1.30220.8004202722701.000.57023.1343
σage0.04240.01331008331.000.01230.0625
σreason0.04050.01321383911.000.02120.0732
cor Intercept,age 0.10330.43331112551.00−0.70420.8303
cor Intercept,reason −0.39020.42041003881.00−0.90110.6212
cor age,reason −0.59220.264551910531.00−0.90420.1407
σe(sigma)3.11020.0333376727481.003.05333.1710
Bayesian hierarchical varying slope convergence diagnosis. Results from the fitted varying slope mode: Model 3. We also noticed in Figure 2 and Figure 3 below that adding any term to the early model showed predictive performance improvements on the fitted models are ordered from Models 1 to 3 (full model). However, such a result may not be interpreted as a universal rule, subsequent adding extra terms to a unique model may also result in overfitting, which corresponds to a condition in which the fitted model is over-specified about the data, making the model good at clarifying the sample dataset but poor at predicting no observed data. The model convergence diagnosis plots are hairy caterpillars which showed the model converged. On the other hand, the models have well converged based on the estimated statistical values. This means that the R-hat statistics were close to 1 and the (bulk and tail) ESSs values were sufficiently high when ESS > 100 was chosen as the cutoff (Vehtari et al., 2021). The majority of parameters still showed sufficiently high ESS values when more conservative cutoffs were chosen (i.e., ESS > 400 or even 1,000, see Zitzmann and Hecht, 2019).
FIGURE 3

Bayesian hierarchical varying slope convergence diagnosis (Continuous).

Bayesian hierarchical varying slope convergence diagnosis (Continuous). Based on the fitted varying slope model, which accounted for six predictors from the data, fixed effects showed that age, gender, reasoning ability, and booster training were significant predictors of verbal learning and memory test scores, whereas random-effect showed that much of the variation in test scores occurred within-subjects (between measurement time points) than between subjects. After we have built the three different models, it is necessary to identify relatively the best model that can be used to predict the outcome variable and make inferences. However, choosing the model that has the best predictive and a better fit on the actual data is complicated with diverse information criteria since all selected models on the actual data might not essentially achieve as fit on a different dataset. In its place, it is necessary to decide on a model that fits best in terms of predicting new data which had not been practiced. In case of the non-existence of extra information, cross-validation methods such as WAIC and LOO-CV can be used. According to Table 4, the varying slope model has the lowest WAIC, LOO-IC, and 10-fold estimates. However, the difference is relatively small when we compare the difference in estimates of criteria for each model and the corresponding standard errors (in the column SE).
TABLE 4

Model comparisons based on predictive performance.

Model typeModel selection criteria from BRMS package
WAIC
LOO-IC
10-fold
EstimateSEEstimateSEEstimateSE
Null model (Model 1)33638.0134.633744.1136.233923.8136.4
Varying Intercept model (Model 2)33494.5139.533574.9140.633717.0141.4
Varying slopes model (Model 3)33488.4141.833567.5143.033685.2140.8
Model comparisons based on predictive performance. Among the fitted models above, it looks like the final model (Model 3) in the HMC algorithm is the best model. Therefore, as a function of the six explanatory variables and the random coefficient for age and reasoning ability, Model 3 has the best predictive performance for the cognitive HVLT. According to Figure 4, the varying slope and intercept model fit well and produced nearly identical posterior observed density and posterior predictive distribution plots of the outcome variable of THVLTS from the ACTIVE study.
FIGURE 4

Bayesian hierarchical varying slope fitted model on the observed and predicted outcomes.

Bayesian hierarchical varying slope fitted model on the observed and predicted outcomes. Furthermore, the marginal effect of each predictor variable revealed (Figure 5) that age and reasoning skills are the most significant explanatory variables that predict the THVLTS of the ACTIVE study.
FIGURE 5

Bayesian hierarchical varying slope model marginal prediction effects.

Bayesian hierarchical varying slope model marginal prediction effects.

4. Discussion

Based on the selected sample participants in the ACTIVE study dataset (Willis et al., 2015), the Bayesian hierarchical linear models of three types were fitted by considering only six explanatory variables as predictors of the cognitive verbal learning test. The null model without any predictor effect but with only the intercept term was fitted, and it shows a mass of cognitive verbal learning ability variability across subjects. The varying intercept model with the addition of all predictor variables was fitted; and getting booster training, age, and reasoning ability were significant predictor of verbal test scores (Duff, 2016). The varying coefficient/slope model (i.e., Model 3) is the best-fitted model than the other fitted models since it had the lowest WAIC, LOO-IC, and 10-fold estimates (Bafumi and Gelman, 2011). A bulk of participants’ cognitive verbal test scores variations were observed between subjects (Ryoo, 2011). The full hierarchical model with varying intercepts and slopes has the best performance for predicting verbal learning tests (from ACTIVE study dataset) using the hybrid Hamiltonian Markov Chain Monte Carlo approach. Socio-demographic and training-related characteristics influence elderly verbal learning tests that can be measured in multiple occupations (Welsh-Bohmer et al., 2009).

5. Conclusion

Total Hopkins Verbal Learning Test Score from the ACTIVE study can be used as a measure of elderly adults’ cognitive verbal learning ability. Four demographic characteristics of adults, such as age, gender, educational status, and cognitive status (MMSE score), were measured at the baseline, and characteristics measured after cognitive training such as reasoning ability and booster training were considered. THVLTS from the ACTIVE study can be used as a measure of elderly adults’ cognitive verbal learning ability. According to the findings, the varying intercept and slope model fit best, and age, gender, booster, and reasoning ability are the main significant predictors for THVLTS, which measures cognitive verbal learning. Taking booster training had a positive significant predictive effect, while years of education (edu) had a negative significant predictive effect on THVLTS.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/supplementary material.

Ethics Statement

Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements. Written informed consent was not obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article. This is because this quantitative analysis and modeling paper used open-access secondary data on repeated measurements.

Author Contributions

EE participated in all aspects of the study: designing the study, performing data management, conducting the data analysis, writing the first draft of the manuscript, and discussing with MC to improve the manuscript, as it is a part of the first author’s Ph.D. dissertation. MC participated in revising the manuscript, commenting, and proofreading. Both authors listed have made a substantial, direct, and intellectual contribution to the manuscript and approved it for publication.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
APPENDIX TABLE 1

Posterior estimates with the verity of priors: Sensitivity analysis results.

Alternative priorsParameter/CovariatesEstimate (SD)Median (50%)5–95% HDPDefault estimate (SD)Percentage deviation
Alternative prior I: Half- Cauchy (0,1)Intercept9.8412 (1.521)9.83315.8523, 13.93449.8321 (1.932)0.0926

Age−0.1211 (0.021)−0.1201−0.1513, −0.0823−0.1223 (0.423)−0.9812
Edu (education)−0.0033 (0.001)−0.0033−0.0122, 0.0012−0.0034 (0.005)−2.9412
Booster0.1412 (0.102)0.1413−0.2145, 0.48310.1411 (0.623)0.0709
Gender2.5505 (0.112)2.55042.1712, 2.93312.5487 (0.222)0.0706
Reason0.1444 (1.902)0.14430.1313, 0.44020.1443 (2.081)0.0693
MMSE0.5803 (0.028)0.59210.4822, 0.68400.5801 (0.082)0.0345
σintercept,subject1.92221.92210.0724, 4.31111.92120.0521
σintercept,time1.30221.31020.5702, 3.13431.3032−0.0767
σage0.04240.04030.0123, 0.06250.04250.0126
σreason0.04050.04020.0212, 0.07320.0401−0.9975
cor Intercept,age 0.10330.1032−0.7042, 0.83030.10310.1040
cor Intercept,reason −0.3902−0.8902−0.9011, 0.6212−0.3904−0.0512
cor age,reason −0.5922−0.5887−0.9042, 0.1407−0.5923−0.0169
σe(sigma)3.11023.20413.0533, 3.17103.1112−0.0321
ParameterEstimate (SD)Median (50%)5–95% HDPDefault estimate (SD)Percentage deviation

Alternative prior II: Normal (5, 0.01)Intercept9.8423 (1.543)9.82316.1415, 13.65529.8321 (1.932)−0.0112

Age−0.1212 (0.034)−0.1212−0.1514, −0.0855−0.1223 (0.423)−0.0825
Edu (education)−0.0034 (0.011)−0.0034−0.0124, 0.0015−0.0034 (0.005)−2.9412
Booster0.1413 (0.124)0.1412−0.2165, 0.48710.1411 (0.623)−0.0708
Gender2.5514 (0.142)2.55052.1722, 2.93712.5487 (0.222)−0.0353
Reason0.1445 (2.013)0.14440.1453, 0.45620.1443 (2.081)−0.0692
MMSE0.5802 (0.035)0.58720.4852, 0.69510.5801 (0.082)0.0172
σintercept,subject1.92131.92210.0724, 4.34131.92120.0468
σintercept,time1.30331.31020.5622, 3.13441.3032−0.0844
σage0.04250.04030.0123, 0.06270.0425−0.2353
σreason0.04060.04020.0212, 0.07330.0401−0.2463
cor Intercept,age 0.10340.1032−0.7044, 0.83040.10310.0969
cor Intercept,reason −0.3903−0.8902−0.9021, 0.6217−0.3904−0.0256
cor age,reason −0.5923−0.5987−0.8045, 0.1404−0.5923−0.0169
σe(sigma)3.21153.21413.0533, 3.57103.1112−3.1543

The relative percentage deviation can be computed as: {[(estimate using new alternative prior)–(estimate using default/reference prior)]/ (estimate using default/reference prior)}*100. Interpreting percentage deviation results is largely subjective and dependent on the metric of the parameters. However, percentage deviation under 10% would likely be considered negligible (

  16 in total

1.  Improving transparency and replication in Bayesian statistics: The WAMBS-Checklist.

Authors:  Sarah Depaoli; Rens van de Schoot
Journal:  Psychol Methods       Date:  2015-12-21

2.  Model Selection with the Linear Mixed Model for Longitudinal Data.

Authors:  Ji Hoon Ryoo
Journal:  Multivariate Behav Res       Date:  2011-07-29       Impact factor: 5.923

3.  Bayesian Hierarchical Models.

Authors:  Anna E McGlothlin; Kert Viele
Journal:  JAMA       Date:  2018-12-11       Impact factor: 56.272

4.  Random-effects models for longitudinal data.

Authors:  N M Laird; J H Ware
Journal:  Biometrics       Date:  1982-12       Impact factor: 2.571

5.  Demographically corrected normative data for the Hopkins Verbal Learning Test-Revised and Brief Visuospatial Memory Test-Revised in an elderly sample.

Authors:  Kevin Duff
Journal:  Appl Neuropsychol Adult       Date:  2015-10-23       Impact factor: 2.248

6.  The California Verbal Learning Test and other standard clinical neuropsychological tests to predict conversion from mild memory impairment to dementia.

Authors:  Françoise Lekeu; Delphine Magis; Patricia Marique; Xavier Delbeuck; Sophie Bechet; Bénédicte Guillaume; Stéphane Adam; Jean Petermans; Gustave Moonen; Eric Salmon
Journal:  J Clin Exp Neuropsychol       Date:  2009-05-20       Impact factor: 2.475

7.  Neuropsychological performance in advanced age: influences of demographic factors and Apolipoprotein E: findings from the Cache County Memory Study.

Authors:  Katheen A Welsh-Bohmer; Truls Ostbye; Linda Sanders; Carl F Pieper; Kathleen M Hayden; JoAnn T Tschanz; Maria C Norton
Journal:  Clin Neuropsychol       Date:  2008-06-10       Impact factor: 3.535

8.  Markov chain Monte Carlo methods for hierarchical clustering of dynamic causal models.

Authors:  Yu Yao; Klaas E Stephan
Journal:  Hum Brain Mapp       Date:  2021-04-07       Impact factor: 5.038

9.  Limitations of Bayesian Leave-One-Out Cross-Validation for Model Selection.

Authors:  Quentin F Gronau; Eric-Jan Wagenmakers
Journal:  Comput Brain Behav       Date:  2018-09-27

10.  The utility of the Hopkins Verbal Learning Test (Chinese version) for screening dementia and mild cognitive impairment in a Chinese population.

Authors:  Jing Shi; Jinzhou Tian; Mingqing Wei; Yingchun Miao; Yongyan Wang
Journal:  BMC Neurol       Date:  2012-11-07       Impact factor: 2.474

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.