BACKGROUND: Data-adaptive approaches to confounding adjustment may improve performance beyond expert knowledge when analyzing electronic healthcare databases and have additional practical advantages for analyzing multiple databases in rapid cycles. Improvements seemed possible if outcome predictors were reliably identified empirically and adjusted. METHODS: In five cohort studies from diverse healthcare databases, we implemented a base-case high-dimensional propensity score algorithm with propensity score decile-adjusted outcome models to estimate treatment effects among prescription drug initiators. The original variable selection procedure based on the estimated bias of each variable using unadjusted associations between confounders and exposure (RRCE) and disease outcome (RRCD) was augmented by alternative strategies. These included using increasingly adjusted RRCD estimates, including models considering >1,500 variables jointly (Lasso, Bayesian logistic regression); using prediction statistics or likelihood-ratio statistics for covariate prioritization; directly estimating the propensity score with >1,500 variables (Lasso, Bayesian regression); or directly fitting an outcome model using all covariates jointly (Lasso, Ridge). RESULTS: In five example studies, most tested augmentations of the base-case hdPS did not meaningfully change estimates in light of wide confidence intervals except for Bayesian regression and Lasso to estimate RRCD, which moved estimates minimally closer to the expectation in three of five examples. The direct outcome estimation with Lasso performed worst. CONCLUSION: Overall, the basic heuristic of variable reduction in high-dimensional propensity score adjustment performed, as well as alternative approaches in diverse settings. Minor improvements in variable selection may be possible using Bayesian outcome regression to prioritize variables for propensity score estimation when outcomes are rare. See video abstract at, http://links.lww.com/EDE/B162.
BACKGROUND: Data-adaptive approaches to confounding adjustment may improve performance beyond expert knowledge when analyzing electronic healthcare databases and have additional practical advantages for analyzing multiple databases in rapid cycles. Improvements seemed possible if outcome predictors were reliably identified empirically and adjusted. METHODS: In five cohort studies from diverse healthcare databases, we implemented a base-case high-dimensional propensity score algorithm with propensity score decile-adjusted outcome models to estimate treatment effects among prescription drug initiators. The original variable selection procedure based on the estimated bias of each variable using unadjusted associations between confounders and exposure (RRCE) and disease outcome (RRCD) was augmented by alternative strategies. These included using increasingly adjusted RRCD estimates, including models considering >1,500 variables jointly (Lasso, Bayesian logistic regression); using prediction statistics or likelihood-ratio statistics for covariate prioritization; directly estimating the propensity score with >1,500 variables (Lasso, Bayesian regression); or directly fitting an outcome model using all covariates jointly (Lasso, Ridge). RESULTS: In five example studies, most tested augmentations of the base-case hdPS did not meaningfully change estimates in light of wide confidence intervals except for Bayesian regression and Lasso to estimate RRCD, which moved estimates minimally closer to the expectation in three of five examples. The direct outcome estimation with Lasso performed worst. CONCLUSION: Overall, the basic heuristic of variable reduction in high-dimensional propensity score adjustment performed, as well as alternative approaches in diverse settings. Minor improvements in variable selection may be possible using Bayesian outcome regression to prioritize variables for propensity score estimation when outcomes are rare. See video abstract at, http://links.lww.com/EDE/B162.
Authors: Elizabeth Wall-Wieler; Leslie L Roos; James Bolton; Marni Brownell; Nathan Nickel; Dan Chateau Journal: Can J Psychiatry Date: 2017-10-29 Impact factor: 4.356
Authors: Patrick Blin; Caroline Dureau-Pournin; Yves Cottin; Jacques Bénichou; Patrick Mismetti; Abdelilah Abouelfath; Regis Lassalle; Cécile Droz; Nicholas Moore Journal: Br J Clin Pharmacol Date: 2018-12-16 Impact factor: 4.335
Authors: Alejandro Schuler; David A Wulf; Yun Lu; Theodore J Iwashyna; Gabriel J Escobar; Nigam H Shah; Vincent X Liu Journal: Crit Care Med Date: 2018-06 Impact factor: 7.598
Authors: Hasmik Grigoryan; Courtney Schiffman; Marc J Gunter; Alessio Naccarati; Silvia Polidoro; Sonia Dagnino; Sandrine Dudoit; Paolo Vineis; Stephen M Rappaport Journal: Cancer Res Date: 2019-10-22 Impact factor: 12.701
Authors: Cheng Ju; Richard Wyss; Jessica M Franklin; Sebastian Schneeweiss; Jenny Häggström; Mark J van der Laan Journal: Stat Methods Med Res Date: 2017-12-11 Impact factor: 3.021
Authors: Lauren M Petrick; Courtney Schiffman; William M B Edmands; Yukiko Yano; Kelsi Perttula; Todd Whitehead; Catherine Metayer; Craig E Wheelock; Manish Arora; Hasmik Grigoryan; Henrik Carlsson; Sandrine Dudoit; Stephen M Rappaport Journal: Cancer Lett Date: 2019-03-20 Impact factor: 8.679
Authors: Elizabeth Wall-Wieler; Thalia K Robakis; Carolyn E Cesta; Reem Masarwa; Deirdre J Lyell; Can Liu; Robert W Platt; Suzan L Carmichael Journal: Can J Psychiatry Date: 2020-05-21 Impact factor: 4.356