Marine Corbin1,2, Stephen Haslett1,3,4, Neil Pearce1,5, Milena Maule2, Sander Greenland6. 1. Centre for Public Health Research, Massey University, Wellington, New Zealand. 2. Unit of Cancer Epidemiology, Department of Medical Sciences, University of Turin, Turin, Italy. 3. Statistical Consulting Unit, Australian National University, Camberra, Australia. 4. Institute of Fundamental Sciences, Massey University, Palmerston North, New Zealand. 5. Department of Medical Statistics, London School of Hygiene and Tropical Medicine, London, UK. 6. Department of Epidemiology and Department of Statistics, University of California, Los Angeles, CA, USA.
Abstract
Purpose: Measurement error is an important source of bias in epidemiological studies. We illustrate three approaches to sensitivity analysis for the effect of measurement error: imputation of the 'true' exposure based on specifying the sensitivity and specificity of the measured exposure (SS); direct imputation (DI) using a regression model for the predictive values; and adjustment based on a fully Bayesian analysis. Methods: We deliberately misclassify smoking status in data from a case-control study of lung cancer. We then implement the SS and DI methods using fixed-parameter (FBA) and probabilistic (PBA) bias analyses, and Bayesian analysis using the Markov-Chain Monte-Carlo program WinBUGS to show how well each recovers the original association. Results: The 'true' smoking-lung cancer odds ratio (OR), adjusted for sex in the original dataset, was OR = 8.18 [95% confidence limits (CL): 5.86, 11.43]; after misclassification, it decreased to OR = 3.08 (nominal 95% CL: 2.40, 3.96). The adjusted point estimates from all three approaches were always closer to the 'true' OR than the OR estimated from the unadjusted misclassified smoking data, and the adjusted interval estimates were always wider than the unadjusted interval estimate. When imputed misclassification parameters departed much from the actual misclassification, the 'true' OR was often omitted in the FBA intervals whereas it was always included in the PBA and Bayesian intervals. Conclusions: These results illustrate how PBA and Bayesian analyses can be used to better account for uncertainty and bias due to measurement error.
Purpose: Measurement error is an important source of bias in epidemiological studies. We illustrate three approaches to sensitivity analysis for the effect of measurement error: imputation of the 'true' exposure based on specifying the sensitivity and specificity of the measured exposure (SS); direct imputation (DI) using a regression model for the predictive values; and adjustment based on a fully Bayesian analysis. Methods: We deliberately misclassify smoking status in data from a case-control study of lung cancer. We then implement the SS and DI methods using fixed-parameter (FBA) and probabilistic (PBA) bias analyses, and Bayesian analysis using the Markov-Chain Monte-Carlo program WinBUGS to show how well each recovers the original association. Results: The 'true' smoking-lung cancer odds ratio (OR), adjusted for sex in the original dataset, was OR = 8.18 [95% confidence limits (CL): 5.86, 11.43]; after misclassification, it decreased to OR = 3.08 (nominal 95% CL: 2.40, 3.96). The adjusted point estimates from all three approaches were always closer to the 'true' OR than the OR estimated from the unadjusted misclassified smoking data, and the adjusted interval estimates were always wider than the unadjusted interval estimate. When imputed misclassification parameters departed much from the actual misclassification, the 'true' OR was often omitted in the FBA intervals whereas it was always included in the PBA and Bayesian intervals. Conclusions: These results illustrate how PBA and Bayesian analyses can be used to better account for uncertainty and bias due to measurement error.
Authors: Nastaran Hajizadeh; Ahmad Reza Baghestani; Mohamad Amin Pourhoseingholi; Sara Ashtari; Hadis Najafimehr; Luca Busani; Mohammad Reza Zali Journal: Galen Med J Date: 2019-07-09
Authors: Tricia L Larose; Florence Guida; Anouar Fanidi; Arnulf Langhammer; Kristian Kveem; Victoria L Stevens; Eric J Jacobs; Stephanie A Smith-Warner; Edward Giovannucci; Demetrius Albanes; Stephanie J Weinstein; Neal D Freedman; Ross Prentice; Mary Pettinger; Cynthia A Thomson; Qiuyin Cai; Jie Wu; William J Blot; Alan A Arslan; Anne Zeleniuch-Jacquotte; Loic Le Marchand; Lynne R Wilkens; Christopher A Haiman; Xuehong Zhang; Meir J Stampfer; Allison M Hodge; Graham G Giles; Gianluca Severi; Mikael Johansson; Kjell Grankvist; Renwei Wang; Jian-Min Yuan; Yu-Tang Gao; Woon-Puay Koh; Xiao-Ou Shu; Wei Zheng; Yong-Bing Xiang; Honglan Li; Qing Lan; Kala Visvanathan; Judith Hoffman Bolton; Per Magne Ueland; Øivind Midttun; Neil Caporaso; Mark Purdue; Howard D Sesso; Julie E Buring; I-Min Lee; J Michael Gaziano; Jonas Manjer; Hans Brunnström; Paul Brennan; Mattias Johansson Journal: Int J Epidemiol Date: 2018-12-01 Impact factor: 7.196
Authors: Franya Hutchins; Robert Krafty; Samar R El Khoudary; Janet Catov; Alicia Colvin; Emma Barinas-Mitchell; Maria M Brooks Journal: Epidemiology Date: 2021-03-01 Impact factor: 4.860
Authors: Rebecca M Joseph; Tjeerd P van Staa; Mark Lunt; Michal Abrahamowicz; William G Dixon Journal: Pharmacoepidemiol Drug Saf Date: 2018-09-28 Impact factor: 2.890
Authors: Sophie A Hamilton; Prashant Jarhyan; Daniela Fecht; Nikhil Srinivasapura Venkateshmurthy; Neil Pearce; Kabayam M Venkat Narayan; Mohammed K Ali; Viswanathan Mohan; Nikhil Tandon; Dorairaj Prabhakaran; Sailesh Mohan Journal: Environ Epidemiol Date: 2021-09-24