Dunlei Cheng1, Adam J Branscum, James D Stamey. 1. Institute for Health Care Research and Improvement, Dallas, TX 75206, USA. dunleic@baylorhealth.edu <dunleic@baylorhealth.edu>
Abstract
PURPOSE: To quantify the impact of ignoring misclassification of a response variable and measurement error in a covariate on statistical power, and to develop software for sample size and power analysis that accounts for these flaws in epidemiologic data. METHODS: A Monte Carlo simulation-based procedure is developed to illustrate the differences in design requirements and inferences between analytic methods that properly account for misclassification and measurement error to those that do not in regression models for cross-sectional and cohort data. RESULTS: We found that failure to account for these flaws in epidemiologic data can lead to a substantial reduction in statistical power, over 25% in some cases. The proposed method substantially reduced bias by up to a ten-fold margin compared to naive estimates obtained by ignoring misclassification and mismeasurement. CONCLUSIONS: We recommend as routine practice that researchers account for errors in measurement of both response and covariate data when determining sample size, performing power calculations, or analyzing data from epidemiological studies. 2010 Elsevier Inc. All rights reserved.
PURPOSE: To quantify the impact of ignoring misclassification of a response variable and measurement error in a covariate on statistical power, and to develop software for sample size and power analysis that accounts for these flaws in epidemiologic data. METHODS: A Monte Carlo simulation-based procedure is developed to illustrate the differences in design requirements and inferences between analytic methods that properly account for misclassification and measurement error to those that do not in regression models for cross-sectional and cohort data. RESULTS: We found that failure to account for these flaws in epidemiologic data can lead to a substantial reduction in statistical power, over 25% in some cases. The proposed method substantially reduced bias by up to a ten-fold margin compared to naive estimates obtained by ignoring misclassification and mismeasurement. CONCLUSIONS: We recommend as routine practice that researchers account for errors in measurement of both response and covariate data when determining sample size, performing power calculations, or analyzing data from epidemiological studies. 2010 Elsevier Inc. All rights reserved.
Authors: Sheng Luo; Wenyaw Chan; Michelle A Detry; Paul J Massman; Rachelle S Doody Journal: Stat Methods Med Res Date: 2012-03-15 Impact factor: 3.021
Authors: David Goldberg; Benjamin French; James Trotter; Kirti Shetty; Thomas Schiano; K Rajender Reddy; Scott D Halpern Journal: Transplantation Date: 2013-07-27 Impact factor: 4.939