Yue Zhao1. 1. The University of Hong Kong, Pokfulam, Hong Kong SAR, China. myzhao@hku.hk.
Abstract
PURPOSE: In patient-reported outcome research that utilizes item response theory (IRT), using statistical significance tests to detect misfit is usually the focus of IRT model-data fit evaluations. However, such evaluations rarely address the impact/consequence of using misfitting items on the intended clinical applications. This study was designed to evaluate the impact of IRT item misfit on score estimates and severity classifications and to demonstrate a recommended process of model-fit evaluation. METHODS: Using secondary data sources collected from the Patient-Reported Outcome Measurement Information System (PROMIS) wave 1 testing phase, analyses were conducted based on PROMIS depression (28 items; 782 cases) and pain interference (41 items; 845 cases) item banks. The identification of misfitting items was assessed using Orlando and Thissen's summed-score item-fit statistics and graphical displays. The impact of misfit was evaluated according to the agreement of both IRT-derived T-scores and severity classifications between inclusion and exclusion of misfitting items. RESULTS: The examination of the presence and impact of misfit suggested that item misfit had a negligible impact on the T-score estimates and severity classifications with the general population sample in the PROMIS depression and pain interference item banks, implying that the impact of item misfit was insignificant. CONCLUSIONS: Findings support the T-score estimates in the two item banks as robust against item misfit at both the group and individual levels and add confidence to the use of T-scores for severity diagnosis in the studied sample. Recommendations on approaches for identifying item misfit (statistical significance) and assessing the misfit impact (practical significance) are given.
PURPOSE: In patient-reported outcome research that utilizes item response theory (IRT), using statistical significance tests to detect misfit is usually the focus of IRT model-data fit evaluations. However, such evaluations rarely address the impact/consequence of using misfitting items on the intended clinical applications. This study was designed to evaluate the impact of IRT item misfit on score estimates and severity classifications and to demonstrate a recommended process of model-fit evaluation. METHODS: Using secondary data sources collected from the Patient-Reported Outcome Measurement Information System (PROMIS) wave 1 testing phase, analyses were conducted based on PROMIS depression (28 items; 782 cases) and pain interference (41 items; 845 cases) item banks. The identification of misfitting items was assessed using Orlando and Thissen's summed-score item-fit statistics and graphical displays. The impact of misfit was evaluated according to the agreement of both IRT-derived T-scores and severity classifications between inclusion and exclusion of misfitting items. RESULTS: The examination of the presence and impact of misfit suggested that item misfit had a negligible impact on the T-score estimates and severity classifications with the general population sample in the PROMIS depression and pain interference item banks, implying that the impact of item misfit was insignificant. CONCLUSIONS: Findings support the T-score estimates in the two item banks as robust against item misfit at both the group and individual levels and add confidence to the use of T-scores for severity diagnosis in the studied sample. Recommendations on approaches for identifying item misfit (statistical significance) and assessing the misfit impact (practical significance) are given.
Authors: Bryce B Reeve; Ron D Hays; Jakob B Bjorner; Karon F Cook; Paul K Crane; Jeanne A Teresi; David Thissen; Dennis A Revicki; David J Weiss; Ronald K Hambleton; Honghu Liu; Richard Gershon; Steven P Reise; Jin-shei Lai; David Cella Journal: Med Care Date: 2007-05 Impact factor: 2.983
Authors: Honghu Liu; David Cella; Richard Gershon; Jie Shen; Leo S Morales; William Riley; Ron D Hays Journal: J Clin Epidemiol Date: 2010-08-05 Impact factor: 6.437
Authors: David Cella; William Riley; Arthur Stone; Nan Rothrock; Bryce Reeve; Susan Yount; Dagmar Amtmann; Rita Bode; Daniel Buysse; Seung Choi; Karon Cook; Robert Devellis; Darren DeWalt; James F Fries; Richard Gershon; Elizabeth A Hahn; Jin-Shei Lai; Paul Pilkonis; Dennis Revicki; Matthias Rose; Kevin Weinfurt; Ron Hays Journal: J Clin Epidemiol Date: 2010-08-04 Impact factor: 6.437
Authors: Dagmar Amtmann; Karon F Cook; Mark P Jensen; Wen-Hung Chen; Seung Choi; Dennis Revicki; David Cella; Nan Rothrock; Francis Keefe; Leigh Callahan; Jin-Shei Lai Journal: Pain Date: 2010-07 Impact factor: 6.961
Authors: David Cella; Seung Choi; Sofia Garcia; Karon F Cook; Sarah Rosenbloom; Jin-Shei Lai; Donna Surges Tatum; Richard Gershon Journal: Qual Life Res Date: 2014-06-18 Impact factor: 4.147
Authors: C S Cleeland; R Gonin; A K Hatfield; J H Edmonson; R H Blum; J A Stewart; K J Pandya Journal: N Engl J Med Date: 1994-03-03 Impact factor: 91.245
Authors: David Cella; Susan Yount; Nan Rothrock; Richard Gershon; Karon Cook; Bryce Reeve; Deborah Ader; James F Fries; Bonnie Bruce; Mattias Rose Journal: Med Care Date: 2007-05 Impact factor: 2.983
Authors: Laura Maldonado-Murciano; Halley M Pontes; Maite Barrios; Juana Gómez-Benito; Georgina Guilera Journal: Int J Ment Health Addict Date: 2021-12-06 Impact factor: 3.836
Authors: Niels Smits; Oğuzhan Öğreden; Mauricio Garnier-Villarreal; Caroline B Terwee; R Philip Chalmers Journal: Stat Methods Med Res Date: 2020-03-11 Impact factor: 3.021