OBJECTIVE: To cocalibrate the Mini-Mental State Examination, the Modified Mini-Mental State, the Cognitive Abilities Screening Instrument, and the Community Screening Instrument for Dementia using item response theory (IRT) to compare screening cut points used to identify cases of dementia from different studies, to compare measurement properties of the tests, and to explore the implications of these measurement properties on longitudinal studies of cognitive functioning over time. STUDY DESIGN AND SETTING: We used cross-sectional data from three large (n>1000) community-based studies of cognitive functioning in the elderly. We used IRT to cocalibrate the scales and performed simulations of longitudinal studies. RESULTS: Screening cut points varied quite widely across studies. The four tests have curvilinear scaling and varied levels of measurement precision, with more measurement error at higher levels of cognitive functioning. In longitudinal simulations, IRT scores always performed better than standard scoring, whereas a strategy to account for varying measurement precision had mixed results. CONCLUSION: Cocalibration allows direct comparison of cognitive functioning in studies using any of these four tests. Standard scoring appears to be a poor choice for analysis of longitudinal cognitive testing data. More research is needed into the implications of varying levels of measurement precision.
OBJECTIVE: To cocalibrate the Mini-Mental State Examination, the Modified Mini-Mental State, the Cognitive Abilities Screening Instrument, and the Community Screening Instrument for Dementia using item response theory (IRT) to compare screening cut points used to identify cases of dementia from different studies, to compare measurement properties of the tests, and to explore the implications of these measurement properties on longitudinal studies of cognitive functioning over time. STUDY DESIGN AND SETTING: We used cross-sectional data from three large (n>1000) community-based studies of cognitive functioning in the elderly. We used IRT to cocalibrate the scales and performed simulations of longitudinal studies. RESULTS: Screening cut points varied quite widely across studies. The four tests have curvilinear scaling and varied levels of measurement precision, with more measurement error at higher levels of cognitive functioning. In longitudinal simulations, IRT scores always performed better than standard scoring, whereas a strategy to account for varying measurement precision had mixed results. CONCLUSION: Cocalibration allows direct comparison of cognitive functioning in studies using any of these four tests. Standard scoring appears to be a poor choice for analysis of longitudinal cognitive testing data. More research is needed into the implications of varying levels of measurement precision.
Authors: K H Masaki; K G Losonczy; G Izmirlian; D J Foley; G W Ross; H Petrovitch; R Havlik; L R White Journal: Neurology Date: 2000-03-28 Impact factor: 9.910
Authors: Sarah E Vermeer; Niels D Prins; Tom den Heijer; Albert Hofman; Peter J Koudstaal; Monique M B Breteler Journal: N Engl J Med Date: 2003-03-27 Impact factor: 91.245
Authors: Paul K Crane; Adam Carle; Laura E Gibbons; Philip Insel; R Scott Mackin; Alden Gross; Richard N Jones; Shubhabrata Mukherjee; S McKay Curtis; Danielle Harvey; Michael Weiner; Dan Mungas Journal: Brain Imaging Behav Date: 2012-12 Impact factor: 3.978
Authors: Emily H Trittschuh; Paul K Crane; Eric B Larson; Brenna Cholerton; Wayne C McCormick; Susan M McCurry; James D Bowen; Laura D Baker; Suzanne Craft Journal: J Alzheimers Dis Date: 2011 Impact factor: 4.472