OBJECTIVE: To measure variation among four different Electronic Health Record (EHR) system documentation locations versus 'gold standard' manual chart review for risk stratification in patients with multiple chronic illnesses. METHODS: Adults seen in primary care with EHR evidence of at least one of 13 conditions were included. EHRs were manually reviewed to determine presence of active diagnoses, and risk scores were calculated using three different methodologies and five EHR documentation locations. Claims data were used to assess cost and utilization for the following year. Descriptive and diagnostic statistics were calculated for each EHR location. Criterion validity testing compared the gold standard verified diagnoses versus other EHR locations and risk scores in predicting future cost and utilization. RESULTS: Nine hundred patients had 2,179 probable diagnoses. About 70% of the diagnoses from the EHR were verified by gold standard. For a subset of patients having baseline and prediction year data (n=750), modeling showed that the gold standard was the best predictor of outcomes on average for a subset of patients that had these data. However, combining all data sources together had nearly equivalent performance for prediction as the gold standard. CONCLUSIONS: EHR data locations were inaccurate 30% of the time, leading to improvement in overall modeling from a gold standard from chart review for individual diagnoses. However, the impact on identification of the highest risk patients was minor, and combining data from different EHR locations was equivalent to gold standard performance. The reviewer's ability to identify a diagnosis as correct was influenced by a variety of factors, including completeness, temporality, and perceived accuracy of chart data.
OBJECTIVE: To measure variation among four different Electronic Health Record (EHR) system documentation locations versus 'gold standard' manual chart review for risk stratification in patients with multiple chronic illnesses. METHODS: Adults seen in primary care with EHR evidence of at least one of 13 conditions were included. EHRs were manually reviewed to determine presence of active diagnoses, and risk scores were calculated using three different methodologies and five EHR documentation locations. Claims data were used to assess cost and utilization for the following year. Descriptive and diagnostic statistics were calculated for each EHR location. Criterion validity testing compared the gold standard verified diagnoses versus other EHR locations and risk scores in predicting future cost and utilization. RESULTS: Nine hundred patients had 2,179 probable diagnoses. About 70% of the diagnoses from the EHR were verified by gold standard. For a subset of patients having baseline and prediction year data (n=750), modeling showed that the gold standard was the best predictor of outcomes on average for a subset of patients that had these data. However, combining all data sources together had nearly equivalent performance for prediction as the gold standard. CONCLUSIONS: EHR data locations were inaccurate 30% of the time, leading to improvement in overall modeling from a gold standard from chart review for individual diagnoses. However, the impact on identification of the highest risk patients was minor, and combining data from different EHR locations was equivalent to gold standard performance. The reviewer's ability to identify a diagnosis as correct was influenced by a variety of factors, including completeness, temporality, and perceived accuracy of chart data.
Entities:
Keywords:
Multiple Chronic Conditions; Health Information Systems; Risk Stratification; Forecasting; Data Quality
Authors: David A Dorr; Spencer S Jones; Laurie Burns; Steven M Donnelly; Cherie P Brunker; Adam Wilcox; Paul D Clayton Journal: J Am Geriatr Soc Date: 2006-04 Impact factor: 5.562
Authors: Stacey E Jolly; Sankar D Navaneethan; Jesse D Schold; Susana Arrigain; John W Sharp; Anil K Jain; Martin J Schreiber; James F Simon; Joseph V Nally Journal: Am J Nephrol Date: 2014-04-01 Impact factor: 3.754
Authors: Adam Wright; Allison B McCoy; Thu-Trang T Hickman; Daniel St Hilaire; Damian Borbolla; Watson A Bowes; William G Dixon; David A Dorr; Michael Krall; Sameer Malholtra; David W Bates; Dean F Sittig Journal: Int J Med Inform Date: 2015-07-17 Impact factor: 4.046
Authors: Gillian C Bell; Kristine R Crews; Mark R Wilkinson; Cyrine E Haidar; J Kevin Hicks; Donald K Baker; Nancy M Kornegay; Wenjian Yang; Shane J Cross; Scott C Howard; Robert R Freimuth; William E Evans; Ulrich Broeckel; Mary V Relling; James M Hoffman Journal: J Am Med Inform Assoc Date: 2013-08-26 Impact factor: 4.497
Authors: Robert W Voss; Teresa D Schmidt; Nicole Weiskopf; Miguel Marino; David A Dorr; Nathalie Huguet; Nate Warren; Steele Valenzuela; Jean O'Malley; Ana R Quiñones Journal: J Am Med Inform Assoc Date: 2022-04-13 Impact factor: 4.497
Authors: David A Dorr; Rachel L Ross; Deborah Cohen; Devan Kansagara; Katrina Ramsey; Bhavaya Sachdeva; Jonathan P Weiner Journal: BMC Med Inform Decis Mak Date: 2021-03-18 Impact factor: 2.796
Authors: Joseph W Schaefer; Joshua M Riley; Michael Li; Dianna R Cheney-Peters; Chantel M Venkataraman; Chris J Li; Christa M Smaltz; Conor G Bradley; Crystal Y Lee; Danielle M Fitzpatrick; David B Ney; Dina S Zaret; Divya M Chalikonda; Joshua D Mairose; Kashyap Chauhan; Margaret V Szot; Robert B Jones; Rukaiya Bashir-Hamidu; Shuji Mitsuhashi; Alan A Kubey Journal: J Med Virol Date: 2021-12-08 Impact factor: 20.693
Authors: Shelley-Ann M Girwar; Robert Jabroer; Marta Fiocco; Stephen P Sutch; Mattijs E Numans; Marc A Bruijnzeels Journal: Health Sci Rep Date: 2021-07-23