BACKGROUND: Explicit chart review was an integral part of an ongoing national cooperative project, "Using Achievable Benchmarks of Care to Improve Quality of Care for Outpatients with Depression," conducted by a large managed care organization (MCO) and an academic medical center. Many investigators overlook the complexities involved in obtaining high-quality data. Given a scarcity of advice in the quality improvement (QI) literature on how to conduct chart review, the process of chart review was examined and specific techniques for improving data quality were proposed. METHODS: The abstraction tool was developed and tested in a prepilot phase; perhaps the greatest problem detected was abstractor assumption and interpretation. The need for a clear distinction between symptoms of depression or anxiety and physician diagnosis of major depression or anxiety disorder also became apparent. In designing the variables for the chart review module, four key aspects were considered: classification, format, definition, and presentation. For example, issues in format include use of free-text versus numeric variables, categoric variables, and medication variables (which can be especially challenging for abstraction projects). Quantitative measures of reliability and validity were used to improve and maintain the quality of chart review data. Measuring reliability and validity offers assistance with development of the chart review tool, continuous maintenance of data quality throughout the production phase of chart review, and final documentation of data quality. For projects that require ongoing abstraction of large numbers of clinical records, data quality may be monitored with control charts and the principles of statistical process control. RESULTS: The chart review module, which contained 140 variables, was built using MedQuest software, a suite of tools designed for customized data collection. The overall interrater reliability increased from 80% in the prepilot phase to greater than 96% in the final phase (which included three abstractors and 465 unique charts). The mean time per chart was calculated for each abstractor, and the maximum value was 13.7 +/- 13 minutes. CONCLUSIONS: In general, chart review is more difficult than it appears on the surface. It is also project specific, making a "cookbook" approach difficult. Many factors, such as imprecisely worded research questions, vague specification of variables, poorly designed abstraction tools, inappropriate interpretation by abstractors, and poor or missing recording of data in the chart, may compromise data quality.
BACKGROUND: Explicit chart review was an integral part of an ongoing national cooperative project, "Using Achievable Benchmarks of Care to Improve Quality of Care for Outpatients with Depression," conducted by a large managed care organization (MCO) and an academic medical center. Many investigators overlook the complexities involved in obtaining high-quality data. Given a scarcity of advice in the quality improvement (QI) literature on how to conduct chart review, the process of chart review was examined and specific techniques for improving data quality were proposed. METHODS: The abstraction tool was developed and tested in a prepilot phase; perhaps the greatest problem detected was abstractor assumption and interpretation. The need for a clear distinction between symptoms of depression or anxiety and physician diagnosis of major depression or anxiety disorder also became apparent. In designing the variables for the chart review module, four key aspects were considered: classification, format, definition, and presentation. For example, issues in format include use of free-text versus numeric variables, categoric variables, and medication variables (which can be especially challenging for abstraction projects). Quantitative measures of reliability and validity were used to improve and maintain the quality of chart review data. Measuring reliability and validity offers assistance with development of the chart review tool, continuous maintenance of data quality throughout the production phase of chart review, and final documentation of data quality. For projects that require ongoing abstraction of large numbers of clinical records, data quality may be monitored with control charts and the principles of statistical process control. RESULTS: The chart review module, which contained 140 variables, was built using MedQuest software, a suite of tools designed for customized data collection. The overall interrater reliability increased from 80% in the prepilot phase to greater than 96% in the final phase (which included three abstractors and 465 unique charts). The mean time per chart was calculated for each abstractor, and the maximum value was 13.7 +/- 13 minutes. CONCLUSIONS: In general, chart review is more difficult than it appears on the surface. It is also project specific, making a "cookbook" approach difficult. Many factors, such as imprecisely worded research questions, vague specification of variables, poorly designed abstraction tools, inappropriate interpretation by abstractors, and poor or missing recording of data in the chart, may compromise data quality.
Authors: Robert C Smith; Elie Korban; Mohammed Kanj; Robert Haddad; Judith S Lyles; Catherine Lein; Joseph C Gardiner; Annemarie Hodges; Francesca C Dwamena; John Coffey; Clare Collins Journal: Psychother Psychosom Date: 2004 Jan-Feb Impact factor: 17.659
Authors: Lisa L Willett; Katri Palonen; Jeroan J Allison; Gustavo R Heudebert; Catarina I Kiefe; F Stanford Massie; Terry C Wall; Thomas K Houston Journal: J Gen Intern Med Date: 2005-09 Impact factor: 5.128
Authors: Brenda M Booth; Joann E Kirchner; Stacy M Fortney; Xiaotong Han; Carol R Thrush; Michael T French Journal: J Behav Health Serv Res Date: 2006-04 Impact factor: 1.505
Authors: Shawn Murphy; Susanne Churchill; Lynn Bry; Henry Chueh; Scott Weiss; Ross Lazarus; Qing Zeng; Anil Dubey; Vivian Gainer; Michael Mendis; John Glaser; Isaac Kohane Journal: Genome Res Date: 2009-07-14 Impact factor: 9.043
Authors: Otto S Lin; Danielle La Selva; Jae-Myung Cha; Michael Gluck; Andrew Ross; Michael Chiorean; Richard A Kozarek Journal: Surg Endosc Date: 2015-07-15 Impact factor: 4.584
Authors: Brock Polnaszek; Andrea Gilmore-Bykovskyi; Melissa Hovanes; Rachel Roiland; Patrick Ferguson; Roger Brown; Amy J H Kind Journal: Med Care Date: 2016-10 Impact factor: 2.983