Literature DB >> 35351223

Hospital COVID-19 Public Health Reporting: Lessons from Validation of an Automated Surveillance Tool to Facilitate Data Collection.

Shantini D Gamage1,2, Martin E Evans1,3,4, Brian P McCauley1, Karen R Lipscomb1, Linda Flarida1, Makoto M Jones5,6, Michael Baza5,6, Jeremy Barraza5,6, Loretta A Simbartl1, Gary A Roselle1,2,7.   

Abstract

A comparison of computer-extracted and facility-reported counts of hospitalized COVID-19 patients for public health reporting at 36 hospitals found 42% of days with matching counts between the data sources. Mis-categorization of suspect cases was a primary driver of discordance. Clear reporting definitions and data validation facilitate emerging disease surveillance.

Entities:  

Year:  2022        PMID: 35351223      PMCID: PMC9021587          DOI: 10.1017/ice.2022.55

Source DB:  PubMed          Journal:  Infect Control Hosp Epidemiol        ISSN: 0899-823X            Impact factor:   6.520


The coronavirus disease (COVID-19) pandemic has had a significant impact on patient health and logistics in healthcare systems globally. The Veterans’ Health Administration (VHA) developed a surveillance system using electronic health record (EHR) data to monitor the impact of COVID-19 at the 170 VHA medical centers across the United States in real time. The VHA leveraged this new system to report daily data for each VHA medical center to the Centers for Disease Control and Prevention (CDC) National Healthcare Safety Network (NHSN) COVID-19 Acute-Care Module. Although EHR data has been used for other VHA surveillance reporting, application of centralized EHR data for emerging disease reporting can be challenging. Here, we compared the COVID-19 data extracted centrally to data collected by individual VHA facilities to validate the EHR extractions. Our findings provide insight into data collection for public health reporting of novel diseases.

Methods

Data collection was conducted for 10 business days in June 2020 to compare the daily computer-extracted data from the COVID-19 surveillance system (ie, computer-extracted data) being sent to NHSN with data manually collected from volunteer VHA medical centers (ie, facility-reported data) for the same period and using the same NHSN definition for the count of hospitalized patients with suspected and confirmed COVID-19 (see the Supplementary Information online for definitions and details on data collection and analysis). Total counts of hospitalized patients determined from computer extraction and facility reporting were compared by t test (SAS version 9.4 software, SAS Institute, Cary, NC). For both data sources, personally identifiable information (ie, patient first name, last name, birthdate, and last 4 digits of the Social Security number) corresponding to the daily patients counted was collected to compare accuracy at the patient level (ie, even when counts match, the patients reported can be different). Additionally, for facilities that had zero or only 1 day of total counts matching between the 2 sources, the medical charts of patients identified by both sources were reviewed to determine whether they should have been included in the count based on the NHSN definition. This work was approved by VHA Central Office as a validation effort for operational systems improvement.

Results

Of 170 VHA facilities, 36 (21%) volunteered to participate in the project, resulting in 356 days of facility data (Table 1 and Fig. 1; see Supplementary Table S1 online for facility characteristics). Overall, the study included 1,472 patient days for computer-extracted data and 1,353 patient days for facility-reported data (P = .34) (Table 1). The count of hospitalized patients was the same for the 2 data sources in 151 (42%) of the 356 facility days reported. For 139 (92.1%) of these 151 days, there were complete patient matches of personally identifiable information between the 2 sources (Table 1 and Fig. 1). When counts from the 2 sources were compared allowing for discordance by 1 or 2 patients, the percentage of facility days included increased from 42% to 73% (261 of 356) and 85% (301 of 356), respectively (Table 1).
Table 1.

Summary Statistics for All Participating Facilities

CharacteristicTotal
Total no. of days for all facilities356 a
Hospitalized patients for all facility days, no.
 Computer-extracted b 1,472
 Facility-reported c 1,353
Unique hospitalized patients for all facility days, no.
 Computer-extracted559
 Facility-reported474
 Unique patients from computer-extracted and facility-reported813
Comparison of computer-extracted and facility-reported hospitalized patient counts each facility day, no. (%)
 Days counts matched151/356 (42.4)
 Count-matching days with 100% patient match d 139/151 (92.1)
 Days counts matched or were discordant by 1 patient261/356 (73.3)
 Days counts matched or were discordant by up to 2 patients301/356 (84.6)

Note. VHA, Veterans Health Administration.

34 (94%) of 36 facilities reported data for all 10 days of the project; 2 facilities reported data for a portion of the 10 days of the project (see Fig. 1 for more details).

Data extracted by VHA automated data extraction system for daily reporting of VHA medical facility counts to the National Healthcare Safety Network.

Data reported by VHA medical facilities participating in this review based on manual assessment of daily counts by facility staff.

Patient matching at a facility was determined by comparison of patient identifiers (names, dates of birth, and last four digits in social security numbers) between the computer-extracted and facility-reported counts for each day.

Fig. 1.

Summary of daily count and personally identifiable information (PII) matching between computer-extracted and facility-reported data for the 36 participating medical facilities. In the last row of boxes, each box shows the extent of personally identifiable information matching between the 2 data sources for each facility day, by quartiles.

Summary Statistics for All Participating Facilities Note. VHA, Veterans Health Administration. 34 (94%) of 36 facilities reported data for all 10 days of the project; 2 facilities reported data for a portion of the 10 days of the project (see Fig. 1 for more details). Data extracted by VHA automated data extraction system for daily reporting of VHA medical facility counts to the National Healthcare Safety Network. Data reported by VHA medical facilities participating in this review based on manual assessment of daily counts by facility staff. Patient matching at a facility was determined by comparison of patient identifiers (names, dates of birth, and last four digits in social security numbers) between the computer-extracted and facility-reported counts for each day. Summary of daily count and personally identifiable information (PII) matching between computer-extracted and facility-reported data for the 36 participating medical facilities. In the last row of boxes, each box shows the extent of personally identifiable information matching between the 2 data sources for each facility day, by quartiles. Daily counts were also assessed for each facility, and the review of personally identifiable information showed variability in the extent of patient matching (Supplementary Table S2 online). The difference in counts between the 2 sources on nonmatching days was low overall; the median difference in count was 1 for 23 (68%) of 34 facilities. On days when patient counts matched on the facility level, the personally identifiable information of patients also tended to match; the mean for all facilities was 95% ±16%. However, on days with nonmatching counts between data sources, the mean matching of personally identifiable information for all facilities was only 48% ±30% (Supplementary Table S2; see also Figure 1 for variability in matching personally identifiable information). Overall, 12 (33%) facilities met criteria for patient chart reviews. For 8 facilities, chart reviews showed that the data source with higher daily counts overcounted cases (Supplementary Table S3 online). Reasons for facility-reported overcounting included reporting patients who had a past positive test result but no symptoms on admission, having a protocol designating certain patients as suspect without symptoms (eg, transfers from nursing homes), universal laboratory screening of all patients for COVID-19 regardless of symptoms or reason for admission, and errors in following instructions. Computer-extracted overcounting occurred with reporting patients who had a past positive result but no symptoms on admission, patients with no COVID-19 testing or a negative result, and patients without symptoms but who were tested based on local universal screening policy.

Discussion

Electronic health record data are increasingly being used for public health surveillance. Our findings offer insights for designing emerging disease surveillance systems, and they highlight the importance of periodic validation by those collecting data for reporting. The hospitalized patient-count match between computer-extracted and facility-reported data was ∼42%. For the days that counts did match, the individual patients from the 2 sources almost always matched, indicating a key level of accuracy. The most common reason for count mismatching in both sources, substantiated by comparing personally identifiable information and selected chart reviews, was determination of “suspect” cases for inclusion in the daily report though the impact varied by facility. Symptoms of COVID-19 are broad and nonspecific and are especially difficult to discern in older people who may have atypical presentation and multiple comorbidities. Therefore, identifying suspected cases by automated extraction was limited to those who had a specific tag in their electronic record such as being a “person under investigation.” From a facility-reporting perspective, local criteria for suspected cases sometimes differed from the NHSN definition. These findings lead to several observations to improve emerging disease surveillance: (1) early standardization of healthcare system laboratory and health-record terms in the EHR for an emerging disease can facilitate data extraction; (2) capturing suspect cases is often necessary for emerging disease reporting systems, and clear definitions and instructions for suspect cases can prevent overreporting; and (3) having separate data elements for confirmed cases and suspect cases will lessen data interpretation issues. Although we identified areas for improvement, computer extraction was useful for assessing the change in patient counts over time and often was more accurate than reporting by facilities. Allowing for a limited discrepancy in count matching between sources increased the percentage of days matching from 42% to 85%. And, when counts at facilities were discordant, the difference in counts between sources was usually low. In the early months of an emerging disease, leveraging EHRs for automated extraction can relieve the data-collection burden from staff with sufficient accuracy for monitoring changes in case counts. This study had several limitations. Facility participation was voluntary, and several of the facilities had low hospitalized-patient counts. Our results may not reflect validation in high-incidence areas. Nonetheless, in an emerging biological event, even capturing small numbers of cases is critical, and validation in low-incidence areas is valuable. Another limitation is the unknown generalizability of the VHA extraction system for public health reporting in other healthcare systems; however, the reported challenges related to designation of suspect cases is informative for the development of any such system. Finally, we did not collect detailed information on how facilities were determining their counts; interpretations had to be drawn from selected patient chart reviews. Surveillance during an emerging infectious disease depends on data collection definitions and reporting systems. Here, 2 data-collection mechanisms, computer extraction and manual facility reporting demonstrated comparable results for surveillance of disease incidence, with similar pitfalls related to the sometimes ambiguous nature of case classifications. Validation studies are critical for identifying areas for improvement when developing data-collection platforms for emerging diseases.
  9 in total

1.  Web-based electronic health records improve data completeness and reduce medical discrepancies in employee vaccination programs.

Authors:  Monica Salazar; Kaylan E Stinson; Stefan H Sillau; Linda Good; Lee S Newman
Journal:  Infect Control Hosp Epidemiol       Date:  2011-11-30       Impact factor: 3.254

2.  Massive impact of coronavirus disease 2019 pandemic on gastroenterology and hepatology departments and doctors in Spain.

Authors:  Javier Crespo; Carlos Fernández Carrillo; Paula Iruzubieta; Marta Hernández-Conde; Laura Rasines; Francisco Jorquera; Agustín Albillos; Rafael Bañares; Pedro Mora; Inmaculada Fernández Vázquez; Manuel Hernández-Guerra; Juan Turnes; José Luis Calleja
Journal:  J Gastroenterol Hepatol       Date:  2020-12-14       Impact factor: 4.029

3.  In Data We Trust? Comparison of Electronic Versus Manual Abstraction of Antimicrobial Prescribing Quality Metrics for Hospitalized Veterans With Pneumonia.

Authors:  Barbara E Jones; Candace Haroldsen; Karl Madaras-Kelly; Matthew B Goetz; Jian Ying; Brian Sauer; Makoto M Jones; Molly Leecaster; Tom Greene; Scott K Fridkin; Melinda M Neuhauser; Matthew H Samore
Journal:  Med Care       Date:  2018-07       Impact factor: 2.983

Review 4.  Pathophysiology, Transmission, Diagnosis, and Treatment of Coronavirus Disease 2019 (COVID-19): A Review.

Authors:  W Joost Wiersinga; Andrew Rhodes; Allen C Cheng; Sharon J Peacock; Hallie C Prescott
Journal:  JAMA       Date:  2020-08-25       Impact factor: 56.272

5.  Atypical presentation of COVID-19 in a frail older person.

Authors:  Hui Sian Tay; Rowan Harwood
Journal:  Age Ageing       Date:  2020-07-01       Impact factor: 10.668

6.  Impact of coronavirus disease 2019 (COVID-19) on US Hospitals and Patients, April-July 2020.

Authors:  Mathew R P Sapiano; Margaret A Dudeck; Minn Soe; Jonathan R Edwards; Erin N O'Leary; Hsiu Wu; Katherine Allen-Bridson; Agasha Amor; Rashad Arcement; Sheri Chernetsky Tejedor; Ray Dantes; Cindy Gross; Kathryn Haass; Rebecca Konnor; Seth R Kroop; Denise Leaptrot; Kent Lemoine; Allan Nkwata; Kelly Peterson; Lauren Wattenmaker; Lindsey M Weiner-Lastinger; Daniel Pollock; Andrea L Benin
Journal:  Infect Control Hosp Epidemiol       Date:  2021-02-19       Impact factor: 3.254

7.  Harnessing electronic health records for public health surveillance.

Authors:  Michael Klompas; Michael Murphy; Julie Lankiewicz; Jason McVetta; Ross Lazarus; Emma Eggleston; Patricia Daly; Paul Oppedisano; Brianne Beagan; Chaim Kirby; Richard Platt
Journal:  Online J Public Health Inform       Date:  2011-12-22

8.  Projecting hospital utilization during the COVID-19 outbreaks in the United States.

Authors:  Seyed M Moghadas; Affan Shoukat; Meagan C Fitzpatrick; Chad R Wells; Pratha Sah; Abhishek Pandey; Jeffrey D Sachs; Zheng Wang; Lauren A Meyers; Burton H Singer; Alison P Galvani
Journal:  Proc Natl Acad Sci U S A       Date:  2020-04-03       Impact factor: 11.205

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.