Literature DB >> 35559129

Infectious Disease Providers' Knowledge of and Engagement in Quality Improvement.

Kathleen R Sheridan1, Michael A Lane2,3, Thomas J Kim4, Joshua C Eby5.   

Abstract

Background: Although engagement of infectious disease physicians has been demonstrated to improve clinical outcomes in a variety of disease states, the extent of infectious disease (ID) physician engagement in quality improvement (QI) or their knowledge of QI has not been assessed.
Methods: A 12-question, web-based survey was distributed to members of the Infectious Diseases Society of America (IDSA) between August and October 2019 to assess knowledge of and engagement in QI. The survey link was sent to IDSA members who self-identified patient care as their primary professional activity.
Results: Responses were received from 200 individuals (5.4% response rate, which is just below the standard IDSA survey response rate of 6%), consisting of 175 adult infectious disease physicians (IDPs). Most respondents were employed in a hospital or clinic (41%), private or group practice (25%), or university/medical center (24%). Fifty-eight percent of respondents currently participate in QI projects, while 38% serve on QI oversight committees. Among respondents, 27% reported not being engaged in QI. Infection prevention/hospital epidemiology (77%), stewardship (72%), and antimicrobial resistance (56%) were the most commonly reported measure types. Respondents reported barriers that limited participation in QI, including cost (61%), lack of time (56%), lack of data collection resources (48%), and lack of an ID-specific registry (46%). IDPs report significant interest in additional training in QI and new quality measures. Conclusions: Although IDPs participate in QI, there are gaps in QI knowledge and measurement systems. The low response rate of our survey also suggests a lack of engagement in QI among IDPs. Closing these gaps will benefit ID in a value-driven health care economy.
© The Author(s) 2021. Published by Oxford University Press on behalf of Infectious Diseases Society of America.

Entities:  

Keywords:  clinical practice; measurement; quality; value

Year:  2021        PMID: 35559129      PMCID: PMC9088501          DOI: 10.1093/ofid/ofab515

Source DB:  PubMed          Journal:  Open Forum Infect Dis        ISSN: 2328-8957            Impact factor:   4.423


Infectious diseases (ID) specialists in the United States have expressed concern about the future of the specialty [1, 2] due to a decline in the number of ID fellowship trainees [3] and low reimbursement compared with other specialties [4, 5]. One of the ways to improve the appeal of ID to trainees and to promote compensation comparable to the impact and work of infectious disease physicians (IDPs) is to demonstrate the value of IDPs, a top priority of the Infectious Diseases Society of America (IDSA) [6]. Demonstrating value requires measurement of the impact of IDPs on patient care and institutional success. Thus, promoting a culture of quality improvement and measurement within the specialty of ID is critical to the future of the specialty. Literature has demonstrated that involvement of IDPs in patient care for a multitude of diseases improves patient outcomes, improves quality of patient care, reduces cost of care, and benefits population health [7-10]; in order to translate this evidence into measurement of quality and performance for the care of patients, both well-defined measures and engagement of IDPs in quality and performance measurement are needed. Other medicine specialties have made this successful transition from evidence that an intervention is beneficial to quality and performance improvement in the workplace, and subsequently to demonstration of value within an institution. For example, literature-based quality measures were developed for cardiology, implemented through quality improvement, and ultimately reflected in national institutional reputation [11-13]. With the shift to value-based reimbursement, properly designed measures also provide the benefit of use for pay-for-performance reimbursement [14], another way to demonstrate IDP value. In sum, developing a culture of measurement in infectious diseases requires both measures of IDP-impacted quality and engagement of IDPs in quality improvement and measurement. The current level of interest or involvement in quality improvement among IDPs has not been adequately assessed. A review of the literature indicates that IDPs are involved in local and collaborative quality improvement efforts [7-10]. In addition, many of the institutional positions of IDPs such as hospital epidemiology, antimicrobial stewardship, and outpatient parenteral antibiotic therapy require some degree of understanding of quality improvement. Due to the paucity of existing national ID-specific measures and lack of benchmarked goals for the few existing measures, IDPs do not appear superficially to be heavily involved in quality improvement nationally [14, 15]. To more thoroughly assess the current involvement of IDPs in QI and measurement, we surveyed IDP members of the Infectious Diseases Society of America. Herein, we review the survey results and propose strategies for addressing gaps so that IDPs can move toward a culture of measurement and quality improvement that will promote the success of the specialty.

METHODS

The target population for the survey was physician IDSA members based in the United States whose primary professional activity was patient care, primary employment affiliation was at a university/medical school, hospital/clinic, private/group practice, federal government, state/local government, military, correctional facility, or other, and specialization was in adult infectious diseases. The target population was contacted by direct email with a personalized SurveyMonkey Inc. (San Mateo, CA, USA) link that contained a custom identifier that enabled linkage to the IDSA membership database to include demographic data in our analysis. The target population amounted to 3685 contacts out of ~12000 total IDSA members (October 2019). Additional survey responses were solicited through the IDSA’s Volunteer Manager web portal, which lists all IDSA volunteer opportunities. The survey collection period was from August 2019 to October 2019, with a total of 200 respondents (5.4% response rate). As a comparator, the average response rate for IDSA member surveys is ~6%. The survey has been reviewed by the authors’ affiliated institutional review boards and was designated as an exempt study. The survey asked respondents a total of 12 questions. The complete list of questions and answer options is available in Table 1.
Table 1.

Survey Questions and Answers

QuestionAnswer Options
1. How are you currently involved in quality measurement or improvement work at your hospital/practice? (select all that apply)A.\tNot actively involved; I am not sure if my practice/facility collects quality data on my care
B.\tNot actively involved, but I know that my practice/facility collects data on my care
C.\tMy practice/hospital uses an outside vendor to collect data
D.\tI receive reports about the quality of care that I deliver
E.\tI participate in quality improvement projects
F.\tI serve on a quality improvement oversight or review committee
2. Select all the barriers to your participation in quality improvement. If you participate in quality improvement, select the barriers that prevent you from greater or more successful participation.A.\tMortality for all patients you provide care to
B.\tMortality for patients with specific infectious disease conditions (eg, Staphylococcus bacteremia)
C.\tReadmission rates for all patients you provide care to
D.\tReadmission rates for patients with specific infectious disease conditions E.\tAdherence to guidelines from professional societies
F.\tUtilization of hospital-specific treatment pathways/order sets
G.\tAntimicrobial utilization
H.\tHealth maintenance (eg, smoking cessation counseling)
I.\tAdherence to treatment standards for non-ID-related conditions (eg, beta-blocker use after MI, smoking cessation)
J.\tOther
3. Select all quality data that you or your primary practice/hospital collect to measure the safety or quality of care that you deliver.A.\tOngoing professional practice evaluation
B.\tCMS, HRSA, or private payer quality programs
C.\tHospital-based incentive programs
D.\tHospital-based quality improvement effort
E.\tSelf-directed care improvement
F.\tI don’t know
G.\tOther
4. If you or your hospital/practice collects quality data, what are the data used for? (Select all that apply.)A.\tLack of validated metrics that reflect ID care
B.\tLack of access to data
C.\tLack of support for data collection and/or reporting
D.\tUnclear benefit to participating in QI
E.\tLack of time to complete QI work
F.\tLack of knowledge on how to improve quality
G.\tCost of resources for reporting
H.\tLack of ID-specific data registry
I.\tInfectious disease quality measurement is not an institutional priority
5. How likely are you to use the following safety and quality metrics if they existed? (1—very unlikely, 5—very likely)A.\tAntimicrobial use/stewardship
B.\tOPAT use and safety
C.\tClostridium difficile guideline adherence
D.\tStaphylococcus aureus bacteremia guideline adherence
E.\tAdherence to other IDSA-endorsed guideline
F.\tAdherence to HIV care standards (eg, antiretroviral use, OI prophylaxis)
G.\tDevelopment of antimicrobial resistance
H.\tOther (specify)
6. Would you utilize any of the following resources to improve your knowledge of quality improvement if provided by IDSA? (Select all that apply.)A.\tWebinars/trainings
B.\tFellowship educational tools
C.\tData registry
D.\tIDWeek lectures/workshop
7. Select all measures that you/your practice reports on either externally or internally.A.\tStewardship
B.\tOPAT
C.\tHIV
D.\tInfection prevention/hospital epidemiology
E.\tAntimicrobial resistance
F.\tStaphylococcus aureus bacteremia
8. Rate your familiarity with these quality programs or methods. (1—not at all familiar, 5—extremely familiar)A.\tMerit-based Incentive Payment Program (MIPS)
B.\tAlternative Payment Models (APMs)
C.\tModel for Improvement
D.\tLean Methodology
E.\tSix Sigma
9. Please describe the location where you deliver the majority of your ID care.A.\tInpatient
B.\tOutpatient
C.\tOther (specify)
10. How many years have you been in practice?A.\tMedical student
B.\tResident/Fellow
C.\t<5 years out of fellowship
D.\t5–10 years out of fellowship
E.\t10–15 years out of fellowship
F.\t15–20 years out of fellowship
G.\t20–25 years out of fellowship
H.\t≥25 years out of fellowship
11. What is the size of the facility where you spend the majority of your time?A.\t<50 beds
B.\t50–100 beds
C.\t100–200 beds
D.\t200–400 beds
E.\t400–600 beds
F.\t600–800 beds
G.\t800–1000 beds
H.\t>1000 beds
12. Would you be willing to participate in a follow-up survey about quality & safety reporting metrics?A.\tYes
B.\tNo

Abbreviations: CMS, Centers for Medicare and Medicaid Services; HRSA, Heath Resources and Services Administration; ID, infectious disease; IDSA, Infectious Diseases Society of America; MI, myocardial infarction; OI, opportunistic infection; OPAT, outpatient parenteral antibiotic therapy.

Survey Questions and Answers Abbreviations: CMS, Centers for Medicare and Medicaid Services; HRSA, Heath Resources and Services Administration; ID, infectious disease; IDSA, Infectious Diseases Society of America; MI, myocardial infarction; OI, opportunistic infection; OPAT, outpatient parenteral antibiotic therapy.

RESULTS

We surveyed members of the IDSA through their member registry system. A total of 200 respondents completed the survey, with 199 completing all 12 questions. Of the respondents, 175 practice adult ID, 13 practice pediatric ID, and 2 practice internal medicine. Specialty field data were not available from 9 respondents. Most of the respondents were employed in a hospital or clinic (41%), private or group practice (25%), or in a university/medical center (24%). Ninety-four percent of the respondents reported patient care as their primary job role, with inpatient practice being the most common (78%). Teaching (34%) followed by clinical research (19%) and hospital epidemiology (17%) were the most common secondary job roles. Fifty-eight percent of the respondents reported that they participate in quality improvement projects, with 38% serving on a quality improvement oversight committee. Twenty-seven percent of the respondents are not actively involved in quality improvement (Figure 1).
Figure 1.

Current level of involvement in quality measurement or improvement, by years of practice.

Current level of involvement in quality measurement or improvement, by years of practice. Most respondents reported that their facility collects data to be used for hospital-based quality improvement efforts (54%) and Centers for Medicare and Medicaid Services (CMS), Heath Resources and Services Administration, or private payer quality programs (46%). Data may also be used to promote improvement in the quality of patient care delivered either via ongoing professional performance evaluation (OPPE; 29%) or via self-directed improvement (16%). Thirty-three percent of the respondents reported that the data are also used for hospital-based incentive programs. However, 23% of the respondents did not know how the data collected by their hospital/practice are used (Figure 2).
Figure 2.

Reported use of quality data, by employment type. Abbreviations: CMS, Centers for Medicare and Medicaid Services; HRSA, Heath Resources and Services Administration.

Reported use of quality data, by employment type. Abbreviations: CMS, Centers for Medicare and Medicaid Services; HRSA, Heath Resources and Services Administration. Infection prevention/hospital epidemiology (77%), stewardship (72%), and antimicrobial resistance (56%) are the most commonly reported measure types. Antimicrobial utilization (73%) was the most commonly collected data point. Mortality rates and readmission rates for patients who are provided care are also commonly reported, though these are not specific to patients with infectious disease conditions such as Staphylococcus bacteremia. Guideline & treatment standard adherence are commonly measured as well. Those working in a federal or state/local government position were more likely to report also being measured on adherence to treatment standards for non-ID-related disease states, such as tobacco cessation. Several barriers to participation in quality improvement were identified. A deficiency in knowledge on how to collect quality data was not a major barrier for any subgroup. The lack of support for data collection, access to the data, and lack of time to complete QI work were major barriers for all (Figure 3). Specifically, the lack of ID-specific measures and the lack of an ID data registry were barriers, with the majority of respondents reporting that they would be very likely to use ID-specific quality metrics if more existed (Figure 4). Many respondents also reported that they would utilize QI resources through the IDSA, such as webinar trainings (77%), IDWeek lectures (73%), a data registry (51%), or fellowship-directed educational tools (29%).
Figure 3.

Barriers to participating in quality improvement. Abbreviations: ID, infectious disease; QI, quality improvement.

Figure 4.

Likeliness to use potential ID quality and safety metrics (Likert scale). Abbreviations: ID, infectious disease; IDSA, Infectious Diseases Society of America; OI, opportunistic infection; OPAT, outpatient parenteral antibiotic therapy.

Barriers to participating in quality improvement. Abbreviations: ID, infectious disease; QI, quality improvement. Likeliness to use potential ID quality and safety metrics (Likert scale). Abbreviations: ID, infectious disease; IDSA, Infectious Diseases Society of America; OI, opportunistic infection; OPAT, outpatient parenteral antibiotic therapy.

CONCLUSIONS

Health care payment models in the United States have historically been based on payment for volume. However, the growing trend toward value-based care provides incentive for health care providers to engage in quality measurement and improvement activities that allow for the demonstration of value as well as improvements in care. In order to demonstrate value, there must be systems of measurement that accurately reflect care delivery and participation of clinicians to engage in data-driven care improvement. This survey assesses infectious disease physicians’ knowledge of and engagement with quality measurement and improvement. Our survey indicates that many IDPs (58%) already participate in quality improvement projects; an additional 38% of respondents serve on quality improvement oversight or review committees. However, there is variation in engagement in quality improvement projects based on employer, with only 33% (n = 17) of those in private or group practices reporting participation in quality improvement projects. This is in contrast to those employed directly by a hospital (66% of respondents) or university/medical school (69% of respondents). This difference in engagement with QI projects may be a reflection of how participation in these activities is compensated in different settings. Compensation models that are fully supported by the volume of patients seen will disincentivize participation in uncompensated activities, like participation in QI initiatives. Conversely, direct employment by a university or hospital may weaken the link between patient volume and compensation, thus reducing the disincentive to participate in activities that do not directly generate revenue. Participation in QI projects is not surprising given the demonstrated benefits that the engagement of an IDP has on infection prevention and antimicrobial stewardship programs [16]. Respondents indicated that infection prevention (77%), antimicrobial stewardship (72%), and antimicrobial resistance (56%) were the most frequently reported infectious disease–related measures collected by their facilities. By comparison, a minority of respondents indicated that their facility reported data on adherence to treatment guidelines (21%), hospital-based treatment pathways or order sets (33%), or health maintenance measures (33%). Although the CMS reports 30-day readmissions and 30-day mortality for selected conditions, including pneumonia, few IDPs were aware that data were collected on these measures. Only 23% of respondents indicated that mortality data on all patients to whom they provided care were collected, while only 11% reported knowledge that mortality for specific infectious disease conditions were collected. Thirty-four percent of respondents were aware that data on readmission rates were collected for patients to whom they had provided care. IDPs reported that the most common use for quality data was hospital-based improvement (54%), while only 46% reported that the data were used for CMS or other payer quality programs. Twenty-three percent of respondents were unsure how quality data were used. IDPs reported modest knowledge of the 2 components of CMS’ Quality Payment program, the Merit-Based Incentive Payment System (MIPS) and Alternate Payment Models (APMs). Among respondents, 50.6% and 71.3% were slightly or not at all familiar with MIPS and APMs, respectively. Additionally, there was limited knowledge of QI methodologies; 77.1%, 63.5%, and 60.5% were slightly or not familiar at all with the Model for Improvement, lean methodology, or Six Sigma, respectively. Although both internal medicine residency [17] and infectious disease fellowship programs [18] are required to provide training in quality improvement, these findings suggest that this goal may not be sufficiently met. A systematic review of QI curricula for physician trainees found that quality of QI training varied widely [19]. Lickhus et al. found that QI training did not improve family medicine residents’ ability to lead QI initiatives [20]. However, there are examples of QI training improving an individual’s knowledge of QI [21, 22]. A majority of respondents to our survey would participate in additional training in QI through webinars (77%) or lectures/workshops during IDWeek (73%), suggesting an interest in acquiring more knowledge about QI. While knowledge deficits about QI can be overcome by making training more available, there are other barriers that IDPs identified that limit broader participation in QI activities (Figure 3). Notably, the cost of reporting (78%), the lack of time to participate in QI initiatives (56%), and lack of support for data collection (48%) were identified as significant barriers. The lack of an ID-specific registry was reported as a barrier by 46% of respondents. Some medical specialties have developed their own registries for collecting quality data. The American College of Rheumatology has created the Rheumatology Informatics System for Effectiveness (RISE), which captures data electronically [23]. The RISE registry has been used to benchmark care practices and identify opportunities for improvement. For example, 28802 patients who were initiated on biologic or new synthetic disease-modifying antirheumatic drugs (DMARDs) were assessed for receipt of appropriate health screenings before DMARD initiation; only 15.5% of all patients received appropriate screening for all relevant infections before DMARD initiation [24]. The RISE registry has also been able to demonstrate improvements in most measures of rheumatoid arthritis care at the clinic level [25]. Likewise, the American Heart Association (AHA) supports multiple registries that track the quality of care and outcomes of patients with cardiovascular disease and stroke [26]. Patients treated at hospitals participating in the AHA’s Get with the Guidelines–Stroke Registry and program were more likely to receive care consistent with evidence-based guidelines than those treated at hospitals that did not participate in the registry [27]. The lack of an infectious disease–specific registry was identified by 46% of respondents as a significant barrier to engaging further in QI. While the development of an ID-specific or condition-based registry would require significant investment, creation of a registry would address a significant barrier for IDPs. Respondents indicated a high degree of support for multiple quality metrics if they were to be developed (Figure 4). Although IDPs have interest in new quality metrics, there has been an explosion of quality metrics that health care organizations must devote resources to. More than one-third of respondents to our survey indicated that ID quality measures were not an organizational priority (Figure 3). In addition to providing insight into an ID physicians’ care, new quality measures must provide value to the health care organizations, which are often balancing the demands of multiple quality measurement systems with the limited resources available for data collection, reporting, and improvement. Our study does have limitations. We had a low response rate, at only 5.4%, which may be reflective of a low level of engagement in QI among IDPs. Although survey respondents represented a broad array of ID practice types and care settings, it is possible that participants are not representative of all ID providers. Additionally, the survey largely assesses knowledge of inpatient quality measures. There are several HIV-related quality measures that assess the quality of ambulatory HIV care. Additionally, the survey was targeted toward individuals who reported patient care as their primary professional activity. ID physicians who serve in safety- or quality-related administrative roles as their primary activity may have been excluded from this study. The results reflect the knowledge and engagement in QI of a more broadly generalizable IDP population when compared with those who are predominantly in safety or quality administrative roles. Our study suggests that while there is interest in more ID-related quality measures, there currently are knowledge and other barriers that limit the ability of IDPs to participate in quality initiatives. Initiatives to make quality improvement training more available may benefit IDPs. Access to an ID registry may also be beneficial to advancing the role of IDPs in QI.
  25 in total

Review 1.  The value of infectious diseases specialists: non-patient care activities.

Authors:  Daniel P McQuillen; Russell M Petrak; Ronald B Wasserman; Ronald G Nahass; Jason A Scull; Lawrence P Martinelli
Journal:  Clin Infect Dis       Date:  2008-10-15       Impact factor: 9.079

Review 2.  Methodological rigor of quality improvement curricula for physician trainees: a systematic review and recommendations for change.

Authors:  Donna M Windish; Darcy A Reed; Romsai T Boonyasai; Chayan Chakraborti; Eric B Bass
Journal:  Acad Med       Date:  2009-12       Impact factor: 6.893

Review 3.  Charting the Future of Infectious Disease: Anticipating and Addressing the Supply and Demand Mismatch.

Authors:  Rochelle P Walensky; Carlos Del Rio; Wendy S Armstrong
Journal:  Clin Infect Dis       Date:  2017-05-15       Impact factor: 9.079

Review 4.  Charting the Path Forward: Development, Goals and Initiatives of the 2019 Infectious Diseases Society of America Strategic Plan.

Authors:  Cynthia L Sears; Thomas M File; Barbara D Alexander; Daniel P McQuillen; Ann T MacIntyre; Upton D Allen; Jonathan A Colasanti; Javeed Siddiqui; Kelly R Reveles; Chris Busky
Journal:  Clin Infect Dis       Date:  2019-11-27       Impact factor: 9.079

5.  Impact of Infectious Disease Consultation on Quality of Care, Mortality, and Length of Stay in Staphylococcus aureus Bacteremia: Results From a Large Multicenter Cohort Study.

Authors:  Anthony D Bai; Adrienne Showler; Lisa Burry; Marilyn Steinberg; Daniel R Ricciuto; Tania Fernandes; Anna Chiu; Sumit Raybardhan; Michelle Science; Eshan Fernando; George Tomlinson; Chaim M Bell; Andrew M Morris
Journal:  Clin Infect Dis       Date:  2015-02-20       Impact factor: 9.079

6.  Quality improvement initiative and its impact on the management of patients with acute myocardial infarction.

Authors:  R H Mehta; S Das; T T Tsai; E Nolan; G Kearly; K A Eagle
Journal:  Arch Intern Med       Date:  2000-11-13

7.  X + Y = Time for QI: Meaningful Engagement of Residents in Quality Improvement During the Ambulatory Block.

Authors:  Krista M Johnson; Wendy Fiordellisi; Ethan Kuperman; Alexis Wickersham; Carly Kuehn; Aparna Kamath; Joseph Szot; Manish Suneja
Journal:  J Grad Med Educ       Date:  2018-06

8.  2017 Infectious Diseases Society of America Physician Compensation Survey: Results and Analysis.

Authors:  Robin Trotman; Alice I Kim; Ann T MacIntyre; Jethro Trees Ritter; Anurag N Malani
Journal:  Open Forum Infect Dis       Date:  2018-11-22       Impact factor: 3.835

9.  Infectious diseases subspecialty: declining demand challenges and opportunities.

Authors:  Pranatharthi Chandrasekar; Daniel Havlichek; Leonard B Johnson
Journal:  Clin Infect Dis       Date:  2014-08-22       Impact factor: 9.079

10.  Participation in Get With The Guidelines-Stroke and Its Association With Quality of Care for Stroke.

Authors:  George Howard; Lee H Schwamm; John P Donnelly; Virginia J Howard; Adam Jasne; Eric E Smith; J David Rhodes; Brett M Kissela; Gregg C Fonarow; Dawn O Kleindorfer; Karen C Albright
Journal:  JAMA Neurol       Date:  2018-11-01       Impact factor: 18.302

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.