Literature DB >> 33855102

How Are Emerging Data Translated Into Clinical Practice? A Mixed Methods Investigation of Coronavirus Disease 2019 Institutional Treatment Protocols.

Archana Asundi1,2, Jack Resnik3,4, Peter A Benedict5, Marlena Shin3,4, A Rani Elwy3,4,6, Westyn Branch-Elliman3,4,7.   

Abstract

BACKGROUND: Early in the coronavirus disease 2019 (COVID-19) pandemic, there was minimal data to guide treatment, and we lacked understanding of how clinicians translated this limited evidence base for potential therapeutics to bedside care. Our objective was to systematically determine how emerging data about COVID-19 treatments was implemented by analyzing institutional treatment protocols.
METHODS: Treatment protocols from North American healthcare facilities and recommendations from guideline-issuing bodies were collected. Qualitative data on treatment regimens and their applications were extracted using an adapted National Institutes of Health/US Food and Drug Administration experimental therapeutics framework. Structured data on risk factor and severity of illness scoring systems were extracted and analyzed using descriptive statistics.
RESULTS: We extracted data from 105 independent protocols. Guideline-issuing organizations published recommendations after the initial peak of the pandemic in many regions and generally recommended clinical trial referral, with limited additional guidance. Facility-specific protocols favored offering some treatment (96.8%, N = 92 of 95), most commonly, hydroxychloroquine (90.5%), followed by remdesivir and interleukin-6 inhibitors. Recommendation for clinical trial enrollment was limited largely to academic medical centers (19 of 52 vs 9 of 43 community/Veterans Affairs [VA]), which were more likely to have access to research studies. Other themes identified included urgent protocol development, plans for rapid updates, contradictory statements, and entirely missing sections, with section headings but no content other than "in process."
CONCLUSIONS: In the COVID-19 pandemic, emerging information was rapidly implemented by institutions into clinical practice and, unlike recommendations from guideline-issuing bodies, heavily favored administering some form of therapy. Understanding how and why evidence is translated into clinical care is critical to improve processes for other emerging diseases.
© The Author(s) 2021. Published by Oxford University Press on behalf of Infectious Diseases Society of America.

Entities:  

Keywords:  COVID-19; implementation science; treatment protocols

Year:  2021        PMID: 33855102      PMCID: PMC8026152          DOI: 10.1093/ofid/ofab072

Source DB:  PubMed          Journal:  Open Forum Infect Dis        ISSN: 2328-8957            Impact factor:   3.835


The emergence and rapid spread of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) led to an urgent push to develop and implement clinical treatment strategies to optimize outcomes for this new and frequently morbid disease [1]. However, early in the coronavirus disease 2019 (COVID-19) pandemic, there was limited human evidence to guide clinical decision making. Clinical innovations are postulated to spread in a slow but predictable manner: first, preclinical studies, then human efficacy and effectiveness studies, and finally studies that focus on dissemination, implementation, and sustainability [2]. Under typical conditions, there is an average 17-year lag from the time an intervention is proven effective until its implementation [3]. Implementation frameworks highlight the importance of the external environment, organizational structure, innovation characteristics, and processes for facilitating the ultimate adoption of evidence-based treatment [4-6]. Although laboratory and anecdotal evidence were published and available to guide treatment options, major guideline-issuing bodies, including the Centers for Disease Control and Prevention (CDC) [7], the World Health Organization (WHO) [8, 9], the National Institutes of Health (NIH) [10], and the Infectious Diseases Society of America (IDSA) [11], were delayed in releasing guidelines relative to the start of the first surge in many regions in North America. Thus, although evidence from human clinical trials was forthcoming, healthcare systems across the world faced pressure to provide frontline providers with guidance, despite the limited evidence base on which to base those recommendations. Evidence on how evolving data are rapidly disseminated and translated into clinical practice are lacking, but it is critical that we learn more about how this process occurs and what drives decision making for emerging diseases with a limited evidence base. Thus, the objective of this study was to develop and use a concurrent mixed methods approach to categorize, classify, and qualitatively and quantitatively analyze the variation present in these locally and rapidly developed treatment protocols in medical centers across North America.

METHODS

Overview

After obtaining institutional treatment protocols, we used a concurrent analysis mixed methods approach to collect quantitative and qualitative data in parallel [12]. This analytic approach is novel in its application to clinical documents; however, it is a standard mixed methods approach to data analysis that can yield insights into factors driving medical decision making and translation of evidence into practice.

Patient Consent Statement

This study did not include factors necessitating patient consent.

Protocol Identification, Collection, and Recruitment

To ensure diverse representation of institutional protocols across North America, we recruited protocols developed during the period from March 15, 2020 to May 15, 2020 using multiple strategies, including the following: (1) email from the Society of Healthcare Epidemiology of America (SHEA) Research Network [13] to the 74 participating institutions requesting protocol sharing (follow-up email was sent to the facilities that did not initially respond); (2) outreach to personal contacts; (3) postings on Twitter, a major social media platform, by the authors with requests for sharing of protocols (A.A., W.B.-E.); (4) requests for protocol sharing on major national listservs (the National Antimicrobial Stewardship listserv hosted by Northwestern and the VA Society of Infectious Diseases Providers (VASPID); (5) request for protocol sharing posted to the IDSA Message Board with over 13 000 members active in the system; (6) Internet and intranet (Veterans Affairs [VA]) searches for COVID protocols; and (7) review of protocols posted to major national organizations, such as those posted to the AAMC website.

Inclusion/Exclusion Criteria

Inclusion criteria for protocols included (1) those from North American institutions that referenced guidance pertaining to the use and application of COVID-19 therapeutics for adults and (2) published guidance from major guideline-issuing bodies. Protocols without a section on therapeutics (eg, testing or infection control protocols only) and those pertaining to pediatric patients only were excluded. In some cases, city-wide treatment consortia with similar protocols were developed and were coded separately among members of the consortium, because they included individual site specifications. Centers that copied verbatim other sites’ protocols were not coded in duplicate. Protocols from affiliated institutions were included in data analysis under the parent institution, and exact duplicates of protocols were excluded. Guidelines from national and international organizations not originating from a specific institution (eg, CDC, WHO, NIH) were classified as “reference protocols.”

Data Extraction and Coding

Coding was completed by authors using a structured summary template to extract relevant qualitative and structured quantitative data (Supplementary Methods 1) [14]. Institutions were asked to submit copies of their protocols but did not fill out any survey or form. All protocols were coded by clinician investigators on the study team using a standardized template (W.B.-E., A.A., P.B.) after an initial quality assurance process to ensure interrater consistency. To ensure data quality, coded protocols were reviewed by a second clinician investigator; discrepancies were resolved through a third review. Final agreements were determined via discussion and consensus [15]. Codes were grouped according to type of treatment (eg, which medications selected), how treatments were applied (eg, to all patients with a confirmed test, to all patients with a suspected test, to high risk patients, etc), whether a disease severity index was used to guide decision making and what the disease severity index and/or risk stratification for severe/complicated COVID-19 infection scores were based on (eg, vital signs, laboratory findings, medical history), and whether laboratory results were included to guide decision making (eg, interleukin [IL]-6 level to guide tocilizumab use). Authors also collected and classified information about the facility that developed the specific institutional guideline (eg, size, academic vs community, VA vs non-VA, region etc), to identify facility factors associated with different treatment recommendations, access to various treatments and limited resources, and access to clinical trials.

Analysis

For qualitative variables, we developed an a priori coding frame from categories established by the NIH and the US Food and Drug Administration (FDA) (Supplementary Methods) [16]. We used this coding frame in a directed content approach for analysis of qualitative data, allowing for new codes to emerge [17]. The coded data were organized through matrix displays in Microsoft Excel [15]. These matrices were then used to compare and contrast evidence across the treatment protocols and between institutions to identify common themes. Some variables were collected as numerical values (eg, age, obesity definition, laboratory cutoffs). For these quantitative variables, such as cutoffs for abnormal vital signs data, body mass index (BMI) cutoffs, and laboratory results, we extracted structured numerical variables from the protocols (eg, obesity defined as BMI >30) [18-20]. Quantitative variables were analyzed using descriptive statistics, such as t tests and Fisher’s exact tests, and the range of variability of these numerical variables was calculated.

RESULTS

Protocol Characteristics

Protocols (N = 168) representing 123 healthcare institutions, systems, or entities were collected (Supplementary Figure 1); of these, 105 met criteria for inclusion. Among these, 52 were academic (ie, university-affiliated teaching hospitals) centers, 12 were community/private hospitals or healthcare systems, and 31 were VA healthcare systems or networks (se Table 1 and Figure 1 for regional distribution). Protocols varied in data presentation: 46.3% included a reference list and 12.6% referenced another facility’s protocol (Table 1). Protocols (10 of 105) were from guideline-issuing bodies and were commonly cited in other protocols.
Table 1.

Protocol Characteristics and Quantitative Results

Protocol CharacteristicsTotalAcademicCommunityVA
Totala9552 (54.7)12 (12.6)31 (32.6)
Region/Location (%)
 Northeast26 (27.4)1727
 Midwest15 (15.8)834
 South26 (27.4)1529
 West24 (25.2)10311
 Nationwide2 (2.1)020
 Canada2 (2.1)200
Median beds (IQR)-751 (527–890)429 (253–579)244 (144–304)
Median ICU beds (IQR)-88 (67–120)1 (1–2)41 (14–56)
Average no. of facilities (range)-5.8 (1–40)38.5 (1–186)1.5 (1–3)
SHEA Research Network sites2420 -4
Protocol Presentation
Data presentation
 Algorithm or flow chart41 (43.2)---
 Table62 (65.3)---
 Text58 (61.1)---
Version Updates
 1 update3019011
 ≥2 updates12804
Resource limitations (% of subcategory total)32 (33.7)14 (26.9)3 (25.0)15 (48.4)
Has a reference list (% of subcategory total)44 (46.3)26 (50.0)6 (50.0)12 (38.7)
Refers to another sites protocol (% of subcategory total)12 (12.6)3 (5.8)2 (16.7)7 (22.6)
Gatekeeping of Decision Making
Decision for treatment based on disposition (% of subcategory total) 51 (53.7)27 (51.9)8 (66.7)16 (51.6)
Recommended ID consult (% of subcategory total)43 (45.3)27 (51.9)4 (33.3)12 (38.7)
Required ID approval/consult (% of subcategory total)49 (51.6)23 (44.2)4 (33.3)22 (71.0)
Antibiotic steward approval required (% of subcategory total)18 (18.9)7 (13.5)3 (25.0)8 (25.8)
Clinical Trial
Trial Available (% of subcategory total)28 (29.5)19 (36.5)2 (16.7)7 (22.6)
By Facility Sizeb per Number Of Beds (Trial Available/N Institutions in Size Category, %) ---
 1–99 0/4 (0.0)---
 100–1992/9 (22.2)---
 200–4997/30 (23.3)---
 500–99914/39 (35.9)---
 1000+4/9 (44.4)
By Top 20 Metropolisc
 Yes (% of subcategory total)13/26 (50.0) 9 (56.3)0 (0.0)4 (44.4)
 No (% of subcategory total)14/65 (21.5)10 (28.6)2 (22.2)2 (9.5)
Trial open for enrollment (%)22 (23.3)15 (15.8)2 (2.1)5 (5.3)
Recommended trial enrollment (%)43 (42.6)25 (48.1)5 (41.7)13 (41.9)
By Top 20 Metropolis
 Yes (% of subcategory total)15 (57.7)10 (62.5)0 (0.0)5 (55.5)
 No (% of subcategory total)25 (39.1)14 (38.9)4 (36.4)7 (31.8)

Abbreviations: ICU, intensive care unit; ID, Infectious Diseases; IQR, interquartile range; VA, Veterans Affairs.

aOnly includes unique protocols from parent institutions; however, this does include different protocols from institutions with overlapping catchment areas.

bFour protocols were from healthcare institutions where facility size and location in top 20 metropolis could not be categorized.

cLargest 20 cities based on US Census Bureau 2019 population estimates.

Figure 1.

Map of distribution of facilities. This figure was created with the assistance of Google Maps, 2020. Blue, VA healthcare facilities; Yellow, Academic healthcare facilities; Orange, Community healthcare facilities.

Protocol Characteristics and Quantitative Results Abbreviations: ICU, intensive care unit; ID, Infectious Diseases; IQR, interquartile range; VA, Veterans Affairs. aOnly includes unique protocols from parent institutions; however, this does include different protocols from institutions with overlapping catchment areas. bFour protocols were from healthcare institutions where facility size and location in top 20 metropolis could not be categorized. cLargest 20 cities based on US Census Bureau 2019 population estimates. Map of distribution of facilities. This figure was created with the assistance of Google Maps, 2020. Blue, VA healthcare facilities; Yellow, Academic healthcare facilities; Orange, Community healthcare facilities.

Recommendations From Guideline-Issuing Bodies

The 10 reference protocols from guideline-issuing organizations, such as IDSA [11], NIH [10], CDC [7], and WHO [8, 9], were published during the period from March 21, 2020 to April 11, 2020 and encouraged clinical trial enrollment but did not offer specific recommendations about treatments administered outside of a clinical trial, citing a lack of high-quality evidence. An exception to this was the Thoracic Society Guideline (developed using a modified Delphi technique and an expert panel) that was released in March 2020 and included a statement of support for unproven therapies in critically ill patients [21].

Overarching Themes

Two themes present throughout almost every institutional protocol were (1) a clear and strong suggestion to provide some form of treatment (92 of 95 protocols, 96.8%), particularly to patients with more severe illness (78 of 95 protocols, 82.1%), and with risk factors for progression for severe disease (69 of 95, 72.6%) (Figure 2), and (2) the desire to create a “standard” treatment for patients presenting to their institution when a broader standard of care did not exist. Acknowledging a potential for harm associated with unproven treatments, many institutional guidelines included detailed indications for safety monitoring, such as recommendations for electrocardiograms in patients receiving hydroxychloroquine and baseline human immunodeficiency virus testing before antiretrovirals.
Figure 2.

(a) Use of severity of illness (SOI) and risk factors (RF) criteria to guide treatment decisions (% of protocols), and (b) severity of illness and risk factor measures to guide therapeutic decisions. BPM, beats per minute.

(a) Use of severity of illness (SOI) and risk factors (RF) criteria to guide treatment decisions (% of protocols), and (b) severity of illness and risk factor measures to guide therapeutic decisions. BPM, beats per minute.

Treatments and Their Application

The most commonly recommended therapy within the included institutional protocols was hydroxychloroquine (Figure 3), cited in 90.5% of protocols. Of these, 80.0% recommended it in at least some circumstances, most frequently based on a locally developed risk factor and/or severity of illness scoring scheme. A total of 2.1% specifically recommended against its use. The combination of hydroxychloroquine and azithromycin was often specifically “not recommended.” Other commonly included treatment recommendations included remdesivir and IL-6 inhibitors, particularly among facilities with an open and enrolling clinical trial. Criteria for remdesivir use were typically based on a combination of risk factors and severity scores, due to the drug’s non-FDA-approved status and gatekeeping by specific industry-determined eligibility criteria for access (initially via a compassionate use mechanism and later through expanded use access).
Figure 3.

Forest plot of coronavirus disease 2019 (COVID-19) therapeutics by strength of recommendation according to facility type. RF, risk factor; SOI, severity of illness; VA, Veterans Affairs.

Forest plot of coronavirus disease 2019 (COVID-19) therapeutics by strength of recommendation according to facility type. RF, risk factor; SOI, severity of illness; VA, Veterans Affairs. Ancillary treatments, such as steroids and nonsteroidal anti-inflammatory drugs (NSAIDs), were listed in 60.0% and 45.3% of protocols, respectively. Most commonly, these medications were listed with a recommendation against their use for COVID-19 treatment (46.3% and 28.4% of protocols) (Figure 4). Angiotensin-converting enzyme (ACE) inhibitors were also commonly included (48.4%), typically with a recommendation to “continue but not start.” Anticoagulation was included in 23.2%, often with a recommendation to administer standard prophylactic treatment doses.
Figure 4.

Ancillary treatments. ACE-I, angiotensin-converting enzyme inhibitors; ARB, angiotension II receptor blocker; NSAID, nonsteroidal anti-inflammatory drugs.

Ancillary treatments. ACE-I, angiotensin-converting enzyme inhibitors; ARB, angiotension II receptor blocker; NSAID, nonsteroidal anti-inflammatory drugs. A total of 45.3% of protocols included a recommendation for infectious diseases consultation for all COVID-19 patients (Table 1). Approval of any COVID-19 medication was required in 55.8% (pharmacy approval in 18.9% and infectious diseases approval in 51.6%). Requirement for approval was highly correlated with status as an academic facility (Supplementary Figure 2). Although the application of risk factor for severe/complicated COVID-19 disease and severity of illness scoring to guide treatment was a recurring theme, we found substantial interfacility variation in how these scores were constructed and applied (Figure 2). Almost all protocols listed heart and pulmonary disease as risk factors; immunosuppression was also often included, although this term was variably defined. Many protocols (31.5%) included various terms related to renal disease, including “chronic kidney disease”, “end stage renal disease”, or “kidney disease”; however, this latter term was often not further defined and lacked specific cutoffs, potentially creating challenges for application in clinical settings. Lack of standard definitions were also found for other variables, including “respiratory disease” and “heart disease.” For numerical variables, such as BMI cutoffs to define “obesity,” there was similar lack of standardization. Severity of illness was defined in many ways; however, some form of oxygenation status was included in almost all (89.7%). Disposition status (eg, inpatient floor vs intensive care unit, 17.9%) and respiratory rates were also frequently included (46.2%). Similar to findings with numerical variables used in risk factor scoring, there was wide variation in how these variables were defined. For example, oxygen saturations across a wide range to define moderate or severe disease (eg, 90%–95% Oxygen Saturation). Likewise, the respiratory rate used to define severe disease ranged from 20 to 30, encompassing the a range of values that would all be consistent with clinical definitions of tachypnea.

Emerging Themes and Association With Facility Factors

A recommendation to enroll patients in a clinical trial was included in 45.3% (43 of 95) of institutional protocols, with an open trial available at the institution referenced in 29.5% (28 of 95 protocols). Clinical trial availability strongly correlated with institutional size and region (Supplementary Figure 3) and was more common among large academic institutions (36.5%, 19 of 52) versus community centers (16.7%, 2 of 12) and VA facilities (22.6%, 7 of 31). Clinical trials access was also higher in major metropolitan areas (50.0% in top 20 metropolises versus 21.5% in other areas) (Table 1, Supplementary Figures 3 and 4). Although some smaller facilities had academic affiliations and referenced trial availability at a parent institution, most institutions without strong academic ties did not have open and enrolling trials. Institutions without an open trial generally issued stronger recommendations for off-label therapeutics, despite explicit warnings included in many protocols that these agents were “not FDA-approved” for COVID-19. Many protocols highlighted a sense of urgency regarding the development and operationalization of treatment protocols and included statements about “frequent updates” and “do not print” given the “rapidly changing” nature of recommendations (Table 2). Reference to “living documents” being updated in real-time and subject to change were common, including statements such updates pending “further information becom[ing] available from CDC and WHO.” However, possibly as an extension of the speed of development and frequent changes, several protocols included ambiguous or contradictory recommendations, such as recommendations to administer a treatment in one section and cautionary statements recommending against the use of the same treatment in another section of the same document (Table 2). Another emergent theme was the interdisciplinary nature of many of the protocols, which in some cases included substantial coordination efforts across the entire healthcare system. Many were “jointly developed” with clinicians from several departments.
Table 2.

Emergent Themes From Institutional Protocols and Illustrative Examples

Emergent ThemeExamples From Protocols
Protocol Development and Use
Developed rapidly“We built the first iteration of these guidelines “from the bottom up” in less than a week with the input of over 50 people. With the help of our readers, we expect to correct and revise as we as a society learn about COVID-19.”
Developed collaboratively“This protocol was jointly-developed with input from clinicians across multiple departments.”
Discussion of updates/real-time changes“This is a living document that will be updated in real time as more data emerge.”
Language Use in Protocols
Cautionary “Given the lack of clear evidence to support hydroxychloroquine, medical experts have asked clinicians to exercise caution and to consider the risk of the medication—notably the potential cardiac complications. Because the data is still unclear, there are several ongoing trials of hydroxychloroquine.”
Clinical ambiguityRisk factors for severe disease: age ≥65 years, chronic cardiovascular, pulmonary, hepatic, renal, hematologic, or neurologic conditions, immunocompromised, pregnant women, residents of nursing homes or long-term care facilities.
Symptomatic individuals who are older adults (age >65 years), immunocompromised state, chronic medical conditions (eg, diabetes, CAD, chronic lung, or kidney disease)
Urgency“URGENT! Please circulate as widely as possible. It is crucial that every pulmonologist, every critical care doctor and nurse, every hospital administrator, every public health official receive this information immediately.”
Contradictory/conflicting recommendationsLopinavir/ritonavir listed under “not recommended” treatments in main table, however, included criteria for use in table footnote.
Treatment/Care Delivery
Strength of treatment recommendations“Please note there are NO FDA approved treatment options for COVID-19. These medications are experimental in nature and utilization may change with newly discovered clinical trials and results.”
Recommendation to participate in a clinical trial“[Site Name] is committed to participation in randomized controlled clinical trials to facilitate the generation of robust evidence concerning the effectiveness of products in treating COVID-19 and to appropriately delineate risk-vs-benefit assessments for various treatment strategies.”
Resource limitations as an element of decision making“Consider available resources and pump availability when ordering.”
Integration of different clinical services/interdisciplinary care team management “In a surge situation we will work with palliative care to provide palliative services for older patients.”

Abbreviations: CAD, coronary artery disease; COVID-19, coronavirus disease 2019; FDA, US Food and Drug Administration.

Emergent Themes From Institutional Protocols and Illustrative Examples Abbreviations: CAD, coronary artery disease; COVID-19, coronavirus disease 2019; FDA, US Food and Drug Administration. Other notable emerging themes involved the inclusion of “warning” or cautionary statements that therapeutics recommended were not FDA-approved or evidence-based; however, how these cautionary statements were presented varied from inclusion at the beginning of the document in bolded, red font, to a footnote in small font at the bottom of a table (Table 2). Thus, there was strong variation in how visible cautionary statements were and in the probability that the cautionary statement would be digested. Similarly, several protocols included phrases like “Do not share” or “Do not distribute,” potentially conveying discomfort with recommending unproven therapies, yet also acknowledging some pressure to provide local guidance about management strategies. Recommendations were often justified using statements about deference to physicians’ “sound clinical judgement.” Reference to “resource limitations” was a common theme, but this was more frequently found in VA facilities versus non-VA facilities and higher in the Northeast and Southern regions (Supplementary Figure 5). Protocols included phrasing such as “availability of medications may change” due to ongoing drug shortages or that medications were in “limited supply”; therefore, a “strategic utilization” approach was used in the protocol. Due to resource limitations, some institutions implemented crisis standards of care, resulting in variability in which patients were triaged to receive critical care. A small number of facilities developed palliative care treatment guidelines and order sets for patients with COVID-19 and included strategies for referring patients for palliative care consultation upon hospital presentation. Risk factors and severity of illness scores were also often applied as a means of triaging limited medication stockpiles to the patients at highest risk of disease progression and/or most severe disease.

DISCUSSION

Although there is typically a multiyear lag from preclinical evidence to human data and dissemination and implementation, the postulated progression from preclinical evidence to human studies and finally to dissemination and implementation was upended for COVID-19 therapy [2, 22]. Driven by pressure to offer “something” to patients with few options, and often differing from later recommendations from national guideline-issuing bodies, almost every institution endorsed unproven treatments for at least some patients, particularly those with severe disease or risk factors for disease progression. Recommendations from international groups (eg, Lombardy Infectious Diseases Society), which heavily favored treatment, were available by March 2020. National guideline-issuing bodies, such as the CDC and the IDSA, did not release their recommendations until mid-to-late April 2020. Thus, guidelines from US national organizations were not available until after the initial surge in some regions, including the Northeastern United States. This may have contributed to our finding that facility protocols seemed to be more congruent with international recommendations, and advice from more experienced overseas colleagues, than with US national guidelines. Challenges of confounding by indication were present in many of the early data released, for example, suggesting first benefit and then harm associated with hydroxychloroquine [23-25]. Hydroxychloroquine was later found in a meta-analysis of randomized trials and observational studies to be ineffective in reducing mortality [26]. Recognizing that not every question can be answered in a randomized clinical trial, we must find ways to improve the quality and reliability of observational studies to speed answers to questions about COVID-19 management. Incorporation of facility factors, such as treatment recommendations and severity of illness scores, provides a mechanism for including important confounding factors into these analyses. To this end, data matrices are available, so that critical facility factors can be incorporated into future observational studies to enhance their quality. Implementation Science Frameworks postulate that major factors impacting the adoption of innovations include the external environment, organizational structure, and characteristics of the innovation—including the strength of the evidence supporting it, and the processes used [4, 5]. However, during the early pandemic, we found that the 2 constructs that most heavily influenced institutional recommendations were the external environment and provider and organizational factors. Strength of the underlying evidence played a relatively minor role in driving adoption of practices. Preclinical, observational data, and data extrapolated from other respiratory viruses including Middle East respiratory syndrome (MERS) and SARS-CoV-1 led to the common recommendation for hydroxychloroquine and against the routine use of steroids as treatments for SARS-CoV-2 in many early treatment protocols. However, despite early enthusiasm about hydroxychloroquine, subsequent randomized trials found it to be ineffective [25, 27–30]. In contrast, human data extrapolated from other respiratory viral syndromes was found not to apply to COVID-19, and use of steroids for hospitalized patients with COVID-19 has now been widely adopted [31, 32]. These data demonstrate the importance of human clinical data to guide recommendations and the importance of viewing treatment guidelines for emerging diseases as “living documents” that will necessarily evolve. Application of evidence-based data to guide decision making is critical; however, this must be balanced against the near-universal instinct of frontline clinicians to provide some form of treatment—even if human evidence is lacking. Several local protocols included in this study were widely disseminated, including integration into references in other guidelines and commonly used clinician resources such as UpToDate [33]. For medications with a long track record of safety, prescribing something may be preferable to clinicians when the alternative is to offer nothing to patients with a potentially fatal disease—and this feature of physician culture to favor “doing something” over “doing nothing” should be taken into consideration when developing recommendations about best practices for novel or rare diseases without a strong evidence basis to support one treatment strategy or another. Despite being developed by discrete institutions, many of same therapeutic agents (hydroxychloroquine, remdesivir) were included, potentially indicative of some degree of informal information sharing between physician peers as a key component of information dissemination, particularly when guideline-issuing bodies issued recommendations that may not be timely or even feasible for all sites. Guideline-issuing bodies focused on suggestions for clinical trial enrollment [8-11]; however, access to clinical trials primarily available at large, urban, academic medical centers with existing research infrastructure and access at VA medical centers and community medical centers, particularly those in rural settings, was rare. Thus, smaller facilities, which on an individual basis have fewer patients but where the majority of clinical care is provided [34], typically did not have the option of offering treatment as part of a clinical trial. Thus, these facilities faced the option of offering their patients no treatment or developing protocols based on the best evidence available to them at the time, which is what many opted to do. Our findings highlight the importance of innovating mechanisms to expand clinical trials access. Guideline-issuing bodies should recognize the unintended consequences of blanket recommendations about “only in the context of a clinical trial,” without coupling the recommendation with a mechanism to support access to experimental studies for a larger segment of the population. Potential strategies for improving clinical trial access include strategic decisions about availability made at a regional or national level and not by individual institutions or sponsors, which may promote disparities by favoring well connected organizations with existing research infrastructure. As we move toward a Learning Healthcare Model [35], consideration should be given to technological innovations to support expansion of clinical trial access to patients receiving care at smaller, nonacademic institutions. This might include increasing use of embedded clinical trials within integrated electronic medical records systems [36, 37], a rapid referral mechanism to sister institutions that do have clinical trial availability, or pre-existing agreements about transfers if trial enrollment is indicated.

Limitations

We relied on informal networks and established societies and list serves to gather protocols, which may have led to sampling bias. We also do not have data about how many institutions developed and implemented their own protocols. However, the protocols analyzed in this study represent a wide range of facility types and regions facing the brunt of the pandemic during the study period, thus partially mitigating this limitation. In addition, due to the rapidly evolving nature of the pandemic, different protocols were developed at different times, and thus some facilities may have had more data available to them when developing their initial protocol than others. We attempted to mitigate this limitation by collecting protocols developed during a relatively short window from the early part of the pandemic in North America, but we may not have fully accounted for these differences in timing. A final limitation was the potential for bias in any qualitative analysis; however, we attempted to address this by using multiple coders and a coding check, to ensure accuracy and reproducibility.

CONCLUSIONS

Although randomized controlled clinical trials will remain the gold standard for establishing treatment options for COVID-19, not every research question can or will be answered through these typically resource- and time-intensive study designs. Thus, it is critical that we find ways to enhance the quality of observational research to improve our understanding of COVID-19. Viewing treatment through the lens of implementation science, we may be able to enhance the quality of COVID-19 observational data by more fully accounting for patient and facility-level factors that may have driven clinical decision making. Facing a novel pathogen with rapid spread and high mortality, and no high-quality evidence about effective treatment options, many centers generated institutional protocols to locally standardize care, operationalize decision making, and triage limited resources to the sickest patients. Recommendations were developed with minimal input from guideline-issuing bodies, which were slow to publish guidance and which focused on referrals to clinical trials, without addressing the impact of a blanket recommendation for “in the context of a clinical trial” on facilities with limited research infrastructure.

Data Sharing Statement

In line with the Open Science Framework [38], a deidentified version of the matrix display, including relevant implementation variables and clinical definitions stratified by facility, is publicly accessible through a data-sharing resource such that others may include these variables in their investigations. The data-sharing resource is available at: www.covidresourcecenter.com. Release of the institutional name associated with the random study number will be made available upon written request to the study authors.

Supplementary Data

Supplementary materials are available at Open Forum Infectious Diseases online. Consisting of data provided by the authors to benefit the reader, the posted materials are not copyedited and are the sole responsibility of the authors, so questions or comments should be addressed to the corresponding author. Click here for additional data file. Click here for additional data file.
  27 in total

Review 1.  From best evidence to best practice: effective implementation of change in patients' care.

Authors:  Richard Grol; Jeremy Grimshaw
Journal:  Lancet       Date:  2003-10-11       Impact factor: 79.321

Review 2.  A systematic scoping review of yoga intervention components and study quality.

Authors:  A Rani Elwy; Erik J Groessl; Susan V Eisen; Kristen E Riley; Meghan Maiya; Jennifer P Lee; Andrew Sarkin; Crystal L Park
Journal:  Am J Prev Med       Date:  2014-07-02       Impact factor: 5.043

3.  Effect of Hydroxychloroquine on Clinical Status at 14 Days in Hospitalized Patients With COVID-19: A Randomized Clinical Trial.

Authors:  Wesley H Self; Matthew W Semler; Lindsay M Leither; Jonathan D Casey; Derek C Angus; Roy G Brower; Steven Y Chang; Sean P Collins; John C Eppensteiner; Michael R Filbin; D Clark Files; Kevin W Gibbs; Adit A Ginde; Michelle N Gong; Frank E Harrell; Douglas L Hayden; Catherine L Hough; Nicholas J Johnson; Akram Khan; Christopher J Lindsell; Michael A Matthay; Marc Moss; Pauline K Park; Todd W Rice; Bryce R H Robinson; David A Schoenfeld; Nathan I Shapiro; Jay S Steingrub; Christine A Ulysse; Alexandra Weissman; Donald M Yealy; B Taylor Thompson; Samuel M Brown; Jay Steingrub; Howard Smithline; Bogdan Tiru; Mark Tidswell; Lori Kozikowski; Sherell Thornton-Thompson; Leslie De Souza; Peter Hou; Rebecca Baron; Anthony Massaro; Imoigele Aisiku; Lauren Fredenburgh; Raghu Seethala; Lily Johnsky; Richard Riker; David Seder; Teresa May; Michael Baumann; Ashley Eldridge; Christine Lord; Nathan Shapiro; Daniel Talmor; Thomas O’Mara; Charlotte Kirk; Kelly Harrison; Lisa Kurt; Margaret Schermerhorn; Valerie Banner-Goodspeed; Katherine Boyle; Nicole Dubosh; Michael Filbin; Kathryn Hibbert; Blair Parry; Kendall Lavin-Parsons; Natalie Pulido; Brendan Lilley; Carl Lodenstein; Justin Margolin; Kelsey Brait; Alan Jones; James Galbraith; Rebekah Peacock; Utsav Nandi; Taylor Wachs; Michael Matthay; Kathleen Liu; Kirsten Kangelaris; Ralph Wang; Carolyn Calfee; Kimberly Yee; Gregory Hendey; Steven Chang; George Lim; Nida Qadir; Andrea Tam; Rebecca Beutler; Joseph Levitt; Jenny Wilson; Angela Rogers; Rosemary Vojnik; Jonasel Roque; Timothy Albertson; James Chenoweth; Jason Adams; Skyler Pearson; Maya Juarez; Eyad Almasri; Mohamed Fayed; Alyssa Hughes; Shelly Hillard; Ryan Huebinger; Henry Wang; Elizabeth Vidales; Bela Patel; Adit Ginde; Marc Moss; Amiran Baduashvili; Jeffrey McKeehan; Lani Finck; Carrie Higgins; Michelle Howell; Ivor Douglas; Jason Haukoos; Terra Hiller; Carolynn Lyle; Alicia Cupelo; Emily Caruso; Claudia Camacho; Stephanie Gravitz; James Finigan; Christine Griesmer; Pauline Park; Robert Hyzy; Kristine Nelson; Kelli McDonough; Norman Olbrich; Mark Williams; Raj Kapoor; Jean Nash; Meghan Willig; Henry Ford; Jayna Gardner-Gray; Mayur Ramesh; Montefiore Moses; Michelle Ng Gong; Michael Aboodi; Ayesha Asghar; Omowunmi Amosu; Madeline Torres; Savneet Kaur; Jen-Ting Chen; Aluko Hope; Brenda Lopez; Kathleen Rosales; Jee Young You; Jarrod Mosier; Cameron Hypes; Bhupinder Natt; Bryan Borg; Elizabeth Salvagio Campbell; R Duncan Hite; Kristin Hudock; Autumn Cresie; Faysal Alhasan; Jose Gomez-Arroyo; Abhijit Duggal; Omar Mehkri; Andrei Hastings; Debasis Sahoo; Francois Abi Fadel; Susan Gole; Valerie Shaner; Allison Wimer; Yvonne Meli; Alexander King; Thomas Terndrup; Matthew Exline; Sonal Pannu; Emily Robart; Sarah Karow; Catherine Hough; Bryce Robinson; Nicholas Johnson; Daniel Henning; Monica Campo; Stephanie Gundel; Sakshi Seghal; Sarah Katsandres; Sarah Dean; Akram Khan; Olivia Krol; Milad Jouzestani; Peter Huynh; Alexandra Weissman; Donald Yealy; Denise Scholl; Peter Adams; Bryan McVerry; David Huang; Derek Angus; Jordan Schooler; Steven Moore; Clark Files; Chadwick Miller; Kevin Gibbs; Mary LaRose; Lori Flores; Lauren Koehler; Caryn Morse; John Sanders; Caitlyn Langford; Kristen Nanney; Masiku MdalaGausi; Phyllis Yeboah; Peter Morris; Jamie Sturgill; Sherif Seif; Evan Cassity; Sanjay Dhar; Marjolein de Wit; Jessica Mason; Andrew Goodwin; Greg Hall; Abbey Grady; Amy Chamberlain; Samuel Brown; Joseph Bledsoe; Lindsay Leither; Ithan Peltan; Nathan Starr; Melissa Fergus; Valerie Aston; Quinn Montgomery; Rilee Smith; Mardee Merrill; Katie Brown; Brent Armbruster; Estelle Harris; Elizabeth Middleton; Robert Paine; Stacy Johnson; Macy Barrios; John Eppensteiner; Alexander Limkakeng; Lauren McGowan; Tedra Porter; Andrew Bouffler; J. Clancy Leahy; Bennet deBoisblanc; Matthew Lammi; Kyle Happel; Paula Lauto; Wesley Self; Jonathan Casey; Matthew Semler; Sean Collins; Frank Harrell; Christopher Lindsell; Todd Rice; William Stubblefield; Christopher Gray; Jakea Johnson; Megan Roth; Margaret Hays; Donna Torr; Arwa Zakaria; David Schoenfeld; Taylor Thompson; Douglas Hayden; Nancy Ringwood; Cathryn Oldmixon; Christine Ulysse; Richard Morse; Ariela Muzikansky; Laura Fitzgerald; Samuel Whitaker; Adrian Lagakos; Roy Brower; Lora Reineck; Neil Aggarwal; Karen Bienstock; Michelle Freemer; Myron Maclawiw; Gail Weinmann; Laurie Morrison; Mark Gillespie; Richard Kryscio; Daniel Brodie; Wojciech Zareba; Anne Rompalo; Michael Boeckh; Polly Parsons; Jason Christie; Jesse Hall; Nicholas Horton; Laurie Zoloth; Neal Dickert; Deborah Diercks
Journal:  JAMA       Date:  2020-12-01       Impact factor: 56.272

4.  Pragmatic clinical trials embedded in healthcare systems: generalizable lessons from the NIH Collaboratory.

Authors:  Kevin P Weinfurt; Adrian F Hernandez; Gloria D Coronado; Lynn L DeBar; Laura M Dember; Beverly B Green; Patrick J Heagerty; Susan S Huang; Kathryn T James; Jeffrey G Jarvik; Eric B Larson; Vincent Mor; Richard Platt; Gary E Rosenthal; Edward J Septimus; Gregory E Simon; Karen L Staman; Jeremy Sugarman; Miguel Vazquez; Douglas Zatzick; Lesley H Curtis
Journal:  BMC Med Res Methodol       Date:  2017-09-18       Impact factor: 4.615

5.  A Randomized Trial of Hydroxychloroquine as Postexposure Prophylaxis for Covid-19.

Authors:  David R Boulware; Matthew F Pullen; Ananta S Bangdiwala; Katelyn A Pastick; Sarah M Lofgren; Elizabeth C Okafor; Caleb P Skipper; Alanna A Nascene; Melanie R Nicol; Mahsa Abassi; Nicole W Engen; Matthew P Cheng; Derek LaBar; Sylvain A Lother; Lauren J MacKenzie; Glen Drobot; Nicole Marten; Ryan Zarychanski; Lauren E Kelly; Ilan S Schwartz; Emily G McDonald; Radha Rajasingham; Todd C Lee; Kathy H Hullsiek
Journal:  N Engl J Med       Date:  2020-06-03       Impact factor: 91.245

6.  Effect of Hydroxychloroquine in Hospitalized Patients with Covid-19.

Authors:  Peter Horby; Marion Mafham; Louise Linsell; Jennifer L Bell; Natalie Staplin; Jonathan R Emberson; Martin Wiselka; Andrew Ustianowski; Einas Elmahi; Benjamin Prudon; Tony Whitehouse; Timothy Felton; John Williams; Jakki Faccenda; Jonathan Underwood; J Kenneth Baillie; Lucy C Chappell; Saul N Faust; Thomas Jaki; Katie Jeffery; Wei Shen Lim; Alan Montgomery; Kathryn Rowan; Joel Tarning; James A Watson; Nicholas J White; Edmund Juszczak; Richard Haynes; Martin J Landray
Journal:  N Engl J Med       Date:  2020-10-08       Impact factor: 91.245

7.  Observational Study of Hydroxychloroquine in Hospitalized Patients with Covid-19.

Authors:  Joshua Geleris; Yifei Sun; Jonathan Platt; Jason Zucker; Matthew Baldwin; George Hripcsak; Angelena Labella; Daniel K Manson; Christine Kubin; R Graham Barr; Magdalena E Sobieszczyk; Neil W Schluger
Journal:  N Engl J Med       Date:  2020-05-07       Impact factor: 91.245

8.  Infectious Diseases Society of America Guidelines on the Treatment and Management of Patients with COVID-19.

Authors:  Adarsh Bhimraj; Rebecca L Morgan; Amy Hirsch Shumaker; Valery Lavergne; Lindsey Baden; Vincent Chi-Chung Cheng; Kathryn M Edwards; Rajesh Gandhi; William J Muller; John C O'Horo; Shmuel Shoham; M Hassan Murad; Reem A Mustafa; Shahnaz Sultan; Yngve Falck-Ytter
Journal:  Clin Infect Dis       Date:  2020-04-27       Impact factor: 9.079

9.  Dexamethasone in Hospitalized Patients with Covid-19.

Authors:  Peter Horby; Wei Shen Lim; Jonathan R Emberson; Marion Mafham; Jennifer L Bell; Louise Linsell; Natalie Staplin; Christopher Brightling; Andrew Ustianowski; Einas Elmahi; Benjamin Prudon; Christopher Green; Timothy Felton; David Chadwick; Kanchan Rege; Christopher Fegan; Lucy C Chappell; Saul N Faust; Thomas Jaki; Katie Jeffery; Alan Montgomery; Kathryn Rowan; Edmund Juszczak; J Kenneth Baillie; Richard Haynes; Martin J Landray
Journal:  N Engl J Med       Date:  2020-07-17       Impact factor: 91.245

Review 10.  Updated guidance on the management of COVID-19: from an American Thoracic Society/European Respiratory Society coordinated International Task Force (29 July 2020).

Authors:  Chunxue Bai; Sanjay H Chotirmall; Jordi Rello; George A Alba; Leo C Ginns; Jerry A Krishnan; Robert Rogers; Elisabeth Bendstrup; Pierre-Regis Burgel; James D Chalmers; Abigail Chua; Kristina A Crothers; Abhijit Duggal; Yeon Wook Kim; John G Laffey; Carlos M Luna; Michael S Niederman; Ganesh Raghu; Julio A Ramirez; Jordi Riera; Oriol Roca; Maximiliano Tamae-Kakazu; Antoni Torres; Richard R Watkins; Miriam Barrecheguren; Mirko Belliato; Hassan A Chami; Rongchang Chen; Gustavo A Cortes-Puentes; Charles Delacruz; Margaret M Hayes; Leo M A Heunks; Steven R Holets; Catherine L Hough; Sugeet Jagpal; Kyeongman Jeon; Takeshi Johkoh; May M Lee; Janice Liebler; Gerry N McElvaney; Ari Moskowitz; Richard A Oeckler; Iñigo Ojanguren; Anthony O'Regan; Mathias W Pletz; Chin Kook Rhee; Marcus J Schultz; Enrico Storti; Charlie Strange; Carey C Thomson; Francesca J Torriani; Xun Wang; Wim Wuyts; Tao Xu; Dawei Yang; Ziqiang Zhang; Kevin C Wilson
Journal:  Eur Respir Rev       Date:  2020-10-05
View more
  2 in total

1.  The COVID-19 Hospitalization Metric in the Pre- and Post-vaccination Eras as a Measure of Pandemic Severity: A Retrospective, Nationwide Cohort Study.

Authors:  Jennifer La; Paul Monach; Nathanael R Fillmore; Chunlei Zheng; Shira Doron; Nhan Do; Westyn Branch-Elliman
Journal:  Infect Control Hosp Epidemiol       Date:  2022-01-11       Impact factor: 6.520

2.  Subcutaneous sarilumab for the treatment of hospitalized patients with moderate to severe COVID19 disease: A pragmatic, embedded randomized clinical trial.

Authors:  Westyn Branch-Elliman; Ryan Ferguson; Gheorghe Doros; Patricia Woods; Sarah Leatherman; Judith Strymish; Rupak Datta; Rekha Goswami; Matthew D Jankowich; Nishant R Shah; Thomas H Taylor; Sarah T Page; Sara J Schiller; Colleen Shannon; Cynthia Hau; Maura Flynn; Erika Holmberg; Karen Visnaw; Rupali Dhond; Mary Brophy; Paul A Monach
Journal:  PLoS One       Date:  2022-02-25       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.