Literature DB >> 32943437

Computerised clinical decision support systems and absolute improvements in care: meta-analysis of controlled clinical trials.

Janice L Kwan1,2, Lisha Lo3, Jacob Ferguson4, Hanna Goldberg4, Juan Pablo Diaz-Martinez5, George Tomlinson5, Jeremy M Grimshaw6, Kaveh G Shojania2,3,7.   

Abstract

OBJECTIVE: To report the improvements achieved with clinical decision support systems and examine the heterogeneity from pooling effects across diverse clinical settings and intervention targets.
DESIGN: Systematic review and meta-analysis. DATA SOURCES: Medline up to August 2019. ELIGIBILITY CRITERIA FOR SELECTING STUDIES AND METHODS: Randomised or quasi-randomised controlled trials reporting absolute improvements in the percentage of patients receiving care recommended by clinical decision support systems. Multilevel meta-analysis accounted for within study clustering. Meta-regression was used to assess the degree to which the features of clinical decision support systems and study characteristics reduced heterogeneity in effect sizes. Where reported, clinical endpoints were also captured.
RESULTS: In 108 studies (94 randomised, 14 quasi-randomised), reporting 122 trials that provided analysable data from 1 203 053 patients and 10 790 providers, clinical decision support systems increased the proportion of patients receiving desired care by 5.8% (95% confidence interval 4.0% to 7.6%). This pooled effect exhibited substantial heterogeneity (I2=76%), with the top quartile of reported improvements ranging from 10% to 62%. In 30 trials reporting clinical endpoints, clinical decision support systems increased the proportion of patients achieving guideline based targets (eg, blood pressure or lipid control) by a median of 0.3% (interquartile range -0.7% to 1.9%). Two study characteristics (low baseline adherence and paediatric settings) were associated with significantly larger effects. Inclusion of these covariates in the multivariable meta-regression, however, did not reduce heterogeneity.
CONCLUSIONS: Most interventions with clinical decision support systems appear to achieve small to moderate improvements in targeted processes of care, a finding confirmed by the small changes in clinical endpoints found in studies that reported them. A minority of studies achieved substantial increases in the delivery of recommended care, but predictors of these more meaningful improvements remain undefined. © Author(s) (or their employer(s)) 2019. Re-use permitted under CC BY-NC. No commercial re-use. See rights and permissions. Published by BMJ.

Entities:  

Mesh:

Year:  2020        PMID: 32943437      PMCID: PMC7495041          DOI: 10.1136/bmj.m3216

Source DB:  PubMed          Journal:  BMJ        ISSN: 0959-8138


Introduction

Although the first electronic health record (EHR) was introduced almost six decades ago,1 dissemination has occurred surprisingly slowly.2 In the United States, as recently as 2008, fewer than 10% of hospitals had EHRs, and only 17% had computerised order entry of drugs. By 2015, however, 75% of US hospitals had adopted at least basic EHR systems.3 This rapid uptake reflects the financial incentives in the Health Information Technology for Economic and Clinical Health (HITECH) Act passed in 2009.4 Although EHR adoption has been slower in the UK,5 the NHS long term plan, released in 2019, sets a goal for all trusts and providers to move to digital health records by 2024.6 Clinical decision support systems embedded within EHRs—from pop-up alerts about serious drug allergies to more sophisticated tools incorporating clinical prediction rules—prompt clinicians to deliver evidence based processes of care,7 discourage non-indicated care,8 9 optimise drug orders,10 11 and improve documentation.12 13 Despite optimism over the effects of these support systems,14 15 16 a systematic review published by our group in 201017 found that clinical decision support systems typically improved the proportion of patients who received target processes of care by less than 5%. The subsequent decade has seen a dramatic rise in the application and evaluation of clinical decision support systems. Systematic reviews of an increasingly large number of publications, however, have typically looked only at the features of these support systems that predicted improvements in care,14 18 and reported odds ratios or risk ratios,19 20 21 without quantifying the actual sizes of the improvements achieved. In our systematic review and meta-analysis, we sought firstly, to estimate the typical improvement in processes of care—and thus the potential for clinical effect—conferred by clinical decision support systems delivered at the point of care; and secondly, to identify any study characteristics or features of these support systems consistently associated with larger effects.

Methods

We followed established methods recommended by Cochrane22 and report our findings in accordance with PRISMA (preferred reporting items for systematic reviews and meta-analyses).23

Search strategy and selection criteria

We searched Medline from the earliest available date to August 2019 without language restrictions (supplementary appendix 1) and scanned reference lists from included studies and relevant systematic reviews. We did not search Embase, CINAHL, and Cochrane Central Register of Controlled Trials, as these databases did not increase the yield of eligible studies included in our previous review.17 24 We did, however, conduct fresh study screening and data abstraction for all articles, even those covered by the previous review, to accommodate new and modified data elements reflecting changes in technology in the intervening decade.17 24 Eligible studies evaluated the effects of clinical decision support systems on processes or outcomes of care using a randomised or quasi-randomised controlled design (allocation on the basis of an arbitrary but not truly random process, such as even or odd patient identification numbers). Patients in control arms received “usual care” contemporaneous with care delivered in the intervention arm (that is, we excluded head-to-head comparisons of different clinical decision support systems). We defined a clinical decision support system as any on-screen tool designed to improve adherence of physicians to a recommended process of care. Eligible studies delivered the support system intervention within a clinical information system routinely used by the provider (not a computer application separate from the EHR) at the time of providing care to the targeted patient (eg, while entering an order or a clinical note). We excluded specialised diagnostic decision support systems (eg, in medical imaging) and systems not directly related to patient care, such as decision support for billing or health record coding. We also excluded studies using simulated patients and studies in which fewer than half of participants were physicians or physician trainees. We focused on improvements in processes of care (eg, prescribing drugs, immunisations, test ordering, documentation), rather than clinical outcomes, because we sought to determine the degree to which clinical decision support systems achieve their immediate goal—namely, changing provider behaviour. The extent to which such changes ultimately improve patient outcomes will vary depending on the strength of the relation between targeted processes and clinical outcomes. Nonetheless, we did capture clinical outcomes where reported, including intermediate endpoints, such as haemoglobin A1c and blood pressure. We coded all results so that larger numbers always corresponded to improvements in care. For example, if a study reported the proportion of patients who received inappropriate drugs,8 25 we recorded the complementary proportion of patients who did not receive inappropriate drugs. Two investigators independently evaluated the eligibility of all identified studies based on titles and abstracts. Studies not excluded in this first step were independently assessed for inclusion after full text evaluation by two investigators. For articles that met all inclusion criteria, two investigators independently extracted the following information: clinical setting, participants, methodological details, characteristics of the design and content of the clinical decision support system, presence of educational and non-educational co-interventions, and outcomes. We extracted studies with more than one eligible intervention arm as separate trials (that is, comparisons). Two investigators also independently assessed risk of bias for each study using criteria outlined in the Cochrane Handbook for Systematic Reviews of Interventions.26 27 Specific biases considered for each study included selective enrolment, attrition bias, similarity of baseline characteristics, unit of analysis errors, performance bias (systematic differences between groups with respect to the care provided or exposures other than the interventions of interest), and detection bias (systematic differences between groups in outcome ascertainment). We resolved all discrepancies and disagreements by discussion and consensus among the study team.

Data analysis

We used a multilevel meta-analysis model to estimate the absolute improvements (risk differences) in processes of care between intervention and control groups, using the number of patients receiving the target process of care and the total sample size for each reported outcome. This approach allowed us to account for heterogeneity between studies, and for clustering of multiple outcomes reported for the same patients within a given study. Most trials used clustered designs, assigning intervention status to the provider or provider group rather than to the individual patient, but did not always report cluster adjusted estimates. We accounted for clustering by multiplying the standard error of the risk differences by the square root of the design effect.22 Intraclass correlation coefficients were abstracted directly from the study when reported. To impute intraclass correlation coefficients in studies that did not report them, we used a published database of intraclass correlation coefficients28 stratified by type of setting (eg, hospital versus ambulatory care) and endpoint (eg, process versus outcome). We calculated the median intraclass correlation coefficient for hospital and ambulatory process measures across the 200 studies in this database and applied the relevant value to a given study. For clinical endpoints, we quantified the median improvement and interquartile range across all studies that reported dichotomous clinical endpoints, such as the percentage of patients who achieved a target blood pressure or the percentage of patients who experienced a clinical event (eg, a critical laboratory result, adverse drug event, or venous thromboembolism). This method29 for summarising the effects of improvement interventions on disparate clinical endpoints was first developed in a large systematic review of implementation strategies for clinical practice guidelines,30 and subsequently applied in other systematic reviews.31 32 We also calculated the median improvement and interquartile range for changes in blood pressure, the most commonly reported continuous clinical endpoint. We performed univariate meta-regression analyses to explore the extent to which effects varied according to the study characteristics and features of the clinical decision support system. These analyses estimated the difference in absolute improvements reported between studies with and without each feature. Most features are easy to understand (eg, the setting of the intervention as hospital or ambulatory, the presence of co-interventions other than clinical decision support system). Table 1 defines those features with less obvious meanings.
Table 1

Study and clinical decision support system features

Study or CDSS feature Definition
Acknowledgement and documentationRequiring the user to register receipt of the information (eg, clicking on a YES or OK button) and to record the rationale for the clinical decision (eg, selecting from a dropdown list of guideline based indications when ordering a broad spectrum antibiotic)
AmbushA CDSS that is likely to appear when not relevant to the immediate clinical task being performed by the user (eg, a pop-up prompting the user to order vaccinations upon opening a patient’s record at all visits, not just preventative visits)
Baseline adherenceThe proportion of patients in the control group receiving the desired process of care. In studies that reported pre-intervention data, the mean percentage of patients in the pre-intervention control and intervention groups was calculated
Complex decision supportA CDSS that incorporates two or more pieces of clinical or demographic information to guide clinical decision making (eg, a tool that recommends deep vein thrombosis prophylaxis for a newly admitted inpatient over the age of 65). This complex decision support is in contrast to simple decision support, which incorporates no pieces, or one piece, of clinical or demographic information
Consideration of alert fatigueStudies that specifically mentioned alert fatigue in designing or delivering the CDSS intervention under study
InterruptiveA CDSS that appears on screen and stops the user from performing any further tasks until it is either “acknowledged”, or a button is pressed, such as the Escape key
PushA CDSS is fully presented to the user without requiring additional clicks versus “pull”, in which the user is required to click a link or icon to receive information contained within the CDSS
Target underuseA CDSS that aims to increase the percentage of patients who receive a recommended process of care (eg, adherence to guideline recommended care). Systems that target underuse are in contrast to those systems that target overuse, whereby improvements correspond to reductions in the percentage of patients receiving inappropriate or unnecessary processes of care (eg, daily bloodwork in a clinically stable patient)

CDSS=clinical decision support system.

Study and clinical decision support system features CDSS=clinical decision support system. Finally, we fitted a multivariable meta-regression model that included covariates with P<0.1 in the univariate analyses, and simplified the model by stepwise selection.33 We used the I2 statistic to summarise statistical heterogeneity (that is, the degree to which trials exhibited non-random variation in effect sizes). We expected substantial heterogeneity in effect sizes given the range of care settings, targeted clinician behaviours, and design features of the clinical decision support system. The multivariable meta-regression model was used to identify study and clinical decision support system features that predicted larger effects and to determine if heterogeneity could be reduced. We performed all statistical analyses using R Software, version 3.5.1 (R Foundation for Statistical Computing, Vienna). Rma.mv function from metafor library was used to fit all models.

Patient and public involvement

Patients or the public were not involved in the design, conduct, reporting, or dissemination plans of this study. There were no funds or time allocated for patient and public involvement at the time of the study so we were unable to involve patients.

Results

Our search identified 5895 citations, of which 5428 were excluded at the initial stage of screening and an additional 352 on full text review, yielding a total of 115 studies that met all inclusion criteria (fig 1 and supplementary appendix 2). Nine studies contained two trials (that is, two comparisons of a clinical decision support system intervention with a control group) 7 34 35 36 37 38 39 40 41 and one study contained six such trials,42 resulting in 129 included trials.
Fig 1

Flow of studies through the review process

Flow of studies through the review process Of the 129 included trials (table 2 and supplementary table 1), most used a true randomised design, with only 16 (12%) of the 129 trials involving a quasi-random allocation process. Most trials (113/129; 88%) had a clustered design, allocating intervention status to providers or provider groups rather than patients. Most trials (93/129; 72%) occurred in outpatient settings, and 85 of 129 trials (66%) came from US centres. Moreover, 93 (72%) of all 129 trials were published from 2009 onwards, when the US HITECH Act became law. The period since 2009 also included significantly more trials using commercial EHRs (39/93 (42%) v 2/36 (6%), P<0.05 with Bonferroni correction for multiple comparisons). Epic accounted for 26 (63%) of all 41 interventions on commercial EHR platforms. As shown in supplementary figures 1a-b, the risk of bias was generally low. The one exception involved unit of analysis errors, which occurred in 18 (16%) of 113 clustered studies and thus possessed a high risk of bias. (We included these studies because our use of intraclass correlation coefficients and a multilevel model avoided replicating unit of analysis errors from the primary studies.)
Table 2

Summary of characteristics of included trials, according to publication year. Data are number (%) of trials unless stated otherwise

Characteristics of trialsBefore 2009 (n=36)2009 onwards (n=93)Total (n=129)
Targeted process of care:
 Prescribing24 (67)44 (47)68 (53)
 Test ordering12 (33)18 (19)30 (23)
 Documentation3 (8)22 (24)25 (19)
 Vaccination5 (14)7 (8)12 (9)
 Other7 (19)28 (30)35 (27)
Study features:
 Randomised controlled trial30 (83)83 (89)113 (88)
 United States23 (64)62 (67)85 (66)
 Outpatient23 (64)70 (75)93 (72)
 No of providers (median (IQR))89 (41-151)79 (36-171)81 (36-171)
 No of patients (median (IQR))2254 (452-6275)2204 (450-11 282)2237 (450-8574)
 Baseline adherence (median (IQR))29 (21-50)45 (17-64)39 (17-62)
 Co-interventions (education)12 (33)48 (52)60 (47)
 Co-interventions (non-education)8 (22)28 (30)36 (28)
CDSS features:
 Push22 (61)72 (77)94 (73)
 Complex decision support3 (8)40 (43)43 (33)
 Acknowledgement and documentation2 (6)18 (19)20 (16)
 Execute action through CDSS14 (39)46 (49)60 (47)
 Ambush3 (8)30 (32)33 (26)
 Interruptive14 (39)45 (48)59 (46)
 Other concurrent CDSS2 (6)12 (13)14 (11)
 Consideration of alert fatigue4 (11)15 (16)19 (15)
 Target underuse27 (75)57 (61)84 (65)
 Commercial EHR (eg, Epic)*2 (6)39 (42)41 (32)

CDSS=clinical decision support system; EHR=electronic health record; IQR=interquartile range.

Proportion of studies involving commercial systems increased significantly on or after 2009 compared with earlier (Bonferroni adjusted P<0.05).

Summary of characteristics of included trials, according to publication year. Data are number (%) of trials unless stated otherwise CDSS=clinical decision support system; EHR=electronic health record; IQR=interquartile range. Proportion of studies involving commercial systems increased significantly on or after 2009 compared with earlier (Bonferroni adjusted P<0.05). Across the 122 trials that provided analysable data from 1 203 053 patients and 10 790 providers (fig 2), clinical decision support systems produced an average absolute improvement of 5.8% (95% confidence interval 4.0% to 7.6%) in the percentage of patients receiving the desired process of care. For specific types of physician behaviours, prescribing improved by 4.4% (95% confidence interval 2.6% to 6.2%; 68 trials); test ordering by 6.8% (4.9% to 8.6%; 30 trials); documentation adherence by 7.1% (5.4% to 8.9%; 25 trials; vaccination by 5.9% (4.1% to 7.7%; 12 trials); and other process measures by 6.8% (5.0% to 8.6%; 35 trials). Other process measures included referral for specialty consultations,43 44 45 overall guideline concordance,46 47 48 and documenting key elements of the diagnostic process.49 50 51 Supplementary figures 2a-e present the multilevel models for each category.
Fig 2

Absolute improvements in desired care by different categories of clinical care. Results from the multilevel random effects meta-analysis are shown. The diamond shows the summary overall absolute improvement and 95% confidence interval across all types of outcomes; the squares with lines represent estimates and their 95% confidence intervals for different categories of clinical care. *Other process outcomes included referrals for specialty consultations, overall guideline concordance, and diagnosis

Absolute improvements in desired care by different categories of clinical care. Results from the multilevel random effects meta-analysis are shown. The diamond shows the summary overall absolute improvement and 95% confidence interval across all types of outcomes; the squares with lines represent estimates and their 95% confidence intervals for different categories of clinical care. *Other process outcomes included referrals for specialty consultations, overall guideline concordance, and diagnosis In univariate meta-regression analyses (table 3), clinical decision support systems requiring acknowledgement and documentation of a reason for not adhering to the recommended action achieved improvements 4.8% larger than support systems without this feature (95% confidence interval 0.1% to 9.6%). The ability to execute the desired action through the clinical decision support system was also associated with larger effects than interventions without this feature, with an incremental difference of 4.4% (0.9% to 7.9%). Only 17 (14%) of 122 trials considered alert fatigue in designing or delivering the clinical decision support system, and the associated incremental increase for this feature was small (1.7%) and non-significant (95% confidence interval −3.5% to 6.8%; P=0.52).
Table 3

Absolute incremental improvements in desired care by CDSS feature

CDSS featureCategory (No of trials)Absolute incremental improvement (%; 95% CI)*P value
Acknowledgement and documentation required Yes (20) v no (102)4.8 (0.1 to 9.6)0.05
Execute desired action through CDSS Yes (58) v no (64)4.4 (0.9 to 7.9)0.01
Behaviour targeted, when reported Underuse (82) v overuse (37)4.0 (0.1 to 7.9)0.05
EHR platform Epic (25) v other (97)3.8 (−0.5 to 8.2)0.09
Interruptive Yes (55) v no (67)3.2 (−0.4 to 6.7)0.08
Mode of delivery, when reported Push (90) v pull (13)2.4 (−3.5 to 8.2)0.43
Other concurrent CDSS Yes (14) v no (108)2.1 (−3.4 to 7.6)0.46
Considered alert fatigue in design Yes (17) v no (105)1.7 (−3.5 to 6.8)0.52
User workflow specifically considered in design Yes (35) v no (87)1.2 (−2.7 to 5.2)0.53
Level of decision support Complex (40) v simple (82)0.9 (−2.9 to 4.7)0.64
Developed in consultation with users Yes (7) v no (115)0.1 (−7.3 to 7.5)0.98
Makes recommendation for care Yes (98) v no (24)0 (−4.4 to 4.5)0.98
Appearance differed based on urgency Yes (3) v no (119)−0.5 (−11.4 to 10.4)0.93
Inclusion of supporting information on screen Yes (53) v no (69)−1.1 (−4.7 to 2.5)0.55
Ambush Yes (32) v no (90)−1.3 (−5.4 to 2.7)0.51
Developed by study investigators Yes (62) v no (60)−1.5 (−5.1 to 2.0)0.40
Requires input of clinical information Yes (9) v no (113)−1.9 (−8.8 to 4.9)0.58
Conveys patient-specific information Yes (107) v no (15)−4.4 (−9.8 to 1.0)0.11

CDSS=clinical decision support system; EHR=electronic health record.

The third column shows the estimated pooled difference (and 95% confidence interval) in the percentage of desired care between the trials classified according to the CDSS features in the first column. Results from 122 trials (108 studies) reporting dichotomous outcomes were each estimated in a univariate meta-regression model.

Absolute incremental improvements in desired care by CDSS feature CDSS=clinical decision support system; EHR=electronic health record. The third column shows the estimated pooled difference (and 95% confidence interval) in the percentage of desired care between the trials classified according to the CDSS features in the first column. Results from 122 trials (108 studies) reporting dichotomous outcomes were each estimated in a univariate meta-regression model. Improvements seen for clinical decision support systems in paediatric settings exceeded those in other patient populations by 14.7% (95% confidence interval 8.4% to 21.0%; table 4). Additional study features associated with significant incremental effects included small studies (lower than the median patient sample size), with an incremental improvement of 3.7% (0.2% to 7.3%) greater than larger studies, and studies that took place in the US, with an incremental improvement of 3.7% (0% to 7.4%).
Table 4

Absolute incremental improvements in desired care by study feature

Study feature Category (No of trials)Absolute incremental improvement (%; 95% CI)*P value
Clinical specialty Paediatrics (9) v other (113)14.7 (8.4 to 21.0)<0.001
Patient sample size† Small (<median) v large (≥median)3.7 (0.2 to 7.3)0.04
Country US (82) v other (40)3.7 (0 to 7.4)0.05
Baseline adherence‡ <Median v ≥median3.3 (−0.2 to 6.9)0.06
Co-interventions beyond clinician education Yes (35) v no (87)1.3 (−2.6 to 5.1)0.53
Provider sample size§ Small (<median) v large (≥median)0.2 (−5.7 to 6.2)0.94
Study design Quasi-RCT (16) v RCT (106)0.1 (−5.4 to 5.6)0.96
Educational co-interventions Yes (57) v no (65)−0.8 (−4.4 to 2.8)0.66
Publication year 2009 onwards (87) v before 2009 (35)−0.8 (−4.8 to 3.2)0.70
Care setting Outpatient (90) v other (32)−3.7 (−7.8 to 0.5)0.08

RCT=randomised controlled trial.

The third column shows the estimated pooled difference (and 95% confidence interval) in the percentage of desired care between the two study types listed in the first column. Results from 122 trials (108 studies) reporting dichotomous outcomes were each estimated using a univariate meta-regression model.

Defined as small or large relative to the median patient sample size. Across all trials, the median was 2237 (interquartile range 450 to 8574).

The proportion of patients in control groups who received the desired process of care. Across all trials, the median was 39.4% (interquartile range 17.3% to 62.3%).

Defined as small and large relative to the median provider sample size. Across all trials, the median was 81 (interquartile range 36 to 171).

Absolute incremental improvements in desired care by study feature RCT=randomised controlled trial. The third column shows the estimated pooled difference (and 95% confidence interval) in the percentage of desired care between the two study types listed in the first column. Results from 122 trials (108 studies) reporting dichotomous outcomes were each estimated using a univariate meta-regression model. Defined as small or large relative to the median patient sample size. Across all trials, the median was 2237 (interquartile range 450 to 8574). The proportion of patients in control groups who received the desired process of care. Across all trials, the median was 39.4% (interquartile range 17.3% to 62.3%). Defined as small and large relative to the median provider sample size. Across all trials, the median was 81 (interquartile range 36 to 171). As expected, the meta-analytic improvement in processes of care across all trials exhibited substantial heterogeneity (I2=76%), with a minority of studies reporting much larger improvements than the meta-analytic average. The top quartile of trials reported improvements in process adherence ranging from 10% to 62%. Using multivariable meta-regression (table 5), the final model identified paediatric studies as achieving incremental improvements of 13.6% (95% confidence interval 7.4% to 19.8%), and those trials with low baseline adherence (relative to the median across all studies) reported incrementally greater improvements by 3.2% (0% to 6.4%). Even when these characteristics were incorporated in the meta-regression model, heterogeneity remained high and essentially unchanged.
Table 5

Multivariable meta-regression model for the absolute incremental improvements in desired care by study and CDSS features

Study and CDSS features*Category (No of trials)Absolute incremental improvement (%; 95% CI)†P value
Clinical specialty Paediatrics (9) v other (113)13.6 (7.4 to 19.8)<0.001
Baseline adherence‡ <Median v ≥median3.2 (0 to 6.4)0.05
Patient sample size§ Small (<median) v large (≥median)2.4 (−0.8 to 5.7)0.14

CDSS=clinical decision support system.

The covariates listed in the table include all those that were included in the final multivariable meta-regression model, fitted through a backwards stepwise procedure that initially included all study and CDSS feature covariates found to have P<0.1 in the univariate meta-regression models.

Incremental improvements refer to the additional increases in the percentages of patients receiving the desired care beyond those reported in interventions without this feature.

The proportion of patients in control groups who received the desired process of care. Across all trials, the median was 39.4% (interquartile range 17.3% to 62.3%).

Defined as small and large relative to the median patient sample size. Across all trials, the median was 2237 (interquartile range 450 to 8574).

Multivariable meta-regression model for the absolute incremental improvements in desired care by study and CDSS features CDSS=clinical decision support system. The covariates listed in the table include all those that were included in the final multivariable meta-regression model, fitted through a backwards stepwise procedure that initially included all study and CDSS feature covariates found to have P<0.1 in the univariate meta-regression models. Incremental improvements refer to the additional increases in the percentages of patients receiving the desired care beyond those reported in interventions without this feature. The proportion of patients in control groups who received the desired process of care. Across all trials, the median was 39.4% (interquartile range 17.3% to 62.3%). Defined as small and large relative to the median patient sample size. Across all trials, the median was 2237 (interquartile range 450 to 8574). Thirty trials reported dichotomous clinical endpoints (supplementary table 1). These endpoints included achieving guideline based targets for blood pressure52 53 54 and lipid levels53 54; adverse events, such as bleeding55 56; in-hospital pulmonary embolism57 58; hospital readmission59 60; and mortality.58 60 For these various endpoints, clinical decision support systems increased the proportion of patients achieving guideline based targets by a median of 0.3% (interquartile range −0.7% to 1.9%). Twenty trials reported continuous clinical endpoints (supplementary table 1), including laboratory test values (eg, haemoglobin A1c54 61), and questionnaire scores, such as the 36-Item Short Form Survey.62 63 Blood pressure was the most commonly reported continuous clinical endpoint.52 53 54 64 65 Patients in intervention groups had a median reduction in diastolic blood pressure of 1.0 mm Hg (interquartile range 1.0 mm Hg decrease to 0.2 mm Hg increase). Median systolic blood pressure for patients receiving an intervention, however, increased by 1.0 mm Hg (interquartile range 0.3 mm Hg decrease to 1.0 mm Hg increase). We conducted sensitivity analyses by reanalysing the percentage of patients receiving desired processes of care using the largest improvement from each trial. Instead of each trial contributing data for all eligible processes of care, we used only the largest improvement among the outcomes reported in a given trial for these sensitivity analyses. These “best case analyses” produced a pooled absolute improvement in patients receiving desired care of 8.5% (95% confidence interval 6.8% to 10.2%). We also repeated our analyses using odds ratios rather than risk differences, since absolute risk differences can produce less stable meta-analytic estimates than relative risks or odds ratios across varying levels of background risk.66 The pooled odds ratio for improvement in adherence was 1.43 (95% confidence interval 1.30 to 1.58). Overall, when odds ratios were used, our results remained qualitatively the same for important study and clinical decision support system features, types of outcome, and level of heterogeneity.

Discussion

Principal findings

Across 122 controlled, mostly randomised, trials involving 1 203 053 patients and 10 790 providers, clinical decision support systems improved the average percentage of patients receiving the desired element of care by 5.8% (95% confidence interval 4.0% to 7.6%). As expected, these trials exhibited substantial heterogeneity (I2=76%). Although it is generally not advisable to perform a meta-analysis with this degree of heterogeneity, we have reported these results to show that the current literature, despite its substantial size, provides little guidance for identifying the circumstances under which clinical decision support system interventions produce worthwhile improvements in care.

Implications

On the one hand, the extreme heterogeneity indicates non-random variation in effect sizes, such that a minority of interventions might have achieved significantly larger effects than the 95% confidence intervals around the meta-analytic average. Indeed, 25% of studies reported absolute improvements greater than 10%, with one as high as 62%.67 Yet, even with the identification of two significant predictors of larger effects—namely, paediatric studies and those with low baseline adherence—the meta-regression model still showed extreme heterogeneity. Thus even when these characteristics were taken into account, a wide, non-random variation remained in the improvements seen with clinical decision support systems. The reason for this remains largely unknown. Other systematic reviews have similarly reported extreme heterogeneity, with I2 as high as 97% in one instance.18 Moja and colleagues reported more moderate heterogeneity (I2=41-64%),21 but their review focused specifically on measures of morbidity and mortality. Other reviews alluded to high heterogeneity without reporting formal analyses,19 whereas others did not mention it.14 20 Previous reviews of clinical decision support systems have typically looked only at predictors of improvements in care,14 18 or reported odds or risk ratios19 20 21 for receiving recommended care. We sought to characterise the typical improvement achieved—namely, a 5.8% increase in the percentage of patients receiving the desired process of care. To put the magnitude of this improvement into perspective, for the control groups in the included trials, a median of 39.4% of patients received care recommended by the clinical decision support system. Thus, in the typical intervention group, about 45% would receive this recommended process of care. Even if the meta-analytic result of 8.5% from our sensitivity analysis incorporating the largest improvement reported in each trial were used, the typical clinical decision support system intervention would still leave over 50% of patients not receiving recommended care. We would characterise these absolute increases of 5.8% to 8.5% in the percentage of patients receiving recommended care as a small to moderate effect, but this does not imply that all such improvements are unimportant. For some processes of care, such as vaccinations and evidence based cancer screening, even a small increase in the percentage of patients who receive this care will translate into worthwhile benefits at the population level. For many other processes of care, where the recommendation has a weaker connection with important outcomes, a small increase in adherence may not justify its implementation and subsequent contribution to clinicians’ frustrations with EHRs.68 Clinical decision support systems embody one of the eagerly awaited applications of artificial intelligence and machine learning for patient care.69 By leveraging the power of “big data,” these technologies promise earlier recognition of sepsis70 71 and impending clinical deterioration.72 73 Even with the most effective artificial intelligence algorithms,74 however, the decision support tools alerting clinicians to their complex outputs will still largely depend on—and therefore be limited by—the small to moderate effect sizes typically obtained in our analysis. Notably, studies that targeted a paediatric population were associated with the largest absolute improvement in process adherence. Many of these studies occurred at large health centres, affiliated to universities, with mature clinical information systems. In our earlier review,17 we noted that studies conducted at institutions with longstanding experience in clinical informatics showed significantly larger improvements. In this updated review, we performed similar analyses and found no association between effect size and specific institutions or mature homegrown systems. Although much of the early work on EHRs and clinical decision support systems took place on inpatients, 93 (72%) of the 129 trials included in this analysis took place in the outpatient setting. Across both settings, most trials (66%; 85/129) came from the US. Nonetheless, we believe our findings are relevant elsewhere. For instance, primary care EHRs in the UK increasingly record more and more coded demographic, diagnostic, and therapeutic data, providing a basis for future clinical decision support systems. Several factors could explain the disappointingly small effect sizes typically achieved by clinical decision support systems. Firstly, researchers and leaders in clinical informatics and human factors have recognised for years the importance of informing this work with a rich sociotechnical model. This model includes the human-computer interface, hardware and software computing infrastructure, clinical content, people, workflow and communication, internal organisational culture, external regulations, and system measurement.75 76 77 78 79 80 81 Yet, those who develop and implement clinical decision support systems typically do not take into account the full range of these dimensions, or the complex interplay between them. In addition to using such considerations to inform the design and evaluation of clinical decision support systems, future studies should report specific potential effect modifiers, including physician EHR training, number of alerts for each visit, physician visit volume, and broader contextual factors, such as validated measures of burnout and local safety culture. A recent publication guideline may foster such improved reporting of trials of clinical decision support systems.82 Alert fatigue is a second possible explanation for the small to moderate effect sizes typically achieved in our analysis. Clinicians could encounter the same alert for many patients, or a given clinical information system could have several clinical decision support systems operating concurrently. In either case, clinicians could become less receptive to alerts from these support systems. Surprisingly, only 19 (15%) of 129 trials in this review mentioned alert fatigue in their design. Although only a minority of trials considered alert fatigue, studies focusing on this topic outside of controlled trials have highlighted the problem, and strategies to minimise the burden to users from clinical decision support systems are under investigation. Growing concerns that alert fatigue contributes to dissatisfaction with EHRs, and even clinician burnout, underscore the importance of considering this problem in future studies of support systems.68 Finally, as has happened with other common improvement interventions, from bundles and checklists to performance report cards and financial incentives, clinical decision support systems have become a “go to” off-the-shelf intervention. Many reflexively reach for these interventions without considering the way they work or the degree to which they would be likely to help with the cause of the target problem.83 84 The decision to employ a clinical decision support system often reflects ease of deployment rather than its usefulness in dealing with the problem. A pop-up computer screen reminding clinicians of the approved indications for a broad spectrum antibiotic may ignore the psychological reality of the clinician’s primary concern at that moment—avoiding further deterioration of a very sick patient, rather than the public health consequences of excessive antibiotic use. An antimicrobial stewardship programme85 could hold far greater promise in achieving this goal than an interruptive alert, which many clinicians will probably ignore, or worse, could result in alert fatigue, thereby undermining the potential effectiveness of other support system interventions in a given healthcare setting. Future interventions may also seek to harness potential synergies between clinical decision support systems and other well known interventions, such as performance feedback.86

Risk of bias

The only included studies with high risk of bias were 18 (16%) of 113 clustered trials exhibiting unit of analysis errors, which typically underestimate the width of confidence intervals. We were able to include such studies without replicating this bias because we used their primary data, rather than the reported effect sizes, together with reported or imputed intraclass correlation coefficients for each clustered trial in a multilevel model. The lack of other trials with a high risk of bias among the included studies probably reflects the nature of the intervention (that is, clinical decision support systems delivered at the point-of-care) and the endpoints in our analysis (that is, the degree to which patients received recommended care as documented in the EHR), as many of the methodological shortcomings that can undermine evaluations of other interventions are easier to avoid. For instance, systematic differences in outcome ascertainment can hardly arise. There is also no clear way in which exposures other than the intervention of interest (the clinical decision support system) could systematically differ between intervention and control patients. Losses to follow-up (that is, attrition bias) are also virtually impossible given that the EHR captures whether or not the clinician’s orders or documentation complied with recommended care in response to the clinical decision support system triggered. Attrition due to loss of entire clusters is the one possible exception and occurred in several trials in which one or more clinics assigned to the control arm dropped out of the trial. In a representative example of such a trial,87 participating clinics received an EHR system with or without additional guideline based decision support. Some clinics assigned to the control group dropped out owing to lack of motivation to implement a new system with no chance of benefitting their patients. The decision of clinics in the control arm not to follow through with an information technology intervention, which is well known to take personnel time and cause frustration to clinicians, did not seem to us to carry a clear risk of bias. We labelled the few such trials as “unclear risk.”

Limitations

Heterogeneity among the included studies is the main limitation to our analysis. Recommendations to refrain from meta-analysis when the I2 statistic exceeds 50%22 stem from the desire to avoid spurious precision in the estimated effect size. For instance, the extreme heterogeneity we report—again, indicating substantial variation in effects across trials beyond those expected from chance alone—suggests that some subsets of trials might have achieved substantially larger (or smaller) effects than denoted by the 95% confidence interval for the pooled effect across all trials. Indeed, this is one of the central findings in our study. Yet, even with a meta-regression model incorporating objective features of either the clinical decision support systems or the studies evaluating them, heterogeneity remained unchanged. Thus we conducted this meta-analysis despite extreme levels of heterogeneity because the results highlight that, while clinical decision support systems sometimes achieve large effects, current literature does not adequately identify the circumstances under which worthwhile improvements occur.

Conclusions

Despite publication of over 100 randomised and quasi-randomised trials involving over one million patients and 10 000 providers, the observed improvements display extreme, unexplained variation. Achieving worthwhile improvements therefore remains largely a case of trial and error. Future research must identify new ways of designing clinical decision support systems that reliably confer larger improvements in care while balancing the threat of alert fatigue. Head-to-head trials comparing design features of different support systems will also be important.88 In the meantime, a critical consideration in deciding to implement clinical decision support systems is the strength of the connection between the targeted process of care and patient outcomes. Achieving small improvements in care with only a presumptive connection to patient outcomes may not be worth the risk of contributing to alert fatigue or the growing concern of physician burnout attributable to EHRs.68 89 Clinical decision support systems embedded in electronic health records prompt clinicians to deliver recommended processes of care Despite enthusiasm over the potential for clinical decision support systems to improve care, a previous systematic review in 2010 found that such systems typically improved the proportion of patients receiving target processes of care by less than 5% The number of trials published over the ensuing decade has grown considerably, but subsequent systematic reviews have focused only on identifying features of clinical decision support systems associated with positive results, rather than quantifying the actual sizes of improvements achieved Most clinical decision support system interventions achieve small to moderate improvements in the percentage of patients receiving recommended processes of care The pooled effect size exhibited extreme heterogeneity (that is, variation across trials beyond that expected from chance alone), which did not diminish with a meta-regression model using significant predictors of larger effect sizes Thus although a minority of studies have shown that clinical decision support systems deliver clinically worthwhile increases in recommended care, the circumstances under which such improvements occur remain undefined
  84 in total

Review 1.  Effectiveness and efficiency of guideline dissemination and implementation strategies.

Authors:  J M Grimshaw; R E Thomas; G MacLennan; C Fraser; C R Ramsay; L Vale; P Whitty; M P Eccles; L Matowe; L Shirran; M Wensing; R Dijkstra; C Donaldson
Journal:  Health Technol Assess       Date:  2004-02       Impact factor: 4.014

Review 2.  Effects of computerized clinical decision support systems on practitioner performance and patient outcomes: a systematic review.

Authors:  Amit X Garg; Neill K J Adhikari; Heather McDonald; M Patricia Rosas-Arellano; P J Devereaux; Joseph Beyene; Justina Sam; R Brian Haynes
Journal:  JAMA       Date:  2005-03-09       Impact factor: 56.272

3.  Automated clinical reminders for primary care providers in the care of CKD: a small cluster-randomized controlled trial.

Authors:  Khaled Abdel-Kader; Gary S Fischer; Jie Li; Charity G Moore; Rachel Hess; Mark L Unruh
Journal:  Am J Kidney Dis       Date:  2011-10-07       Impact factor: 8.860

4.  Exploring the sociotechnical intersection of patient safety and electronic health record implementation.

Authors:  Derek W Meeks; Amirhossein Takian; Dean F Sittig; Hardeep Singh; Nick Barber
Journal:  J Am Med Inform Assoc       Date:  2013-09-19       Impact factor: 4.497

5.  Impact of a Comprehensive Antimicrobial Stewardship Program on Institutional Burden of Antimicrobial Resistance: A 14-Year Controlled Interrupted Time-series Study.

Authors:  Christine Peragine; Sandra A N Walker; Andrew Simor; Scott E Walker; Alexander Kiss; Jerome A Leis
Journal:  Clin Infect Dis       Date:  2020-12-31       Impact factor: 9.079

6.  Guided medication dosing for inpatients with renal insufficiency.

Authors:  G M Chertow; J Lee; G J Kuperman; E Burdick; J Horsky; D L Seger; R Lee; A Mekala; J Song; A L Komaroff; D W Bates
Journal:  JAMA       Date:  2001-12-12       Impact factor: 56.272

7.  Physician and patient tools to improve chronic kidney disease care.

Authors:  Thomas D Sequist; Alison M Holliday; E John Orav; David W Bates; Bradley M Denker
Journal:  Am J Manag Care       Date:  2018-04-01       Impact factor: 2.229

8.  Use of physician education and computer alert to improve targeted use of gastroprotection among NSAID users.

Authors:  Gregory A Coté; John P Rice; William Bulsiewicz; John P Norvell; Keri Christensen; Anne Bobb; Michael Postelnick; Colin W Howden
Journal:  Am J Gastroenterol       Date:  2008-05       Impact factor: 10.864

9.  The at-risk registers in severe asthma (ARRISA) study: a cluster-randomised controlled trial examining effectiveness and costs in primary care.

Authors:  Jane Rebecca Smith; Michael J Noble; Stanley Musgrave; Jamie Murdoch; Gill M Price; Garry R Barton; Jennifer Windley; Richard Holland; Brian Dw Harrison; Amanda Howe; David B Price; Ian Harvey; Andrew M Wilson
Journal:  Thorax       Date:  2012-08-31       Impact factor: 9.139

10.  Point-of-care cluster randomized trial in stroke secondary prevention using electronic health records.

Authors:  Alex Dregan; Tjeerd P van Staa; Lisa McDermott; Gerard McCann; Mark Ashworth; Judith Charlton; Charles D A Wolfe; Anthony Rudd; Lucy Yardley; Martin C Gulliford
Journal:  Stroke       Date:  2014-06-05       Impact factor: 7.914

View more
  44 in total

Review 1.  Electronic Medical Records for (Visceral) Medicine: An Overview of the Current Status and Prospects.

Authors:  Sven Kernebeck; Theresa Sophie Busse; Chantal Jux; Ulrich Bork; Jan P Ehlers
Journal:  Visc Med       Date:  2021-09-22

2.  Pediatric Provider Utilization of a Clinical Decision Support Alert and Association with HIV Pre-exposure Prophylaxis Prescription Rates.

Authors:  Carrie T Chan; Megen Vo; Jennifer Carlson; Tzielan Lee; Marcello Chang; Geoffrey Hart-Cooper
Journal:  Appl Clin Inform       Date:  2022-01-12       Impact factor: 2.342

3.  Perspectives: Envisioning healthcare quality and safety in 2030.

Authors:  Richard Ricciardi
Journal:  J Res Nurs       Date:  2021-03-22

4.  Using the Electronic Health Record User Context in Clinical Decision Support Criteria.

Authors:  Hojjat Salmasian; David Rubins; David W Bates
Journal:  Appl Clin Inform       Date:  2022-09-28       Impact factor: 2.762

5.  Integration of Risk Scores and Integration Capability in Electronic Patient Records.

Authors:  Ann-Kathrin Heider; Harald Mang
Journal:  Appl Clin Inform       Date:  2022-09-07       Impact factor: 2.762

Review 6.  Clinical Decision Support Stewardship: Best Practices and Techniques to Monitor and Improve Interruptive Alerts.

Authors:  Juan D Chaparro; Jonathan M Beus; Adam C Dziorny; Philip A Hagedorn; Sean Hernandez; Swaminathan Kandaswamy; Eric S Kirkendall; Allison B McCoy; Naveen Muthu; Evan W Orenstein
Journal:  Appl Clin Inform       Date:  2022-05-25       Impact factor: 2.762

7.  Electronic Health Record-Embedded, Behavioral Science-Informed System for Smoking Cessation for the Parents of Pediatric Patients.

Authors:  Brian P Jenssen; Dean J Karavite; Shannon Kelleher; Ekaterina Nekrasova; Jeritt G Thayer; Raj Ratwani; Judy Shea; Emara Nabi-Burza; Jeremy E Drehmer; Jonathan P Winickoff; Robert W Grundmeier; Robert A Schnoll; Alexander G Fiks
Journal:  Appl Clin Inform       Date:  2022-05-18       Impact factor: 2.762

8. 

Authors:  Kaveh G Shojania
Journal:  CMAJ       Date:  2022-05-30       Impact factor: 16.859

9.  Medication risk management and health equity in New Zealand general practice: a retrospective cross-sectional study.

Authors:  Sharon Leitch; Jiaxu Zeng; Alesha Smith; Tim Stokes
Journal:  Int J Equity Health       Date:  2021-05-11

Review 10.  Scoping review exploring the impact of digital systems on processes and outcomes in the care management of acute kidney injury and progress towards establishing learning healthcare systems.

Authors:  Clair Ka Tze Chew; Helen Hogan; Yogini Jani
Journal:  BMJ Health Care Inform       Date:  2021-07
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.