Literature DB >> 26290894

CDS, UX, and System Redesign - Promising Techniques and Tools to Bridge the Evidence Gap.

Thomas McGinn1.   

Abstract

INTRODUCTION: In this special issue of eGEMs, we explore the struggles related to bringing evidence into day-to-day practice, what I define as the "evidence gap." We are all aware of high quality evidence in the form of guidelines, randomized clinical trials for treatments and diagnostic tests, and clinical prediction rules, which are all readily available online. We also know that electronic health records (EHRs) are now ubiquitous in health care and in most practices across the country. How we marry this high quality evidence and the practice of medicine through effective decision support is a major challenge. ABOUT THE ISSUE: All of the articles in this issue explore, in some fashion, CDS systems and how we can best bring providers and their work environment to the evidence. We are at the very early stages of the science of usability. Much more research and funding is needed in this area if we hope to improve the dissemination and implementation of evidence in practice. While the featured examples, techniques, and tools in the special issue are a promising start to improving usability and CDS, many of the papers highlight current gaps in knowledge and a great need for generalizable approaches. The great promise is for "learning" approaches to generate new evidence and to integrate this evidence in reliable, patient-centered ways at scale using new technology. Closing the evidence gap is a real possibility, but only if the community works together to innovate and invest in research on the best ways to disseminate, communicate, and implement evidence in practice.

Entities:  

Keywords:  clinical decisions support; patient-reported outcomes; systems redesign; user experience; user inferface

Year:  2015        PMID: 26290894      PMCID: PMC4537145          DOI: 10.13063/2327-9214.1184

Source DB:  PubMed          Journal:  EGEMS (Wash DC)        ISSN: 2327-9214


In this special issue of eGEMs, we explore the struggles related to bringing evidence into day-to-day practice, what I define as the “evidence gap.” We are all aware of high quality evidence in the form of guidelines, randomized clinical trials for treatments and diagnostic tests, and clinical prediction rules, which are all readily available online. We also know that electronic health records (EHRs) are now ubiquitous in health care and in most practices across the country. How we marry this high quality evidence and the practice of medicine through effective decision support is a major challenge. There are multiple steps to this process of closing the evidence gap, which are part of the larger concept of usability and clinical decision support (CDS). First and foremost is the proper evaluation of the quality of the evidence. Only after this step can we discuss the following steps of workflow analysis, usability testing, and full integration into EHR that enhance adoption of evidence by health care providers. While outside the scope of this issue, the systems for grading evidence are critical and agreement upon how to grade is important. In the era of “big data” and learning health systems, this first step needs special consideration. Arguably, many health systems and providers are overly impressed with large quantities of data or big data. However, big data does not necessarily equal evidence or at least high quality evidence. To illustrate this point, Derek Corrigan and colleagues propose a multistep maturity model for clinical prediction rules, which describes six incremental steps required to evolve an organizations’ infrastructure from the lowest level use of evidence (literature-based clinical prediction rules) to a fully electronic and computable service-oriented general model.1 Once the evidence is considered to be of high enough quality for integration into the EHR, the challenge then becomes adoption of the evidence through CDS. Without adoption of the evidence offered through effective CDS, providers can’t be influenced. Early attempts to offer providers evidence in the EHR have had poor adoption rates and hence little influence on the clinical environment. CDS implementation now incorporates a formative assessment and understanding of the facilitators and barriers in the environment in which the CDS will be integrated. Wyer et al. shed light on the activities needed to change and develop practitioner behavior toward more evidence-based practice, drawing on lessons learned from five years of running the Teaching Evidence Assimilation for Collaborative Health Care (TEACH) program.2 This poor adoption of CDS has created the need for appropriate usability studies where testing the user interface (UI), user experience (UX), and functionality of the tool become integral parts of the tool’s success. Usability testing is most effective when conducted in clinical environments where it is possible to observe behaviors and interactions between providers, patients, the EHR, and the clinical microcosm and workflow (emergency room, primary care clinics, inpatient setting, etc.)—all of which should be formally evaluated when developing and testing a CDS intervention. The special issue has a unique focus on the science and methods of CDS and usability. Understanding and measuring how all these factors interface with each other enables us to determine the most effective means of providing decision support and, hence, improving the clinical process. Formally analyzing the workflow environment, performing think-aloud studies, performing simulation scenarios with the CDS tool, and performing near live clinical testing prior to launching the decision support tool are all critical steps to improving and bridging the evidence gap. Several authors in this issue address CDS and usability. Dr. Edward Melnick and colleagues describe the user-centered design (UCD) approach to developing a decision support tool for deciding whether or not to obtain a CT scan for minor head injury in an emergency department. Through shared use of a bedside tablet computer by the patient and provider, the tool promotes their shared decision-making.3 Along similar lines, Foraker et al. share the implementation strategy for an EHR-based CDS visualization tool to enhance patient-provider communication about cardiovascular health in the outpatient setting.4 Dr. Andrea Hartzler and colleagues from the University of Washington share the human-centered design process used to develop a dashboard displaying patient-reported pain and disability outcomes following spinal surgery.5 The collection of patient-reported outcomes (PROs) draws attention to issues important to patients—physical function and quality of life. Integrating this data to enhance patient care, however, requires thoughtful consideration of the user experience and the data presentation. In addition, developing techniques to measure various forms of usability and reporting on the success and failure of these tests are important in this phase of the evolution of the EHR. Several examples of new approaches to integrate evidence at the point of care and changing clinical behavior are shared, as well as the related challenges of these new processes. A multistep model for integrating evidence at the point of care is reviewed in the article by Corrigan and colleagues mentioned above. Dr. Emily Patterson and colleagues share a case example from the Veteran’s Health Administration, detailing insights from a workflow analysis conducted by the National Institute of Standards and Technology of an innovation prototype.6 And Dr. Joseph Kannry and colleagues review a unique randomized trial with usability testing that includes key factors for successful CDS.7 Finally, Baier et al., Sheehan et al., and Kharbanda et al. explore methods to improve UCD by designing technology to support the health consumer and priority populations, specifically older adults and children. Baier et al. test an electronic, personal health-record based, self-management system for falls prevention, using an approach that analyzes four components: tasks, users, representations, and functions.8 Sheehan et al. share a multiphased approach to create a Web application for home health consumers seeking to choose among providers.9 Dr. Kharbanda et al. discuss a usability study to explore health record interfaces that help providers better identify teen-related blood pressure problems.10 All of the articles in this issue explore, in some fashion, CDS systems and how we can best bring providers and their work environment to the evidence. We are at the very early stages of the science of usability. Much more research and funding is needed in this area if we hope to improve the dissemination and implementation of evidence in practice. While the featured examples, techniques, and tools in the special issue are a promising start to improving usability and CDS, many of the papers highlight current gaps in knowledge and a great need for generalizable approaches. The great promise is for “learning” approaches to generate new evidence and to integrate this evidence in reliable, patient-centered ways at scale using new technology. Closing the evidence gap is a real possibility, but only if the community works together to innovate and invest in research on the best ways to disseminate, communicate, and implement evidence in practice.
  10 in total

1.  Integrating Patient-Reported Outcomes into Spine Surgical Care through Visual Dashboards: Lessons Learned from Human-Centered Design.

Authors:  Andrea L Hartzler; Shomir Chaudhuri; Brett C Fey; David R Flum; Danielle Lavallee
Journal:  EGEMS (Wash DC)       Date:  2015-03-13

2.  Using Qualitative Methods to Create a Home Health Web Application User Interface for Patients with Low Computer Proficiency.

Authors:  Rosa R Baier; Emily Cooper; Andrea Wysocki; Stefan Gravenstein; Melissa Clark
Journal:  EGEMS (Wash DC)       Date:  2015-05-13

3.  A Multistep Maturity Model for the Implementation of Electronic and Computable Diagnostic Clinical Prediction Rules (eCPRs).

Authors:  Derek Corrigan; Ronan McDonnell; Atieh Zarabzadeh; Tom Fahey
Journal:  EGEMS (Wash DC)       Date:  2015-07-07

4.  Back to the Bedside: Developing a Bedside Aid for Concussion and Brain Injury Decisions in the Emergency Department.

Authors:  Edward R Melnick; Kevin Lopez; Erik P Hess; Fuad Abujarad; Cynthia A Brandt; Richard N Shiffman; Lori A Post
Journal:  EGEMS (Wash DC)       Date:  2015-06-29

5.  EHR-based Visualization Tool: Adoption Rates, Satisfaction, and Patient Outcomes.

Authors:  Randi E Foraker; Bobbie Kite; Marjorie M Kelley; Albert M Lai; Caryn Roth; Marcelo A Lopetegui; Abigail B Shoben; Michael Langan; Nicole L Rutledge; Philip R O Payne
Journal:  EGEMS (Wash DC)       Date:  2015-06-18

6.  TeenBP: Development and Piloting of an EHR-Linked Clinical Decision Support System to Improve Recognition of Hypertension in Adolescents.

Authors:  Elyse O Kharbanda; James D Nordin; Alan R Sinaiko; Heidi L Ekstrom; Jerry M Stultz; Nancy E Sherwood; Patricia L Fontaine; Steve E Asche; Steven P Dehmer; Jerry H Amundson; Deepika X Appana; Anna R Bergdall; Marcia G Hayes; Patrick J O'Connor
Journal:  EGEMS (Wash DC)       Date:  2015-07-09

7.  A Framework for Usable and Effective Clinical Decision Support: Experience from the iCPR Randomized Clinical Trial.

Authors:  Joseph Kannry; Lauren McCullagh; Andre Kushniruk; Devin Mann; Daniel Edonyabo; Thomas McGinn
Journal:  EGEMS (Wash DC)       Date:  2015-07-09

8.  Teaching Evidence Assimilation for Collaborative Health Care (TEACH) 2009-2014: Building Evidence-Based Capacity within Health Care Provider Organizations.

Authors:  Peter C Wyer; Craig A Umscheid; Stewart Wright; Suzana A Silva; Eddy Lang
Journal:  EGEMS (Wash DC)       Date:  2015-07-06

9.  Initial Usability and Feasibility Evaluation of a Personal Health Record-Based Self-Management System for Older Adults.

Authors:  Barbara Sheehan; Robert J Lucero
Journal:  EGEMS (Wash DC)       Date:  2015-05-07

10.  Improving Clinical Workflow in Ambulatory Care: Implemented Recommendations in an Innovation Prototype for the Veteran's Health Administration.

Authors:  Emily S Patterson; Svetlana Z Lowry; Mala Ramaiah; Michael C Gibbons; David Brick; Robert Calco; Greg Matton; Anne Miller; Ellen Makar; Jorge A Ferrer
Journal:  EGEMS (Wash DC)       Date:  2015-07-09
  10 in total
  2 in total

1.  Knowledge translation of an online tool to determine candidacy for epilepsy surgery evaluation.

Authors:  Khara M Sauro; Jayna Holroyd-Leduc; Samuel Wiebe; Hude Quan; Lara Cooke; J Helen Cross; Gary W Mathern; Heather Armson; Julie Stromer; Nathalie Jetté
Journal:  Neurol Clin Pract       Date:  2016-08

2.  Concordium 2016: Data and Knowledge Transforming Health.

Authors:  Beth Devine
Journal:  EGEMS (Wash DC)       Date:  2017-04-20
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.