Literature DB >> 26576324

Quality improvement and practice-based research in neurology using the electronic medical record.

Demetrius M Maraganore1, Roberta Frigerio1, Nazia Kazmi1, Steven L Meyers1, Meredith Sefa1, Shaun A Walters1, Jonathan C Silverstein1.   

Abstract

We describe quality improvement and practice-based research using the electronic medical record (EMR) in a community health system-based department of neurology. Our care transformation initiative targets 10 neurologic disorders (brain tumors, epilepsy, migraine, memory disorders, mild traumatic brain injury, multiple sclerosis, neuropathy, Parkinson disease, restless legs syndrome, and stroke) and brain health (risk assessments and interventions to prevent Alzheimer disease and related disorders in targeted populations). Our informatics methods include building and implementing structured clinical documentation support tools in the EMR; electronic data capture; enrollment, data quality, and descriptive reports; quality improvement projects; clinical decision support tools; subgroup-based adaptive assignments and pragmatic trials; and DNA biobanking. We are sharing EMR tools and deidentified data with other departments toward the creation of a Neurology Practice-Based Research Network. We discuss practical points to assist other clinical practices to make quality improvements and practice-based research in neurology using the EMR a reality.

Entities:  

Year:  2015        PMID: 26576324      PMCID: PMC4634157          DOI: 10.1212/CPJ.0000000000000176

Source DB:  PubMed          Journal:  Neurol Clin Pract        ISSN: 2163-0402


The American Academy of Neurology (AAN) has proposed quality measures, but they have not been incorporated into electronic medical records (EMRs) by vendors.[1-5] Quality measures documented as unstructured text are not captured discretely, making it difficult to report performance. Neurology is also hampered by a lack of comparative effectiveness research. There are several approved treatments for common neurologic disorders, but it is unknown which are superior in efficacy and tolerability and for which patient subgroups. Traditional clinical trials enroll selected patients, use surrogate measures, follow patients for short periods, and generalize poorly to clinical practice.[6-9] Data captured in the EMR could be used to identify eligible patients, assign treatments, and measure outcomes at the point of care.[10] To address these unmet needs, we created a quality improvement and practice-based research initiative in neurology using the EMR. We present a step-by-step description of our quality journey for 10 neurologic disorders (brain tumors, epilepsy, migraine, memory disorders, mild traumatic brain injury [mTBI], multiple sclerosis [MS], neuropathy, Parkinson disease [PD], restless legs syndrome, and stroke) and for brain health (risk assessments and interventions to prevent Alzheimer disease [AD] in targeted populations).

Overview: A quality journey

The Department of Neurology at NorthShore University HealthSystem (NorthShore) includes 40 neurologists practicing at 4 hospitals and 8 outpatient sites in the north suburbs of Chicago, IL. The neurologists include generalists and subspecialists in epilepsy, neurodegenerative disorders, MS, neuromuscular disorders, neuro-oncology, sleep disorders, and stroke. Figure 1 illustrates our stepwise process of quality improvement and practice-based research using the EMR. Figure 2 illustrates our progress for 11 projects (10 neurologic disorders plus brain health) as of April 1, 2015.
Figure 1

Our quality journey

Quality improvement and practice-based research in neurology using the electronic medical record (EMR) consists of a stepwise progression from the development and implementation of structured clinical documentation support (SCDS) toolkits (including note writing and electronic data capture); to enrollment reports; to data quality reports (and data cleaning); to descriptive reports of cohort characteristics; to quality improvement projects (including the creation of benchmark data and quality improvement dashboards); to the use of clinical decision support tools (to hardwire patient safety and improved outcomes); to the use of subgroup-based adaptive assignments (in support of pragmatic clinical trials and personalized medicine); to other research (e.g., biobanking of DNA and the association of genotypes with longitudinal outcomes). The EMR provides a framework for measuring and impacting the 3 dimensions of quality improvement: structure, process, and outcomes.

Figure 2

Quality stages by disorder

The x-axis lists each of the quality improvement and practice-based research projects in order of implementation at the NorthShore site. The y-axis indicates the quality improvement and practice-based research stages. The bars indicate the stage of completion as of April 1, 2015, for each disorder. CDS = clinical decision support; NPBRN = neurology practice-based research network; QI = quality improvement; SCDS = structured clinical documentation support; SUBA = subgroup-based adaptive assignments.

Our quality journey

Quality improvement and practice-based research in neurology using the electronic medical record (EMR) consists of a stepwise progression from the development and implementation of structured clinical documentation support (SCDS) toolkits (including note writing and electronic data capture); to enrollment reports; to data quality reports (and data cleaning); to descriptive reports of cohort characteristics; to quality improvement projects (including the creation of benchmark data and quality improvement dashboards); to the use of clinical decision support tools (to hardwire patient safety and improved outcomes); to the use of subgroup-based adaptive assignments (in support of pragmatic clinical trials and personalized medicine); to other research (e.g., biobanking of DNA and the association of genotypes with longitudinal outcomes). The EMR provides a framework for measuring and impacting the 3 dimensions of quality improvement: structure, process, and outcomes.

Quality stages by disorder

The x-axis lists each of the quality improvement and practice-based research projects in order of implementation at the NorthShore site. The y-axis indicates the quality improvement and practice-based research stages. The bars indicate the stage of completion as of April 1, 2015, for each disorder. CDS = clinical decision support; NPBRN = neurology practice-based research network; QI = quality improvement; SCDS = structured clinical documentation support; SUBA = subgroup-based adaptive assignments.

Step 1: Structured clinical documentation support

We optimized our EMR (Epic Systems Corporation) by building structured clinical documentation support (SCDS) toolkits that standardize office visits, write progress notes, and capture data. The process of creating the 11 SCDS toolkits (1 per project) is as follows.

Content building

We conducted meetings with the neurologists in each program every 2 weeks to select a disorder that is prevalent, variable, and unpredictable and to standardize the content of progress notes (to define the disorder, specify its outcomes using validated measures, and document associated factors). We consulted the medical literature, AAN guidelines,[11] National Institute of Neurological Diseases and Stroke Common Data Elements,[12] and subspecialty guidelines.[13] We envisioned standard workflows (the order and assignment of tasks to a care team that included a medical assistant, sometimes a nurse, and a neurologist) and progress notes (the order and layout in which the content would write). We limited the medical assistant and/or nurse assessments to 15 minutes each and the neurologist assessment to 60 minutes. Content building required 3 months per project.

EMR building

We then conducted meetings with programmers from NorthShore's EMR Optimization team every 2 weeks. They built SCDS toolkits that included navigators (a sidebar list of activities assigned to a workflow, with links to electronic forms), electronic forms (modules that included discrete data fields with mouse-click selection of variables represented by radio buttons, tabs, or drop-down menus and additional branching logic, autoscoring, and auto-interpretation features), and summary flow sheets (of prior test results or office visit information). We included optional free text fields (for typing additional comments or narrative history or the impression and plan). The toolkits were designed to support initial and long follow-up visits (conducted annually per our Best Practices, except for malignant brain tumors [every 6 months] and mTBI [at 2 weeks and 3 months]). The toolkits were also designed for optional use at interval visits (e.g., for medication adjustments or review of test results). The SCDS toolkits were designed to support clinical practice and research. For example, an electronic pop-up box was built to prompt neurologists to enroll eligible patients in a DNA biobanking initiative. When the neurologist selected enrollment within the pop-up box, an electronic notice was sent to the research team to prompt consenting and blood drawing at the point of care or to document nonparticipation (and why). EMR building required 3 months per project.

Implementation

The neurologists tested the SCDS toolkits in the EMR's development environment. Revisions were made and the toolkits were moved to the EMR's production environment. The project team continued to meet every 2 weeks to make revisions based on patient encounters. Implementation required 3 months per project. In total, SCDS toolkit development required 9 months per project, and we worked on 3 toolkits a year (one undergoing content building, another EMR building, and another implementation at any time). Our toolkits focused on outpatient visits using the EMR's ambulatory environment (except for brain tumors, which used the EMR's hospital outpatient visit environment). As of April 1, 2015, 8 SCDS toolkits have been used several thousand times at routine office visits. Appendix e-1 at Neurology.org/cp provides screenshots of our toolkit for PD.

Step 2: Enrollment reports

After each SCDS toolkit implementation, neurologists met every 2 weeks with programmers specialized in extracting, transforming, and loading data from the EMR's relational data repository to project-specific data marts in NorthShore's Enterprise Data Warehouse (EDW). The data marts provide an interface for analytic tools. Up to 1,000 fields of data were captured per office visit. Neurologists provided input as to which tasks were required at initial and long follow-up visits and data were binned accordingly. Within 1 month of project implementation, the EDW programmers created reports to track research enrollment in our DNA biobanking initiative. The enrollment reports are generated monthly. Tables indicate the number of participants by ethnicity and race or by month and year; the number of initial or follow-up visits per year; a listing of enrollees (including dates of consent and initial and follow-up visits, and annotations regarding death, withdrawal from the study, invalid consent, screen failure, or pending blood draw); and a summary of longitudinal follow-up (including numbers of patients actively followed, past due, pending due, or not due; and follow-up rates). Once enrollment reports were produced, project team meetings transitioned to monthly intervals.

Step 3: Data quality reports

After project-specific data marts were created and enrollment reports were available, the EDW programmers created data quality reports (as early as 3 months after an SCDS toolkit was implemented). These reports indicated which of the required data were missing for each office visit. Data quality reports are distributed to the care team monthly. Data not cleaned within 3 months were archived as permanently missing, and those data were not listed on subsequent reports. The care team learned where they were error prone from the data quality reports, and they remediated their use of the toolkits. When systematic errors occurred for many providers, the teams had the opportunity to improve their use of the toolkits or to request optimizations or a change in data requirements. The monthly reports produced only a few or no data checks per provider once projects were established.

Step 4: Descriptive reports (cohorts)

Once we enrolled 100 patients into a project and cleaned the data, a statistician created a descriptive report of the cohort. These reports include frequencies, medians, and ranges and bar and box plots for multiple fields of data and score tests (in the sample overall and in strata). These reports are generated quarterly, and by reviewing descriptive data at the cohort level, neurologists identify opportunities for quality improvement. Appendix e-2 provides a sample descriptive cohort report for migraine.

Step 5: Quality improvement projects and dashboards

Each neurology team envisioned 3 quality improvement projects using the SCDS toolkits, data, and reports. They developed projects using experience, medical literature searches, and additional Web site resources.[11] Some quality improvement projects and measures were crosscutting but others were disorder-specific. Table 1 provides a list of quality improvement projects, disorders, and measures. We are analyzing baseline data for the first 100 patients enrolled per disorder and for the first year of follow-up to define quality benchmarks. The neurologists will review quality improvement “dashboards” quarterly.
Table 1

Examples of quality improvement projects using the electronic medical record

Examples of quality improvement projects using the electronic medical record

Next steps

Clinical decision support

We expect to improve quality by standardizing office visits using SCDS toolkits and by regularly reviewing enrollment, data quality, descriptive cohort, and quality improvement dashboard reports. As an added safeguard we are creating electronic pop-up boxes (best practice advisories [BPAs]) that fire at the point of care whenever a quality improvement opportunity is identified, based on data captured by the EMR. For example, if a patient has PD and the neurologist documents recent falls or a positive pull test and the patient has not seen physical therapy in the past year, upon accepting the note, a BPA will fire and present the mouse-click options “order physical therapy” (which will place the order) or “defer physical therapy” (which will cascade and prompt selection of a reason for deferral). After implementing the BPA, we will track improvements in referral and fall rates vs benchmark data. We empowered our neurologists to design the toolkits, reports, quality improvement projects, and BPAs to mitigate “physician burnout.”[14,15]

Subgroup-based adaptive assignments

Neurologists will conduct pragmatic trials using the EMR for 10 common neurologic disorders. Table 2 lists the disorders, compared treatments, and outcomes. Our captured data will trigger BPAs prompting reassignment when a “to be compared” treatment is prescribed. For example, a neurologist might prescribe pramipexole to a patient with PD but be prompted to consider ropinirole or rotigotine instead. We will integrate with the EMR subgroup-based adaptive design (SUBA) software that uses data captured from the patients previously enrolled in a given trial to identify subgroup effects and to assign patients newly enrolled in the trial to treatments that are expected to be more effective in real time.[16] The statistical features of SUBA include the continuous learning of patient subgroups based on a random partition model and the adaptive allocation of patients to the best treatment based on posterior predictive probabilities.[17-24] SUBA has desirable performance in computer-simulated trials with a sample size of 300.
Table 2

Examples of pragmatic trials using the electronic medical record

Examples of pragmatic trials using the electronic medical record

Neurology practice-based research network

We created a neurology practice-based research network (NPBRN) using the EMR. We are sharing SCDS toolkits and data for quality improvement and practice-based research. We obtained a list of Epic clients and invited neurology practices to join the NPBRN. The AAN is providing guidance regarding integration of quality measures into the EMR and assisting with recruitment of sites and dissemination of findings. Details are as follows (see also appendix e-4). A. Sharing of SCDS toolkits. NorthShore is sharing its toolkits for 10 neurologic disorders with the other neurologists under a free license sharing agreement. Sharing of tools will be achieved via the EMR's application exchange. This will substantially reduce the work of installing toolkits from one health system to another. B. Sharing of deidentified data. The neurologists will share deidentified data captured by the SCDS toolkits (minus protected health information). The NPBRN will govern the use of the shared data collaboratively. We expect no overlapping patients (1 site per state). C. Data management. NorthShore will manage the data in collaboration with other NPBRN sites. Data submitted to the EDW will be the object of constant monitoring. We will provide monthly encounter (akin to enrollment) reports and data quality reports, and quarterly descriptive cohort and quality improvement dashboard reports to participating neurologists. D. Analyses of the data. For each quality improvement initiative (table 1), we will describe the baseline demographic and clinical characteristics for the patients by site and combined. Several statistical packages interface with the EDW and will be used for comparisons. Neurologists are expected to assess thousands of unique patients per year at each NPBRN site separately and tens of thousands of new and established patients per year combined. We have registered the NPBRN (8 sites presently).[25]

Biobanking and research informatics initiative

A driver of our quality journey at NorthShore has been our department's biobanking and research informatics initiative. Neurologists are enrolling 1,000 patients in each of 11 SCDS cohorts. Following informed consent and with institutional review board approval, enrollees provide a blood sample for DNA and plasma storage and permission to associate information in their blood with information in their medical records. We will use this resource for molecular prognostic studies. Appendix e-5 summarizes biobanking as of April 1, 2015.

Practical points

From the point of view of practicing neurologists, our use of the EMR may seem futuristic. Many neurologists do not use an EMR or have limited capabilities for its optimization or data sharing. Nevertheless, we conclude by addressing many practical points to assist other clinical practices in making quality improvements and practice-based research using the EMR a reality.

How to promote buy-in

Physicians and staff may resist the implementation of SCDS toolkits and related clinical workflows. To build buy-in we regularly communicated benefits of SCDS, including (1) care navigation, (2) note writing, (3) timely communications, (4) value-based payment, (5) patient safety, (6) quality improvement, (7) comparative effectiveness, (8) personalized medicine, (9) scholarly activities, and (10) innovation. We engaged the neurologists in the design, building, and implementation of toolkits and the related reports and quality improvement initiatives and met regularly to improve tools and processes. We incorporated bonuses for group performance in quality goals into our compensation model. In other practices the data captured might also be used to meet federal quality reporting mandates such as the Physician Quality Reporting System.[26,27] Our objective was to stay “time neutral.” We encouraged neurologists to pilot test the SCDS toolkits in the EMR development environment and we placed frequent requests for optimizations post implementation. Our new office visits and long follow-up visits are 60 minutes each (neurologist component) and we support our neurologists with medical assistants or sometimes nurses, but the time and resources allocated can easily be reduced because of the modular build of our toolkits. Other neurology practices may choose to designate some toolkits (navigators) or modules (electronic forms) as required and others as optional or to implement neurology office visit workflows including students, residents, fellows, physician assistants, or advanced practice nurses (in addition to or instead of medical assistants and nurses). We also anticipate challenges to the conduct of practice-based research, such as pragmatic trials, using the EMR.[28] Neurologists or patients may find it cumbersome to engage in research at the point of care. Ethicists determined that pragmatic trials comparing standard therapies via the EMR with no additional burden or risk to patients might be exempted from informed consent.[29] Alternatively, the burden of informed consent can be reduced by using electronic consent forms. Consenting electronically at the point of care is standard practice for risk-bearing clinical procedures at some institutions, and others have reported a favorable experience with electronic consenting for research.[30,31]

How to promote collaboration

We are licensing the use of our SCDS toolkits to other neurology practices at no direct cost, provided they are willing to join the NPBRN and share deidentified data. The installation costs for the toolkits will be nominal compared with the installation costs of the EMR (and a fraction of the cost of building toolkits from scratch). Practices that use EMRs are accustomed to some annual cost associated with upgrades. All of the NPBRN principal investigators are department leaders, and we will regularly discuss engagement tactics. We will accommodate shorter office visits at some sites by designating a minimal set of SCDS toolkits or modules as required and others as recommended. We are exploring workflows in which patients may self-enter some information into the electronic forms using an online portal or computer kiosks or tablet computers at check-in to accommodate clinical practices with limited staff support. We will empower neurologists at each NPBRN site to propose changes to the toolkits and to have access to data. We will provide reports at the patient encounter (rather than research enrollment) level. NorthShore will leverage its experience in the conduct of EMR-based data sharing in a consortium.[32] We also hope that EMR vendors will incorporate SCDS toolkits similar to ours and make them broadly available to clients as part of their standard products. Once common data elements are extracted, transformed, and loaded into a data warehouse (regardless of the EMR platform), the same enrollment/encounter, descriptive, and quality reports can be used and the quality improvement initiatives can be the same. The AAN is creating a registry (the Axon Registry) to demonstrate the quality and value of neurologic care, to support improvements to care, and to alleviate the administrative burdens of quality reporting for payment and maintenance of certification.[33] Data captured using SCDS toolkits such as ours can be easily exported to the AAN registry, forgoing the need for extraction software, natural language processing, or chart abstraction.
  28 in total

1.  Practical clinical trials: increasing the value of clinical research for decision making in clinical and health policy.

Authors:  Sean R Tunis; Daniel B Stryer; Carolyn M Clancy
Journal:  JAMA       Date:  2003-09-24       Impact factor: 56.272

2.  A Bayesian adaptive design for multi-dose, randomized, placebo-controlled phase I/II trials.

Authors:  Fang Xie; Yuan Ji; Lothar Tremmel
Journal:  Contemp Clin Trials       Date:  2012-03-09       Impact factor: 2.226

3.  Adaptive dose insertion in early phase clinical trials.

Authors:  Bo Hu; B Nebiyou Bekele; Yuan Ji
Journal:  Clin Trials       Date:  2010-09-06       Impact factor: 2.486

4.  Informed versus uninformed consent for prostate surgery: the value of electronic consents.

Authors:  Muta M Issa; Erin Setzer; Christine Charaf; Alexandra L B Webb; Rachel Derico; I Jane Kimberl; Aaron S Fink
Journal:  J Urol       Date:  2006-08       Impact factor: 7.450

5.  Adaptive randomization for multiarm comparative clinical trials based on joint efficacy/toxicity outcomes.

Authors:  Yuan Ji; B Nebiyou Bekele
Journal:  Biometrics       Date:  2009-01-23       Impact factor: 2.571

6.  Quality improvement in neurology: dementia management quality measures.

Authors:  Germaine Odenheimer; Soo Borson; Amy E Sanders; Rebecca J Swain-Eng; Helen H Kyomen; Samantha Tierney; Laura N Gitlin; Mary Ann Forciea; John Absher; Joseph Shega; Jerry Johnson
Journal:  Neurology       Date:  2013-09-25       Impact factor: 9.910

7.  Physician burnout: A neurologic crisis.

Authors:  Bruce Sigsbee; James L Bernat
Journal:  Neurology       Date:  2014-11-05       Impact factor: 9.910

8.  Challenges to ethics and professionalism facing the contemporary neurologist.

Authors:  James L Bernat
Journal:  Neurology       Date:  2014-08-29       Impact factor: 9.910

9.  A Bayesian Dose-finding Design for Oncology Clinical Trials of Combinational Biological Agents.

Authors:  Chunyan Cai; Ying Yuan; Yuan Ji
Journal:  J R Stat Soc Ser C Appl Stat       Date:  2014-01-01       Impact factor: 1.864

10.  A phase I/II seamless dose escalation/expansion with adaptive randomization scheme (SEARS).

Authors:  Haitao Pan; Fang Xie; Ping Liu; Jielai Xia; Yuan Ji
Journal:  Clin Trials       Date:  2013-10-17       Impact factor: 2.486

View more
  18 in total

1.  Configuring electronic health records to meet the needs of neurologists: A case study.

Authors:  Eric M Cheng
Journal:  Neurol Clin Pract       Date:  2015-10

2.  Using EHRs to advance epilepsy care.

Authors:  Juma S Mbwana; Zachary M Grinspan; Russell Bailey; Madison Berl; Jeffrey Buchhalter; Adrian Bumbut; Zach Danner; Tracy Glauser; Angie Glotstein; Howard Goodkin; Brian Jacobs; Lisa Jones; Barbara Kroner; Gardiner Lapham; Tobias Loddenkemper; Demetrius M Maraganore; Doug Nordli; William D Gaillard
Journal:  Neurol Clin Pract       Date:  2019-02

3.  Pregnant? Validity of the pregnancy checkbox on death certificates in four states, and characteristics associated with pregnancy checkbox errors.

Authors:  Andrea Catalano; Nicole L Davis; Emily E Petersen; Christopher Harrison; Lyn Kieltyka; Mei You; Elizabeth J Conrey; Alexander C Ewing; William M Callaghan; David A Goodman
Journal:  Am J Obstet Gynecol       Date:  2019-10-19       Impact factor: 8.661

4.  Conceptualization and implementation of an interdisciplinary clinic for children with drug-resistant epilepsy during the COVID-19 pandemic.

Authors:  Debopam Samanta; Vimala Elumalai; Vidya C Desai; Megan Leigh Hoyt
Journal:  Epilepsy Behav       Date:  2021-10-29       Impact factor: 2.937

5.  Tumor type, epilepsy burden, and seizure documentation: experiences at a single center neuro-oncology clinic.

Authors:  Omar Bushara; Alex Guzner; Elizabeth Bachman; Roger Stupp; Rimas V Lukas; Jessica W Templer
Journal:  Neurooncol Pract       Date:  2021-06-11

6.  From Brain Disease to Brain Health: Primary Prevention of Alzheimer's Disease and Related Disorders in a Health System Using an Electronic Medical Record-Based Approach.

Authors:  A M Fosnacht; S Patel; C Yucus; A Pham; E Rasmussen; R Frigerio; S Walters; D Maraganore
Journal:  J Prev Alzheimers Dis       Date:  2017

7.  Improving thiamine prescribing at an academic hospital network using the computerized provider order entry system: a cohort study.

Authors:  Gregory S Day; Safiya Ladak; C Martin Del Campo
Journal:  CMAJ Open       Date:  2020-05-15

8.  Creation and implementation of an electronic health record note for quality improvement in pediatric epilepsy: Practical considerations and lessons learned.

Authors:  Jeffrey R Buchhalter; Morris H Scantlebury; Sabrina D'Alfonso; Juan Pablo Appendino; Luis Bello Espinosa; Brian L Brooks; Curtis Claassen; Jane Corbeil; David Czank; Stafford Dean; Alice W Ho; Julia Jacobs; Maarit Mackay; Joka McMahon; Aleksandra Mineyko; Jong M Rho; Trina Roberts; Sonia Rothenmund; Gary Ruta; Tyson Sawchuk; Brett A Simms; Kim Smyth; Tammy Still; Nancy Thornton
Journal:  Epilepsia Open       Date:  2021-03-10

9.  Structured clinical documentation in the electronic medical record to improve quality and to support practice-based research in epilepsy.

Authors:  Jaishree Narayanan; Sofia Dobrin; Janet Choi; Susan Rubin; Anna Pham; Vimal Patel; Roberta Frigerio; Darryck Maurer; Payal Gupta; Lourdes Link; Shaun Walters; Chi Wang; Yuan Ji; Demetrius M Maraganore
Journal:  Epilepsia       Date:  2016-11-19       Impact factor: 5.864

10.  Standardizing Care of Neuro-oncology Patients Using a Customized Electronic Medical Record Toolkit.

Authors:  Ryan T Merrell; Kelly Claire Simon; Nina Martinez; Rosa Maria Vazquez; Bryce Hadsell; Alexander Epshteyn; Gary Wilk; Roberta Frigerio; Demetrius M Maraganore
Journal:  Mayo Clin Proc Innov Qual Outcomes       Date:  2021-06-05
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.