Literature DB >> 24384231

QNOTE: an instrument for measuring the quality of EHR clinical notes.

Harry B Burke1, Albert Hoang2, Dorothy Becher2, Paul Fontelo3, Fang Liu3, Mark Stephens4, Louis N Pangaro5, Laura L Sessums5, Patrick O'Malley1, Nancy S Baxi6, Christopher W Bunt4, Vincent F Capaldi6, Julie M Chen6, Barbara A Cooper6, David A Djuric7, Joshua A Hodge7, Shawn Kane6, Charles Magee5, Zizette R Makary6, Renee M Mallory6, Thomas Miller4, Adam Saperstein4, Jessica Servey4, Ronald W Gimbel2.   

Abstract

BACKGROUND AND
OBJECTIVE: The outpatient clinical note documents the clinician's information collection, problem assessment, and patient management, yet there is currently no validated instrument to measure the quality of the electronic clinical note. This study evaluated the validity of the QNOTE instrument, which assesses 12 elements in the clinical note, for measuring the quality of clinical notes. It also compared its performance with a global instrument that assesses the clinical note as a whole.
MATERIALS AND METHODS: Retrospective multicenter blinded study of the clinical notes of 100 outpatients with type 2 diabetes mellitus who had been seen in clinic on at least three occasions. The 300 notes were rated by eight general internal medicine and eight family medicine practicing physicians. The QNOTE instrument scored the quality of the note as the sum of a set of 12 note element scores, and its inter-rater agreement was measured by the intraclass correlation coefficient. The Global instrument scored the note in its entirety, and its inter-rater agreement was measured by the Fleiss κ.
RESULTS: The overall QNOTE inter-rater agreement was 0.82 (CI 0.80 to 0.84), and its note quality score was 65 (CI 64 to 66). The Global inter-rater agreement was 0.24 (CI 0.19 to 0.29), and its note quality score was 52 (CI 49 to 55). The QNOTE quality scores were consistent, and the overall QNOTE score was significantly higher than the overall Global score (p=0.04).
CONCLUSIONS: We found the QNOTE to be a valid instrument for evaluating the quality of electronic clinical notes, and its performance was superior to that of the Global instrument. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

Entities:  

Keywords:  Clinical Note; Clinical Quality; Electronic Health Record; Note Quality; QNOTE

Mesh:

Year:  2014        PMID: 24384231      PMCID: PMC4147610          DOI: 10.1136/amiajnl-2013-002321

Source DB:  PubMed          Journal:  J Am Med Inform Assoc        ISSN: 1067-5027            Impact factor:   4.497


Introduction

The clinical note documents the clinician's information collection, problem assessment, and patient management. It is a record of the clinician's care of the patient, it assists other clinicians in their management of the patient, and it is a legal document.1 2 Outpatient medical records during the first half of the 20th century were typically index cards stored in an envelope.3 During the second half of the 20th century, there was an exponential increase in the complexity of medicine and the recognition that detailed, complete, and accurate medical documentation was necessary for high-quality care.4 5 Proposals designed to improve the clinical note through the use of a problem-based approach were adopted.6–8 In addition, starting in the 1960s, there was a great deal of interest in creating computer-based notes.9 But until the recent widespread adoption of electronic health records (EHRs), assessing the quality of clinical notes was an arduous task. It required medical records personnel to manually find and pull the patients’ paper charts, the physicians’ notes had to be legible, raters had to be colocated with the charts, and the raters had to manually review and summarize each of the charts.10 These difficulties delayed efforts to develop an instrument for assessing clinical note quality. The EHR makes it easier to evaluate clinical notes because they are legible, they can be accessed instantly, and they can be reviewed anywhere. This provides an opportunity for a properly designed instrument that systematically assesses note quality.10–12 A global instrument assesses the entire note with one judgment of its quality. This approach has been used to assess inpatient notes.13 14 However, there is currently no validated non-global instrument for measuring the quality of a clinical note. QNOTE is an instrument that assesses a clinical note by judging each of its evaluative elements. The note's quality score is the combination of its element scores. In a previous study, we assessed the external validity of the QNOTE clinical note quality evaluation.15 In this study, we conducted a validation of the QNOTE instrument and we compared the QNOTE with a global instrument. We hypothesized that, in addition to its prior external validity, the QNOTE instrument would demonstrate high internal validity and be superior to a global measure of note quality.

Methods

Instruments

QNOTE is an instrument that assesses the quality of the clinical note in terms of 12 clinical elements (figure 1). The external validation of QNOTE has been described.15 Briefly, the 12 elements and their evaluative components were determined by a group of 61 clinicians privileged to write clinical notes. A set of organizing themes was identified: note characteristics, content, and its relation to the healthcare system. From these themes, a set of note elements was developed and the evaluative components of the elements were developed (figure 1). The 12 elements in the clinical note are chief complaint, history of present illness, problem list, past medical history, medications, adverse drug reactions, social and family history, review of systems, physical findings, assessment, plan of care, and follow-up information. The seven evaluative components are clear, complete, concise, current, organized, prioritized, and sufficient information. Not all of the components were used to evaluate all the elements, but every element was evaluated using at least one component. The result was QNOTE (figure 2), an instrument that scores the note based on the evaluative criteria (components) of each element.
Figure 1

Qualitative assessment of clinical notes. Column A is a list of the evaluative elements, and column B is a list of the components of each element.

Figure 2

QNOTE, an instrument that scores the note based on the evaluative criteria (components) of each element.

Qualitative assessment of clinical notes. Column A is a list of the evaluative elements, and column B is a list of the components of each element. QNOTE, an instrument that scores the note based on the evaluative criteria (components) of each element. Global scores have been used for rating the quality of clinical notes.13 14 Essentially, raters review the entire clinical note de novo and provide a judgment of the quality of the entire note. We created a global instrument to score the note (figure 3). It provides an overall note score based on the rater's judgment of whether the note supports the diagnosis and the planned work-up or planned treatment and is consistent with the diagnosis.
Figure 3

Global quality instrument.

Global quality instrument.

Clinical notes

Patient records were collected from five US military medical outpatient primary care clinics located in Texas, Mississippi, Nebraska, Nevada, and Virginia populated by predominately civilian patients. An independent contractor blinded to the study was tasked with collecting the charts of at least 100 patients with type 2 diabetes mellitus from each site. Type 2 diabetes was selected because it is the second most common reason for an ambulatory visit (hypertension is the most common)16 and patients with this condition exhibit multiple comorbid conditions and present with sufficient clinical complexity that they require a proper clinical note. The contractor was instructed to select the notes of outpatients who were treated by a physician, who had a diagnosis of type 2 diabetes mellitus (patients could have comorbid conditions), and who had been seen in clinic on at least three occasions: once approximately 6 months prior to EHR adoption, once approximately 6 months after EHR adoption, and once approximately 5 years after EHR adoption, resulting in three notes per patient. The contractor collected 568 civilian diabetic patients. One-third of the notes were handwritten (pre-EHR) and two-thirds were electronic (post-EHR). The pre-EHR notes were free text, and the post-EHR notes were templates structured by the physicians. An independent contractor photocopied and deidentified the notes prior to their distribution to the investigators. The deidentified notes were reviewed, and 31 patients were removed because at least one of their notes was either incomplete or illegible, resulting in 537 patients. From the pool of 537 patients, 100 study patients were randomly selected, resulting in 300 study notes. Each of the patient's set of three notes was randomly placed in one of four patient groups (figure 4). The result was four patient groups each with 25 patients, with each patient having three clinical notes, resulting in 75 notes per patient group. The presentation of the clinical notes to the raters was randomized within each of the four patient groups. The Uniformed Services University of the Health Sciences Institutional Review Board approved the study.
Figure 4

Study design.

Study design.

Physician raters

We recruited eight general internal medicine and eight family medicine practicing physicians in the national capital medical region (District of Columbia metropolitan area). There were no inclusion or exclusion criteria. We accepted the first eight physicians from each specialty who responded to the announcement. Ten were military physicians and six were civilian physicians, and none had prior experience assessing the quality of clinical notes. Study investigators met the recruited physicians to explain the review process, to describe the Global note and the QNOTE elements and components, and to allow the reviewers to practice on up to three notes. There was no proficiency testing and all the volunteer physicians participated in the study. Within their specialty, the physicians were randomly assigned to one of the four rater groups, resulting in two internal medicine and two family medicine physicians in each of the four-rater groups. Each rater rated two different sets of 75 photocopied deidentified notes. For the QNOTE, raters were asked to score the components of each of the elements of the note as fully acceptable, partially acceptable, or unacceptable. After the raters completed their component ratings of the elements of the note, we assigned a score to each rating: fully acceptable was scored 100, partially acceptable was scored 50, and unacceptable was scored 0. The average of the component scores for an element was the score for that element. A missing element that was required but not present—that is, an element that was not optional—was assigned a score of zero. Several elements could be rated as not applicable, and they would be omitted from the scoring. For the Global instrument, raters were given the note, asked to read it, and then provide an overall rating of fully acceptable, partially acceptable, or unacceptable. After the raters completed their global rating, we assigned a score to each rating: fully acceptable was scored 100, partially acceptable was scored 50, and unacceptable was scored 0. After the reviewer scoring was completed, it was determined that one reviewer's scores were significantly different from every one of the other reviewers’ scores and that reviewer was dropped from the study. Reviewers were provided with unique user names and passwords and they scored the notes online using the Global instrument and/or the QNOTE instrument. The results were directly entered into an encrypted database. The reviewers were not contacted during the review process and they were not compensated for participating in the study.

Study design

There were eight study groups (figure 4). Each study group consisted of four raters, each of whom rated two sets of 75 notes. All the study groups used the Global quality instrument to score their notes; four study groups (2, 3, 6, 7) used just the Global quality instrument (‘Global 1’) and four study groups (1, 4, 5, 8) used the Global quality instrument after using the QNOTE instrument on the same notes (‘Global 2’).

Statistical analysis

The inter-rater agreement for the QNOTE elements and across elements was assessed by the intraclass correlation coefficient (ICC), and the inter-rater agreement for the Global instrument was assessed by the Fleiss κ score. The Pearson correlation coefficient was used to assess the quality scores across groups. Differences between study groups were tested by Student's t test and the bootstrap method. Quality mean scores and 95% CIs, Fleiss κ mean scores, ICCs, and t tests were calculated using SAS 9.3. The Fleiss κ CI bootstraps and the ICC CIs were calculated using R (R Project, http://www.r-project.org).

Results

QNOTE inter-rater agreement results are shown in table 1. The QNOTE mean ICC for each study group was: group 1 score, 0.84 (CI 0.81 to 0.87); group 4 score, 0.80 (CI 0.76 to 0.85); group 5 score, 0.85 (CI 0.83 to 0.88); and group 8 score, 0.78 (CI 0.73 to 0.83). The four study groups were not significantly different. The grand mean QNOTE ICC score was 0.82 (CI 0.80 to 0.84). The QNOTE instrument demonstrated a high level of inter-rater agreement, which suggests that it is a reliable instrument for assessing note quality.
Table 1

QNOTE element intraclass correlation coefficients (CI) by rater group

ElementsGroup 1Group 4Group 5Group 8All groups
CC0.71 (0.59 to 0.81)0.77 (0.61 to 0.88)0.62 (0.46 to 0.75)0.68 (0.49 to 0.80)0.70 (0.58 to 0.80)
HPI0.90 (0.85 to 0.93)0.83 (0.72 to 0.91)0.81 (0.73 to 0.87)0.80 (0.69 to 0.88)0.84 (0.78 to 0.91)
Problems0.74 (0.60 to 0.83)0.53 (0.21 to 0.75)0.84 (0.72 to 0.92)0.57 (0.25 to 0.75)0.67 (0.57 to 0.70)
Med history0.92 (0.89 to 0.94)0.87 (0.79 to 0.93)0.93 (0.89 to 0.95)0.89 (0.82 to 0.93)0.90 (0.88 to 0.93)
Medications0.93 (0.87 to 0.94)0.94 (0.90 to 0.97)0.94 (0.91 to 0.96)0.88 (0.81 to 0.93)0.92 (0.88 to 0.96)
Allergies0.71 (0.59 to 0.82)0.75 (0.55 to 0.88)0.89 (0.84 to 0.92)0.79 (0.69 to 0.87)0.79 (0.72 to 0.80)
Soc/family0.88 (0.83 to 0.92)0.85 (0.75 to 0.92)0.84 (0.77 to 0.89)0.85 (0.76 to 0.91)0.86 (0.81 to 0.90)
ROS0.88 (0.83 to 0.92)0.90 (0.83 to 0.95)0.93 (0.90 to 0.95)0.82 (0.72 to 0.89)0.88 (0.85 to 0.92)
Exam0.88 (0.83 to 0.92)0.83 (0.71 to 0.91)0.94 (0.91 to 0.96)0.88 (0.80 to 0.92)0.88 (0.83 to 0.93)
Assess0.79 (0.69 to 0.86)0.81 (0.68 to 0.90)0.84 (0.77 to 0.89)0.81 (0.70 to 0.88)0.81 (0.75 to 0.88)
Plan0.88 (0.82 to 0.92)0.78 (0.63 to 0.88)0.85 (0.79 to 0.90)0.79 (0.67 to 0.87)0.82 (0.76 to 0.89)
Follow-up0.84 (0.77 to 0.89)0.74 (0.57 to 0.86)0.82 (0.75 to 0.88)0.63 (0.42 to 0.78)0.76 (0.68 to 0.84)
Mean0.84 (0.81 to 0.87)0.80 (0.76 to 0.85)0.85 (0.83 to 0.88)0.78 (0.73 to 0.83)0.82 (0.80 to 0.84)

CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information.

QNOTE element intraclass correlation coefficients (CI) by rater group CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information. Each of QNOTE's 12 evaluative elements was scored in terms of the mean score of its components. The mean scores for each group, as well as their grand means, are shown in table 2. The QNOTE instrument's mean quality score for each study group was: group 1 score, 61 (CI 60 to 64); group 4 score, 66 (CI 64 to 68); group 5 score, 67 (CI 65 to 69); and group 8 score, 64 (CI 62 to 66). The four study groups were not significantly different. The grand mean QNOTE quality score was 65 (CI 64 to 66). The mean element scores across the four groups correlated highly (r=0.89).
Table 2

QNOTE element quality scores (CI) by rater group

ElementsGroup 1Group 4Group 5Group 8All groups
CC75 (70 to 81)80 (75 to 85)72 (67 to 77)54 (47 to 61)71 (68 to 73)
HPI62 (54 to 69)74 (67 to 81)72 (66 to 78)71 (64 to 77)70 (66 to 73)
Problems39 (31 to 46)40 (34 to 46)48 (40 to 57)34 (26 to 42)39 (37 to 44)
Med history50 (41 to 60)53 (44 to 62)50 (41 to 60)51 (42 to 60)51 (47 to 56)
Medications68 (60 to 76)71 (63 to 79)75 (67 to 83)75 (68 to 83)72 (68 to 76)
Allergies78 (72 to 85)78 (71 to 86)73 (66 to 80)71 (64 to 79)75 (72 to 79)
Soc/family44 (36 to 55)42 (35 to 50)56 (49 to 63)35 (28 to 43)45 (41 to 57)
ROS47 (38 to 55)58 (49 to 67)55 (46 to 63)53 (45 to 62)53 (49 to 57)
Exam71 (54 to 75)77 (71 to 83)74 (67 to 80)75 (69 to 81)74 (71 to 77)
Assess69 (64 to 75)75 (70 to 81)78 (73 to 83)81 (77 to 86)76 (73 to 78)
Plan68 (62 to 74)76 (72 to 81)76 (71 to 81)82 78 to 86)75 (73 to 78)
Follow-up68 (62 to 74)72 (67 to 77)76 (71 to 82)78 (72 to 83)74 (71 to 76)
Mean61 (60 to 64)66 (64 to 68)67 (65 to 69)64 (62 to 66)65 (64 to 66)

CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information.

QNOTE element quality scores (CI) by rater group CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information. In order to determine whether the QNOTE instrument could reliably assess both handwritten and electronic notes, the ICCs for the two types of notes were assessed. The results are shown in table 3. The QNOTE mean ICC was 0.79 (CI 0.75 to 0.84) for handwritten notes and 0.80 (CI 0.75 to 0.86) for electronic notes, and the two groups were not significantly different. The QNOTE instrument was equally reliable in assessing the quality of handwritten and electronic notes.
Table 3

Mean intraclass correlation coefficients (CI) for the QNOTE elements for the four groups by type of note

ElementsHandwrittenElectronic
CC0.63 (0.37 to 0.90)0.76 (0.55 to 0.97)
HPI0.78 (0.69 to 0.87)0.87 (0.78 to 0.96)
Problems0.63 (0.37 to 0.90)0.59 (0.35 to 0.83)
Med history0.90 (0.82 to 0.97)0.84 (0.75 to 0.94)
Medications0.93 (0.87 to 0.99)0.89 (0.75 to 1.00)
Allergies0.80 (0.68 to 0.92)0.74 (0.43 to 1.00)
Soc/family0.77 (0.58 to 0.96)0.81 (0.66 to 0.96)
ROS0.81 (0.68 to 0.94)0.89 (0.82 to 0.96)
Exam0.86 (0.77 to 0.96)0.92 (0.83 to 1.00)
Assess0.81 (0.72 to 0.91)0.79 (0.62 to 0.96)
Plan0.86 (0.78 to 0.94)0.81 (0.65 to 0.96)
Follow-up0.75 (0.65 to 0.85)0.74 (0.48 to 1.00)
Mean0.79 (0.75 to 0.84)0.80 (0.75 to 0.86)

CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information.

Mean intraclass correlation coefficients (CI) for the QNOTE elements for the four groups by type of note CC, chief complaint; HPI, history of the present illness; Problems, problem list; Med history, past medical history; Medication, medication list; Allergies, adverse drug reactions and allergies; Soc/family, social and family history; ROS, review of systems; Exam, physical examination; Assess, assessment; plan, plan of care; Follow-up, follow-up information. The Global instrument assessed the overall quality of the note, where the Global instrument was used either alone (Global 1) or after the QNOTE instrument (Global 2). Each of the four rater groups’ global quality mean scores and their inter-rater agreement scores, as well as their grand means, is shown in table 4.
Table 4

Global 1 and global 2 mean quality scores (CI) and Fleiss κ inter-rater agreement (CI) by rater group

Global 1
Study group 2Study group 3Study group 6Study group 7Grand mean
Quality score50 (43 to 56)50 (43 to 57)64 (58 to 69)44 (38 to 51)52 (49 to 55)
Fleiss κ0.26 (0.16 to 0.37)0.25 (0.14 to 0.38)0.22 (0.13 to 0.34)0.22 (0.12 to 0.36)0.24 (0.19 to 0.29)

Global 2

Rater group 1Rater group 4Rater group 5Rater group 8Grand mean

Quality score44 (38 to 51)57 (50 to 60)54 (47 to 60)48 (43 to 54)51 (48 to 54)
Fleiss κ0.39 (0.28 to 0.52)0.23 (0.08 to 0.44)0.23 (0.13 to 0.35)0.10 (−0.02 to 0.27)0.24 (0.17 to 0.31)
Global 1 and global 2 mean quality scores (CI) and Fleiss κ inter-rater agreement (CI) by rater group The Global 1 mean Fleiss κ score for each study group was: group 2 score, 0.26 (CI 0.16 to 0.37); group 3 score, 0.25 (CI 0.14 to 0.38); group 6 score, 0.22 (CI 0.13 to 0.34); and group 7 score, 0.22 (CI 0.12 to 0.36). The Global 2 mean Fleiss κ score for each study group was: group 1 score, 0.39 (CI 0.28 to 0.52); group 4 score, 0.23 (CI 0.08 to 0.44); group 5 score, 0.23 (CI 0.13 to 0.35); and group 8 score, 0.10 (CI −0.02 to 0.27). The Global 1 grand mean Fleiss κ score was 0.24 (CI 0.19 to 0.29), the Global 2 grand mean Fleiss κ score was 0.24 (CI 0.17 to 0.31), and the two groups were not significantly different. The Global instrument demonstrated a relatively low level of inter-rater agreement. The Global 1 mean quality score for each study group was: group 2 score, 50 (CI 43 to 56); group 3 score, 50 (CI 43 to 57); group 6 score, 64 (CI 58 to 69); and group 7 score, 44 (CI 38 to 51). The Global 2 mean quality score for each study group was: group 1 score, 44 (CI 38 to 51); group 4 score, 57 (CI 50 to 60); group 5 score, 54 (CI 47 to 60); and group 8 score, 48 (CI 43 to 54). The Global 1 grand mean quality score was 52 (CI 49 to 55), the Global 2 grand mean quality score was 51 (CI 48 to 54), and the two groups were not significantly different. The Global 1 and 2 group scores correlated highly (r=0.86). The fact that there was no difference between the grand means of the two groups and that they correlated highly suggests that the QNOTE scoring did not affect the Global 2 scoring. The QNOTE quality scores were significantly higher than the Global 1 quality scores (p=0.04), and they were significantly higher than the Global 2 quality scores (p=0.02). In other words, an examination of the elements leads to higher quality scores than an overall Global assessment.

Discussion

In this large, retrospective multicenter blinded study, we found QNOTE to be a valid instrument for evaluating the quality of clinical notes. It possesses external validity, its elements are congruent with clinical practice,15 and it possesses internal validity. We found that the systematic examination of the clinical note elements using the QNOTE instrument, compared with a global subjective assessment, resulted in significantly higher quality scores and higher inter-rater agreement. The QNOTE instrument was equally reliable on handwritten and electronic documentation. There was no association between the QNOTE instrument element quality scores and the element inter-rater agreement (r=0.08, NS). This, in conjunction with the high inter-rater agreement, suggests that the QNOTE instrument quality scores reflect true differences in the quality of the notes. The Global ratings that were performed after the QNOTE ratings on the same notes (Global 2) were not affected by the prior QNOTE element-based assessment. This suggests that global assessments and element assessments use two different cognitive evaluation processes. In terms of the global assessment of notes, Stetson et al14 created the Physician Documentation Quality Instrument (PDQI-9) to measure inpatient note quality. The PDQI-9 assesses the entire note as one global entity. It requires nine global judgments regarding a note. The raters determine, on a 1–5 scale where 1 is ‘not at all’ and 5 is ‘extremely’, whether the overall note is, ‘up-to-date, accurate, thorough, useful, organized, comprehensible, succinct, synthesized, and internally consistent’. The scores for each of the nine judgments are summed to create a total score that can range from 9 to 45. An issue with the global approach is that it does not recognize that the clinical note consists of multiple clinical elements, where each element represents a type of clinical information that can have its own set of quality characteristics and must be evaluated in its own right. It is the combination of these elements that forms the basis for a high-quality note. Further, when physicians create and use a note, they do so in terms of its elements. Clinicians are interested in the element information being clinically useful.15 Finally, the elements are necessary for medicolegal and billing documentation. A clinical note can be viewed as a set of elements that inform and support the clinician's reasoning, decision-making, and recommendations to the patient. This study has several limitations. (1) It was performed using Military Health System records, but it is well established that this health system treats its patients similarly to patients treated in the community.17–20 (2) We focused on diabetics because: diabetes is very prevalent (it is the second most common diagnosis in ambulatory medicine); diabetic patients are typical of outpatients with serious diseases; diabetic patients have multiple comorbid conditions so they present with a wide spectrum of diseases. (3) Our results may not be as useful for patients with acute, self-limiting conditions. (4) It is unlikely that another single-item global assessment tool would perform differently from the current instrument because a global assessment is just that, a single overall judgment. This study's strengths are that it is a large multicenter blinded trial that rigorously evaluated the inter-rater agreement of 16 practicing physicians using QNOTE to assess the quality of electronic clinical notes.

Conclusions

The documentation of a clinician–patient encounter serves three important purposes: (1) it documents the clinician's information collection, problem assessment, and plan for the patient; (2) it creates a complete and accurate record that can be used by other clinicians to care for the patient; and (3) it provides substantiation for what was done for medicolegal reasons and reimbursement. These goals are consistent with the American Medical Informatics Association's 2011 Health Policy Meeting that stated that the main purpose of documentation is to support patient care and improved outcomes.21 A high-quality note contains the necessary and sufficient elements required to achieve these goals. The QNOTE instrument can determine whether a clinical note contains the necessary quality elements and is clinically acceptable. The QNOTE instrument has been shown to possess external validity and, in this study, it demonstrated high internal validity. It is an important advance because it provides a valid instrument for determining the quality of a clinical encounter, as documented by the clinician's note. Further, it provides a way for clinicians to assess their electronic notes and to use that knowledge to improve their documentation. Future QNOTE studies will include: (1) changes in note quality before (handwritten notes) and 6 months after EHR adoption; (2) changes in note quality 6 months after EHR adoption and 5 years later; and (3) the relationship between EHR note quality and clinical outcomes. The QNOTE instrument provides a structured, easy to use, transparent, valid and reliable way to assess the quality of electronic clinical notes.
  19 in total

1.  COMPUTER HANDLING OF AMBULATORY CLINIC RECORDS.

Authors:  R P LEVY; M R CAMMARN; M J SMITH
Journal:  JAMA       Date:  1964-12-21       Impact factor: 56.272

2.  Volume and cost of keeping records in a group practice.

Authors:  E V KUENSSBERG
Journal:  Br Med J       Date:  1956-06-09

3.  America's "undiscovered" laboratory for health services research.

Authors:  Ronald W Gimbel; Louis Pangaro; Galen Barbour
Journal:  Med Care       Date:  2010-08       Impact factor: 2.983

4.  Generating Clinical Notes for Electronic Health Record Systems.

Authors:  S Trent Rosenbloom; William W Stead; Joshua C Denny; Dario Giuse; Nancy M Lorenzi; Steven H Brown; Kevin B Johnson
Journal:  Appl Clin Inform       Date:  2010-01-01       Impact factor: 2.342

5.  The truth about doctors' handwriting: a prospective study.

Authors:  D M Berwick; D E Winickoff
Journal:  BMJ       Date:  1996 Dec 21-28

6.  The future state of clinical data capture and documentation: a report from AMIA's 2011 Policy Meeting.

Authors:  Caitlin M Cusack; George Hripcsak; Meryl Bloomrosen; S Trent Rosenbloom; Charlotte A Weaver; Adam Wright; David K Vawdrey; Jim Walker; Lena Mamykina
Journal:  J Am Med Inform Assoc       Date:  2012-09-08       Impact factor: 4.497

7.  Medical records that guide and teach.

Authors:  L L Weed
Journal:  N Engl J Med       Date:  1968-03-21       Impact factor: 91.245

8.  Medical records that guide and teach.

Authors:  L L Weed
Journal:  N Engl J Med       Date:  1968-03-14       Impact factor: 91.245

9.  Improving the electronic health record--are clinicians getting what they wished for?

Authors:  James J Cimino
Journal:  JAMA       Date:  2013-03-13       Impact factor: 56.272

10.  Quality of medical care.

Authors:  W P DEARING
Journal:  Calif Med       Date:  1963-06
View more
  22 in total

Review 1.  Clinical Natural Language Processing in 2014: Foundational Methods Supporting Efficient Healthcare.

Authors:  A Névéol; P Zweigenbaum
Journal:  Yearb Med Inform       Date:  2015-08-13

2.  Evaluation of interventions to improve inpatient hospital documentation within electronic health records: a systematic review.

Authors:  Natalie Wiebe; Lucia Otero Varela; Daniel J Niven; Paul E Ronksley; Nicolas Iragorri; Hude Quan
Journal:  J Am Med Inform Assoc       Date:  2019-11-01       Impact factor: 4.497

3.  Bringing Ophthalmic Graduate Medical Education into the 2020s with Information Technology.

Authors:  Emily Cole; Nita G Valikodath; April Maa; R V Paul Chan; Michael F Chiang; Aaron Y Lee; Daniel C Tu; Thomas S Hwang
Journal:  Ophthalmology       Date:  2020-12-24       Impact factor: 12.079

4.  Duly noted: Lessons from a two-site intervention to assess and improve the quality of clinical documentation in the electronic health record.

Authors:  Laura Fanucchi; Donglin Yan; Rosemarie L Conigliaro
Journal:  Appl Clin Inform       Date:  2016-07-06       Impact factor: 2.342

5.  The 9-Item Physician Documentation Quality Instrument (PDQI-9) score is not useful in evaluating EMR (scribe) note quality in Emergency Medicine.

Authors:  Katherine J Walker; Andrew Wang; William Dunlop; Hamish Rodda; Michael Ben-Meir; Margaret Staples
Journal:  Appl Clin Inform       Date:  2017-09-26       Impact factor: 2.342

6.  Development and Establishment of Initial Validity Evidence for a Novel Tool for Assessing Trainee Admission Notes.

Authors:  Danielle E Weber; Justin D Held; Roman A Jandarov; Matthew Kelleher; Ben Kinnear; Dana Sall; Jennifer K O'Toole
Journal:  J Gen Intern Med       Date:  2020-01-28       Impact factor: 5.128

7.  The effectiveness of a bundled intervention to improve resident progress notes in an electronic health record.

Authors:  Shannon M Dean; Jens C Eickhoff; Leigh Anne Bakel
Journal:  J Hosp Med       Date:  2014-11-25       Impact factor: 2.960

8.  Computer versus physician identification of gastrointestinal alarm features.

Authors:  Christopher V Almario; William D Chey; Sentia Iriana; Francis Dailey; Karen Robbins; Anish V Patel; Mark Reid; Cynthia Whitman; Garth Fuller; Roger Bolus; Buddy Dennis; Rey Encarnacion; Bibiana Martinez; Jennifer Soares; Rushaba Modi; Nikhil Agarwal; Aaron Lee; Scott Kubomoto; Gobind Sharma; Sally Bolus; Lin Chang; Brennan M R Spiegel
Journal:  Int J Med Inform       Date:  2015-07-26       Impact factor: 4.046

9.  A Heuristic Evaluation to Assess Use of After Visit Summaries for Supporting Continuity of Care.

Authors:  Patrice Tremoulet; Ramya Krishnan; Dean Karavite; Naveen Muthu; Susan Harkness Regli; Amy Will; Jeremy Michel
Journal:  Appl Clin Inform       Date:  2018-09-12       Impact factor: 2.342

10.  Effect of Outpatient Note Templates on Note Quality: NOTE (Notation Optimization through Template Engineering) Randomized Clinical Trial.

Authors:  Jeremy A Epstein; Joseph Cofrancesco; Mary Catherine Beach; Amanda Bertram; Helene F Hedian; Sara Mixter; Hsin-Chieh Yeh; Gail Berkenblit
Journal:  J Gen Intern Med       Date:  2020-09-08       Impact factor: 5.128

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.