Literature DB >> 26429290

Assessing Reliability of Medical Record Reviews for the Detection of Hospital Adverse Events.

Minsu Ock1, Sang-il Lee1, Min-Woo Jo1, Jin Yong Lee2, Seon-Ha Kim3.   

Abstract

OBJECTIVES: The purpose of this study was to assess the inter-rater reliability and intra-rater reliability of medical record review for the detection of hospital adverse events.
METHODS: We conducted two stages retrospective medical records review of a random sample of 96 patients from one acute-care general hospital. The first stage was an explicit patient record review by two nurses to detect the presence of 41 screening criteria (SC). The second stage was an implicit structured review by two physicians to identify the occurrence of adverse events from the positive cases on the SC. The inter-rater reliability of two nurses and that of two physicians were assessed. The intra-rater reliability was also evaluated by using test-retest method at approximately two weeks later.
RESULTS: In 84.2% of the patient medical records, the nurses agreed as to the necessity for the second stage review (kappa, 0.68; 95% confidence interval [CI], 0.54 to 0.83). In 93.0% of the patient medical records screened by nurses, the physicians agreed about the absence or presence of adverse events (kappa, 0.71; 95% CI, 0.44 to 0.97). When assessing intra-rater reliability, the kappa indices of two nurses were 0.54 (95% CI, 0.31 to 0.77) and 0.67 (95% CI, 0.47 to 0.87), whereas those of two physicians were 0.87 (95% CI, 0.62 to 1.00) and 0.37 (95% CI, -0.16 to 0.89).
CONCLUSIONS: In this study, the medical record review for detecting adverse events showed intermediate to good level of inter-rater and intra-rater reliability. Well organized training program for reviewers and clearly defining SC are required to get more reliable results in the hospital adverse event study.

Entities:  

Keywords:  Adverse event; Inter-rater reliability; Intra-rater reliability; Medical record review; Patient safety

Mesh:

Year:  2015        PMID: 26429290      PMCID: PMC4592027          DOI: 10.3961/jpmph.14.049

Source DB:  PubMed          Journal:  J Prev Med Public Health        ISSN: 1975-8375


INTRODUCTION

Measuring the size of patient safety problem is the first step for enhancing patient safety [1]. Various methods and indicators are used to measure patient safety problems [2,3]. Incidence of adverse event (AE) is a representative indicator widely used for this measurement. The Harvard Medical Practice Study (HMPS), which provided a basis for raising patient safety as an important policy agenda in the US, defined AE as “an injury that was caused by medical management and that prolonged the hospitalization, produced a disability at the time of discharge, or both” [4]. As AEs are related to direct treatment outcomes, such as patient outcomes, length of stay, and medical expenditure, they are useful for measuring patient safety levels [5]. Furthermore, AEs include medical errors such as medication errors, comprehensively representing patient safety levels. Additionally, as AEs have been widely and continuously used as indicators of patient safety, the concept has been well established. The HMPS examined patient medical records in New York State hospitals and evaluated AE occurrence, medical mistake or error occurrence, and patient disability level caused by AEs [4,6]. Many studies have since determined the incidence of AEs based on the HMPS methodology [7]. The biggest commonality is that evidence for AE occurrence was found by examining medical records. Specifically, previous studies applied a 2-stage examination method in which 2 nurses in the first stage and 2 physicians in the second stage examined medical records individually, taking the form of a sequential and independent review. Medical record review is commonly used to identify AE occurrence as medical records are easily accessible. However, it requires much time and necessitates commitment from medical professionals, resulting in higher cost than other methods [8]. Furthermore, unfaithful documentation can lead to underestimation of occurrence of AEs. However, the biggest shortcoming of medical record review for determining AEs is that the results may not be reliable if the consistency among different reviewers or even in a single reviewer is lacking [9]. Therefore, medical record review reliability should be examined and enhanced before conducting a study to measure AE incidence in hospitals [10]. In this study, we evaluated both intra-rater and inter-rater reliability for screening criteria (SC) detection and AE identification through medical record review.

METHODS

Study Design

A retrospective medical record review was conducted in a general hospital with approximately 500 beds. Review for AE identification was based on the HMPS methodology including a 2-stage examination (Figure 1) [4,6]. First, specific dates in 2007 were randomly selected using a random number table, and admissions of all patients discharged on those dates were selected as index admission. Psychiatric department admissions were excluded. Medical records from the entire duration of hospitalization and 1 year before and after were reviewed using a case review form developed in a previous study [11]. The first-stage review conducted by nurses screened medical records for 41 SC. These criteria encompassed events with high possibility of AE occurrence, such as antidote use, or those highly likely to lead to occurrence of additional AEs. The case review form included 41 SC chosen through the modified Delphi method from a combination of previously used SC [11]. These criteria can be divided into 19 items from the HMPS-format studies and 28 items from the Global Trigger Tool (GTT) (Supplemental Table 1). As opposed to the HMPS-type SC, which rely more heavily on reviewer’s clinical judgment, the GTT-type SC are more heavily based on objective clinical examination results [11]. For convenience of physicians while reviewing, nurse reviewers marked the corresponding medical record using a post-it when they found entries meeting the SC.
Figure. 1.

Process and results of medical record review for detecting hospital adverse events.

Two nurse reviewers independently reviewed medical records of the same patients. All cases determined by at least one nurse to meet the SC were included in the second-stage review for AE identification. Furthermore, to determine the occurrence of false negative results in the first-stage review, approximately 10% of cases that did not meet the SC during the first-stage review were randomly selected and included in the second-stage review. Two physician reviewers also independently reviewed medical records of the same patients to identify AE occurrence. Cases identified by both physician reviewers as having AEs were determined as AEs. In cases where opinions of the physician reviewers differed, a panel comprising professionals made the final call. For evaluation of inter-rater reliability, approximately half of the patients from the primary investigation were selected for the secondary investigation using the previously discussed method. However, in the secondary investigation, the second-stage review was not conducted on cases that did not meet the SC during the first-stage review.

Reviewers and Reviewer Training

A total of four reviewers participated in this study: two physicians (P1 and P2) and two nurses (N1 and N2). First-stage reviewers were nurses with five or more years of clinical experience, and second-stage reviewers were specialists with ten or more years of clinical experience. Reviewers were trained for approximately two hours. Educational content consisted of understanding the concept of AEs, instructions on using the case review form, and a practice exercise using an actual medical record.

Statistical Analysis

Inter-rater and intra-rater reliability of reviews of nurses and physicians was examined (Table 1). First, inter-rater reliability of reviews of nurses was analyzed for two different purposes: to assess the presence of SC and the necessity of the second-stage review. In evaluating the SC, agreements and kappa values were calculated for each of the 41 items of the SC. This represents reliability of decision of two nurses on whether or not the same patient meets a specific screening criterion. As patients determined to meet at least one screening criterion are subjected to the second-stage review, inter-rater reliability of nurses was also shown using agreements and kappa value for second-stage review necessity (last rows on Tables 2 and 3). It was then confirmed whether the reliability of reviews of nurses varied depending on SC characteristics. Using the aforementioned analytical method, differences in inter-rater reliability of nurses regarding second-stage review necessity based on SC characteristics were examined. In other words, cases corresponding to more than one screening criterion used in the HMPS-type studies and cases corresponding to more than one GTT-type screening criterion were examined separately. Six overlapping criteria were included in both groups.
Table 1.

Analytic framework for assessing reliability

First stage (nurse review)Second stage (physician review)
Inter-rater reliabilityAgreement and kappa for the presence of each screening criteriaAgreement and kappa for adverse event identification
Agreement and kappa for the determination of second-stage review necessity
Intra-rater reliabilityAgreement and kappa for the presence of each screening criteriaAgreement and kappa for adverse event identification
Agreement and kappa for the determination of second-stage review necessity
Table 2.

Inter-rater reliability of nurse reviewers

Item No.SCSC typePrimary investigation (n=95)
Secondary investigation (n=54)
Agreement (%)Kappa95% CIAgreement (%)Kappa95% CI
1The index admission was an unplanned admission related to previous healthcare managementH99-100-
2Length of index was over 30 daysH990.790.40, 1.00100-
3Unplanned readmission after discharge from index admission[1]H100-100-
4Revisiting emergency room within 72 hours after discharge from index admissionG96.80.710.40, 1.0098.20.850.55, 1.00
5Unplanned transfer to another acute care hospitalH97.90.740.40, 1.0098.20.660.03, 1.00
6Temperature higher than 38.3˚C at the point of dischargeH99-100-
7Unplanned transfer from general care to intensive care[1]H, G100-100-
8Speciality consultG93.70.730.53, 0.9488.90.35-0.04, 0.74
9Cardiac or respiratory arrest, rapid response team activation[1]H100-100-
10Unexpected death[1]H, G100-100-
11Healthcare-associated infectionH, G97.90.49-0.11, 1.0098.2-
12Hospital incurred patient injury[1]H, G100-98.2-
13Over-sedation/hypotensionG1001.00100-
14Restraint use[1]G100-100-
15Acute dialysis[1]G100-100-
16In-unit procedureG96.80.870.73, 1.0092.60.680.39, 0.97
17Treatment for organ damage after an invasive procedure[1]H100-100-
18Acute myocardial infarction, cerebrovascular accident, or pulmonary embolus during or after an invasive procedure[1]H, G100-100-
19Transfusion or use of blood productsG990.790.40, 1.0098.20.660.03, 1.00
20Avil (pheniramine) or Benadryl (diphenhydramine) use by intramuscular or intravenous ruteG95.80.780.57, 0.9998.20.90.70, 1.00
21Abrupt medication stopG94.70.24-0.15, 0.6294.40.550.10, 0.99
22Antidotes useG94.70.26-0.19, 0.7198.20.660.03, 1.00
23Adverse drug reactionH98.90.660.04, 0.1098.2-
24Decrease in hemoglobin or hematocrit of 25% or greaterG97.90.660.22, 0.1094.4-0.02-0.06, 0.01
25Hypoglycemic symptomG99-98.2-
26Bleeding tendencyG98.90.660.04, 1.0098.2-
27Rising BUN or serum creatine greater than 2 times baselineG99-100-
28Clostridium difficile positive stool[1]G100-100-
29Post-op troponin level greater than upper normal limit[1]G100-100-
30Mechanical ventilation greater than 24 hours post-op[1]G100-100-
31Unplanned return to the operating theatre[1]H100-100-
32Unplanned removal, injury or repair of organ during surgeryH99-98.2-
33Intra-op epinephrine, norepinephrine, naloxone, or romazicon[1]G100-100-
34Unplanned change in procedure or surgery[1]H100-100-
35Intubation, reintubation, BiPap in post anesthesia care unit[1]G100-100-
36X-ray in post anesthesia care unit[1]G100-100-
37Terbutaline use in obstetricsG99-100-
38Oxytocic agents in obstetricsG1001.001001.00
39Neonatal complications such as 5-minute Apgar score<6, or complication of abortion, amniocentesis or labor and delivery[1]H, G100-100-
40Documentation or correspondence indicating litigation, dissatisfaction[1]H100-100-
41Any other undesirable outcomes not covered aboveH91.60.16-0.17, 0.5090.70.26-0.15, 0.68
Second-stage review necessity84.20.680.54, 0.8377.80.550.33, 0.77

SC, screening criteria; CI, confidence interval; H, Harvard Medical Practice Study; G, Global Trigger Tool; BUN, blood urea nitrogen; op, operation; BiPap, bilevel positive airway pressure.

Both nurse reviewers could not detect the presence of SC from all records.

Table 3.

Intra-rater reliability of nurse reviewers

Item No.SCSC typeNurse reviewer N1 (n=54)
Nurse reviewer N2 (n=54)
Agreement (%)Kappa95% CIAgreement (%)Kappa95% CI
1The index admission was an unplanned admission related to previous healthcare managementH98.2-100.0-
2Length of index was over 30 daysH100.0-100.0-
3Unplanned readmission after discharge from index admission[1]H100.0-100.0-
4Revisiting emergency room within 72 hours after discharge from index admissionG98.20.850.55, 1.0096.30.650.19, 1.00
5Unplanned transfer to another acute care hospitalH100.01.00100.01.00
6Temperature higher than 38.3˚C at the point of dischargeH100.0-100.0-
7Unplanned transfer from general care to intensive care[1]H, G100.0-100.0-
8Speciality consultG88.90.19-0.22, 0.6192.60.670.37, 0.97
9Cardiac or respiratory arrest, rapid response team activation[1]H100.0-100.0-
10Unexpected death[1]H, G100.0-100.0-
11Healthcare-associated infectionH, G98.2-100.0-
12Hospital incurred patient injury[1]H, G98.2-100.0-
13Over-sedation/hypotensionG100.0-100.0-
14Restraint use[1]G100.0-100.0-
15Acute dialysis[1]G100.0-100.0-
16In-unit procedureG100.01.0096.30.850.66, 1.00
17Treatment for organ damage after an invasive procedure[1]H100.0-100.0-
18Acute myocardial infarction, cerebrovascular accident, or pulmonary embolus during or after an invasive procedure[1]H, G100.0-100.0-
19Transfusion or use of blood productsG100.01.0098.20.660.63, 1.00
20Avil (pheniramine) or Benadryl (diphenhydramine) use by intramuscular or intravenous routeG92.60.470.05, 0.8998.20.900.70, 1.00
21Abrupt medication stopG96.3-96.30.730.38, 1.00
22Antidotes useG98.20.660.03, 1.0096.30.49-0.11, 1.00
23Adverse drug reactionH98.2-100.01.00
24Decrease in hemoglobin or hematocrit of 25% or greaterG96.30.48-0.13, 1.00100.01.00
25Hypoglycemic symptomG100.01.00100.0-
26Bleeding tendencyG100.01.00100.0-
27Rising BUN or serum creatine greater than 2 times baselineG100.0-100.0-
28Clostridium difficile positive stool[1]G100.0-100.0-
29Post-op troponin level greater than upper normal limit[1]G100.0-100.0-
30Mechanical ventilation greater than 24 hours post-op[1]G100.0-100.0-
31Unplanned return to the operating theatre[1]H100.0-100.0-
32Unplanned removal, injury or repair of organ during surgeryH100.0-100.01.00
33Intra-op epinephrine, norepinephrine, naloxone, or romazicon[1]G100.0-100.0-
34Unplanned change in procedure or surgery[1]H100.0-100.0-
35Intubation, reintubation, BiPap in post anesthesia care unit[1]G100.0-100.0-
36X-ray in post anesthesia care unit[1]G100.0-100.0-
37Terbutaline use in obstetricsG98.2-100.0-
38Oxytocic agents in obstetricsG100.01.00100.01.00
39Neonatal complications such as 5-minute Apgar score<6, or complication of abortion, amniocentesis or labor and delivery[1]H, G100.0-100.0-
40Documentation or correspondence indicating litigation, dissatisfaction[1]H100.0-100.0-
41Any other undesirable outcomes not covered aboveH96.3-0.02-0.05, 0.0194.50.700.37, 1.00
Second-stage review necessity77.80.540.31, 0.7783.30.670.47, 0.87

SC, screening criteria; CI, confidence interval; H, Harvard Medical Practice Study; G, Global Trigger Tool; BUN, blood urea nitrogen; op, operation; BiPap, bilevel positive airway pressure.

Both nurse reviewers could not detect the presence of SC from all records.

Next, cases that at least one nurse determined to have at least one screening criterion and five randomly selected cases from those determined to have no screening criterion by reviews of two nurses were selected to investigate inter-rater reliability of physicians. AE occurrence was confirmed when calculating physicians’ inter-rater reliability. When one patient was determined to have 2 AEs, no examinations as to whether the contents were consistent with one another were done. Inter-rater reliability of physicians was calculated using agreement and kappa value for AE identification. Using the results from the primary and secondary investigations, intra-rater reliability between reviews of nurses and physicians was determined. The time period for N1 and N2 was 13 to 14 days; 11 to 12 days for P1; and 14 days for P2. Using the same method as for inter-rater reliability, agreements and kappa values were determined. In all cases, agreements were calculated using the total number of cases in which decisions of both reviewers corresponded as a numerator and the total number of patients as a denominator. SPSS version 21.0 (IBM Corp., Armonk, NY, USA) was used for all statistical analyses. The study was approved by the institutional review board of the National Evidence-based Healthcare Collaborating Agency (NECAIRB12-016-1).

RESULTS

Medical record review results are shown in Figure 1. The review was conducted on ninety-six patients discharged on three dates in 2007. Fifty-two patients were determined to have events satisfying at least one screening criterion by at least one nurse, and forty-four were determined by both nurses to meet none. Including the additional five individuals randomly selected from cases determined to have no screening criterion by reviews of the two nurses, physicians reviewed a total of fifty-seven patients. Eight patients were determined to have experienced AEs by both physician reviewers. Of the total 96 patients, 54 were re-investigated by nurses and 30 by physicians. First, in looking at inter-rater reliability of nurses, agreements for the presence of each screening criterion ranged from 91.6% to 100% and the kappa values from 0.16 to 1.00 (Table 2). Due to the reviewers’ lack of observed frequency, kappa values were not proposed for some SC. Nineteen of the SC were not found in any of the ninety-six patients during the first-stage review. SC yielding kappa values of 0.4 or lower in the primary investigation were No. 41 (kappa, 0.16), No. 21 (kappa, 0.24), and No. 22 (kappa, 0.26). The kappa value for determination of second-stage review necessity in the primary investigation was 0.68 (95% confidence interval [CI], 0.54 to 0.83). In the secondary investigation, SC with kappa values of 0.4 or lower were No. 24 (kappa, -0.02), No. 41 (kappa, 0.26), No. 8 (kappa, 0.35). In the secondary investigation, the kappa value for determination of second-stage review necessity was 0.55 (95% CI, 0.33 to 0.77). For intra-rater reliability, agreements for the presence of each screening criterion ranged from 88.9% to 100% for N1, and the kappa values were -0.02 to 1.00 (Table 3). SC with kappa values lower than 0.4 were No. 41 (kappa, -0.02), No. 8 (kappa, 0.19). The kappa value for second-stage review necessity was 0.54 (95% CI, 0.31 to 0.77). In the case of N2, agreements for the presence of each screening criterion ranged from 90.7% to 100%, and the kappa values ranged from 0.49 to 1.00. No SC had kappa values of 0.4 or less. The kappa value for second-stage review necessity was 0.67 (95% CI, 0.47 to 0.87). Looking at differences in reliability of reviews of nurses based on SC characteristics, the kappa values for reviews of nurses of SC used in the HMPS-type studies were 0.56 (95% CI, 0.32 to 0.80) for the primary investigation and 0.23 (95% CI, -0.11 to 0.57) for the secondary investigation (Table 4). On the other hand, the kappa values for reviews of nurses of the GTT-type SC were 0.72 (95% CI, 0.58 to 0.86) for the primary investigation and 0.62 (95% CI, 0.40 to 0.83) for the secondary investigation (Table 4).
Table 4.

Reliability of nurse reviewers by characteristics of screening criteria (SC)

SC typeTypes of reviewAgreement (%)Kappa95% CI
HMPSPrimary investigation89.50.560.32, 0.80
Secondary investigation83.30.23-0.11, 0.57
GTTPrimary investigation86.30.720.58, 0.86
Secondary investigation81.50.620.40, 0.83
TotalPrimary investigation84.20.680.54, 0.83
Secondary investigation77.80.550.33, 0.77

CI, confidence interval; HMPS, Harvard Medical Practice Study; GTT, Global Trigger Tool.

The kappa values for inter-rater reliability of physicians for detecting AE occurrence were 0.71 (95% CI, 0.44 to 0.97) for the primary investigation and 0.29 (95% CI, -0.16 to 0.75) for the secondary investigation (Table 5). In addition, intra-rater reliability of P1 for detecting AE occurrence for the secondary investigation of 30 individuals was represented by the kappa value of 0.87 (95% CI, 0.62 to 1.00) and that of P2 by the kappa value of 0.37 (95% CI, -0.16 to 0.89) (Table 5).
Table 5.

Reliability of physician reviewers for adverse event identification

Types of reviewAgreement (%)Kappa95% CI
Inter-rater reliabilityPrimary investigation (n = 57)93.00.710.44, 0.97
Secondary investigation (n = 30)86.70.29-0.16, 0.75
Intra-rater reliabilityPhysician reviewer P1 (n = 30)96.70.870.62, 1.00
Physician reviewer P2 (n = 30)90.00.37-0.16, 0.89
TotalPrimary investigation84.20.680.54, 0.83
Secondary investigation77.80.550.33, 0.77

CI, confidence interval.

DISCUSSION

In this study, medical records reviews were performed to identify inter-rater and intra-rater reliability on detection of SC and AE occurrence by reviewers. In the case of inter-rater reliability of nurses, agreements for decision of second-stage review necessity were 84.2% in the primary investigation and 77.8% in the secondary investigation, and the kappa values were 0.68 (95% CI, 0.54 to 0.83) for the primary investigation and 0.55 (95% CI, 0.33 to 0.77) for the secondary investigation. Previous studies looking at inter-rater reliability of nurses using the same method showed an agreement of 84% and kappa value of 0.67 [12], an agreement of 82% and kappa value of 0.62 (95% CI, 0.54 to 0.69) [13], and a kappa value of 0.73 [14], all similar to inter-rater reliability of nurses determined in this study. For inter-rater reliability of physicians, agreements on AE occurrence identification were 93.0% for the primary investigation and 86.7% for the secondary investigation, and the kappa values were 0.71 (95% CI, 0.44 to 0.97) for the primary investigation and 0.29 (95% CI, -0.16 to 0.75) for the secondary investigation. Previous studies analyzing inter-rate reliability of physicians using the same methods showed an agreement of 89% and a kappa value of 0.61 [4]; an agreement of 79% and a kappa value of 0.4 (95% CI, 0.3 to 0.5) [15]; an agreement of 76% and a kappa value of 0.25 (95% CI, 0.05 to 0.45) [13]; and a kappa value of 0.74 [14]. Comparing former results to this study, inter-rater reliability of physicians is shown to be relatively higher in the primary investigation and lower in the secondary investigation. Furthermore, kappa values for the secondary investigation for both nurse and physician raters were lower compared to those for the primary investigation. As CIs overlap, one must be cautious in interpreting such findings as significant, and a repeat of medical record reviews will be needed for future studies. For intra-rater reliability, the kappa values of all reviewers except P2 were 0.4 or higher, showing intermediate to good agreement for N1 and N2, and excellent agreement for P1 [16]. This may be attributed to differences in reviewers’ medical record review experience. The 3 reviewers who showed intermediate to excellent agreement had more experience than the reviewer with a lower agreement level. As such, intra-rater reliability may increase as experience and training on medical record review for AE occurrence identification increase. In general, reliability of a measurement is affected by instrument variability, subject variability, and observer variability [17]. Reliability results in this study were analyzed considering such factors. First, evaluation item characteristics were considered as instrument variability could have affected reliability. Intra-rater reliability of nurses was measured by determining the absence or presence of SC, which is less likely to involve subjective judgment of reviewers. Intra-rater reliability of physicians was determined by the absence or presence of AEs, which involves clinical decisions more so than determining the absence of presence of SC. Therefore, differences in identification of AEs were shown based on the clinical experience and knowledge of the physicians reviewers [18]. Considering such differences, setting specific examples of various AEs will be needed to ensure more consistent clinical decisions in physician reviewer training. Second, as a subject variability, independence of medical record review affected reliability. When cases that met the SC were found, the nurse reviewers marked the corresponding medical records using a post-it for the convenience of physician reviewers. This could have given away clues while the nurses successively reviewed the medical records to find cases meeting the SC. Creating 2 copies of the medical records or utilizing electronic medical records with an anonymous system will solve this problem. Third, as observer variability, experience and training of reviewers could have affected the reliability. Much like intra-rater reliability, differences in experience in reviewing medical records led to a reducing factor of inter-rater reliability. As the concept of AEs or SC are still new to health care professionals, relying only on reviewers with sufficient experience and training on patient safety-related medical record review will increase inter-rater reliability. Overall, it can be indirectly inferred that studies determining AE incidence through medical record review can be conducted in South Korea (hereafter Korea). In addition, AE incidence in hospitals was found to be 8.3% (8/96), similar to the level in other studies [7,14]. This finding suggests that hospitals in Korea must pay more attention to patient safety and conduct additional studies to determine the size of patient safety problems within hospitals using sample patients or hospitals that can represent the whole nation. Moreover, considering criticism on medical record review reliability [9], future studies must apply measures to increase reliability in determining AE occurrence. Such measures include standardizing measurement methods, training reviewers and raising their qualification levels, refining measurement tools, automating measurement tools, and repeating the measurements [17]. Considering these, measures to increase reliability in this study were determined as follows. First, more thorough reviewer training will be needed to ensure their training and qualification levels. This study was thought to lack sufficient reviewer training, compared to other study, which conducted an education program for 3 days [19]. Reviewers with less experience and training in medical record review for detecting AEs are thought to be more greatly affected by it. In future studies, a much more thorough and standardized reviewer training should be conducted, such as reiterating the concept of AEs and repeatedly conducting medical record review exercises. Second, to refine the measuring tool, clearer identification of each screening criterion will be needed. For example, for SC that may involve subjective evaluation, such as “No. 21. Abrupt medication stop”, clarifying operational definitions will help in increasing inter-rater reliability [20]. In this study, the GTT-type SC showed a slightly higher trend of reliability of reviews of nurses compared to the HMPS-type SC relying more heavily on the clinical experience of reviewers. The case review form should be improved by including more GTT-type SC while ensuring no reduction in its sensitivity. Furthermore, adding a question judging AE occurrences in reviews of nurses to aid decision-making of physicians should be considered to increase inter-rater reliability of physicians. Introducing scoring system of short-answer questions to judge AE occurrences may be considered, like the questions on causality and preventability of AE [21]. This study has limitations. First, there is a limitation in evaluating inter-rater reliability of nurses on each screening criterion due to the sample size. Nineteen of the SC were not found even once in the process of reviewing medical records of 96 individuals. It was difficult to determine whether this was due to SC characteristics or the small sample size. However, SC not commonly seen in other studies, such as intra-operative or post-operative death, may not have been witnesses even with a bigger sample size due to their characteristics [20]. Furthermore, it was difficult to evaluate reliability and validity of decisions on specific details regarding AEs due to the sample size. Second, as the medical record review was conducted in a single hospital, hospital characteristics may have had an effect. For example, this study largely lacked SC related to surgeries; therefore, inter-rater reliability for such SC was difficult to evaluate. Future studies should further evaluate inter-rater reliability of nurses for specific SC. Moreover, the quality of medical records in the hospital may have affected the medical record review, which may ultimately impact the inter-rater reliability. In conclusion, although reliability was not good in certain cases, intermediate or higher levels were shown for inter-rater and intra-rater reliability in AE identification through medical record review. In the future, when conducting a large-scale medical record review for AE identification, measures such as reviewer training enhancement and further clarification of the definition of SC should be considered to increase inter-rater reliability in detecting AEs.
  16 in total

1.  Measuring errors and adverse events in health care.

Authors:  Eric J Thomas; Laura A Petersen
Journal:  J Gen Intern Med       Date:  2003-01       Impact factor: 5.128

2.  The nature of adverse events in hospitalized patients. Results of the Harvard Medical Practice Study II.

Authors:  L L Leape; T A Brennan; N Laird; A G Lawthers; A R Localio; B A Barnes; L Hebert; J P Newhouse; P C Weiler; H Hiatt
Journal:  N Engl J Med       Date:  1991-02-07       Impact factor: 91.245

3.  Reliability and validity of judgments concerning adverse events suffered by hospitalized patients.

Authors:  T A Brennan; R J Localio; N L Laird
Journal:  Med Care       Date:  1989-12       Impact factor: 2.983

4.  The Quality in Australian Health Care Study.

Authors:  R M Wilson; W B Runciman; R W Gibberd; B T Harrison; L Newby; J D Hamilton
Journal:  Med J Aust       Date:  1995-11-06       Impact factor: 7.738

5.  The measurement of observer agreement for categorical data.

Authors:  J R Landis; G G Koch
Journal:  Biometrics       Date:  1977-03       Impact factor: 2.571

6.  Incidence of adverse events and negligence in hospitalized patients. Results of the Harvard Medical Practice Study I.

Authors:  T A Brennan; L L Leape; N M Laird; L Hebert; A R Localio; A G Lawthers; J P Newhouse; P C Weiler; H H Hiatt
Journal:  N Engl J Med       Date:  1991-02-07       Impact factor: 91.245

7.  Characteristics associated with the occurrence of adverse events: a retrospective medical record review using the Global Trigger Tool in a fully digitalized tertiary teaching hospital in Korea.

Authors:  Jee-In Hwang; Ho Jun Chin; Yoon-Seok Chang
Journal:  J Eval Clin Pract       Date:  2013-07-29       Impact factor: 2.431

8.  Using patient safety indicators to estimate the impact of potential adverse events on outcomes.

Authors:  Peter E Rivard; Stephen L Luther; Cindy L Christiansen; Susan Loveland; Anne Elixhauser; Patrick S Romano; Amy K Rosen
Journal:  Med Care Res Rev       Date:  2008-02       Impact factor: 3.929

9.  Incidence of adverse events related to health care in Spain: results of the Spanish National Study of Adverse Events.

Authors:  J M Aranaz-Andrés; C Aibar-Remón; J Vitaller-Murillo; P Ruiz-López; R Limón-Ramírez; E Terol-García
Journal:  J Epidemiol Community Health       Date:  2008-12       Impact factor: 3.710

Review 10.  The incidence and nature of in-hospital adverse events: a systematic review.

Authors:  E N de Vries; M A Ramrattan; S M Smorenburg; D J Gouma; M A Boermeester
Journal:  Qual Saf Health Care       Date:  2008-06
View more
  7 in total

1.  Patient and hospital characteristics that influence incidence of adverse events in acute public hospitals in Portugal: a retrospective cohort study.

Authors:  Paulo Sousa; António Sousa Uva; Florentino Serranheira; Mafalda Sousa Uva; Carla Nunes
Journal:  Int J Qual Health Care       Date:  2018-03-01       Impact factor: 2.038

2.  A Systematic Review of Methods for Medical Record Analysis to Detect Adverse Events in Hospitalized Patients.

Authors:  Dorthe O Klein; Roger J M W Rennenberg; Richard P Koopmans; Martin H Prins
Journal:  J Patient Saf       Date:  2021-12-01       Impact factor: 2.243

3.  Feasibility of Capturing Adverse Events From Insurance Claims Data Using International Classification of Diseases, Tenth Revision, Codes Coupled to Present on Admission Indicators.

Authors:  Juyoung Kim; Eun Young Choi; Won Lee; Hae Mi Oh; Jeehee Pyo; Minsu Ock; So Yoon Kim; Sang-Il Lee
Journal:  J Patient Saf       Date:  2021-12-17       Impact factor: 2.243

4.  The Korea National Patient Safety Incidents Inquiry Survey: Characteristics of Adverse Events Identified Through Medical Records Review in Regional Public Hospitals.

Authors:  Min Ji Kim; Hee Jung Seo; Hong Mo Koo; Minsu Ock; Jee-In Hwang; Sang-Il Lee
Journal:  J Patient Saf       Date:  2022-08-01       Impact factor: 2.243

5.  Identifying Adverse Events Using International Classification of Diseases, Tenth Revision Y Codes in Korea: A Cross-sectional Study.

Authors:  Minsu Ock; Hwa Jung Kim; Bomin Jeon; Ye-Jee Kim; Hyun Mi Ryu; Moo-Song Lee
Journal:  J Prev Med Public Health       Date:  2018-01

6.  Interpregnancy weight change and adverse pregnancy outcomes: a systematic review and meta-analysis.

Authors:  Eugene Oteng-Ntim; Sofia Mononen; Olga Sawicki; Paul T Seed; Debra Bick; Lucilla Poston
Journal:  BMJ Open       Date:  2018-06-04       Impact factor: 2.692

Review 7.  The Korea National Patient Safety Incidents Inquiry Survey: Feasibility of Medical Record Review for Detecting Adverse Events in Regional Public Hospitals.

Authors:  Min Ji Kim; Hee Jung Seo; Hong Mo Koo; Minsu Ock; Jee-In Hwang; Sang-Il Lee
Journal:  J Patient Saf       Date:  2022-01-24       Impact factor: 2.243

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.