Literature DB >> 34674415

Automated identification of diagnostic labelling errors in medicine.

Wolf E Hautz1, Moritz M Kündig2, Roger Tschanz2, Tanja Birrenbach1, Alexander Schuster3, Thomas Bürkle2, Stefanie C Hautz1, Thomas C Sauter1, Gert Krummrey1.   

Abstract

OBJECTIVES: Identification of diagnostic error is complex and mostly relies on expert ratings, a severely limited procedure. We developed a system that allows to automatically identify diagnostic labelling error from diagnoses coded according to the international classification of diseases (ICD), often available as routine health care data.
METHODS: The system developed (index test) was validated against rater based classifications taken from three previous studies of diagnostic labeling error (reference standard). The system compares pairs of diagnoses through calculation of their distance within the ICD taxonomy. Calculation is based on four different algorithms. To assess the concordance between index test and reference standard, we calculated the area under the receiver operating characteristics curve (AUROC) and corresponding confidence intervals. Analysis were conducted overall and separately per algorithm and type of available dataset.
RESULTS: Diagnoses of 1,127 cases were analyzed. Raters previously classified 24.58% of cases as diagnostic labelling errors (ranging from 12.3 to 87.2% in the three datasets). AUROC ranged between 0.821 and 0.837 overall, depending on the algorithm used to calculate the index test (95% CIs ranging from 0.8 to 0.86). Analyzed per type of dataset separately, the highest AUROC was 0.924 (95% CI 0.887-0.962).
CONCLUSIONS: The trigger system to automatically identify diagnostic labeling error from routine health care data performs excellent, and is unaffected by the reference standards' limitations. It is however only applicable to cases with pairs of diagnoses, of which one must be more accurate or otherwise superior than the other, reflecting a prevalent definition of a diagnostic labeling error.
© 2021 Wolf E. Hautz et al., published by De Gruyter, Berlin/Boston.

Entities:  

Keywords:  decision support; diagnostic error; quality improvement

Mesh:

Year:  2021        PMID: 34674415      PMCID: PMC9125795          DOI: 10.1515/dx-2021-0039

Source DB:  PubMed          Journal:  Diagnosis (Berl)        ISSN: 2194-802X


Introduction

Diagnostic error is a frequent health care problem [1], [2], [3], [4] with major medical [4], [5], [6], legal [7], [8], [9] and economic consequences [10]. On average, every patient in the U.S. experiences one major diagnostic error throughout his or her lifetime, often with devastating consequences for themselves, their families and their health care providers [11]. Thus, research into causes and predictors of diagnostic error as well as interventions to achieve diagnostic excellence is urgently needed. In the existing research into the phenomenon, two conceptions of diagnostic error prevail [12, 13]. One strand of research evaluates lapses in the diagnostic processes and often aims to identify missed opportunities [14]. Indeed, record reviews have found that in both, primary care and internal medicine, errors in the diagnostic process are common [15, 16] and often not limited to just one process error per case. For example, Singh and colleagues, in a review of cases identified through an automated flagging of electronic health records (a technique called e-triggers [17]), found that 43.7% of erroneous cases in primary care involved more than one type of process breakdown [15]. Similarly, in a review of cases identified through autopsy discrepancies, quality assurance activities, and voluntary reports, Graber and colleagues identified 5.9 different process errors per case [16]. However, only some of these errors resulted in a wrong diagnostic label, and/or harm from delayed or wrong treatment. In his conceptual model of missed opportunities in diagnosis, Singh accounts for this phenomenon and distinguishes between four types of errors [18]. These are missed opportunities in diagnosis that did not result in patient harm, those that did (an entity he consequently terms “preventable diagnostic harm”), and delayed/wrong diagnosis without clear evidence of a missed opportunity, either with or without resulting harm [18]. This latter group of wrong diagnosis without clear evidence of a missed opportunity can be referred to as diagnostic labelling errors – the second major conception of diagnostic error (i.e. naming a disease A while in fact, it is B) [12]. A labelling error does not necessarily imply existence of a causative process error [12], but may instead also result from what Graber labelled “no-fault errors”, including masked or unusual disease presentations, or error due to uncooperative or deceptive patients [16]. Identification of a diagnostic labelling error should thus trigger a case review, because arguably, only those labelling errors that result from a process error are readily susceptible to interventions. However, the labelling error is what ultimately affects treatment and prognosis and thus results in suboptimal patient outcome [13]. Consequently, identification (and ultimately prevention) of labelling errors is of major importance for the patient. It is, however, only a first step on the way to identify addressable underlying causes. In that regard, identification of diagnostic labelling errors can be viewed as another trigger within Singh’s concept of them (other triggers being e.g. unscheduled revisits [17] or elevated PSA values without timely follow-up [19]). Each case identified by such a trigger should then lead to an in-depth review to identify and potentially rectify underlying process errors. Such a strategy would substantially lower the workload of chart review, because labelling errors have been identified in between 2 and 35% of cases, depending on medical speciality and context [20]. Existing research into labelling errors often compares first diagnoses to final, confirmed or otherwise more secure diagnoses. Next to triggering chart reviews, studies evaluating concordance between first and later diagnoses have also been used to identify associations between labelling errors and hospital length of stay [4, 21], mortality [4, 22], costs [23], or type and size of hospital [24]. Such comparisons are typically conducted by a group of expert raters, a method with several disadvantages: The method is tedious, requires substantial expert contribution and thus does not scale well to larger datasets. Rater based identification of labelling errors are subject to hindsight-, and outcome bias [25]. Rater agreement is often limited [4, 25, 26]. The procedure introduces a substantial temporal delay between a patient being seen and the evaluation of his or her diagnosis, because it requires rater recruitment, training and the actual ratings. Taken together, all the above disadvantages explain why measurement of diagnostic labelling errors is hardly incorporated into routine quality evaluations in healthcare and has not been established as another trigger tool in diagnostic safety management and research. Beyond the field of labelling errors in clinical practice, rater based comparison of actual vs. correct diagnoses is also used in competence assessments of health care professionals, research into such assessments and research into the cognitive psychology of diagnostic reasoning. Here, case vignettes are frequently used (e.g. Refs. [27], [28], [29], [30]) and a known, correct diagnosis for each vignette is then compared to the actual diagnosis of study participants. Similar designs are used when evaluating the accuracy of diagnostic decision support systems (for a review see Ref. [31]) and patient employed symptom checkers (for a review, see Ref. [32]). All of the above research requires comparison of two diagnoses and their classification as either identical or discrepant. Thus, the shortcomings of rater based comparisons of two diagnoses are relevant to a large audience, and to fields of research as different as diagnostic quality management, health professions assessment and medical decision support. To address the shortcomings of rater based identification of diagnostic labelling errors, we developed an automated scoring system. The system (i) compares pairs of diagnoses, (ii) quantifies their degree of discrepancy and (iii) allows classifying them as either similar or discrepant, based on pre-specified values for either sensitivity or specificity of such classifications. Here, we describe and validate this automated scoring system.

Materials and methods

Classification of disease similarity

We developed a web based computer program that compares pairs of diagnoses coded according to WHOs international classification of diseases (ICD). The similarity between two given ICD-codes is determined in four different ways, described in the following. The current ICD version 10 is a uniaxial and non-hierarchical classification system, which contains 22 disease chapters (e.g. chapter IV: metabolic diseases), subdivided into 261 disease groups (e.g. E10 to 14: diabetes mellitus). These are further subdivided into 2,037 categories (e.g. E 10.-: primary insulin dependent diabetes mellitus), finally classifying each disease by one of 12,161 specific four-digit codes (e.g. E10.1: primary insulin dependent diabetes mellitus with ketoacidosis). The whole classification system can be depicted as a graph (Figure 1). A simple way (termed steps in the following) to determine the similarity of two given diseases is to count the edges on the shortest path connecting them (depicted as a dashed line in Figure 1).
Figure 1:

Example of an ICD taxonomy.

Dotted line: distance between two given exemplary diagnosis within this taxonomy is determined by counting the edges of the shortest path connecting them. Bold line: alternatively, the edges on the shortest connecting path between any two diagnoses can be determined by summing the weights of the edges connecting them.

Example of an ICD taxonomy. Dotted line: distance between two given exemplary diagnosis within this taxonomy is determined by counting the edges of the shortest path connecting them. Bold line: alternatively, the edges on the shortest connecting path between any two diagnoses can be determined by summing the weights of the edges connecting them. A slightly more complex approach allows adjusting for medical similarity by assigning different weights for each type of edge (for an example, see Figure 1). For example, one could argue that misdiagnosing one metabolic disease as another metabolic disease is less problematic then classifying a metabolic disease as a neoplasm. This medical similarity can be represented by assigning the edges connecting chapters of the ICD a higher weight than the edges connecting diseases within a chapter (depicted by the bold line in Figure 1). Similarity of two diseases is then determined by summing the weights of the edges on the shortest path connecting the two diseases (an approach termed weights below). Two more approaches originate from research into natural language processing. The approach devised by Wu and Palmer (thus labelled WuPalmer below) accounts for the position of two given concepts in the classification hierarchy relative to the position of their least common subsumer (i.e. their most specific common ancestor) [33, 34]. The score devised by Li and colleagues (termed Lietal) adapts this approach by combining the shortest path length and the depth of the subsumer nonlinearly, thus better reflecting human similarity ratings [35, 36].

Validation

To assess the performance of each of the four different algorithms steps, weights, WuPalmer and Lietal, we applied each one to three different datasets, two of which have been previously published. Each of these three datasets contains pairs of diagnoses that were previously rated by at least two experts as either similar or discrepant, a rating we used as reference standard to evaluate the four algorithms against. Notably, it is the free text diagnoses that were assessed by raters as similar or discrepant, not their ICD codes. The first dataset, termed clinical, contains first and final primary diagnoses of 755 patients that presented to an emergency room and were subsequently hospitalized to an internal medicine ward [4]. Patients presented with a broad variety of specific and non-specific chief complaints [26] and were diagnosed with a multitude of conditions across the spectrum of internal medicine. Three trained raters independently classified the pair of ER admittance diagnosis and internal medicine discharge diagnosis as either similar or discrepant, based on a previously validated scheme [37]. Discrepancies between raters were subsequently discussed and resolved. The second and third dataset resulted from studies of clinical reasoning in advanced medical students and are thus termed educational. The second dataset stems from a study of 20 advanced medical students who each diagnosed six virtual patients with shortage of breath for a variety of reasons in a previously validated computerized assessment of diagnostic reasoning [38], resulting in 120 pairs of diagnoses. The true diagnosis of each of these six patients has been validated by 20 expert physicians [38]. Three expert raters scored each pair of diagnoses as either similar or discrepant, disagreements were again resolved by discussion and consensus. The third dataset stems from a study of 51 advanced medical students who each diagnosed eight virtual patients either individually (17 students) or in pairs of two (34 students in 17 pairs), resulting in a total of 34 × 8=272 pairs of diagnoses. Chief complaints and correct diagnoses were dispersed across the spectrum of internal medicine. Again, three expert raters scored each pair of diagnoses as either similar or discrepant, disagreements were resolved by discussion and consensus. All diagnoses used in the study were ICD coded by one of two trained physicians. After coding a random sample of 10% of all diagnoses independent and in duplicate according to a predefined scheme, we assessed their coding agreement as excellent (kappa=0.96). The remaining diagnoses were then evenly split between raters and coded individually. Because this paper reports a reanalysis of previously collected datasets, we did not obtain ethical approval for the conduct of this study. The three original studies collecting the datasets used here have all obtained ethical approval and informed participant consent.

Statistical analysis

The existing rater based evaluations of pairs of diagnoses as either discrepant or similar were used as reference standard. We quantified semantic similarity of the diagnoses within each pair through each of the four approaches detailed above (steps, weights, WuPalmer and Lietal) and computed sensitivity and specificity for each of these approaches in comparison to the reference standard. We calculated receiver operating characteristics (ROC) curves, and estimated the area under the curve (AUC) and its 95% confidence interval. We further tested for a significant difference towards the null hypothesis (i.e. that the approach under evaluation is no better than throwing a coin in determining whether two given diagnoses are discrepant). All analyse were computed in IBM SPSS Statistics version 25 and conducted for the clinical end educational datasets separately as well as for the pooled data from all of the validation datasets.

Results

The web based computer program that compares pairs of diagnoses coded according to the ICD is available for free and public use online at https://simple.right-icd.de. The program supports diagnoses coded in different versions of the ICD, such as versions 9 and 10. The following results were obtained for diagnoses coded in ICD, German modification, version 10. [39], which is similar to other versions of ICD 10 (such as e.g. the “clinical modification” frequently employed in the U.S.) with regard to the four digit codes of diagnoses. A total of 1,172 pairs of diagnoses were available for analysis (Table 1). In the second dataset, 11 students only provided a diagnosis in five out of six cases, resulting in 109 out of 120 pairs of diagnoses being available for analysis.
Table 1:

Datasets used to validate the automated identification of diagnostic labelling errors.

Data-setTypeSizePairs of diagnosesa Cases with diagnosesPrevalence of error
SimilarDiscrepant
1ClinicalSeven hundred and fifty five patients, admission and discharge diagnosis755 (100%)6629312.30%
2EducationalTwenty advanced students diagnosing six virtual patients each109 (90.34%)149587.20%
3EducationalFifty one advanced students diagnosing 8 virtual patients alone [17] or in pairs (34; 17 teams)272 (100%)1809266.2%
Total 1,127 (98.26%) 850 277 24.58%

aPercentages refer to pairs of diagnoses with complete data available for analysis. Bold values signify the column total.

Datasets used to validate the automated identification of diagnostic labelling errors. aPercentages refer to pairs of diagnoses with complete data available for analysis. Bold values signify the column total. The performance (i.e. the ability to discriminate between similar and discrepant diagnoses) of all four algorithms evaluated was high, ranging from an AUC of 0.821–0.835 (Table 2). The corresponding ROC curves are depicted in Figure 2. AUC further increases when educational and clinical datasets are analysed separately. The highest AUC observed was 0.924 for the weights algorithm applied to the educational datasets. No AUC was below 0.82 (Table 2).
Table 2:

Performance of four algorithms overall, and by type of dataset.

AlgorithmArea under the curve AUC and (95% confidence interval)
OverallClinical dataset onlyEducational datasets only
Steps0.835 (0.812–0.859)0.853 (0.822–0.884)0.92 (0.88–0.959)
Weights0.830 (0.806–0.854)0.856 (0.826–0.887)0.924 (0.887–0.962)
Lietal0.821 (0.797–0.845)0.856 (0.825–0.887)0.923 (0.855–0.962)
WuPalmer0.821 (0.797–0.846)0.856 (0.825–0.887)0.923 (0.855–0.962)
Figure 2:

Sensitivity and specificity to identify diagnostic labelling errors, by algorithm and type of dataset.

Performance of four algorithms overall, and by type of dataset. Sensitivity and specificity to identify diagnostic labelling errors, by algorithm and type of dataset. All four investigated algorithms calculate continuous values, such as the number of steps required within the ICD taxonomy to get from one diagnosis to another. Conversion of these numerical values into a dichotomous decision (i.e. discrepant or similar) requires a cut-off value, below which two diagnoses are classified as similar and above which they are classified as discrepant (Note that this statement is only true for the algorithms steps and weights, where smaller numbers indicate higher similarity. For WuPalmer and Lietal, larger numbers indicate higher similarity. Consequently, for the latter two algorithms, the cut off represents the value below which two diagnoses are classified as discrepant). Table 3 allows to choose appropriate cut-off values based on a choice of sensitivity (or specificity). Tables S1 and S2 in the Supplementary Material provide similar cut off values for clinical and educational datasets separately.
Table 3:

Cut-off values for all for algorithms to achieve a given sensitivity (or specificity).

SensitivitySpecificityClassify as discrepant if larger or equal toClassify as discrepant if smaller or equal to
StepsWeightsWuPalmerLietal
0.001.0011.0026.00−1.00−1.00
0.011.009.5024.85
0.080.978.5024.55
0.380.887.5023.65
0.510.8012.850.1
0.510.8117.800.05
0.660.806.50
0.670.8012.550.240.15
0.800.785.5011.650.270.18
0.820.7611.40.310.22
0.830.764.505.850.420.28
0.890.733.504.650.540.44
0.910.712.501.950.710.60
0.970.670.500.450.930.89
1.000.00−1.00−1.002.002.00
Cut-off values for all for algorithms to achieve a given sensitivity (or specificity).

Discussion

“Improving the diagnostic process […] represents a moral, professional, and public health imperative.” [11] At present, however, “our health care systems are unable to systemically measure diagnostic performance […], which limits the ability to quantify performance and guide improvements” [40]. Furthermore, because diagnostic error is rarely measured outside of research studies, we may underestimate the magnitude of the problem. In effect, health care organizations do not have insights into the diagnostic quality they provide, and quality improvement initiatives lack measures of their efficacy and efficiency. The approach presented here represents a step towards routine measurement of diagnostic labelling errors at a systemic level and scale. In this study, we present and validate an approach that allows to automatically identify diagnostic labelling errors from routinely available healthcare data, namely ICD coded diagnoses. The approach performs quite well when compared to human classifications of diagnostic labelling errors, resulting in an AUC ranging from 0.821 to 0.924. Its performance is largely independent of the prevalence of labelling error in the data evaluated, which ranged from 12.3% in the clinical to 87.2% in one of the educational the datasets used for validation. Furthermore, it can be applied to datasets of any size, and the sensitivity (or specificity) for the identification of diagnostic labelling errors can be tailored to the specific task at hand. Manual identification of diagnostic labelling errors would only allow to adjust sensitivity or specificity to a given purpose, if raters graded the amount of discrepancy, e.g. using Likert scales, which would further add to their workload. Thus, most studies of diagnostic labelling errors employ dichotomous ratings instead (see e.g. Refs. [4, 21], [22], [23]). However, adjusting sensitivity (or specificity) to the studies purpose may be beneficial. For example, when used as a trigger in clinical quality assurance or research into diagnostic error in a clinical setting, one would most likely prioritize sensitivity over specificity, in order to identify all cases with potential for improvements. In summative assessments however (i.e. assessments resulting in pass or fail decisions), one could argue that coding of diagnostic labelling errors should prioritize specificity over sensitivity to support defensible fail decisions. Because of the wide spread use of Grabers’ definition of a diagnostic error as “a diagnosis that was unintentionally delayed […], wrong […], or missed […], as judged from the eventual appreciation of more definitive information”, [16] diagnostic errors are conceptualized as a discrepancy between a first and a more definitive diagnosis across a wide range of research and applications. This makes our approach applicable to many fields, from routine quality assurance to decision-making research, and from the evaluation of decision support systems to the scoring of assessments. Our approach complements existing methods used in research into diagnostic error. The limitations of rater based assessments of diagnostic error, including hindsight-, and outcome bias [25], limited rater agreement [4, 25, 26] and their limitation to small sample sizes have been pointed out in the introduction. None of these shortcomings applies to the approach suggested here. It should however be noted that simply only identifying a discrepancy between two diagnostic labels is insufficient to identify an underlying diagnostic process error. Instead, a discrepancy between diagnostic labels should be interpreted as a trigger to review the case in more detail to identify potentially preventable process errors and because a discrepancy between a first and a later diagnostic label may also result from disease evolution, or development of a secondary complication. Such discrepancies may not represent an error (not even a labelling error) at all. To account for these problems, combining our automated identification of diagnostic labelling errors with approaches such as the Symptom-Disease Pair Analysis of Diagnostic Error (SPADE) may improve the specificity of our trigger: in brief, SPADE looks out for “well-known diagnostic error dyads” such as dizziness – stroke or back pain – aortic aneurysm [41]. Limiting comparison of diagnostic labels with our approach to those labels that share a common symptom or chief complaint may potentially flag more diagnostic discrepancies as a labelling error where an underlying process error can be identified. In general, electronic trigger tools watch out for predefined signals in electronic health records, indicative of diagnostic error or of suboptimal diagnostic performance [42]. Triggers can be generic (such as unscheduled revisits or visits followed by an unplanned hospitalization within a specified time frame [17]) or disease specific (such as screening for abnormal prostate-specific antigen (PSA) measurements without a timely follow-up action [19]). Such trigger tools can be applied to very large datasets and substantially reduce the number of cases to then manually check for diagnostic process error (e.g. from 292,587 records screened to 426 abnormal PSA measurements without follow-up [19]). However, manual work-up of the cases identified is still required. While the extent to which these triggers are associated with missed diagnostic opportunities (or process errors) has been established for the two aforementioned triggers [17], similar work on the predictive value of labelling discrepancies remains to be done. Nevertheless, diagnostic labelling errors have been associated with adverse outcomes such as increased mortality or length of hospital stay before [4, 21, 22, 43], likely because it ultimately is the diagnostic label assigned to a patient that determines (in-)adequate therapy and prognosis [13]. Researchers and quality managers who consider to use the approach and/or program presented here can easily validate it for their specific purpose, provided they do possess a dataset for validation, i.e. containing a reference classification against which to evaluate the approach and program. By running their datasets of pairs of ICD coded diagnoses through the web based program provided at https://simple.right-icd.de, they can not only calculate the AUC the different algorithms achieve for their validation datasets, but can also determine cutoff values to achieve a given sensitivity (or specificity) for their specific classification task. Publication of their findings can help to empirically determine the use cases for which the suggested approach and program is suitable and where reasonable borders of its applicability are. In addition, such revalidation attempts may help to identify differences between the four different algorithms evaluated here and implemented on the project website. In the three datasets used in the current study, all four algorithms performed remarkably similar. We do not see any operational advantage of one over the other, but would recommend to use the steps algorithm as a default, simply because it is the least complicated and most comprehensible approach.

Limitations

The approach presented in this paper and its validation have several limitations, which warrant consideration. First, the approach is limited to datasets of pairs of diagnoses, of which one should be more certain, of better quality or otherwise superior than the other is. Second, currently the program is limited to comparing pairs of single diagnoses. In theory, the approach can be applied to the comparison of two sets of one to many diagnoses each, and we are currently developing the program to account for that scenario. However, definition of a reference standard and thus validation become much more complex (and much less certain) in that scenario. Third, the program is currently limited to the comparison of pairs of diagnoses coded in any of the more recent versions of ICD, making it susceptible to all the limitations that are inherent to ICD coding, and requiring ICD coding of all diagnoses to start with. Last, we only validated our approach and program on medical cases. Weather our approach also works in e.g. neurological or paediatric cases remains to be tested. As pointed out above, researchers in possession of an ICD coded set of pair of diagnoses from any field can easily check the performance of our approach online.

Conclusions

In comparison to the reference standard, the trigger system developed to automatically identify diagnostic labelling error from routine health care data performs excellent, and is unaffected by the reference standards’ limitations. It is however only applicable to cases with pairs of diagnoses, of which one must be better, more accurate or otherwise superior than the other, reflecting the prevalent definition of a diagnostic labeling errors in medicine. Supplementary Material Details Click here for additional data file. Supplementary Material Details Click here for additional data file. Supplementary Material Details Click here for additional data file.
  36 in total

1.  Tort claims analysis in the Veterans Health Administration for quality improvement.

Authors:  W B Weeks; T Foster; A E Wallace; E Stalhandske
Journal:  J Law Med Ethics       Date:  2001 Fall-Winter       Impact factor: 1.718

2.  Assessing clinical reasoning (ASCLIRE): Instrument development and validation.

Authors:  Olga Kunina-Habenicht; Wolf E Hautz; Michel Knigge; Claudia Spies; Olaf Ahlers
Journal:  Adv Health Sci Educ Theory Pract       Date:  2015-03-01       Impact factor: 3.853

3.  A unified conceptual model for diagnostic errors: underdiagnosis, overdiagnosis, and misdiagnosis.

Authors:  David E Newman-Toker
Journal:  Diagnosis (Berl)       Date:  2014-01-08

4.  Symptom-Disease Pair Analysis of Diagnostic Error (SPADE): a conceptual framework and methodological approach for unearthing misdiagnosis-related harms using big data.

Authors:  Ava L Liberman; David E Newman-Toker
Journal:  BMJ Qual Saf       Date:  2018-01-22       Impact factor: 7.035

5.  Is bias in the eye of the beholder? A vignette study to assess recognition of cognitive biases in clinical case workups.

Authors:  Laura Zwaan; Sandra Monteiro; Jonathan Sherbino; Jonathan Ilgen; Betty Howey; Geoffrey Norman
Journal:  BMJ Qual Saf       Date:  2016-01-29       Impact factor: 7.035

Review 6.  Understanding missed opportunities for more timely diagnosis of cancer in symptomatic patients after presentation.

Authors:  G Lyratzopoulos; P Vedsted; H Singh
Journal:  Br J Cancer       Date:  2015-03-31       Impact factor: 7.640

7.  Factors predicting a change in diagnosis in patients hospitalised through the emergency room: a prospective observational study.

Authors:  Stefanie C Hautz; Luca Schuler; Juliane E Kämmer; Stefan K Schauber; Meret E Ricklin; Thomas C Sauter; Volker Maier; Tanja Birrenbach; Aristomenis Exadaktylos; Wolf E Hautz
Journal:  BMJ Open       Date:  2016-05-11       Impact factor: 2.692

8.  Discrepancy between emergency department admission diagnosis and hospital discharge diagnosis and its impact on length of stay, up-triage to the intensive care unit, and mortality.

Authors:  Manish Bastakoti; Mohamad Muhailan; Ahmad Nassar; Tariq Sallam; Sameer Desale; Ragai Fouda; Hussam Ammar; Carmella Cole
Journal:  Diagnosis (Berl)       Date:  2021-07-05

Review 9.  The Effectiveness of Electronic Differential Diagnoses (DDX) Generators: A Systematic Review and Meta-Analysis.

Authors:  Nicholas Riches; Maria Panagioti; Rahul Alam; Sudeh Cheraghi-Sohi; Stephen Campbell; Aneez Esmail; Peter Bower
Journal:  PLoS One       Date:  2016-03-08       Impact factor: 3.240

10.  Application of electronic trigger tools to identify targets for improving diagnostic safety.

Authors:  Daniel R Murphy; Ashley Nd Meyer; Dean F Sittig; Derek W Meeks; Eric J Thomas; Hardeep Singh
Journal:  BMJ Qual Saf       Date:  2018-10-05       Impact factor: 7.035

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.