| Literature DB >> 35487728 |
Jawad Al-Khafaji1,2, Ryan F Townsend3, Whitney Townsend4, Vineet Chopra5, Ashwin Gupta6,2.
Abstract
OBJECTIVES: To apply a human factors framework to understand whether checklists reduce clinical diagnostic error have (1) gaps in composition; and (2) components that may be more likely to reduce errors.Entities:
Keywords: Key Words; checklists; diagnostic error; human factors
Mesh:
Year: 2022 PMID: 35487728 PMCID: PMC9058772 DOI: 10.1136/bmjopen-2021-058219
Source DB: PubMed Journal: BMJ Open ISSN: 2044-6055 Impact factor: 3.006
Figure 1Study flow diagram.
Categorisation of checklists components based on SEIPS 2.0*
| Study ID | Study design | Checklist | Work systems | Processes | Outcomes |
| Bahrami 2009 | Expert opinion | Radiology interpretation checklist for brain | Tasks | ||
| Bello 2019 | Expert opinion | Radiology interpretation checklist for skull base | Tasks | ||
| Chew 2016 | Quasi-experimental | Mnemonic tool (TWED) meant to facilitate metacognition | Internal environment | Cognitive | Professional |
| Chew 2017 | Quasi-experimental | Mnemonic tool (TWED) meant to facilitate metacognition (Chew 2016) | Internal environment | Cognitive | Professional |
| Chew 2019 | Focus groups | Mnemonic tool (TWED) meant to facilitate metacognition (Chew 2016) | Internal environment | Cognitive | Professional |
| Ely 2011 | Expert opinion | General debiasing checklist | Tasks | Cognitive | |
| Ely’s differential diagnosis (DDx) checklists | Cognitive | ||||
| Disease-specific cognitive forcing checklist | Tasks | Cognitive | |||
| Ely 2015 | RCT | Ely’s DDx checklists (Ely 2011) | Cognitive | ||
| Ely 2016 | Expert opinion | General checklist for mental pause | Internal environment, Tasks, Organisation, Persons | Cognitive | Professional |
| Ely’s DDx checklists (Ely 2011) | Cognitive | ||||
| Graber 2014 | Pretest and post-test (interviews/user perception) | Checklist for high-risk diagnostic error | Internal environment, Persons | Cognitive | |
| Ely’s DDx checklists (Ely 2011) | Cognitive | ||||
| Hess 2008 | Expert opinion | Lower extremity ulcer checklist | Cognitive | ||
| Huang 2017 | Pretest and post-test, focus groups, chart audits | Diagnostic pause tool | Tasks | Cognitive | |
| Kilian 2019 | Pretest and post-test | Mnemonic tool (ACT) meant to elicit diagnostic reflection | Cognitive | ||
| Kok 2017 | RCT | Chest radiograph interpretation checklist | Tasks | Cognitive | |
| Li 2022 | Expert opinion | Checklist of causes of abdominal pain | Cognitive | ||
| Lv 2022 | Case examples | Checklist of causes of abdominal pain (Li 2022) | Cognitive | ||
| Nedorost 2018 | Observational and survey | Dermatitis checklist | Tasks | Social and Behavioural | |
| Nickerson 2019 | RCT | Electrocardiogram (ECG) syncope checklist | Cognitive | ||
| Nordick 2020 | Expert opinion | Diagnostic and Reasoning Tool | Tasks | Cognitive | |
| O’Sullivan 2018 | Expert opinion | Debiasing checklist | Cognitive | ||
| O’Sullivan 2019 | RCT | Mnemonic tool (SLOW) meant to slow down reasoning and counter bias | Cognitive | ||
| Pan 2021 | Retrospective cohort | Abdominal pain checklist and algorithm | Cognitive | ||
| Rush 2017 | Expert opinion | Mnemonic tool (CARE) meant to counter bias | Tasks | Cognitive, Social and Behavioural | |
| Shimizu 2013 | Pretest and post-test | General debiasing checklist (Ely 2011) | Tasks | Cognitive | |
| Symptom-specific DDx checklist (similar to Ely’s 2011 DDx checklists) | Cognitive | ||||
| Sibbald 2013 | Pretest and post-test | ECG interpretation checklist | Tasks | ||
| Sibbald 2013 | Pretest and post-test | Checklist for cardiac exam | Tasks | ||
| Sibbald 2014 | Pretest and post-test | ECG interpretation checklist (Sibbald 2013–1) | Tasks | ||
| Sibbald 2015 | RCT | ECG interpretation checklist (Sibbald 2013–1) | Tasks | ||
| Sibbald 2019 | RCT | General debiasing checklist (Ely 2011) | Tasks | Cognitive | |
| ECG interpretation checklist (Sibbald 2013–1) | Tasks | ||||
| Weber 1997 | Expert opinion | Checklist for orbital and periorbital swelling | Cognitive | ||
| Yung 1983 | Expert opinion | Flow chart | Tasks | Cognitive |
*The three primary components of SEIPS 2.0 include: Work System (eg, Person (s), Tasks, Internal Environment, External Environment, Tools and Technologies, and Organisation), Processes (eg, Cognitive, Physical, and Social and Behavioural) and Outcomes (eg, Patient, Professional and Organisational).
ACT, Alternatives, Consequences, Traits; CARE, communicate, assess, reconsider, enact; RCT, randomised controlled trial; SEIPS, Systems Engineering Initiative for Patient Safety; SLOW, Sure, Look, Opposite, Worst; TWED, Threat, Wrong/What else, Evidence, Dispositional.
Categorisation of checklists that did and did not (italics) improve diagnostic error
| Study ID | Study design | Checklist | Work systems | Processes | Outcomes |
| Chew 2016 | Quasi-experimental | Mnemonic tool (TWED) meant to facilitate metacognition | Internal environment | Cognitive | Professional |
| Kok 2017 | RCT | Chest radiograph interpretation checklist | Tasks | Cognitive | |
| Pan 2021 | Retrospective cohort | Abdominal pain checklist and algorithm | Cognitive | ||
| Shimizu 2013 | Pretest and post-test | Symptom-specific differential diagnosis (DDx) checklist (similar to Ely’s 2011 DDx checklists) | Cognitive | ||
| Sibbald 2013 | Pretest and post-test | Electrocardiogram (ECG) interpretation checklist | Tasks | ||
| Sibbald 2013 | Pretest and post-test | Checklist for cardiac exam | Tasks | ||
| Sibbald 2014 | Pretest and post-test | ECG interpretation checklist (Sibbald 2013) | Tasks | ||
| Sibbald 2015 | RCT | ECG interpretation checklist (Sibbald 2013) | Tasks | ||
|
| Quasi-experimental | Mnemonic tool (TWED) meant to facilitate metacognition (Chew 2016) | Internal environment | Cognitive | Professional |
|
| RCT | Ely’s DDx checklists (Ely 2011) | Cognitive | ||
|
| Pretest and post-test | Mnemonic tool (ACT) meant to elicit diagnostic reflection | Cognitive | ||
|
| RCT | ECG syncope checklist | Cognitive | ||
|
| RCT | Mnemonic tool (SLOW) meant to slow down reasoning and counter bias | Cognitive | ||
|
| Pretest and post-test | General debiasing checklist (Ely 2011) | Tasks | Cognitive | |
|
| RCT | General debiasing checklist (Ely 2011) | Tasks | Cognitive | |
| ECG interpretation checklist (Sibbald 2013) | Tasks |
*Shimizu et al reported outcomes on two different checklists, one that demonstrated improvement in diagnostic error and one that did not.
ACT, Alternatives, Consequences, Traits; RCT, randomised controlled trial; SLOW, Sure, Look, Opposite, Worst; TWED, Threat, Wrong/What else, Evidence, Dispositional.
Characteristics of studies evaluating effectiveness of checklists
| Study ID | Checklist | Participants | Setting | Outcome |
| Chew 2016 | Mnemonic tool (TWED) meant to facilitate metacognition | Medical Students | Experimental | Checklist group scored significantly higher on a five-case scenario test compared with control group without the checklist (18.50 vs 12.50, respectively) |
| Chew 2017 | Mnemonic tool (TWED) meant to facilitate metacognition | Medical Students | Experimental | No significant difference with or without the checklist in a script concordance test consisting of 10 cases with three response items per case. There was only a significant difference in the checklist group score when looking at the first 5 cases compared with the group without the checklist (9.15 vs 8.18, respectively). |
| Chew 2019 | Mnemonic tool (TWED) meant to facilitate metacognition | Medical Students and Medical Doctors | Clinical | Findings from four separate focus groups suggest that the TWED mnemonic was easy to use and effective in promoting metacognition. |
| Ely 2015 | Ely’s differential diagnosis (DDx) checklists (Ely 2011) | Primary Care Physicians | Clinical | No significant difference in diagnostic error rate between physicians using checklist and those not (11.2% and 17.8% respectively), but checklist did prompt consideration of a greater number of diagnoses per patient (6.5 with checklist vs 3.4 without). |
| Graber 2014 | Checklist for high-risk diagnostic error | Emergency Room Physicians | Clinical | Interviews demonstrated that the majority of checklist use was to help confirm original considerations and had no major impact on the final diagnosis (only 10% of usages resulted in change to the working diagnosis). One-third of usages prompted consideration of novel diagnoses. |
| Ely’s DDx checklists (Ely 2011) | Emergency Room Physicians | Clinical | ||
| Huang 2017 | Diagnostic pause tool | Primary Care Physicians and Nurse Practitioners | Clinical | Diagnostic pause evoked new diagnostic actions in 13% of alerts and resulted in 13% of alerted cases showing diagnostic discrepancies at a 6 month chart audit. Participants reported good integration and minimal interruption of using tool. |
| Kilian 2019 | Mnemonic tool (ACT) meant to elicit diagnostic reflection | Emergency Medicine Residents | Experimental | Emergency medicine residence reviewing eight vignettes altered their provisional diagnosis 13% after using the ACT checklist; however, this did not demonstrate any change in diagnostic error between the provisional diagnosis and the post-checklist diagnosis. |
| Kok 2017 | Chest radiograph interpretation checklist | Medical Students | Experimental | Medical students using the checklist found more abnormalities on chest radiographs with multiple abnormalities (50.1%) compared with the group without the checklist (41.9%). Of note, there was no difference between groups in images containing no abnormalities or a single abnormality. |
| Nedorost 2018 | Dermatitis checklist | Dermatologist (Principal Investigator) | Clinical | Surveys were used to gauge clinician experience after using the checklist. 8 of 15 clinicians surveyed indicated increased efficiency of diagnostic work-up. 10 patients were shown the checklist and in 6 of these instances clinicians reported improved patient engagement. In at least 2 cases, checklist lead to definitive diagnosis on the first visit. |
| Nickerson 2019 | Electrocardiogram (ECG) syncope checklist | Emergency Medicine Residents | Experimental | No significant difference (p=0.19) was found in overall score of residents who read the ECGs with the checklists (median score 7.2; SD 1.4) vs those who read without the checklist (median score 6.8; SD 1.6). There were some significant improvements with checklist use in post-hoc assessment of recognition of Brugada, long QT, heart block and hypertrophic obstructive cardiomyopathy (HOCM) in the ECG readings. Checklist group was more likely to overread normal ECGs as abnormal. |
| O’Sullivan 2019 | Mnemonic (SLOW) meant to slow down reasoning and counter bias | Medical Professionals | Experimental | No significant difference in error rates between checklist and non-checklist groups (2.8 and 3.1 cases correct respectively out of 10 cases total). |
| Pan 2021 | Abdominal pain checklist and algorithm | Deputy Chief Physicians | Clinical | Retrospectively, diagnostic outcomes were assessed for patients presenting to the Emergency Department (ED) with acute abdominal pain for the first time. Cases that were seen using the checklist were placed in the “processes thinking group” while all others went to the “traditional group.” It was found that for hospitalised patients (emergency level 2 and 3) in the processes thinking group there was a significant improvement in diagnostic accuracy as well as patient outcomes and a reduction in length of stay and average hospital expenses. |
| Shimizu 2013 | General debiasing checklist (Ely 2011) | Medical Students | Experimental | Medical students significantly increased their average proportion of correct diagnosis in five case scenarios after using the Differential Diagnosis Checklist (67% correct) compared with initial diagnosis before any checklist (60% correct) even if they had used the Debiasing Checklist beforehand. There was no statistically significance between initial diagnosis and after use of the Debiasing Checklist (62% correct). |
| Symptom-specific DDx checklist (similar to Ely’s 2011 DDx checklists) | Medical Students | Experimental | ||
| Sibbald 2013 | ECG interpretation checklist | Cardiology Fellows | Experimental | Checklist use resulted in a statistically significant lower error rate of 0.39 per ECG interpreted compared with 1.04 without checklist. Use also increased average interpretation and verification time (94 s vs 83 s without) but did not affect surveyed cognitive load. |
| Sibbald 2013 | Checklist for cardiac exam | Medical Students | Experimental | Statistically significant increase in diagnostic accuracy pre-checklist (46%) compared with post-checklist (51%) in the setting of examining a cardiac simulator; however, this benefit was restricted to residents that were allowed to re-examine the simulator while using the checklist. |
| Sibbald 2014 | ECG interpretation checklist (Sibbald 2013–1) | Medical Students, Internal Medicine Residents, and Cardiology Fellows | Experimental | Participants pre-checklist made an average of 2.9 errors per ECG. After using the checklist participants fixed a statistically significant mean number of 1.6 mistakes. |
| Sibbald 2015 | ECG interpretation checklist (Sibbald 2013–1) | Cardiology Residents | Experimental | Checklist use was associated with higher error correction compared with an analytic prompt (0.27 errors corrected per ECG vs 0.04 respectively) as well as greater scrutiny of key variables of the ECG as found per eye-tracking. |
| Sibbald 2019 | General checklist targeting bias (Ely 2011) | Emergency Medicine Residents, Internal Medicine Residents and Cardiology Fellows | Experimental | No significant difference of error rates between content-specific checklist, process-focused checklist and no checklist groups when interpreting 20 ECGs even when cognitive biases were incorporated into cases. |
| ECG interpretation checklist (Sibbald 2013–1) | Emergency Medicine Residents, Internal Medicine Residents and Cardiology Fellows | Experimental |
ACT, Alternatives, Consequences, Traits; SLOW, Sure, Look, Opposite, Worst; TWED, Threat, Wrong/What else, Evidence, Dispositional.