Literature DB >> 25626747

A systematic review of validity evidence for checklists versus global rating scales in simulation-based assessment.

Jonathan S Ilgen1, Irene W Y Ma, Rose Hatala, David A Cook.   

Abstract

CONTEXT: The relative advantages and disadvantages of checklists and global rating scales (GRSs) have long been debated. To compare the merits of these scale types, we conducted a systematic review of the validity evidence for checklists and GRSs in the context of simulation-based assessment of health professionals.
METHODS: We conducted a systematic review of multiple databases including MEDLINE, EMBASE and Scopus to February 2013. We selected studies that used both a GRS and checklist in the simulation-based assessment of health professionals. Reviewers working in duplicate evaluated five domains of validity evidence, including correlation between scales and reliability. We collected information about raters, instrument characteristics, assessment context, and task. We pooled reliability and correlation coefficients using random-effects meta-analysis.
RESULTS: We found 45 studies that used a checklist and GRS in simulation-based assessment. All studies included physicians or physicians in training; one study also included nurse anaesthetists. Topics of assessment included open and laparoscopic surgery (n = 22), endoscopy (n = 8), resuscitation (n = 7) and anaesthesiology (n = 4). The pooled GRS-checklist correlation was 0.76 (95% confidence interval [CI] 0.69-0.81, n = 16 studies). Inter-rater reliability was similar between scales (GRS 0.78, 95% CI 0.71-0.83, n = 23; checklist 0.81, 95% CI 0.75-0.85, n = 21), whereas GRS inter-item reliabilities (0.92, 95% CI 0.84-0.95, n = 6) and inter-station reliabilities (0.80, 95% CI 0.73-0.85, n = 10) were higher than those for checklists (0.66, 95% CI 0-0.84, n = 4 and 0.69, 95% CI 0.56-0.77, n = 10, respectively). Content evidence for GRSs usually referenced previously reported instruments (n = 33), whereas content evidence for checklists usually described expert consensus (n = 26). Checklists and GRSs usually had similar evidence for relations to other variables.
CONCLUSIONS: Checklist inter-rater reliability and trainee discrimination were more favourable than suggested in earlier work, but each task requires a separate checklist. Compared with the checklist, the GRS has higher average inter-item and inter-station reliability, can be used across multiple tasks, and may better capture nuanced elements of expertise.
© 2015 John Wiley & Sons Ltd.

Mesh:

Year:  2015        PMID: 25626747     DOI: 10.1111/medu.12621

Source DB:  PubMed          Journal:  Med Educ        ISSN: 0308-0110            Impact factor:   6.251


  56 in total

1.  Development and Validation of an Assessment Tool for Competency in Critical Care Ultrasound.

Authors:  Paru Patrawalla; Lewis Ari Eisen; Ariel Shiloh; Brijen J Shah; Oleksandr Savenkov; Wendy Wise; Laura Evans; Paul Mayo; Demian Szyld
Journal:  J Grad Med Educ       Date:  2015-12

2.  Why Content and Cognition Matter: Integrating Conceptual Knowledge to Support Simulation-Based Procedural Skills Transfer.

Authors:  Jeffrey J H Cheung; Kulamakan M Kulasegaram; Nicole N Woods; Ryan Brydges
Journal:  J Gen Intern Med       Date:  2019-06       Impact factor: 5.128

Review 3.  Simulation for competency assessment in vascular and cardiac ultrasound.

Authors:  Florence H Sheehan; R Eugene Zierler
Journal:  Vasc Med       Date:  2018-02-07       Impact factor: 3.239

Review 4.  Cross-species examination of single- and multi-strain probiotic treatment effects on neuropsychiatric outcomes.

Authors:  Jamie M Joseph; Catrin Law
Journal:  Neurosci Biobehav Rev       Date:  2018-11-22       Impact factor: 8.989

5.  Development of a technical checklist for the assessment of suturing in robotic surgery.

Authors:  Ahmad Guni; Nicholas Raison; Ben Challacombe; Shamim Khan; Prokar Dasgupta; Kamran Ahmed
Journal:  Surg Endosc       Date:  2018-09-07       Impact factor: 4.584

6.  Using Expert Consensus to Develop a Tool to Assess Physical Therapists' Knowledge, Skills, and Judgement in Performing Airway Suctioning.

Authors:  Erin Miller; Dina Brooks; Brenda Mori
Journal:  Physiother Can       Date:  2020       Impact factor: 1.037

7.  A novel assessment tool for evaluating competence in video-assisted thoracoscopic surgery lobectomy.

Authors:  Katrine Jensen; René Horsleben Petersen; Henrik Jessen Hansen; William Walker; Jesper Holst Pedersen; Lars Konge
Journal:  Surg Endosc       Date:  2018-03-30       Impact factor: 4.584

8.  Diagnostic flexible pharyngo-laryngoscopy: development of a procedure specific assessment tool using a Delphi methodology.

Authors:  Jacob Melchiors; Mikael Johannes Vuokko Henriksen; Frederik G Dikkers; Javier Gavilán; J Pieter Noordzij; Marvin P Fried; Daniel Novakovic; Johannes Fagan; Birgitte W Charabi; Lars Konge; Christian von Buchwald
Journal:  Eur Arch Otorhinolaryngol       Date:  2018-02-13       Impact factor: 2.503

9.  Procedure-specific assessment tool for flexible pharyngo-laryngoscopy: gathering validity evidence and setting pass-fail standards.

Authors:  Jacob Melchiors; K Petersen; T Todsen; A Bohr; Lars Konge; Christian von Buchwald
Journal:  Eur Arch Otorhinolaryngol       Date:  2018-04-17       Impact factor: 2.503

10.  Performance of Vascular Exposure and Fasciotomy Among Surgical Residents Before and After Training Compared With Experts.

Authors:  Colin F Mackenzie; Evan Garofalo; Adam Puche; Hegang Chen; Kristy Pugh; Stacy Shackelford; Samuel Tisherman; Sharon Henry; Mark W Bowyer
Journal:  JAMA Surg       Date:  2017-06-01       Impact factor: 14.766

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.