Literature DB >> 27651046

Inter-rater variability as mutual disagreement: identifying raters' divergent points of view.

Andrea Gingerich1, Susan E Ramlo2, Cees P M van der Vleuten3, Kevin W Eva4, Glenn Regehr4.   

Abstract

Whenever multiple observers provide ratings, even of the same performance, inter-rater variation is prevalent. The resulting 'idiosyncratic rater variance' is considered to be unusable error of measurement in psychometric models and is a threat to the defensibility of our assessments. Prior studies of inter-rater variation in clinical assessments have used open response formats to gather raters' comments and justifications. This design choice allows participants to use idiosyncratic response styles that could result in a distorted representation of the underlying rater cognition and skew subsequent analyses. In this study we explored rater variability using the structured response format of Q methodology. Physician raters viewed video-recorded clinical performances and provided Mini Clinical Evaluation Exercise (Mini-CEX) assessment ratings through a web-based system. They then shared their assessment impressions by sorting statements that described the most salient aspects of the clinical performance onto a forced quasi-normal distribution ranging from "most consistent with my impression" to "most contrary to my impression". Analysis of the resulting Q-sorts revealed distinct points of view for each performance shared by multiple physicians. The points of view corresponded with the ratings physicians assigned to the performance. Each point of view emphasized different aspects of the performance with either rapport-building and/or medical expertise skills being most salient. It was rare for the points of view to diverge based on disagreements regarding the interpretation of a specific aspect of the performance. As a result, physicians' divergent points of view on a given clinical performance cannot be easily reconciled into a single coherent assessment judgment that is impacted by measurement error. If inter-rater variability does not wholly reflect error of measurement, it is problematic for our current measurement models and poses challenges for how we are to adequately analyze performance assessment ratings.

Entities:  

Keywords:  Inter-rater variability; Mini-CEX; Q methodology; Rater cognition; Rater-based assessment; Workplace-based assessment

Mesh:

Year:  2016        PMID: 27651046     DOI: 10.1007/s10459-016-9711-8

Source DB:  PubMed          Journal:  Adv Health Sci Educ Theory Pract        ISSN: 1382-4996            Impact factor:   3.853


  10 in total

1.  How Do Thresholds of Principle and Preference Influence Surgeon Assessments of Learner Performance?

Authors:  Tavis Apramian; Sayra Cristancho; Alp Sener; Lorelei Lingard
Journal:  Ann Surg       Date:  2018-08       Impact factor: 12.969

2.  Proceed With Caution: Implementing Competency-Based Graduate Medical Education.

Authors:  M Douglas Jones; Tai M Lockspeiser
Journal:  J Grad Med Educ       Date:  2018-06

3.  Teaching residents how to break bad news: piloting a resident-led curriculum and feedback task force as a proof-of-concept study.

Authors:  Joseph Sleiman; David J Savage; Benjamin Switzer; Colleen Y Colbert; Cory Chevalier; Kathleen Neuendorf; David Harris
Journal:  BMJ Simul Technol Enhanc Learn       Date:  2021-06-22

Review 4.  Better Decision-Making: Shared Mental Models and the Clinical Competency Committee.

Authors:  Laura Edgar; M Douglas Jones; Braden Harsy; Morgan Passiment; Karen E Hauer
Journal:  J Grad Med Educ       Date:  2021-04-23

5.  Guidelines: The do's, don'ts and don't knows of direct observation of clinical skills in medical education.

Authors:  Jennifer R Kogan; Rose Hatala; Karen E Hauer; Eric Holmboe
Journal:  Perspect Med Educ       Date:  2017-10

6.  Considering the interdependence of clinical performance: implications for assessment and entrustment.

Authors:  Stefanie S Sebok-Syer; Saad Chahine; Christopher J Watling; Mark Goldszmidt; Sayra Cristancho; Lorelei Lingard
Journal:  Med Educ       Date:  2018-04-19       Impact factor: 6.251

7.  Communication skills supervisors' monitoring of history-taking performance: an observational study on how doctors and non-doctors use cues to prepare feedback.

Authors:  Michaela Wagner-Menghin; Anique B H de Bruin; Jeroen J G van Merriënboer
Journal:  BMC Med Educ       Date:  2020-02-06       Impact factor: 2.463

8.  E-ASSESS: Creating an EPA Assessment Tool for Structured Simulated Emergency Scenarios.

Authors:  Caroline Andler; Sneha Daya; Katie Kowalek; Christy Boscardin; Sandrijn M van Schaik
Journal:  J Grad Med Educ       Date:  2020-04

9.  Determining the influence of different linking patterns on the stability of students' score adjustments produced using Video-based Examiner Score Comparison and Adjustment (VESCA).

Authors:  Peter Yeates; Gareth McCray; Alice Moult; Natalie Cope; Richard Fuller; Robert McKinley
Journal:  BMC Med Educ       Date:  2022-01-17       Impact factor: 2.463

10.  "Languaging" tacit judgment in formal postgraduate assessment: the documentation of ad hoc and summative entrustment decisions.

Authors:  Anneke van Enk; Olle Ten Cate
Journal:  Perspect Med Educ       Date:  2020-12
  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.