Literature DB >> 26757931

Inter-rater reliability and generalizability of patient note scores using a scoring rubric based on the USMLE Step-2 CS format.

Yoon Soo Park1, Abbas Hyderi2, Georges Bordage3, Kuan Xing3, Rachel Yudkowsky3.   

Abstract

Recent changes to the patient note (PN) format of the United States Medical Licensing Examination have challenged medical schools to improve the instruction and assessment of students taking the Step-2 clinical skills examination. The purpose of this study was to gather validity evidence regarding response process and internal structure, focusing on inter-rater reliability and generalizability, to determine whether a locally-developed PN scoring rubric and scoring guidelines could yield reproducible PN scores. A randomly selected subsample of historical data (post-encounter PN from 55 of 177 medical students) was rescored by six trained faculty raters in November-December 2014. Inter-rater reliability (% exact agreement and kappa) was calculated for five standardized patient cases administered in a local graduation competency examination. Generalizability studies were conducted to examine the overall reliability. Qualitative data were collected through surveys and a rater-debriefing meeting. The overall inter-rater reliability (weighted kappa) was .79 (Documentation = .63, Differential Diagnosis = .90, Justification = .48, and Workup = .54). The majority of score variance was due to case specificity (13 %) and case-task specificity (31 %), indicating differences in student performance by case and by case-task interactions. Variance associated with raters and its interactions were modest (<5 %). Raters felt that justification was the most difficult task to score and that having case and level-specific scoring guidelines during training was most helpful for calibration. The overall inter-rater reliability indicates high level of confidence in the consistency of note scores. Designs for scoring notes may optimize reliability by balancing the number of raters and cases.

Entities:  

Keywords:  Patient note; Rater effects; USMLE Step-2 CS; Validity

Mesh:

Year:  2016        PMID: 26757931     DOI: 10.1007/s10459-015-9664-3

Source DB:  PubMed          Journal:  Adv Health Sci Educ Theory Pract        ISSN: 1382-4996            Impact factor:   3.853


  3 in total

1.  Comparing Students' Clinical Grades to Scores on a Standardized Patient Note-Writing Task.

Authors:  Benjamin D Gallagher; Saman Nematollahi; Henry Park; Salila Kurra
Journal:  J Gen Intern Med       Date:  2020-07-13       Impact factor: 5.128

2.  Evaluator Agreement in Medical Student Assessment Across a Multi-Campus Medical School During a Standardized Patient Encounter.

Authors:  Sherri A Braksick; Yunxia Wang; Suzanne L Hunt; William Cathcart-Rake; Jon P Schrage; Gary S Gronseth
Journal:  Med Sci Educ       Date:  2020-02-05

3.  Correlation between physician assistant students' performance score of history taking and physical exam documentation and scores of Graduate Record Examination, clinical year grade point average, and score of Physician Assistant National Certifying Exam in the United States.

Authors:  Sara Lolar; Jamie L McQueen; Sara F Maher
Journal:  J Educ Eval Health Prof       Date:  2020-05-27
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.