Literature DB >> 23270978

Examining rating quality in writing assessment: rater agreement, error, and accuracy.

Stefanie A Wind1, George Engelhard.   

Abstract

The use of performance assessments in which human raters evaluate student achievement has become increasingly prevalent in high-stakes assessment systems such as those associated with recent policy initiatives (e.g., Race to the Top). In this study, indices of rating quality are compared between two measurement perspectives. Within the context of a large-scale writing assessment, this study focuses on the alignment between indices of rater agreement, error, and accuracy based on traditional and Rasch measurement theory perspectives. Major empirical findings suggest that Rasch-based indices of model-data fit for ratings provide information about raters that is comparable to direct measures of accuracy. The use of easily obtained approximations of direct accuracy measures holds significant implications for monitoring rating quality in large-scale rater-mediated performance assessments.

Entities:  

Mesh:

Year:  2012        PMID: 23270978

Source DB:  PubMed          Journal:  J Appl Meas        ISSN: 1529-7713


  3 in total

1.  Adjacent-Categories Mokken Models for Rater-Mediated Assessments.

Authors:  Stefanie A Wind
Journal:  Educ Psychol Meas       Date:  2016-04-18       Impact factor: 2.821

2.  Exploring Rating Quality in Rater-Mediated Assessments Using Mokken Scale Analysis.

Authors:  Stefanie A Wind; George Engelhard
Journal:  Educ Psychol Meas       Date:  2015-09-17       Impact factor: 2.821

3.  An Iterative Parametric Bootstrap Approach to Evaluating Rater Fit.

Authors:  Wenjing Guo; Stefanie A Wind
Journal:  Appl Psychol Meas       Date:  2021-05-22
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.