Literature DB >> 29795895

An Approach to Scoring and Equating Tests With Binary Items: Piloting With Large-Scale Assessments.

Dimiter M Dimitrov1,2.   

Abstract

This article describes an approach to test scoring, referred to as delta scoring (D-scoring), for tests with dichotomously scored items. The D-scoring uses information from item response theory (IRT) calibration to facilitate computations and interpretations in the context of large-scale assessments. The D-score is computed from the examinee's response vector, which is weighted by the expected difficulties (not "easiness") of the test items. The expected difficulty of each item is obtained as an analytic function of its IRT parameters. The D-scores are independent of the sample of test-takers as they are based on expected item difficulties. It is shown that the D-scale performs a good bit better than the IRT logit scale by criteria of scale intervalness. To equate D-scales, it is sufficient to rescale the item parameters, thus avoiding tedious and error-prone procedures of mapping test characteristic curves under the method of IRT true score equating, which is often used in the practice of large-scale testing. The proposed D-scaling proved promising under its current piloting with large-scale assessments and the hope is that it can efficiently complement IRT procedures in the practice of large-scale testing in the field of education and psychology.

Keywords:  equating; item response theory; scaling; test scoring; testing

Year:  2016        PMID: 29795895      PMCID: PMC5965609          DOI: 10.1177/0013164416631100

Source DB:  PubMed          Journal:  Educ Psychol Meas        ISSN: 0013-1644            Impact factor:   2.821


  2 in total

1.  The Rasch model, additive conjoint measurement, and new models of probabilistic measurement theory.

Authors:  G Karabatsos
Journal:  J Appl Meas       Date:  2001

2.  Evaluating the equal-interval hypothesis with test score scales.

Authors:  Ben Domingue
Journal:  Psychometrika       Date:  2013-06-07       Impact factor: 2.500

  2 in total
  7 in total

1.  A Note on the D-Scoring Method Adapted for Polytomous Test Items.

Authors:  Dimiter M Dimitrov; Yong Luo
Journal:  Educ Psychol Meas       Date:  2018-07-04       Impact factor: 2.821

2.  An Application of Reliability Estimation in Longitudinal Designs Through Modeling Item-Specific Error Variance.

Authors:  Georgios D Sideridis; Ioannis Tsaousis; Abdullah Al-Sadaawi
Journal:  Educ Psychol Meas       Date:  2019-04-22       Impact factor: 2.821

3.  Modeling of Item Response Functions Under the D-Scoring Method.

Authors:  Dimiter M Dimitrov
Journal:  Educ Psychol Meas       Date:  2019-06-10       Impact factor: 2.821

4.  Developing Multistage Tests Using D-Scoring Method.

Authors:  Kyung Chris T Han; Dimiter M Dimitrov; Faisal Al-Mashary
Journal:  Educ Psychol Meas       Date:  2019-04-22       Impact factor: 2.821

5.  Testing for Differential Item Functioning Under the D-Scoring Method.

Authors:  Dimiter M Dimitrov; Dimitar V Atanasov
Journal:  Educ Psychol Meas       Date:  2021-03-26       Impact factor: 2.821

6.  The Response Vector for Mastery Method of Standard Setting.

Authors:  Dimiter M Dimitrov
Journal:  Educ Psychol Meas       Date:  2021-07-21       Impact factor: 3.088

7.  On the Choice of the Item Response Model for Scaling PISA Data: Model Selection Based on Information Criteria and Quantifying Model Uncertainty.

Authors:  Alexander Robitzsch
Journal:  Entropy (Basel)       Date:  2022-05-27       Impact factor: 2.738

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.