Literature DB >> 35989727

Non-iterative Conditional Pairwise Estimation for the Rating Scale Model.

Mark Elliott1, Paula Buttery1.   

Abstract

We investigate two non-iterative estimation procedures for Rasch models, the pair-wise estimation procedure (PAIR) and the Eigenvector method (EVM), and identify theoretical issues with EVM for rating scale model (RSM) threshold estimation. We develop a new procedure to resolve these issues-the conditional pairwise adjacent thresholds procedure (CPAT)-and test the methods using a large number of simulated datasets to compare the estimates against known generating parameters. We find support for our hypotheses, in particular that EVM threshold estimates suffer from theoretical issues which lead to biased estimates and that CPAT represents a means of resolving these issues. These findings are both statistically significant (p < .001) and of a large effect size. We conclude that CPAT deserves serious consideration as a conditional, computationally efficient approach to Rasch parameter estimation for the RSM. CPAT has particular potential for use in contexts where computational load may be an issue, such as systems with multiple online algorithms and large test banks with sparse data designs.
© The Author(s) 2021.

Entities:  

Keywords:  Rasch; estimation; pairwise; rating scale model; simulation

Year:  2021        PMID: 35989727      PMCID: PMC9386884          DOI: 10.1177/00131644211046253

Source DB:  PubMed          Journal:  Educ Psychol Meas        ISSN: 0013-1644            Impact factor:   3.088


Introduction

While modern computing speeds mean that most estimation procedures are fast enough for a standard, offline Rasch analysis to be run, there are contexts where computational expense is an important consideration, for example larger scale automated systems where repeated calibration runs may be required alongside other processes, and large adaptive testing banks with hundreds or even thousands of items. For such contexts, some procedures such as conditional maximum likelihood estimation (CMLE) (Andersen, 1970, 1973) may still prove too computationally expensive for practical use. In this context, non-iterative procedures, which do not require repeated runs, possess an advantage. Another desirable property of a Rasch estimation procedure is that it takes advantage of the Rasch property of the existence of sufficient statistics to condition out person parameters (Andrich & Luo, 2003), resulting in consistent estimates (Andersen, 1973), unlike the unconditional joint maximum likelihood estimation procedure (JMLE; Wright & Masters, 1982; Wright & Panchapakesan, 1969), which simultaneously estimates person and item parameters and suffers from the incidental parameter problem (Ghosh, 1994; Lancaster, 2000; Neyman & Scott, 1948) as a result, since the number of parameters to be estimated increases with the sample size. Conditional estimation methods also avoid the need for distributional assumptions as with marginal approaches, which can lead to estimation bias where person abilities are not normally distributed (Zwinderman & van den Wollenberg, 1990). The only established methods which are both conditional and non-iterative are PAIR (Choppin, 1968, 1985; Wright & Masters, 1982), a procedure based on pairwise comparisons of response data for dichotomous items, and the closely related eigenvector method (EVM; Garner & Engelhard, 2009; Garner, 2002), which also extends the PAIR approach fully to the rating scale model (RSM; Andrich, 1978) and many-facet Rasch models (MFRM; Linacre, 1994). Our focus in this study is on the RSM; we will begin by reviewing the formulation of the RSM, PAIR, and EVM, before discussing theoretical issues, proposing solutions and formulating and testing hypotheses using simulated datasets.

Rating Scale Model

The RSM is a specific formulation of a polytomous Rasch model suitable for tests which are marked on an ordinal scale with three or more categories according to a set of rating criteria (items henceforth), for example an essay marked on a 0 to 5 scale for grammar, vocabulary, organization, and content. The model is also used for questionnaires using Likert scales. The RSM is formally identical to the partial credit model (PCM; Masters, 1982), except that the threshold structure is constrained to be the same across all items. Mathematically, the RSM formulation giving the probability P of person achieving a score of on item with a shared maximum available score of is given, in probability ratio form, by: As Equation 1 indicates, the overall item location parameter and the thresholds are independent of each other.

PAIR

Choppin (1968, 1985) developed the conditional pairwise PAIR algorithm, focusing on the dichotomous Rasch model, or simple logistic model (SLM; Rasch, 1960, 1961). The SLM is a further restricted form of the RSM, with one threshold (with location 0) for each item, with the corresponding simplified equation: If we consider a person responding to a pair of items , Equation 2 gives: Assuming that the requirement for local item independence is met, this is equivalent to: This result means that, in the case where a person responds to only one of a pair of items correctly, the probability that the correctly answered item is or is a function only of the difference in the difficulty of the two items: person ability has been conditioned out. Choppin (1968, 1985) developed methods to produce a vector of item difficulties from Equation 4 in both iterative and non-iterative ways. The non-iterative method, PAIR, is based on substituting the observed counts of and for and in Equation 4 to provide an estimate of for each pair of items in a process involving several steps: Create a matrix of conditional category response frequencies for all possible pairs of items . From this matrix , dividing by provides an estimate of based on the responses to items and . However, contains zero values – the leading diagonal is comprised of zeros and there may also be cases where returns a zero value. Choppin (1985) showed algebraically that raising to successive integer powers will eventually remove all zero values provided is well-conditioned (Fischer, 1981), and that the underlying structure of the resulting matrix is maintained—in effect, direct comparisons between a pair of items and are supplemented by additional comparisons mediated by another item in the case of , and further mediations are added with successive powers. From the resulting matrix , generate a new matrix such that . is a positive pairwise reciprocal matrix: , where . Matrix is a matrix of pairwise comparisons of item difficulties once the person parameters have been conditioned out, which is identical to the pairwise comparison model developed by Zermelo (1929) to determine relative player strength from a chess tournament where there may be missing data, the values in the columns of which relate to item difficulties on the ratio scale form of the Rasch model. Bradley and Terry (1952) independently developed an interval scale version of the same model, which is directly equivalent to the standard log-odds form of the Rasch model with person parameters conditioned out. From the matrix , take the arithmetic mean of each row to create a vector of item difficulty estimates; these represent the item difficulty estimates on the ratio scale version of the Rasch model (Rasch, 1960). To convert to the interval scale Rasch model, take the natural logarithm of each element of the difficulty vector and, for convention, subtract the mean difficulty so that . The primary motivation for Choppin’s (1985) work was the sensitivity of previous methods to missing data, and a weakness which extends to CML: since CML is predicated on the probability of complete response vectors across a set of items, and the studies have indicated that it may be sensitive to missing data (Eggen & Verhelst, 2006; Heine & Tarnai, 2015), whereas PAIR is derived purely from the minimal data required to compare a pair of items; missing data only has an effect to the extent that it reduces the number of comparisons available between a pair of items. Heine and Tarnai (2015) compare the estimates (using the PCM) with randomly removed data to the estimates obtained from complete datasets using PAIR, CML, and MML. Heine and Tarnai find that PAIR outperforms CML, which is often regarded as the gold standard of estimation methods, in terms of the stability of the resulting estimates; PAIR and MML were similar in terms of stability. One limitation of Heine and Tarnai’s findings is that, since they are based on authentic data, there is no definitive ground truth of known difficulty values against which to determine performance. For polytomous responses, PAIR can be extended to the PCM (Masters, 1982) by treating each item threshold as an individual item and counting the conditional category frequencies for pairs of thresholds on different items where persons have scored in the adjacent categories either side of the two thresholds (Garner, 2002). It cannot, however, be used in the same manner for the RSM (Andrich, 1978), due to the constraints imposed on the threshold structure, which are identical across items; an alternative formulation is required.

The EVM

Mainly following Choppin’s approach, (Garner 2002) and Garner and Engelhard (2009) apply an approach to the derivation of the item difficulty vector which differs only in the averaging, step, using the eigenvector corresponding to the principal eigenvalue of the matrix rather than the arithmetic means of the rows. This approach, the EVM, has its foundation in the analytical hierarchy process (AHP; Saaty, 1994; Saaty & Bennett, 1977). Garner and Engelhard also extend the applicability of EVM from the SLM and the PCM to the RSM (Garner 2002) and further (Garner and Engelhard 2009) to the MFRM (Linacre, 1994). For the RSM, we can condition out the person ability and Rasch-Andrich thresholds to obtain a formulation for the estimation of the overall item parameters by considering probabilities of scoring 1 more on item than on item or 1 more on item than on item conditional on the scores being in adjacent categories: or, using local item independence: Garner and Engelhard go on to provide a formulation for the estimation of the Rasch–Andrich thresholds: or, again using local item independence and taking exponents: Garner’s (2002) approach to estimating this is to compare the number of pairs of persons with the same total score who scored and on item to the number of pairs of persons who scored and on the same item (p. 115). This assumes the presence of a complete data matrix with no missing data, since the same total score on a different number of items will correspond to a different ability estimate. In order to extend their approach to datasets with missing data, it is necessary to subset the sample into sets of persons who responded to the same items; pairs would then be then counted within each subset and summed across all subsets. There are certain theoretical issues caused by features of this approach to generating threshold estimates, which conditions out both the item difficulties and person abilities simultaneously in order to generate the conditional category frequency matrix for the thresholds. Firstly, for estimation purposes the person parameters cannot be said to have been fully conditioned out, since persons are grouped according to the same total score, meaning that person ability must be taken into account during the estimation procedure; this may result in biased estimates due to the incidental parameter problem (Ghosh, 1994; Lancaster, 2000; Neyman & Scott, 1948). Secondly, while two persons with the same total score on the same set of items will obtain the same ability estimate due to the sufficiency of the raw score statistic for person parameter estimation, this does not mean that they have the same true ability—the model considers the raw score to the outcome of a stochastic process mapping a continuous variable (ability on the latent trait) onto a closed set of discrete points corresponding to the set of possible raw scores. In reality, the only person who can be said with certainty to have the same underlying ability as a person is that same person, and the use of estimates as a proxy for true ability introduces a source of error. Finally, where there is missing data, the dataset must be subsetted further according to items responded to, since ability scores on the same number of responses to different subsets of items will have different ability estimates and therefore cannot be grouped together. This means more responses are unused, which will be detrimental to estimation accuracy. In fact, all persons who have a unique combination of scores and response patterns are unusable, while others may have small numbers of persons with whom comparisons can be made. For these reasons, we can hypothesize that EVM threshold estimates are likely to be more sensitive to missing data than item difficulty estimates, which do not suffer from the same issues since they follow the PAIR approach apart from the averaging step. The subsetting process also adds computational expense. These observations, particularly the third point, indicate that the a theoretically correct approach to pairwise estimation should consider the scores obtained by the same person on two different items, that is, by counting comparisons of pairs of columns, as it is the case for item difficulty estimates—a modified approach is required. Although it is not possible to condition out both person and item parameters simultaneously while counting columns, it is straightforward to condition out the person ability parameter: or, using local item independence and re-arranging terms: Equation 10 would constitute a sufficient statistic, except that it contains the difference between the difficulties, . However, this can be eliminated by considering the item pair instead of ; reversing and in Equation 10 while keeping thresholds and in the same order gives: Combining Equations 10 and 11 gives: Using Equation 12, it is possible to construct an matrix of threshold differences with . This matrix is the equivalent of the element-wise natural logarithm of the reciprocal pairwise matrix in PAIR/EVM. Taking the arithmetic mean of row : due to the identity . This means that the vector of arithmetic row means of is a set of threshold estimates for thresholds . While this modification avoids the theoretical issues with EVM, it brings its own. Where there are several categories, the distance between categories near opposite ends of the range becomes large, which in terms of deriving estimates from finite and particularly small data sets means that there are likely to be many zero or very small counts, even in both classes, since the probabilities of scoring either or are very small. These uninformative comparisons may lead to small or even zero results for individual estimators of the intervals between such pairs of individual thresholds compared to the real value. However, these uninformative estimators are treated the same as more informative ones, since the procedure makes no distinction when aggregating individual estimators into the final estimates. Considering this fact indicates that it would be desirable to account for uninformative estimators by excluding them or weighting them according to their informativeness.

Conditional Pairwise Adjacent Thresholds

An alternative approach to threshold estimation which avoids making comparisons between the extremes of the scale and permits the possibility of weighting estimators is to consider only pairs of adjacent thresholds. Setting in Equation 10, we have: This provides an estimator for from any given pair of items , subject to eliminating the term , which can be done in one of two ways. Firstly, following the logic employed in the modified EVM method, we can consider the item pair rather than and combine the two equations to give: The second approach is simply to incorporate the item difficulty estimates, which can be estimated separately beforehand using PAIR or EVM, directly into Equation 14. Once estimators have been derived from each pair of items, it remains to find a means of combining these estimators across all possible pairs of items in an appropriate way. The simplest approach to combining the estimators is to take the arithmetic mean, providing a final estimate for . From Equation 15, we have: Where is the total number of items. Alternatively, from Equation 14, we have: where is the total number of items/criteria (we only need to consider cases where in Equation 15 since it is symmetrical in and ). Where any of the counts are zero, neither Equation 17 nor Equation 16 will produce a finite estimate, so all such cases must be discarded (and the estimator count adjusted accordingly); this means that cases where all item pairs produce a zero count, the method will fail to produce an estimate for . This is more likely for Equation 16 since it contains three counts which may be zero, rather than two for Equation 17.

Weighting CPAT Estimators

Combining the set of estimators in the form of a simple arithmetic mean has certain disadvantages, as discussed for the modified formulation of EVM. Some estimators will be more stable than others since they contain more information; for instance, estimators drawn from pairs of items of similar difficulty are likely to have more observations than for estimators drawn from pairs of items of very different difficulties since persons are more likely to obtain similar scores on such pairs of items. However, noting that each estimator is a sufficient statistic in its own right, it is possible to take a weighted average of estimators rather than a simple arithmetic mean—in fact, any arbitrary set of weights will theoretically produce a valid estimate. We therefore have, from Equation 16: or, from Equation 17: where is the weight for item pair . A standard approach to this weighting problem is to weight estimators by the inverse of their variance; this approach, known as inverse-variance weighting, draws its justification from the inequality (Silvey, 1975): more information in an estimator corresponds to less variance. To calculate the inverse-variance weighting for an estimator, we must naturally calculate its variance. We note that the sufficient statistic to calculate the estimator is = p, where , with = 1 if person n scored (k, k) or 0 if person n scored (k − 1, k+ 1) among the persons who scored either or on the pair of items . We must calculate the variance of , where . We can approximate this by using the delta method (Doob, 1935 ), which is based on Taylor series expansions: Here, , with and , so we have: which gives the weighting: There is a direct relationship here with the Fisher information function for the SLM, where . Indeed, this observation provides an alternative means of deriving the weighting function: note that , which is identical to dichotomous items of difficulty 0, the total Fisher information of which is . The weights therefore represent weighting each estimator by the Fisher information of . If we now define as the count of and as the count of , we have and . In terms of and , we have, from Equation 14: where is the harmonic mean of and . Since we can multiply the top and bottom of Equations 18 and 19 by any arbitrary constant, we can simply set: Equation 18 contains a further complication, however: there are two terms, each with its own weighting, and these two weights must be further combined into a single weight. We should note that in doing this, we are combining two estimators which may be of different quality together via a straightforward arithmetic mean as per Equation 16, which cannot account for any difference in informativeness. There does not appear to be a clearly indicated way of combining weights, so following the spirit of Equation 24, the weighting is calculated by combining the terms into a single term and taking the harmonic mean of the numerator and denominator of the combined term as with Equation 24: Giving: We name this Rasch–Andrich threshold estimation approach as the conditional pairwise adjacent thresholds method (CPAT); in order to distinguish between the two approaches to conditioning out the item difficulties from Equation 14, for the remainder of this paper we will call the first method (based on Equation 18) CPAT 1 and the second method (based on Equation 19) CPAT 2.

Additive Smoothing

There can be cases where estimation methods fail to produce a complete set of Rasch-Andrich thresholds since the sample is not large enough and/or poorly targeted across the range of item/threshold combinations, meaning that there is insufficient data to generate a full set of estimates—specifically, this is most likely to occur with estimates for extreme thresholds, that is threshold 1 and threshold , where is the maximum score. It is, however, possible to ensure that a set of estimates is returned for all samples by applying Choppin’s (1985) alternative method for avoiding zeros in the PAIR conditional category frequency matrix, which is to add an arbitrary constant to all counts—this constant can be any small positive number, and does not need to be integer. This approach, called additive smoothing (Murphy, 2012), is designed to account for the fact that zero counts observed in a sample do not reflect underlying small but non-zero probabilities and amounts to the imposition of a weak (depending on the magnitude of the additive constant) uniform prior distribution. Although Choppin presents additive smoothing as an alternative to raising the matrix to successive powers, the two approaches may be combined. An additive smoothing constant operates as a hyperparameter (Murphy, 2012), with its effect on estimates depending on the value chosen. Additive smoothing has two effects on threshold estimates, which act in opposite directions. Since it imposes a uniform prior, it brings the quotient closer to one, which, after taking the logarithm, will reduce the magnitude of the threshold distance, which will tend toward reduced threshold distances (indeed, as the additive constant tends to infinity, all threshold distances will tend to 0). At the same time, additive smoothing leads to the inclusion of estimators with zero counts in the final estimates. These estimators, which must be excluded without additive smoothing, are the result of random error where one result is unlikely—for example, an estimator with 0 out of 20 cases of (0, 1) rather than 1 out of 20 where the true probability is .05. Excluding these estimators, and only these estimators, systematically results in bias toward reduced threshold distances since random error only in one direction is removed—estimators with random error in the other direction (2 or 3 out of 20 rather than 1 out of 20, which underestimate threshold distances, are included). In this way, additive smoothing removes a source of reduced threshold distances. The net effect of these two effects is likely to vary with the value of the additive smoothing constant: any non-zero value will cause the second effect, although with information weighting, the effect will small for small additive smoothing constants. As the additive smoothing constant increases, the first effect will increase, suggesting that there may be an optimal value, which may vary across datasets.

Hypotheses for Testing

We hypothesize the following: EVM will perform less well for threshold estimates than for overall item location estimates due to the issues outlined above. CPAT will perform better than EVM. Weighting CPAT estimators as described above will improve the performance of CPAT. The use of an additive smoothing constant will improve performance in general up to optimal value, after which performance will worsen. We do not hold active hypotheses regarding the relative performance of CPAT 1 and CPAT 2, or the relative performance of PAIR and EVM for item estimates; we leave these as exploratory questions to be answered. We will also compare the performance of CPAT to that of CMLE and JMLE.

Method

In order to evaluate the quality of the estimates against an objective baseline, it is necessary to use simulated rather than authentic data since it is only with simulated data that the original item parameters can be known. The recovered estimates can then be compared against the generating parameters using appropriate metrics, in which way direct comparisons between different estimation algorithms can be made. Luecht and Ackerman (2018) outline the broad approach to such simulation studies: select a model, specify item and person parameters, generate responses from parameters according to model, estimate parameters from the generated responses, and compare the estimates to generating parameters. All analyses is conducted using code written by the authors in Python 3. For this study, 10,000 different datasets were generated with varying numbers of persons and items and different generating parameters. The simulation parameters were generated following a seven-step procedure: Generate a number of items from a discrete integer random uniform distribution in the range . Generate a maximum score from a discrete integer uniform distribution in the range (the number of Rasch–Andrich thresholds to be estimated). Generate a set of item difficulties with a range drawn from a continuous uniform distribution in the range . Generate a sample size in the range with a median of 1,000: from a discrete integer distribution of , where is drawn from a truncated normal distribution in the interval [2, 4]. This generates a skewed distribution, the result of which is illustrated in Figure 1.
Figure 1.

Sample size distribution.

Generate a set of person abilities following a continuous normal distribution with a SD drawn from a continuous random uniform distribution in the range [1, 3.5]. Add an offset to the person ability distribution relative to the item difficulties from a continuous uniform distribution in the range [–0.5, 1]. Generate a set of thresholds with a mean category width from a continuous uniform distribution in the range [0.5, 2], with random perturbations from the mean. Only allow (but do not force) disordered thresholds in 50% of cases through the random perturbations, with a maximum disorder from a continuous uniform distribution in the range [0.5, 1]. Sample size distribution. The parameters were then used to generate simulated response sets following a five-step procedure: Generate person abilities, items and Rasch–Andrich thresholds as described above. Calculate the category response probabilities for each person n on each item for each threshold , where . Calculate the cumulative category response probabilities for each person on each item for each threshold , where . Generate a table of numbers from a continuous uniform distribution in the range [ ]. Score the responses for each person on each item : for successive , if is in the interval , set . The data were simulated to fit the RSM without data-model misfit in order to maximize the interpretability of comparisons between the generating parameters and the recovered estimates; the consequences of different kinds of item misfit therefore fall outside the scope of this study. Two further datasets were generated from each dataset. Firstly, a reduced dataset was generated by removing 30% of persons at random. Secondly, a missing data dataset was generated by removing 30% of individual responses. Missing data can be classified according to the relationships between missingness and traits (Rubin, 1976); in the context of tests may be a design feature (computer adaptive tests, for example, always produce missing data). Here we only consider the simplest case of missing completely at random (MCAR). The two additional datasets allow comparison between a reduction in the number of persons and the same reduction in individual responses, which provides an indication of a procedure’s sensitivity to missing data—an algorithm which is sensitive to missing data will see a greater deterioration in accuracy under the missing data conditions, even though the datasets contain the same number of responses. To compare the recovered estimates to generating parameters across a number of simulations, suitable summary statistics are required. Three summary statistics were calculated for this study to capture different aspects of the estimation error. Firstly, the root mean squared error (RMSE) of the point estimates provides a straightforward measure of the overall amount of error (RMSE is preferred here to MSE due to the interpretability of the units). Secondly, the SD ratio of the recovered item difficulty estimates to the original difficulty estimates used to generate the simulated data provides a measure of the relative dispersion of the two sets of estimates: are the SDs of the two sets of estimates. SD ratio provides a means of investigating estimation bias—systemic error in the estimates. The same amount of estimation error with an SD ratio far from 1 is indicative of more estimation bias as opposed to random noise. Finally, to quantify the modeling error resulting from the combined item and threshold estimates, the parameter-estimation residuals (Luecht & Ackerman, 2018) are calculated. Parameter-estimation residuals are a means of separating the residual error due to parameter estimation error from that due to the stochastic noise which is an inherent feature of the model, and are calculated per response according to the following formula: where is the parameter-estimation residual, is the observed score, is the item response function (IRF) for the generating parameters and is the IRF for the estimated parameters; in other words, is the difference between the residual for the IRF of the generating parameters and the IRF of the estimated parameters. depends on the targeting of the item relative to person ability; a very easy item where both and are very high will result in a small value of even when the difference between and is relatively large, due the IRFs’ asymptotic properties, whereas this will not be the case for items close to the person ability. In this way, can be seen as a measure of how well the estimates model the specific dataset since unlike RMSE, which does not take into account the lack of information which can lead to large point estimate errors for items which are poorly targeted for the population, implicitly accounts for this by quantifying the practical effect of estimation error. To provide a summary statistic for evaluation purposes, the RMSE of all values for all responses in each simulation was calculated. For descriptive statistics, the primary reported summary statistic is the median rather than the mean. This is because the median is not sensitive to outliers, which are likely to be one-sided (RMSE and SD ratio both have a lower bound of 0 and SD ratio has a reciprocal relationship around 1 rather than a linear one) To provide a statistical test for differences between distributions of summary statistics, the Wilcoxon (1945) signed-rank test, a non-parametric equivalent to the paired sample t-test which tests the null hypothesis that the medians of the two samples are the same and which does not require any assumptions of normality for the distributions (an assumption which is not met by the data in this study) was calculated. The Wilcoxon result is reported; as well as reporting p-values, which are likely to be trivially significant with such large sample sizes, meaningful interpretation is provided by effect sizes, which are not sensitive to sample size: firstly, the common language effect size (CLES; McGraw & Wong, 1992) is reported, which is equivalent to the area under the curve (AUC; Fawcett, 2003) and calculated according to the simple difference formula (Kerby, 2014). The CLES states the probability that, from a randomly chosen pair of samples, the result from the sample with the higher median will be higher than the result from the sample with the lower median. The standard Cohen’s d effect size (Cohen, 1962) is not calculated directly since it assumes a normal distribution; instead, an equivalent Cohen’s d value is derived from the CLES, following Ruscio’s (2008) method of conversion. As a guide, Sawilowsky’s (2009) expanded rules of thumb for interpretation of effect size are used: d  is reported as trivial, as very small, as small, as medium, as large, as very large and as huge. Results are compared when appropriate, both for the same method applied to different data conditions (full vs. reduced, full vs. missing and reduced vs. missing) and for different methods under the same data conditions (e.g., EVM vs. CPAT under full data conditions). In additional to statistics, results are presented graphically, using scatter plots to compare results per simulation, with the identity line as a point of reference, and boxplots to compare distributions of results. For the comparisons between CPAT and CMLE and JMLE, the same procedure is used, although on a smaller set of 100 simulations. The Winsteps program (Linacre, 2021) is used to generate CMLE and JMLE estimates. JMLE estimates include Wright’s (1988) bias estimation correction.

Results

Overall Item Location Estimates

Table 1 summarizes the item difficulty estimation results for PAIR and EVM across the three data conditions. Figure 2 compares the performance of PAIR and EVM on each of the three data conditions; the Wilcoxon signed-rank test are as shown in Table 2.
Table 1.

Item Difficulty Estimation Results.

Least squaresEVM
DatasetRMSE SD ratioRMSE SD ratio
Full data0.0560.9900.0580.992
Reduced data0.0670.9900.0690.992
Missing data0.0740.9930.0770.996
Figure 2.

Comparison of RMSE of item estimates for PAIR versus EVM methods under full data, reduced data and missing data conditions.

Table 2.

Wilcoxon Signed-Rank Test: Least Squares Versus EVM, Item Difficulties.

DatasetWilcoxon p-ValueCLESCohen’s d
Full data31,138,136<.0010.6230.443
Reduced data31,278,833<.0010.6260.453
Missing data32,506,566<.0010.6500.546
Item Difficulty Estimation Results. Comparison of RMSE of item estimates for PAIR versus EVM methods under full data, reduced data and missing data conditions. Wilcoxon Signed-Rank Test: Least Squares Versus EVM, Item Difficulties. The results suggest that least squares PAIR performs slightly better than EVM, although the absolute magnitude of the difference is small and EVM appears to be a viable alternative. All three effects are statistically significant, with in all three cases, although as previously discussed, this is perhaps a trivial observation for a study with so many datasets. The effect size is similar and small under full and reduced data conditions, and medium under the missing data condition, suggesting that EVM may be slightly more sensitive to missing data than PAIR. Selecting PAIR as the better performing method, the Wilcoxon signed-rank test was conducted between the three possible pairs of data conditions, as shown in Table 3. Again in all three cases in Table 3. The effect size is huge in each case, indicating that although PAIR may be slightly less sensitive to missing data, there is still an effect—the distribution suggests around 10% more error with the same amount of data (30% removed) in the missing data pattern.
Table 3.

Wilcoxon Signed-Rank Test: Least Squares, Item Difficulties.

Dataset comparisonWilcoxon p-ValueCLESCohen’s d
Full data vs. reduced data41,177,061<.0010.8241.314
Full data vs. missing data44,355,857<.0010.8871.714
Reduced data vs. missing data32,622,770<.0010.6530.555
Wilcoxon Signed-Rank Test: Least Squares, Item Difficulties. Figure 3 summarizes the performance of the least squares and EVM methods across the three data conditions, showing box-and-whisker plots featuring the median and 25th and 75th percentiles (boxes) and 2.5th and 97.5th percentiles (whiskers); outliers are not shown for clarity.
Figure 3.

RMSE of item estimates for last squares (LS) and EVM under full data, reduced data, and missing data conditions.

RMSE of item estimates for last squares (LS) and EVM under full data, reduced data, and missing data conditions.

EVM Threshold Estimates

The results for the threshold estimates are shown in Table 4. The results are markedly less accurate than the results for the item difficulty estimates—the median RMSE for the threshold estimates under the full data condition is 0.439 logits, a figure which is relatively stable for the reduced data condition at 0.467 logits but rises as high as 1.166 logits for the missing data condition. A similar picture was observed for the SD ratio: the median is 1.278 for the full data condition; the figure remains stable for the reduced data condition at 1.292 but rises to 1.668 for the missing data condition. Figure 4, which plots the RMSE for each simulation under the three conditions pairwise, shows the deterioration in the quality of the estimates under missing data conditions graphically. Again, this is supported by the Wilcoxon signed-rank test, as shown in Table 5, with effect sizes well above 2 (huge) for comparisons between the missing data condition and both the full and reduced data condition.
Table 4.

EVM Rasch–Andrich Threshold Estimation Results.

DatasetRMSE SD ratio
Full data0.4391.278
Reduced data0.4671.292
Missing data1.1661.668
Figure 4.

Comparison of RMSE of EVM Rasch–Andrich threshold estimates methods under full data, reduced data and missing data conditions.

Table 5.

Wilcoxon Signed-Rank Test: EVM, Rasch–Andrich Thresholds.

Dataset comparisonWilcoxon p-ValueCLESCohen’s d
Full data vs. reduced data28,774,124<.0010.7721.056
Full data vs. missing data37,217,155<.0010.9994.406
Reduced data vs. missing data37,133,334<.0010.9973.860
EVM Rasch–Andrich Threshold Estimation Results. Comparison of RMSE of EVM Rasch–Andrich threshold estimates methods under full data, reduced data and missing data conditions. Wilcoxon Signed-Rank Test: EVM, Rasch–Andrich Thresholds.

CPAT Threshold Estimates

The results from the four CPAT variants on the 10,000 datasets under the three data conditions are shown in Table 6. It is immediately clear that the results in all four cases are far more accurate than EVM: in the full data case, the improvement in median RMSE is between 80.4% and 84.2% of the figure for EVM, falling slightly in the reduced data case to between 77.8% and 81.9%; in the missing data case they rise above 90% to between 90.0% and 91.9%. The results of the Wilcoxon tests for the worst performing CPAT variant (unweighted CPAT 1) versus EVM are shown in Table 7—effect sizes are all huge (between 4.08 and 6.18), and higher still for other CPAT variants.
Table 6.

CPAT Rasch–Andrich Threshold Estimation Results.

CPAT 1, unweightedCPAT 1, weightedCPAT 2, unweightedCPAT 2, weighted
DatasetRMSE SD ratioRMSE SD ratioRMSE SD ratioRMSE SD ratio
Full data0.0861.0000.0760.9850.0831.0010.0690.988
Reduced data0.1040.9930.0940.9760.0990.9960.0850.981
Missing data0.1170.9820.1100.9640.1080.9880.0950.975
Table 7.

Wilcoxon Signed-Rank Test: EVM Versus Unweighted CPAT 1 and, Rasch–Andrich Thresholds.

DatasetWilcoxon p-ValueCLESCohen’s d
Full data37,227,572<.0010.9994.555
Reduced data37,178,813<.0010.9984.082
Missing data37,251,163<.0011.0006.178
CPAT Rasch–Andrich Threshold Estimation Results. Wilcoxon Signed-Rank Test: EVM Versus Unweighted CPAT 1 and, Rasch–Andrich Thresholds. In terms of comparisons between the four CPAT variants, Table 6 suggests that CPAT 2 performs better than CPAT 1 in both unweighted and weighted cases and that weighted methods perform better than unweighted methods for both CPAT 1 and CPAT 2. This is confirmed by the results of the four relevant pairwise Wilcoxon tests, which are shown in Table 8. The best performing variant (weighted CPAT 2) outperforms the worst performing variant (unweighted CPAT 1)with a large effect size across all three data conditions in both unweighted and weighted cases—in both cases, the effect size increases for the missing data case, suggesting that CPAT 2 is less sensitive to missing data than CPAT 1. Weighted variants outperform their unweighted variants with a small effect size across all three data conditions for both CPAT 1 and CPAT 2. The cumulative effect from unweighted CPAT 1 to weighted CPAT 2 results in medium (bordering on large), effect sizes of between 0.768 and 0.779 which remain stable across the data.
Table 8.

Wilcoxon Signed-Rank Test: Rasch-Andrich Thresholds.

Wilcoxon p-ValueCLESCohen’s d
Full
 CPAT 1 unwtd vs. CPAT 2 unwtd25,975,853<.0010.6970.731
 CPAT 1 wtd vs. CPAT 2 wtd23,714,007<.0010.6370.494
 CPAT 1 unwtd vs. CPAT 1 wtd24,075,947<.0010.6460.531
 CPAT 2 unwtd vs. CPAT 2 wtd25,281,707<.0010.6790.656
 CPAT 1 unwtd vs. CPAT 2 wtd26,373,912<.0010.7080.774
Reduced
 CPAT 1 unwtd vs. CPAT 2 unwtd26,520,789.5<.0010.7120.791
 CPAT 1 wtd vs. CPAT 2 wtd24,544,767<.0010.6590.579
 CPAT 1 unwtd vs. CPAT 1 wtd22,970,900<.0010.6170.420
 CPAT 2 unwtd vs. CPAT 2 wtd24,600,318<.0010.6600.585
 CPAT 1 unwtd vs. CPAT 2 wtd26,313,886<.0010.7060.768
Missing
 CPAT 1 unwtd vs. CPAT 2 unwtd28,165,640.5<.0010.7560.981
 CPAT 1 wtd vs. CPAT 2 wtd25,419,824<.0010.6820.671
 CPAT 1 unwtd vs. CPAT 1 wtd21,961,703<.0010.5900.320
 CPAT 2 unwtd vs. CPAT 2 wtd23,095,453<.0010.6200.432
 CPAT 1 unwtd vs. CPAT 2 wtd26,418,621<.0010.7090.779

Note. Unwtd = unweighted; wtd = weighted.

Wilcoxon Signed-Rank Test: Rasch-Andrich Thresholds. Note. Unwtd = unweighted; wtd = weighted. Figure 5 shows the relative performance of weighted CPAT 2—which, as the best performing variant we shall henceforth focus on and refer to simply as CPAT - versus EVM on the three datasets, with the clearly superior performance of CPAT evident in all cases.
Figure 5.

Comparisons of RMSE of Rasch–Andrich threshold estimates for EVM and CPAT under full data, reduced data, and missing data conditions.

Comparisons of RMSE of Rasch–Andrich threshold estimates for EVM and CPAT under full data, reduced data, and missing data conditions. Figure 6 summarizes the performance of weighted CPAT against EVM across the three data conditions, showing box-and-whisker plots as for the item estimates in Figure 3. The 97.5th percentile for CPAT is considerably below the 25th percentile for EVM in all three cases.
Figure 6.

RMSE of Rasch–Andrich threshold estimates for EVM and CPAT under full data, reduced data, and missing data conditions.

RMSE of Rasch–Andrich threshold estimates for EVM and CPAT under full data, reduced data, and missing data conditions.

Additive Smoothing

The percentages of cases where estimation failed to produce a complete set of Rasch–Andrich thresholds are described in Table 9. CPAT 2 has the lowest failure rate for all three data conditions. EVM has a lower failure rate than CPAT 1 for both the full and reduced data conditions (0.99% and 1.38% vs. 3.02% and 4.49%, but there is a marked deterioration in the performance of EVM for the missing data condition, with a failure rate of 11.13% compared to 6.3% for CPAT 1, further underlying EVM’s sensitivity to missing data; CPAT 2 remains relatively unaffected by missing data, with the failure rate rising from 0.74% to 1.29%.
Table 9.

Estimation Failure Rate, Rasch–Andrich Thresholds.

EVM (%)CPAT 1 (%)CPAT 2 (%)
Full0.993.020.74
Reduced1.384.490.91
Missing11.136.31.29
Estimation Failure Rate, Rasch–Andrich Thresholds. In order to investigate the consequences of applying this approach, the simulation study was repeated four times, with different additive smoothing constants: 0.01, 0.1, and 1. The distributions of the RMSE and SD ratios are shown in Figure 7 together with the original CPAT distribution (shown as an additive smoothing constant of 0), for those cases where CPAT returned estimates. The additive smoothing constant functions as a hyperparameter (Murphy, 2012) of the algorithm.
Figure 7.

RMSE and SD ratio distributions for Rasch–Andrich threshold estimates using different additive smoothing constants.

RMSE and SD ratio distributions for Rasch–Andrich threshold estimates using different additive smoothing constants. The results follow the hypothesized pattern—as well as ensuring that estimates are produced, additive smoothing appears to lead to better estimates up to a point, in particular by reducing estimation bias (SD ratio closer to 1). As Figure 7 shows, the median SD ratio approaches 1 as the additive smoothing constant increases, although the interquartile range and 95% ranges increase as the additive constant increases beyond 0.1, at which point RMSE also increases. A Wilcoxon test between the original CPAT estimates and those with the best-performing additive smoothing constant of +0.1 returns and (medium effect size). Figure 8 shows the distributions of RMSE and SD ratio with a 0.1 additive constant, comparing the cases where CPAT failed to return an estimate without the additive constant with those where it was successful for the full data case. It is clear that the error of the estimates for the failed standard CPAT estimations is still large—this should perhaps be unsurprising given that these are all cases where the sample was poor for estimation purposes in size and/or targeting.
Figure 8.

Comparison of distributions using additive smoothing (+0.1) between estimations which were successful and unsuccessful using CPAT with no additive smoothing.

Comparison of distributions using additive smoothing (+0.1) between estimations which were successful and unsuccessful using CPAT with no additive smoothing.

Parameter-Estimation Residuals

The RMS parameter-estimation residual for the combined item and threshold parameter set for EVM and CPAT, both with no additive smoothing and with an additive smoothing constant of 0.1, are shown in Table 10. For the full data and reduced data conditions, CPAT performs almost identically with or without additive smoothing, and considerably better than EVM, which has a median RMS residual between 2.8 and 3.2 times larger. This disparity grows for the missing data condition between 5.7 and 6 times due to EVM’s sensitivity to missing data; there is also a very marginal difference here in the performance of CPAT, although the median reduction in parameter-estimation residuals is only 0.25% for the full data condition, rising to 0.81% for the missing data condition.
Table 10.

Median Parameter-Estimation Residuals.

FullReducedMissing
EVM0.1650.1770.417
CPAT, no AS0.0530.0630.073
CPAT, AS0.0520.0620.069

Note. AS = additive smoothing.

Median Parameter-Estimation Residuals. Note. AS = additive smoothing. Wilcoxon tests, shown in Table 11 support this, showing very large effect sizes between EVM and CPAT for full and reduced data conditions, and huge effect sizes for the missing data condition. The use of additive smoothing with CPAT causes a small effect for full and reduced data conditions and a medium effect size for the missing data condition, although it should be noted that practically the difference in the case of CPAT is almost trivial.
Table 11.

Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals and Method Comparisons.

Wilcoxon p-ValueCLESCohen’s d
Full
 EVM vs. CPAT, no AS35,554,603<.0010.9161.950
 EVM vs. CPAT, AS35,681,605<.0010.9191.981
 CPAT, no AS vs. CPAT, AS25,100,968<.0010.6470.532
Reduced
 EVM vs. CPAT, no AS35,332,685<.0010.9101.899
 EVM vs. CPAT, AS35,523,590<.0010.9151.943
 CPAT, no AS vs. CPAT, AS26,181,113<.0010.6750.640
Missing
 EVM vs. CPAT, no AS37,339,310<.0010.9622.510
 EVM vs. CPAT, AS37,415,428<.0010.9642.544
 CPAT, no AS vs. CPAT, AS27,268,753<.0010.7030.752

Note. AS = additive smoothing.

Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals and Method Comparisons. Note. AS = additive smoothing. Table 12 shows the results of Wilcoxon tests on the parameter-estimation residuals using the same method under different data conditions. The sensitivity of EVM to missing data is highlighted here; the effect size (Cohen’s d equivalent) between the reduced data and missing data conditions, with the same total number of responses, is 2.617 (huge), which compares to 0.611 (no additive smoothing, medium) and 0.550 (additive smoothing medium). The effect sizes between full and reduced data conditions are fairy similar across all three methods, with EVM slightly lower (1.041) than CPAT (1.217 and 1.180); all three effect sizes are large.
Table 12.

Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals, Data Condition Comparisons.

Wilcoxon p-ValueCLESCohen’s d
EVM
 Full vs. reduced29,851,520.0<.0010.7691.041
 Full vs. missing37,796,202.0<.0010.9742.744
 Reduced vs. missing37,565,222.0<.0010.9682.617
CPAT, no AS
 Full vs. reduced31,252,939.0<.0010.8051.217
 Full vs. missing33,864,186.0<.0010.8731.610
 Reduced vs. missing25,894,124.0<.0010.6670.611
CPAT, AS
 Full vs. reduced30,969,200.0<.0010.7981.180
 Full vs. missing33,421,124.0<.0010.8611.535
 Reduced vs. missing25,279,800.0<.0010.6510.550

Note. AS = additive smoothing.

Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals, Data Condition Comparisons. Note. AS = additive smoothing.

CPAT versus CMLE and JMLE

Table 13 shows the median RMSE and SD ratio of the item and threshold estimates for CPAT (with a +0.1 additive smoothing constant), CMLE and JMLE (with estimation bias correction) under full, reduced, and missing data conditions. It is immediately clear that the results for both CPAT and CMLE are better than those for JMLE across all datasets, with the difference being more pronounced for threshold estimation and under missing data conditions. The JMLE estimates for both items and thresholds generally display bias, in opposite directions: item estimates are stretched, while threshold estimates are compressed. Comparing CPAT with CMLE, we see that CMLE estimates are generally more accurate, although the differences are small—the largest difference between median RMSEs is for items under full data conditions and thresholds under missing data conditions, at 0.012 logits.
Table 13.

Median Estimation RMSE and SD Ratio: CPAT, CMLE, and JMLE.

RMSE SD ratio
CPATCMLEJMLECPATCMLEJMLE
Items
 Full0.0570.0450.1011.0041.0211.129
 Reduced0.0620.0530.1061.0031.0221.128
 Missing0.0630.0660.1601.0121.0331.232
Thresholds
 Full0.0750.0710.3241.0150.9790.850
 Reduced0.0920.0850.3311.0170.9790.842
 Missing0.0940.0820.5031.0220.9760.782
Median Estimation RMSE and SD Ratio: CPAT, CMLE, and JMLE. To consider the combined effect of item and threshold estimates, Table 14 shows the median parameter-estimation residuals for the three procedures under the three data conditions, with the distribution illustrated in boxplot form in Figure 9. We see the same distinction, with CPAT and CMLE performing similarly, particularly under reduced data conditions, with slightly better performance for CMLE, while the JMLE estimates are significantly less accurate. This is borne out by Wilcoxon tests, shown in Table 15: all tests between CPAT or CMLE and JMLE have a huge effect size, while those between CPAT and CMLE return medium to large effect sizes. All results are significant at except for CPAT versus CMLE on reduced data, where .
Table 14.

Median Parameter-Estimation Residuals: CPAT, CMLE, and JMLE.

CPATCMLEJMLE
Full0.0480.0360.110
Reduced0.0480.0410.114
Missing0.0550.0520.175
Figure 9.

RMS parameter-estimation residuals: CPAT, CMLE, and JMLE.

Table 15.

Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals: CPAT, CMLE, and JMLE.

Wilcoxon p-ValueCLESCohen’s d
Full
 CPAT vs. CMLE3,973.0<.0010.7871.124
 CPAT vs. JMLE4,912.0<.0010.9732.718
 CMLE vs. JMLE5,047.0<.0010.9994.584
Reduced
 CPAT vs. CMLE3,247.0.0070.6430.518
 CPAT vs. JMLE4,978.0<.0010.9863.097
 CMLE vs. JMLE5,048.0<.0011.0004.745
Missing
 CPAT vs. CMLE3,520.0<.0010.6970.730
 CPAT vs. JMLE4,917.0<.0010.9742.740
 CMLE vs. JMLE5,000.0<.0010.9903.295
Median Parameter-Estimation Residuals: CPAT, CMLE, and JMLE. RMS parameter-estimation residuals: CPAT, CMLE, and JMLE. Wilcoxon Signed-Rank Test: Parameter-Estimation Residuals: CPAT, CMLE, and JMLE.

Discussion

Both PAIR and EVM showed good performance for the estimation of item difficulties. Their performance was similar enough to suggest that practically there is little difference, although the evidence suggests that the least-squares PAIR method marginally outperforms EVM. Both methods have some sensitivity to missing data, with results for the missing data condition containing significantly more error than under the reduced data condition, although this is relative, and from a low base (median RMSE rises from 0.056 logits to 0.074 for PAIR and 0.058 to 0.077 for EVM). This is to be expected from methods based on pairwise comparisons, since removing individual data points from different response strings reduces the number of available comparisons by more than removing the same number of responses in whole strings: for example, removing a string of responses removes comparisons, while removing one response from each of rows removes comparisons—twice as many, with other missing patterns falling between these two extremes. There are other methods of deriving the final difficulty vector, such as using a cosine maximization approach as described by Kou and Lin (2014), who also catalogue a variety of further methods developed for Saaty and Bennett (1977) and Saaty’s (1994) AHP, or Choppin’s (1985) iterative approach using higher matrix powers to add additional indirect comparisons until convergence. The evaluation of these further methods falls outside the scope of this study, but may be an avenue for further research both in terms of their performance and relative computational cost. It is worth noting in this regard that the least-squares approach is perhaps the simplest possible computationally as it requires only taking the arithmetic means of matrix rows. The results showed support for hypothesis 1 in that the performance of EVM was markedly worse for threshold estimation than for item difficulty estimation, with a high degree of estimation bias, due to the formulation used to condition out person abilities and item difficulties simultaneously and the use of person raw scores to group persons, which introduces two issues: the use of person abilities in the estimation procedure introduces the incidental parameter problem, and additional error is introduced since the raw scores used for this are only proxies for person abilities, not the true underlying abilities. The performance of EVM was found to be particularly sensitive to missing data patterns. The results showed support for hypothesis 2 in that CPAT in all its formulations significantly outperformed EVM. Compared to EVM, CPAT produced a median reduction in RMSE of between 84% and 92%, and a median reduction in parameter-estimation residuals of between 58% and 77%, with threshold estimation performance close to that for overall item locations. The CPAT formulation defined by Equation 16, where the overall item difficult estimates are eliminated entirely without inputting item estimates, appears to perform less well than that defined by Equation 17, where the overall item location estimates, with their own error of estimation, are used. The explanation for this likely resides in the issues around combining two estimators of different quality into a single estimator and how to weight these combined estimators, issues which do not arise in the case of CPAT 2. The results showed support for hypothesis 3 in that weighting estimators produced an improvement in performance, with a median reduction of around 10% to 15% in RMSE. It is perhaps worthy of note that although weighting estimators results in a lower RMSE, the SD ratio is slightly farther from 1 in all cases; this suggestion of bias merits investigation. The weighting process interacts with random error distinctly in different directions—when random error leads to a reduced estimate of threshold distance, that is, overestimating the smaller count, this coincides with increased weighting; the converse is true for underestimates of smaller counts. In this way, there will typically be a slight bias in the estimates—specifically, threshold distances will be underestimated, although this particular source of bias should disappear asymptotically as sample size increases. In typical sample sizes, however, a small degree of systemic bias is introduced at the same time as reducing random error, although the trade-off appears to result in a net benefit. The results also showed support for hypothesis 4 in that the use of additive smoothing produced a small improvement in performance, with a best median reduction of around 3% to 7% in RMSE for an additive smoothing constant of 0.1 (although the median reduction in parameter-estimation residuals was marginal, at less than 1%), as well as avoiding algorithm failure where data is sparse. Here it is worth noting that where no additive smoothing would have resulted in estimation failure, the estimates produced using additive smoothing were typically poor, which is perhaps unsurprising since this situation only occurred where the data were poor for estimation purposes. There may be scope for further gains from more sophisticated approaches to smoothing, although this falls outside the scope of this paper. At this point, it may be worth reflecting beyond the immediate context on some general implications for pairwise comparison methods in general, and indeed any methods where observed counts are substituted directly for theoretical probabilities for estimation purposes. The superior performance of CPAT, despite using less data than EVM—only comparisons between adjacent thresholds are used, rather than comparisons across all pairs of thresholds—highlights the differences between theoretical probabilities, relating to infinite universes, and estimates of those probabilities drawn from finite sets of observations. In particular, where probabilities are small and zero observed counts can become a source of bias, even in larger datasets; the CPAT approach, which involves discarding those comparisons where such counts are likely to occur, indicates that the quality of data can in some cases be more important than the quantity. The further use of additive smoothing can also, somewhat paradoxically, prevent compression of the set of estimates, despite its being the imposition of a uniform prior. When comparing CPAT estimates with those obtained from CMLE and JMLE, CPAT comfortably outperformed JMLE in almost all cases, while CMLE estimates were more accurate than CPAT, although not by a large amount. There are, however, cases where this is not clear-cut, particularly as the amount of data is reduced; the item estimates under reduced data conditions were marginally more accurate than those from CMLE, and the CPAT parameter-estimation residuals under reduced data conditions were smaller than those for CMLE in 35.7% of cases. It is worth noting that the results do not support the hypothesis that CMLE estimates are sensitive to missing data (Eggen & Verhelst, 2006; Heine & Tarnai, 2015); rather, they suggest that it is JMLE estimates which display such sensitivity. Further research is required to determine whether the relative improvement in CPAT estimates on smaller datasets (reduced data conditions here) is indicative of a strong tendency for CPAT to perform better relative to CMLE as sample size decreases, or whether missingness is a meaningful factor on the relative performance of CPAT and CMLE. The findings from this limited investigation suggest that the accuracy of CPAT is close to that of CMLE, and that there may be data designs where CPAT matches or even outperforms CMLE, while the trade-off in terms of computational speed versus a small performance penalty makes CPAT an attractive alternative in many contexts. We have not reported speed comparisons as part of this research since, although CPAT was considerably faster (the mean time for this study to load a dataset and compute PAIR item difficulty estimates and CPAT Rasch–Andrich threshold estimates was as little as 23.6 milliseconds), there are confounding factors which make direct numerical comparisons problematic—different software was used, written in different languages and run on different machines, with Winsteps also running additional analyses such as item fit as part of the estimation process. Since all data for this study was generated to fit the model, further research is required to investigate the effect of misfitting data. Likewise, further research is required to investigate the effect of missing data which is not MCAR. The analytic framework presented in this study, with its evaluation of sensitivity to missing data (as opposed to reduced sample sizes) together with these further strands of enquiry, could constitute the basis of a comprehensive framework for the validation of estimation methods. CPAT builds on the work of Choppin (1968, 1985) and Garner and Engelhard (2009) on non-iterative pairwise Rasch estimation procedures, combining the PAIR approach to overall item location estimation with a procedure which builds on the EVM approach to threshold estimation but does not simultaneously condition out person and item thresholds, instead using previously generated item estimates to create a chain of estimates for distances between adjacent threshold pairs. CPAT appears to constitute a highly computationally efficient approach to RSM parameter estimation that generates high quality estimates. One particular attraction of CPAT is its computational efficiency, since it is non-iterative and based on transformations of simple frequency counts. The results of this study provide evidence for CPAT as a fast estimation algorithm which approaches CMLE in its accuracy and represents an alternative to other more established procedures. CPAT would be particularly attractive in contexts where computational efficiency is at a premium, such as continuous item bank monitoring (e.g., tests at scale involving essays drawn from a large pool) and online systems with multiple simultaneous processes, and for the analysis of large datasets such as computer adaptive test banks. It also appears to have high applicability at the other end of the logistical scale where sample sizes may be small, such as preliminary validation studies of new essay questions and Likert-scale questionnaires, since the results of this study indicate that CPAT produces estimates of a comparable quality to those of CMLE, or sometimes even better, for such sample sizes. CPAT may also be attractive for contexts which involve missing data designs, since the estimates are not very sensitive to missing data, and these results suggest that they may be considerably less sensitive than JMLE estimates in particular. As such, there are a wide range of practical contexts where CPAT could be seen as a viable alternative to other established estimation procedures.
  8 in total

1.  An eigenvector method for estimating item parameters of the dichotomous and polytomous Rasch models.

Authors:  Mary Garner
Journal:  J Appl Meas       Date:  2002

2.  Conditional pairwise estimation in the Rasch model for ordered response categories using principal components.

Authors:  David Andrich; Guanzhong Luo
Journal:  J Appl Meas       Date:  2003

3.  The statistical power of abnormal-social psychological research: a review.

Authors:  J COHEN
Journal:  J Abnorm Soc Psychol       Date:  1962-09

4.  A probability-based measure of effect size: robustness to base rates and other factors.

Authors:  John Ruscio
Journal:  Psychol Methods       Date:  2008-03

5.  Probability tables for individual comparisons by ranking methods.

Authors:  F WILCOXIN
Journal:  Biometrics       Date:  1947-09       Impact factor: 2.571

6.  Using paired comparison matrices to estimate parameters of the partial credit Rasch measurement model for rater-mediated assessments.

Authors:  Mary Garner; George Engelhard
Journal:  J Appl Meas       Date:  2009

7.  Loss of Information in Estimating Item Parameters in Incomplete Designs.

Authors:  Theo J H M Eggen; Norman D Verhelst
Journal:  Psychometrika       Date:  2017-02-11       Impact factor: 2.500

8.  Item bank using sample-free calibration.

Authors:  B Choppin
Journal:  Nature       Date:  1968-08-24       Impact factor: 49.962

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.