| Literature DB >> 33006068 |
Sabine Zinn1, Uta Landrock2, Timo Gnambs3,4.
Abstract
Educational large-scale studies typically adopt highly standardized settings to collect cognitive data on large samples of respondents. Increasing costs alongside dwindling response rates in these studies necessitate exploring alternative assessment strategies such as unsupervised web-based testing. Before respective assessment modes can be implemented on a broad scale, their impact on cognitive measurements needs to be quantified. Therefore, an experimental study on N = 17,473 university students from the German National Educational Panel Study has been conducted. Respondents were randomly assigned to a supervised paper-based, a supervised computerized, and an unsupervised web-based mode to work on a test of scientific literacy. Mode-specific effects on selection bias, measurement bias, and predictive bias were examined. The results showed a higher response rate in web-based testing as compared to the supervised modes, without introducing a pronounced mode-specific selection bias. Analyses of differential test functioning showed systematically larger test scores in paper-based testing, particularly among low to medium ability respondents. Prediction bias for web-based testing was observed for one out of four criteria on study-related success factors. Overall, the results indicate that unsupervised web-based testing is not strictly equivalent to other assessment modes. However, the respective bias introduced by web-based testing was generally small. Thus, unsupervised web-based assessments seem to be a feasible option in cognitive large-scale studies in higher education.Entities:
Keywords: Computerized testing; Higher education; Measurement invariance; Mode effect; Selection effect; Web-based testing
Mesh:
Year: 2020 PMID: 33006068 PMCID: PMC8219565 DOI: 10.3758/s13428-020-01480-7
Source DB: PubMed Journal: Behav Res Methods ISSN: 1554-351X
Fig. 1Mixed-mode design of the competence assessment with PBA = standardized and supervised paper-based assessment, CBA = standardized and supervised computer-based assessment, WBA = unstandardized and unsupervised web-based assessment, PART = participation, and NONPART = nonparticipation
Fig. 2Example item of the scientific competence test administered in the NEPS. Copyright Leibniz Institute for Educational Trajectories (LIfBi). Reproduced with permission
Number of students by mode and test participation
| Mode | Group | Number of students with a test assigned | Number of students conducting the test | Response rate |
|---|---|---|---|---|
| PBA | Random assignment | 5371 | 1374 | 25.6% |
| CBA | Random assignment | 3431 | 623 | 18.2% |
| WBA | Random assignment | 8671 | 4701 | 54.2% |
| Non-responders from PBA or CBA | 6804† | 1744 | 25.6% |
Note. PBA = standardized and supervised paper-based assessment; CBA = standardized and supervised computer-based assessment; WBA = unstandardized and unsupervised web-based assessment. † One case is missing because the person was not a non-responder in PBA or CBA mode, but the person gave too few valid responses for the estimation of a valid competence score (see Pohl & Carstensen, 2013)
Fig. 3Estimated mode-specific main effects B (white dots) of self-selection analysis for paper-based assessments (PBA), computer-based assessments (CBA), and web-based assessments with random assignment (WBA) with 95% confidence intervals (horizontal bars). Note. Dependent variable is response (coded as 1 = response and 0 = nonresponse). An effect is significant at the .05 level if the confidence interval does not cross the vertical line
Latent variances and empirical reliabilities by assessment mode
| PBA | CBA | WBA | WBA-switch | |
|---|---|---|---|---|
| Variance | 0.73 [0.69, 0.78] | 0.59 [0.52, 0.66] | 0.63 [0.59, 0.67] | 0.57 [0.53, 0.60] |
| Reliability | .79 | .80 | .71 | .70 |
Note. Latent population variances (with 95% confidence intervals) and empirical WLE reliabilities (see Adams, 2005). PBA = standardized and supervised paper-based assessment, CBA = standardized and supervised computer-based assessment, WBA = unstandardized and unsupervised web-based assessment with random assignment, WBA-switch = unstandardized and unsupervised web-based assessment with non-random assignment (for PBA / CBA non-responders)
Fig. 4Test scoring functions for paper-based assessments (PBA), computer-based assessments (CBA), web-based assessments with random assignment (WBA), and web-based assessment with non-random assignment (WBA-switch)
Average differential test functioning
| CBA | WBA | WBA-switch | ||||
| PBA | 0.39 [0.05, 0.72] | 0.87 [0.56, 1.17] | 0.72 [0.39, 1.05] | |||
| CBA | 0.48 [0.16, 0.80] | 0.34 [– 0.00, 0.67] | ||||
| WBA | – 0.14 [– 0.44, 0.16] | |||||
| WBA-switch | ||||||
| CBA | WBA | WBA-switch | CBA | WBA | WBA-switch | |
| PBA | 0.40 [0.10, 0.71] | 0.87 [0.57, 1.17] | 0.73 [0.41, 1.05] | 1.12% [0.27%, 1.97%] | 2.42% [1.58%, 3.25%] | 2.02% [1.13%, 2.91%] |
| CBA | 0.50 [0.21, 0.78] | 0.37 [0.09, 0.64] | 1.39% [0.59%, 2.18%] | 1.02% [0.25%, 1.79%] | ||
| WBA | 0.22 [0.02, 0.42] | 0.61% [0.06%, 1.16%] | ||||
| WBA-switch | ||||||
Note. Confidence intervals (95%) are given in parentheses. = average difference in test scores between groups, = average absolute difference in test scores between groups, = as percentage of maximum test score (here: 36 points), PBA = standardized and supervised paper-based assessment, CBA = standardized and supervised computer-based assessment, WBA = unstandardized and unsupervised web-based assessment with random assignment, WBA-switch = unstandardized and unsupervised web-based assessment with non-random assignment (for PBA / CBA non-responders). Rows represent the reference groups with positive values indicating higher scores in these groups
Fig. 5Signed differential test functioning (sDTF) for paper-based assessments (PBA), computer-based assessments (CBA), and web-based assessments (WBA) with 95% confidence intervals
Linear regressions evaluating prediction bias
| Criterion: | Grade point average | Academic self-concept | Study-related helplessness | Intention to quit | |||||
|---|---|---|---|---|---|---|---|---|---|
| Predictor | 95% CI | 95% CI | 95% CI | 95% CI | |||||
| Intercept | 0.15* | (0.00, 0.30) | – 0.19* | (– 0.33, – 0.06) | 0.14* | (0.01, 0.28) | – 0.01 | (– 0.17, 0.14) | |
| 1. | Scientific literacy | 0.21* | (0.08, 0.34) | 0.46* | (0.34, 0.59) | – 0.22* | (– 0.34, – 0.09) | – 0.09 | (– 0.25, 0.06) |
| 2. | CBA | – 0.21 | (– 0.44, 0.03) | 0.29* | (0.08, 0.51) | – 0.06 | (– 0.27, 0.16) | 0.07 | (– 0.15, 0.28) |
| 3. | WBA | – 0.07 | (– 0.25, 0.11) | 0.17 | (0.00, 0.34) | – 0.11 | (– 0.28, 0.05) | 0.05 | (– 0.14, 0.24) |
| 4. | WBA-switch | – 0.11 | (– 0.32, 0.09) | 0.09 | (– 0.10, 0.28) | – 0.09 | (– 0.28, 0.10) | 0.18 | (– 0.04, 0.39) |
| 5. | 1. x 2. | – 0.09 | (– 0.30, 0.11) | – 0.33* | (– 0.56, – 0.11) | 0.01 | (– 0.19, 0.20) | 0.09 | (– 0.17, 0.35) |
| 6. | 1. x 3. | – 0.13 | (– 0.29, 0.03) | – 0.36* | (– 0.52, – 0.20) | 0.09 | (– 0.06, 0.23) | – 0.05 | (– 0.23, 0.14) |
| 7. | 1. x 4. | – 0.04 | (– 0.23, 0.15) | – 0.26* | (– 0.45, – 0.08) | 0.09 | (– 0.09, 0.28) | – 0.12 | (– 0.33, 0.09) |
| 8. | Sex | – 0.22* | (– 0.35, – 0.10) | 0.10 | (– 0.02, 0.23) | – 0.05 | (– 0.19, 0.09) | – 0.15* | (– 0.28, – 0.01) |
| 9. | Teacher education | – 0.09 | (– 0.23, 0.05) | – 0.05 | (– 0.18, 0.08) | – 0.01 | (– 0.15, 0.13) | 0.01 | (– 0.12, 0.14) |
| .03 / .01 | .05 / .01 | .03 / .00 | .03 / .00 | ||||||
Note. Limited to students of natural sciences (N = 1,825). Linear regression of outcomes on science literacy, assessment mode (dummy-coded with PBA as reference), respective interactions, and covariates. Sex (– 0.5 = male, 0.5 = female) and teacher education (0.5 = other study, – 0.5 = teacher education) were effect-coded. Outcomes and scientific literacy were z-standardized. R2 / ΔR2 = Explained variance / incremental variance explained by moderating effects. PBA = standardized and supervised paper-based assessment, CBA = standardized and supervised computer-based assessment, WBA = unstandardized and unsupervised web-based assessment with random assignment, WBA-switch = unstandardized and unsupervised web-based assessment with non-random assignment (for PBA / CBA nonresponders)
*p < .05