| Literature DB >> 26136693 |
Marc Dupuis1, Emanuele Meier1, Roland Capel1, Francis Gendre1.
Abstract
The functional method is a new test theory using a new scoring method that assumes complexity in test structure, and thus takes into account every correlation between factors and items. The main specificity of the functional method is to model test scores by multiple regression instead of estimating them by using simplistic sums of points. In order to proceed, the functional method requires the creation of hyperspherical measurement space, in which item responses are expressed by their correlation with orthogonal factors. This method has three main qualities. First, measures are expressed in the absolute metric of correlations; therefore, items, scales and persons are expressed in the same measurement space using the same single metric. Second, factors are systematically orthogonal and without errors, which is optimal in order to predict other outcomes. Such predictions can be performed to estimate how one would answer to other tests, or even to model one's response strategy if it was perfectly coherent. Third, the functional method provides measures of individuals' response validity (i.e., control indices). Herein, we propose a standard procedure in order to identify whether test results are interpretable and to exclude invalid results caused by various response biases based on control indices.Entities:
Keywords: exploratory factor analysis; functional method; psychometrics; response reliability; response validity; self-rated questionnaires
Year: 2015 PMID: 26136693 PMCID: PMC4470441 DOI: 10.3389/fpsyg.2015.00629
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
Functional control indices and their definitions.
| Response coherence | A measure of how coherent and predictable the response strategy is. High values in coherence can suggest that one is faking one's answers, while low values in coherence indicate that one is very atypical in one's response pattern, that one completed the questionnaire carelessly, or that one had problems in understanding the items |
| Response reliability | A measure of how stable the response strategy is, obtained after applying the bisection method. A low value in reliability can be interpreted as a lack of application or problems in understanding the items, and negative values surely highlight random response patterns |
| Response mean | A measure of response centrality. In tests with reverse-coded items, high means can indicate that the participant agreed with most items regardless of their meaning, highlighting an acquiescence bias |
| Response variability | A measure of response dispersion. Response sets with little variability indicate that the participant gave little information due to either extreme or central answers. This also invalidates the results (which are always based on response variance) |
| Response modality | A measure of how often the modal answer is chosen, using Cohen's weighted kappa. High values in modality can suggest that the subject attempted to describe himself or herself as a very banal person. More interestingly, very negative values in modality highlight that the participant might have made a mistake by reporting reversed answers |
| Response normativity | A measure of how much a response set fits general response tendencies, using the correlation between the participant's answers and the means of each item. This measure's main use is to detect reversed answers, but it is also sensitive to socially desirable responding |
| Response positivity and negativity | Measures of how much both positive and negative aspects of personality have been accentuated in one's self-description. Such measures are interesting in order to detect unbalanced self-descriptions: depending on the context (e.g., applying for a job), people can overrate positive dimensions and underrate negative dimensions, which can be highlighted by calculating the difference between positivity and negativity |
Figure 1A decision flowchart for determining whether one's test results are interpretable or not.
A comparison of CCT, IRT, and FMT characteristics.
| The standard error of measurement is consistent across scores, but differs across populations. | ✓ | ✓ | |
| The standard error of measurement varies across scores, but is consistent across populations. | ✓ | ||
| Shorter tests can be as reliable as longer tests. | ✓ | ✓ | |
| The comparison of test scores across different forms does not depend on test parallelism. | ✓ | ✓ | |
| Unbiased estimates of item characteristics can be calculated using non-representative samples. | ✓ | ||
| Trait scores are measured by comparison of distances from the items. | ✓ | (✓) | |
| Trait scores are measured by correlations between the vector of strategy and the items' characteristics. | ✓ | ||
| Individuals, items, and factor scores are expressed in the same metric and are therefore comparable. | ✓ | ✓ | |
| Scores are expressed on a continuous metric. | ✓ | ✓ | |
| The scores are expressed on an absolute metric. | ✓ | ✓ | |
| The absolute metric proposed carries a psychological meaning. | ✓ | ||
| The method is specifically designed for multidimensional tests with polytomous items. | ✓ | ✓ | |
| Each item is assumed to be correlated to every factor. | ✓ | ||
| The prediction of other test scores is maximized by design. | ✓ | ✓ | |
| Item-specific person-fit analyses can be conducted. | ✓ | ✓ | |
| General person-fit adequacy can be measured. | ✓ | ✓ | |
| Specific functional indices (i.e., coherence, reliability, positivity, negativity) can be calculated. | ✓ | ||
| Non-specific indices (i.e., level, variability, modality, normativity) can be calculated. | ✓ | ✓ | ✓ |
| Explicative hypotheses concerning person-fit inadequacy can be empirically supported by indices. | ✓ |
Note. ✓, theoretically assumed; (✓), theoretically assumed and applicable, and not yet exploited.