| Literature DB >> 26903498 |
Todd D Reeves1, Gili Marbach-Ad2.
Abstract
Most discipline-based education researchers (DBERs) were formally trained in the methods of scientific disciplines such as biology, chemistry, and physics, rather than social science disciplines such as psychology and education. As a result, DBERs may have never taken specific courses in the social science research methodology--either quantitative or qualitative--on which their scholarship often relies so heavily. One particular aspect of (quantitative) social science research that differs markedly from disciplines such as biology and chemistry is the instrumentation used to quantify phenomena. In response, this Research Methods essay offers a contemporary social science perspective on test validity and the validation process. The instructional piece explores the concepts of test validity, the validation process, validity evidence, and key threats to validity. The essay also includes an in-depth example of a validity argument and validation approach for a test of student argument analysis. In addition to DBERs, this essay should benefit practitioners (e.g., lab directors, faculty members) in the development, evaluation, and/or selection of instruments for their work assessing students or evaluating pedagogical innovations.Entities:
Mesh:
Year: 2016 PMID: 26903498 PMCID: PMC4803101 DOI: 10.1187/cbe.15-08-0183
Source DB: PubMed Journal: CBE Life Sci Educ ISSN: 1931-7913 Impact factor: 3.325
Figure 1.Categories of evidence used to argue for the validity of test score interpretations and uses (AERA, APA, and NCME, 2014).
Example table of specifications for evolution by means of natural selection test showing numbers of test items pertaining to each content area at each cognitive level and total number of items per content area and cognitive level
| Cognitive process | ||||
|---|---|---|---|---|
| Content (behavioral objective) | Comprehension | Application | Analysis | Total |
| 1. Students will define evolution by means of natural selection. | 1 | 1 | ||
| 2. Students will define key principles of evolution by means of natural selection (e.g., heredity, differential fitness). | 5 | 5 | ||
| 3. Students will compute measures of absolute and relative fitness. | 5 | 5 | ||
| 4. Students will compare evolution by means of natural selection with earlier evolution theories. | 3 | 3 | ||
| 5. Student will analyze phylogenetic trees. | 4 | 4 | ||
| Total | 6 | 5 | 7 | 18 |
Example validity argument and validation approach for a test of students’ ability to analyze the elements of evidence-based scientific arguments showing argument claims and subclaims concerning the validity of the intended test score interpretations and uses and relevant validity evidence used to substantiate those claims
| Validity argument claims and subclaims | Relevant validity evidence based on |
|---|---|
| 1. The overall score represents a student’s current level of argument-element analysis skills, because: | – |
| a single higher-order construct (i.e., argument-element analysis skill) underlies all item responses. | Internal structure |
| the overall score is distinct from content knowledge and thinking dispositions. | Relations with other variables |
| the items represent a variety of arguments and argument elements. | Test content |
| items engage respondents in the cognitive process of argument-element analysis. | Response processes |
| the overall score is highly related to other argument analysis measures and less related to content knowledge and thinking disposition measures. | Relations with other variables |
| 2. A subscore (e.g., claim identification) represents a student’s current level of argument-element identification skill, because: | – |
| each subscore is distinct from other subscores and the total score (the internal structure is multidimensional and hierarchical). | Internal structure |
| the items represent a variety of arguments and particular argument elements (e.g., claims). | Test content |
| the subscore is distinct from content knowledge and thinking dispositions. | Relations with other variables |
| items engage respondents in the cognitive process of identifying a particular element argument (e.g., claims). | Response processes |
| subscores are highly related to other argument analysis measures and less related to content knowledge and thinking disposition measures. | Relations with other variables |
| 3. Error indicators can be interpreted to represent students’ current errors made in identifying particular argument elements, because when students misclassify an element in the task, they are making cognitive errors. | Response processes |
| 4. Use of the test will facilitate improved argument instruction and student learning, because: | – |
| teachers report that the test is useful and easy to use and have positive attitudes toward it. | Consequences of testing |
| teachers report using the test to improve argument instruction. | Consequences of testing |
| teachers report that the provided information is timely. | Consequences of testing |
| teachers learn about argumentation with test use. | Consequences of testing |
| students learn about argumentation with test use | Consequences of testing |
| any unintended consequences of test use do not outweigh intended consequences. | Consequences of testing |