| Literature DB >> 25431359 |
Abstract
BACKGROUND: Written assessments fall into two classes: constructed-response or open-ended questions, such as the essay and a number of variants of the short-answer question, and selected-response or closed-ended questions; typically in the form of multiple-choice. It is widely believed that constructed response written questions test higher order cognitive processes in a manner that multiple-choice questions cannot, and consequently have higher validity. DISCUSSION: An extensive review of the literature suggests that in summative assessment neither premise is evidence-based. Well-structured open-ended and multiple-choice questions appear equivalent in their ability to assess higher cognitive functions, and performance in multiple-choice assessments may correlate more highly than the open-ended format with competence demonstrated in clinical practice following graduation. Studies of construct validity suggest that both formats measure essentially the same dimension, at least in mathematics, the physical sciences, biology and medicine. The persistence of the open-ended format in summative assessment may be due to the intuitive appeal of the belief that synthesising an answer to an open-ended question must be both more cognitively taxing and similar to actual experience than is selecting a correct response. I suggest that cognitive-constructivist learning theory would predict that a well-constructed context-rich multiple-choice item represents a complex problem-solving exercise which activates a sequence of cognitive processes which closely parallel those required in clinical practice, hence explaining the high validity of the multiple-choice format.Entities:
Mesh:
Year: 2014 PMID: 25431359 PMCID: PMC4275935 DOI: 10.1186/s12909-014-0249-2
Source DB: PubMed Journal: BMC Med Educ ISSN: 1472-6920 Impact factor: 2.463
Figure 1Modified bloom’s taxonomy [ [11]].
Figure 2Miller’s pyramid of assessment of clinical skills, competence and performance [ [15]].
Figure 3Stylized depiction of the contrasting ability of the presumed open-ended and multiple-choice formats to assess recognition and recall as opposed to higher forms of cognitive learning. Ideally, multiple-choice and open-ended questions would measure two different abilities (such as recall/recognition versus reasoning/application) – this may be shown as two divergent axes (shown on left). The error variance associated with each type of question is indicated by the shaded blocks, and is much greater for the open-ended question, given its inherent lower reliability. In practice, it appears that the two axes are closely aligned, implying that the two types of questions are measuring essentially the same thing (shown on right). What little additional information the open-ended question might be giving (as shown by a slight divergence in axis) is offset by its wide error variance, which in effect overlaps the information given by the multiple-choice question, thus significantly reducing the value of any additional information it provides.
Adapted and extended from Kim [100]
| Stage | Description | Expected performance |
|---|---|---|
| Novice | Knowledge lacks structure and is essentially context-free. Concepts and relationships are of poor quality | The candidate will be unable to identify or contextualise the problem. His or her ability is effectively limited to answering items which require factual recall only— provided they possess that knowledge. |
| Advanced beginner | Situated knowledge is present but cannot be prioritised appropriately for the problem | Though the problem may be recognised, the candidate will be unable to represent it internally or activate a mental model with sufficient fidelity for problem-solving. |
| Competent learner | Is able to extract the key elements from the problem and possesses many or most of the concepts required for application to the problem, but the relationship between these may not yet be fully mature. | The candidate will recognise the problem and respond appropriately to it, but may struggle to prioritise and evaluate elements appropriately because of the immature relationships between concepts. |
| Proficient learner | Immediately recognises the problem and is able to accommodate it fully in a mental model which permits a solution. | Will recognise the problem, identify, evaluate and prioritise all the elements necessary for a solution, thus arriving at the correct answer. |
| Intuitive expert | Understands and responds to the situation intuitively, using tacit knowledge arising from extensive experience | Is able to answer the question intuitively with minimal analysis or thinking. |
The first column comprises the stages of learning proposed by Dreyfus and Dreyfus [16]; in this context, we shall regard stage of learning as synonymous with level of proficiency or expertise, which is a measure of the effectiveness of problem-solving skill. The second column contains descriptors for each stage chosen for their relevance to complex problem-solving posed by a well-constructed context-rich multiple-choice item. The third column contains a description of the likely performance on that item of a candidate at that stage of proficiency. The relationship between proficiency and performance in a complex multiple-choice item is in fact remarkably direct.