| Literature DB >> 26086658 |
Annwesa P Dasgupta1, Trevor R Anderson2, Nancy Pelaez3.
Abstract
It is essential to teach students about experimental design, as this facilitates their deeper understanding of how most biological knowledge was generated and gives them tools to perform their own investigations. Despite the importance of this area, surprisingly little is known about what students actually learn from designing biological experiments. In this paper, we describe a rubric for experimental design (RED) that can be used to measure knowledge of and diagnose difficulties with experimental design. The development and validation of the RED was informed by a literature review and empirical analysis of undergraduate biology students' responses to three published assessments. Five areas of difficulty with experimental design were identified: the variable properties of an experimental subject; the manipulated variables; measurement of outcomes; accounting for variability; and the scope of inference appropriate for experimental findings. Our findings revealed that some difficulties, documented some 50 yr ago, still exist among our undergraduate students, while others remain poorly investigated. The RED shows great promise for diagnosing students' experimental design knowledge in lecture settings, laboratory courses, research internships, and course-based undergraduate research experiences. It also shows potential for guiding the development and selection of assessment and instructional activities that foster experimental design.Entities:
Mesh:
Year: 2014 PMID: 26086658 PMCID: PMC4041504 DOI: 10.1187/cbe.13-09-0192
Source DB: PubMed Journal: CBE Life Sci Educ ISSN: 1931-7913 Impact factor: 3.325
Figure 1.The process for developing and validating the RED involved (A) a systematic review of the literature to identify experimental design difficulties documented by research, (B) testing three published assessments by looking at more than 1100 responses to see how well they probe for difficulties consistent with research on experimental design difficulties from the literature, and (C) recruiting four cohorts of students to take the assessments to develop a RED based on their responses to published assessments collected before and after an introductory biology course. The assessments are used with permission from: #, SRI International (2003) and the College Board (*, 2006; **, 2009).
Experimental design difficulties classified on the four-level framework and how they relate to what the three published assessments measure
| Published assessmentsd | ||||||
|---|---|---|---|---|---|---|
| Difficultya | Levelb | Demographic populationc | Shrimp | Drug | Bird | |
| I | Identifying the experimental subject ( | Partially established | UN | x | x | x |
| II. Variables: a variable property of an experimental subject | ||||||
| A | Categorical (discrete) variable ( | Partially established | UN | |||
| B | Quantitative (continuous) variable ( | Established | UB | |||
| C | Treatment (independent) variable ( | Established | MS, HS, UN, UB | x | x | x |
| D | Outcome (dependent) variable ( | Established | MS, UN, UB | x | x | |
| E | Control (comparison) group ( | Established | ES, MS, U | x | ||
| F | Combinatorial reasoning (Karplus by | Established | MS, HS, U | x | x | x |
| III | Measurement of results ( | Established | MS, UB | x | x | x |
| IV. How to deal with variability | ||||||
| A | Recognition of natural variation within a biological sample ( | Established | MS, UB | x | ||
| B | Random (representative) sample ( | Established | UB | x | ||
| C | Randomization of treatments ( | Established | UB | x | x | x |
| D | Replication of treatments ( | Established | MS, UB | x | x | x |
| E | Reducing effect of unrelated variables ( | Established | ES, MS, UB | x | x | x |
| V. Interpretation of experimental conclusions | ||||||
| A | Scope of inference/generalizability of results ( | Established | ES, MS, U | x | x | x |
| B | Cause and effect conclusions ( | Established | ES, MS, U | x | x | |
aA review of the literature revealed that student difficulties with experimental design knowledge could be organized into five categories I–V. For definitions of the terms under I–V refer to the glossary of terms in the Supplemental Material (p. 20).
bBased on the four-level framework (Grayson ), “Level” refers to how much insight there is about a particular difficulty. Difficulties found across different populations of students at multiple educational levels are classified as established; others that require further research are classified as partially established.
cU: undergraduate students; UN: undergraduate science nonmajors; UB: undergraduate biology students; ES: elementary school students; MS: middle school students; HS: high school students.
dx represents cases in which scoring materials from the publishers claim the assessment measures knowledge consistent with the difficulty documented by past research.
Rubric for experimental design—REDa
| Areas of difficulty | Propositional statements/completely correct ideas | Typical evidence of difficulties |
|---|---|---|
| 1. Variable property of an experimental subject | Experimental subject or units: The individuals to which the specific variable treatment or experimental condition is applied. An experimental subject has a variable property. | a. An experimental subject was considered to be a variable. |
| A variable is a certain property of an experimental subject that can be measured and that has more than one condition. | b. Groups of experimental subjects were considered based on a property | |
| c. Variable property of experimental subject considered is not consistent throughout a proposed experiment. | ||
| 2. Manipulation of variables | Testable hypothesis: A hypothesis is a testable statement that carries a predicted association between a treatment and outcome variable ( | a. Only the treatment and/or outcome variable is present in the hypothesis statement. |
| b. Hypothesis does not clearly indicate the expected outcome to be measured from a proposed experiment. | ||
| Treatment group: A treatment group of experimental subjects or units is exposed to experimental conditions that vary in a specific way ( | c. Haphazard assignment of treatments to experimental units in a manner inappropriate for the goal of an experiment. | |
| d. Treatment conditions proposed are unsuitable physiologically for the experimental subject or inappropriate according to the goal of an investigation. | ||
| Combinatorial reasoning: In experimental scenarios, when two or more treatment (independent) variables are present simultaneously, all combined manipulations of both together are examined to observe combinatorial effects on an outcome. | e. Independent variables are applied haphazardly in scenarios when the combined effects of two independent variables are to be tested simultaneously. | |
| f. Combining treatments in scenarios where the effect of two different treatments are to be determined individually | ||
| Controlling outside variables: The control and treatment groups are required to be matched as closely as possible to equally reduce the effect of lurking variables on both groups ( | g. Variables unrelated to the research question (often showing a prior knowledge bias) are mismatched across treatment and control groups. | |
| Control group: A control group of experimental subjects or units, for comparison purposes, measures natural behavior under a normal condition instead of exposing them to experimental treatment conditions. Parameters other than the treatment variables are identical for both the treatment and control conditions ( | h. The control group does not provide natural behavior conditions, because absence of the variable being manipulated in the treatment group results in conditions unsuitable for the experimental subject. | |
| i. Control group treatment conditions are inappropriate for the stated hypothesis or experimental goal. | ||
| j. Experimental subjects carrying obvious differences are assigned to treatment vs. control group. | ||
| 3. Measurement of outcome | Treatment and outcome variables should match up with proposed measurements or outcome can be categorical and/or quantitative variables treatments | a. No coherent relationship between a treatment and outcome variable is mentioned. |
| A categorical variable sorts values into distinct categories. | b. The treatment and outcome variables are reversed. | |
| A quantitative or continuous variable answers a “how many?” type question and usually would yield quantitative responses. | c. An outcome variable that is quantitative is treated as a categorical variable. | |
| Outcome group: The experimental subject carries a specific outcome (dependent variable) that can be observed/measured in response to the experimental conditions applied as part of the treatment ( | d. Outcome variables proposed are irrelevant for the proposed experimental context provided or with the hypothesis. | |
| e. Stated outcome not measurable. | ||
| f. No measure was proposed for the outcome variable. | ||
| g. An outcome variable was not listed for an investigation. | ||
| h. There is a mismatch between what the investigation claims to test and the outcome variable. | ||
| 4. Accounting for variability | Experimental design needs to account for the variability occurring in the natural biological world. Reducing variability is essential to reduce effect of nonrelevant factors in order to carefully observe effects of relevant ones ( | a. Claims that a sample of experimental subjects will eliminate natural variability with those subjects. |
| Selection of a random (representative) sample: A representative sample is one where all experimental subjects from a target demographic have an equal chance of being selected in the control or treatment group. An appropriate representative sample size is one that averages out any variations not controlled for in the experimental design ( | b. Criteria for | |
| c. Criteria for selecting experimental subjects for investigation are different in a way that is not representative of the target population. | ||
| Randomized design of an experiment: Randomizing the order in which experimental subjects or units experience treatment conditions as a way to reduce the chance of bias in the experiment ( | d. Decisions to | |
| e. Random assignment of treatments is not considered. | ||
| Randomization can be complete or restricted. One can restrict randomization by using block design, which accounts for known variability in the experiment that cannot be controlled. | f. Random assignment of treatments is incomplete, as they show random assignment of the experimental subjects, but what is needed instead is random assignment of treatments. | |
| Replication of treatments to experimental units or subjects: Replication is performed to assess natural variability, by repeating the same manipulations to several experimental subjects (or units carrying multiple subjects), as appropriate under the same treatment conditions ( | g. Replication means repeating the entire experiment | |
| h. No evidence of replication or suggested need to replicate as a method to access variability or to increase validity/power of an investigation. | ||
| 5. Scope of inference of findings | Scope of inference: Recognizing the limit of inferences that can be made from a small characteristic sample of experimental subjects or units, to a wider target population and knowing to what extent findings at the experimental subject level can be generalized. | a. The inference from a sample is to a different target population. Usually students under- or overestimate their findings beyond the scope of the target population. |
| b. No steps are carried out to randomly select experimental subjects’ representative of the target population about which claims are made. | ||
| Cause-and-effect conclusions: A cause-and-effect relationship can be established as separate from a mere association between variables only when the effect of lurking variables is reduced by random assignment of treatments and matching treatment and control group conditions as closely as possible. Appropriate control groups also need to be considered also in comparison to the treatment group ([ | c. A causal relationship is claimed even though the data show only association between variables. Correlation does not establish causation. ( |
aRefer to the glossary of terms in the Supplemental Material (p. 20).
Examples of student responses with the RED areas of difficulty across three assessments
| 1. Variable property of an experimental subject |
|---|
| Shrimp assessment |
| Correct (C) idea from Anna: |
| Difficulty (D) from Beth: |
| Drug assessment |
| Correct (C) idea from Josh: |
| Difficulty (D) from Ken: |
| Bird assessment |
| Correct (C) idea from Rita: |
| Difficulty (D) from Sara: |
| 2. Manipulation of variables |
| Shrimp assessment |
| Correct (C) idea from Anna: |
| Difficulty (D) from Beth: |
| Drug assessment |
| Correct (C) idea from Josh: “[Administration of] new drug … |
| Difficulty (D) idea from Ken: (i) |
| (ii) |
| (iii) |
| Bird assessment |
| Correct (C) idea from Rita: (i) |
| (ii) |
| Difficulty (D) idea from Sara: (i) |
| (ii) |
| 3. Measurement of outcome |
| Shrimp assessment |
| Correct (C) idea from Anna: |
| Difficulty (D) from Beth: |
| Drug assessment |
| Correct (C) idea from Josh: |
| Difficulty (D) from Ken: |
| Bird assessment |
| Correct (C) idea from Rita: |
| Difficulty (D) from Sara: |
| 4. Accounting for variability |
| Shrimp assessment |
| Correct (C) idea from Anna: “Using only tiger shrimps reduces variance.” |
| Difficulty (D) from Beth: (i) |
| (ii) |
| Correct (C) idea from Josh: |
| Difficulty (D) idea from Ken: (i) |
| (ii) |
| Bird assessment |
| Correct (C) idea from Rita: |
| Difficulty (D) from Sara: |
| 5. Scope of inference |
| Shrimp assessment |
| Correct (C) idea from Anna: |
| Difficulty (D) from Beth: |
| Drug assessment |
| Correct (C) idea from Josh: |
| Difficulty (D) from Ken: |
| Bird assessment |
| Correct (C) idea from Rita: |
| Difficulty (D) from Sara: |
Figure 2.Proportions of students who had correct ideas (dark gray), difficulties (medium gray), and LOE (light gray) for knowledge of experimental abilities as probed by three assessments administered at the beginning and at the end of a semester. The shrimp assessment was given as a posttest during 2009 to cohort A (panel B; n = 40) and as pretest during 2010 to cohort B (panel A; n = 40). The drug assessment was used as a posttest in 2011 for cohort C (panel D; n = 40) and as a pretest in 2012 for cohort D (panel C; n = 31). The bird assessment was assigned as a posttest in 2010 to cohort B (panel F; n = 40) and as a pretest in 2011 to cohort C (panel E; n = 40). The y-axis topics are areas of difficulty from Table 2. Fisher's exact test was applied to compare responses at the beginning and at the end of a semester to detect differences in correct knowledge and difficulties in each area of difficulty for each assessment. *, p < 0.1 significance level; **, p < 0.05 significance level; ***, p < 0.01 significance level.