| Literature DB >> 25574258 |
Alison L Antes1, James M DuBois1.
Abstract
Efforts to advance research integrity in light of concerns about misbehavior in research rely heavily on education in the responsible conduct of research (RCR). However, there is limited evidence for the effectiveness of RCR instruction as a remedy. Assessment is essential in RCR education if the research community wishes to expend the effort of instructors, students, and trainees wisely. This article presents key considerations that instructors and course directors must consider in aligning learning objectives with instructional methods and assessment measures, and it provides illustrative examples. Above all, in order for RCR educators to assess outcomes more effectively, they must align assessment to their learning objectives and attend to the validity of the measures used.Entities:
Year: 2014 PMID: 25574258 PMCID: PMC4278457 DOI: 10.1128/jmbe.v15i2.852
Source DB: PubMed Journal: J Microbiol Biol Educ ISSN: 1935-7877
Aligning instructional objectives, methods, and assessments.
| Foster ethical problem-solving skills in the conduct of research | Increase the ability to recognize ethical issues in the design and conduct of research | Identify and understand research ethics regulations, policies, and resources | Cultivate constructive attitudes towards research ethics and compliance | |
| Researchers confront complex problems involving ethical, regulatory, and interpersonal dimensions. Specific strategies can be taught to improve the quality of decisions. | Researchers must recognize the presence of an ethical issue to engage in problem-solving. Researchers may also require sensitivity toward compliance, professionalism, and broader interpersonal issues to be fully effective. Ethical sensitivity skills are inter-twined with knowledge, problem-solving skills, and attitudes about research ethics. | Researchers require foundational knowledge about the rules and regulations of the research enterprise. This knowledge provides a basis for ethical sensitivity and problem-solving. | To motivate ethical action, individuals must appreciate the importance of RCR and fostering research integrity. Attitudes influence action subsequent to instruction and influence the learning process itself through motivation and engagement. Attitudes are closely linked to values and biases, and researchers may not be fully aware of them or their influence. | |
| Activities must activate the multiple, complex skills associated with ethical problem-solving, such as considering the impact of actions on others, predicting downstream consequences, and applying relevant ethical principles and regulatory rules. Instruction should involve practicing skills through active case discussion or role plays. Case scenarios should not describe flagrant misbehavior, but present complex, “gray” areas that require problem-solving. | Activities and instruction should encourage creative thinking. Students should engage “what if” scenarios to explore multiple possibilities. The notion of particular “correct” answers should be suspended in favor of a focus on multiple competing principles, goals, and concerns. The learning environment must feel open and accepting so that all learners are comfortable sharing ideas. | Traditional lecture format may be effective to deliver key content; however, engaging students in discussions to reinforce concepts and make the topics more personally relevant facilitates learning. For this learning outcome, it may be appropriate for the instructor to think about the traditional model of an expert “delivering” content. However, for the other learning outcomes, the instructor is a facilitator or guide. | The instruction must challenge people to question and test their assumptions about the world, themselves, and others. Activities should challenge students to engage in self-assessment or self-reflection about their values, assumptions, or beliefs. Discussions should engage classmates in debates and sharing related to attitudes toward research ethics and the responsibilities of researchers. Instructors and mentors should model core values and positive attitudes. | |
| Written case analysis; small and large group discussion; role-play, video case analysis, student-generated case writing; online/video simulations | Written case analysis; small and large case discussion; role-play, video case analysis, student-generated case writing | Readings; informational lectures; PowerPoint slides; question-and-answer sessions; quizzes (graded or ungraded); independent study and research; student-led lectures/teaching others; individual or group written reports; work-sheets; concept mapping | Perspective-focused lectures; reflective writing; debate; discussion; blogging; service learning; role modeling; interaction with non-experts (e.g., community members); self-assessments/awareness exercises; peer feedback; creative exercises such as drawing or acting; interviewing others; films; storytelling | |
| Engage the learner in the psychological activities that would underlie real-world ethical problem-solving by presenting scenarios that are interesting, relevant, and engaging. Objective tests should present response options that are all plausible, with some better and some worse. Qualitative approaches should develop detailed coding guides that reflect criteria for good decision making. | Present a realistic scenario followed by an open-ended prompt asking participants to indicate issues within the scenario; trained raters code the responses according to the issues identified. | Multiple-choice items with one best response or fill in the blanks. True/false items are generally not as effective as multiple-choice items in validly discriminating between those who know and do not know material. “Tricky” items should be avoided, as well as response options that are not plausible. | Brief statements followed by Likert-type scale responses to indicate agreement or disagreement with statements. Presentation of value statements or value names that can be rank ordered. Projective measures may involve picking a number of values from a longer list and placing them inside concentric circles. |
Sample assessment measures in the four domains.
| Ethical Decision-Making Measure (EDM) ( | 25 vignettes specific to biological, health, or social sciences; pick two of eight options; about 45 minutes to complete. Produces multiple scores: four ethicality scores across four domains of research behavior—data management, the conduct of human or animal research, professional practices (e.g., treatment of staff and peer review), and business practices (e.g., conflict of interest). Also produces seven scores that reflect use of sensemaking strategies. Items may also be scored for endorsement of social-behavioral responses, such as deception and retaliation. | Beta version validated in sample of 102 doctoral students; demonstrated adequate reliability and correlated appropriately with the other psychological measures (e.g., intelligence, narcissism, self-deceptive enhancement) included to examine construct validity. Subsequent research using this measure in a sample of 252 doctoral students demonstrated that scores on the EDM were related, as expected, to environmental variables, such as laboratory climate and exposure to unethical behavior ( |
| Professional Decision-Making in Research Measure (PDR) | 16 vignettes relevant across human subjects, animal subjects, and translational research; pick two of six options; about 20 minutes to complete. This research is recent and ongoing, but preliminary evidence provides solid support for the validity of the measure ( | Preliminary validation study with 300 NIH-funded researchers using a battery of measures to examine convergent validity. This stage of validation research demonstrated promising evidence for its validity—scores were not correlated with socially desirable responding, they were moderately correlated with narcissism and cynicism, and they were strongly correlated with a measure of moral disengagement in research. Ongoing research will seek to collect normative data in a sample of 400 NIH-funded researchers to establish “typical” scores. |
| Test for Ethical Sensitivity in Science (TESS) ( | Adapted from Bebeau’s Dental Ethical Sensitivity Test ( | No inter-rater agreement estimates provided. A sample of students in an ethics program (n = 133) was compared to a control group (n = 134) using a pre/post design. The training sample scores increased after the course, and the control group scores went down on the posttest. |
| Test of Ethical Sensitivity in Science and Engineering (TESSE) ( | Seven scenarios related to professional practice in science and engineering followed by open-ended space to comment on professional ethical issues and a set of eight statements. Participants were asked to rate each statement on a Likert-type scale according to whether they agree/disagree that it corresponds to an ethical issue in the scenario. Three of the seven scenarios are ethically neutral, and each scenario includes distractor responses that sound important, but are not relevant to the scenario. Authors aim to remove the open-ended portion after initial pilot studies. | No reliability estimates provided. Analyses using a pre/post test design indicated no change in scores from pretest to posttest in the control or experimental groups. Authors recommend instrument revision and further validation studies. |
| Research Ethics Knowledge and Analytical Skills Assessment (REKASA) ( | 33 multiple-choice, true-false, and short-answer items mapped to research ethics knowledge (e.g., IRB procedures, regulatory requirements), in addition to two cases with four open-ended ethical analysis questions each (for 41 items total). | Content validity established by extracting 271 available quiz items and mapping items to testing domains and to learning objectives. An initial pilot of 74 items (split into two assessment tools) was given to a group of 58 researchers before and after a research ethics course. Item discrimination was calculated for each item, and item discrimination greater than 0.2 allowed an item to be retained for the final version. The final version, consisting of 41 items, produced a Cronbach’s alpha reliability coefficient of 0.84. The reliability coefficients of the shortened versions of the test without the case questions (α = 0.72) and the short-answer knowledge questions (α = 0.67) were also estimated. |
| RCR knowledge items indexed to Delphi topics | 125 multiple-choice items with one best choice among four options. Content of items indexed to specific topics within seven core areas of RCR instruction identified by a Delphi panel ( | Items developed to cover core RCR content areas. Correct answers were indexed to five leading RCR textbooks or online courses. Preliminary reliability testing was conducted by dividing the 125 items into five test booklets consisting of 25 items and administering to 232 graduate students at the University of Oklahoma from 2009 to 2011 following RCR training. The average Cronbach’s alpha across the five test booklets was good (0.71) and the Spearman Brown correction for test length provided a stronger reliability estimate (0.92). The average number of participants answering an item correctly was 67%. |
| The How I Think about Research (HIT-Res) | Assesses the use of cognitive distortions (e.g., assuming the worst, blaming others, minimizing, and self-centered thinking) to disengage from research integrity and compliance ( | Preliminary validation data from 300 NIH-funded investigators and trainees indicate excellent internal reliability and that the HIT-Res is strongly correlated with a general measure of moral disengagement. |
| Norms and Counter-norms of Science Survey ( | Presents 16 items, each representing a norm or counter-norm in science (e.g., “Scientists openly share new findings with colleagues” vs. “Scientists protect their newest findings to ensure priority in publishing, patenting, or applications”). Using three sets of three-point scales, participants indicate the degree to which the norms should represent behavior of scientists, do represent the behavior of scientists, and represent their own behavior. | Content validity established through literature reviews and focus groups. Items administered to approximately 3,650 participants to examine variation of norms across disciplines and career stage. However, focus was not on item reliability or measure validation. Reported data focus on frequencies and differences between groups. |
Measure developed by James DuBois and Holly Bante. Measure is owned by the U.S. Office of Research Integrity but may be made available by contacting the lead author at jdubois@wustl.edu.
Articles on the HIT-Res and PDM validation studies are currently in preparation. Further information available by contacting the lead author at jdubois@wustl.edu.