| Literature DB >> 33337579 |
Kate L Lapane1, Catherine E Dube1.
Abstract
In the Spring of 2020, we launched a rigor and reproducibility curriculum for medical students in research training programs. This required class consisted of eight, 2-h sessions, which transitioned to remote learning in response to the coronavirus disease 2019 (COVID-19) epidemic. The class was graded as pass/fail. Flipped classroom techniques, with multiple hands-on exercises, were developed for first-year medical students (MD/PhD [n = 9], Clinical and Translational Research Pathway (CTRP) students [n = 9]). Four focus groups (n = 13 students) and individual interviews with the two instructors were conducted in May 2020. From individual interviews with instructors and focus groups with medical students, the course and its components were favorably reviewed. Students thought the course was novel, important, relevant, and practical-and teaching strategies were effective (e.g., short lectures, interactive small group exercises, and projects). Most students expressed concerns about lack of time for course preparation. Sharper focus and streamlining of preparation work may be required. Pre- and post-student self-assessments of rigor and reproducibility competencies showed average post-scores ranging from high/moderate to strong understanding (n = 11). We conclude that rigor and reproducibility can be taught to first-year medical students in research pathways programs in a highly interactive and remote format. Study Highlights WHAT IS THE CURRENT KNOWLEDGE ON THE TOPIC? The rigor and reproducibility crisis calls for robust training of scientists in best practices for enhancing the research rigor. WHAT QUESTION DID THIS STUDY ADDRESS? We evaluated a curriculum to develop physician-scientists skilled at documenting research workflow from idea generation to publication with reproducibility in mind. WHAT DOES THIS STUDY ADD TO OUR KNOWLEDGE? Highly interactive exercises, coupled with a hands-on replication group project provide a pathway for students to gain competencies important to the improvement of rigor and reproducibility in scientific research. Rigor and reproducibility can be taught in a highly interactive format and using a remote format. HOW MIGHT THIS CHANGE CLINICAL PHARMACOLOGY OR TRANSLATIONAL SCIENCE? Formal training is needed to raise awareness of the reproducibility crisis and improve the rigor of research conducted. If techniques taught are used, the transparency and reproducibility of clinical and translational science will be improved.Entities:
Mesh:
Year: 2021 PMID: 33337579 PMCID: PMC8212706 DOI: 10.1111/cts.12966
Source DB: PubMed Journal: Clin Transl Sci ISSN: 1752-8054 Impact factor: 4.689
Goals, learning objectives, estimated time commitment, and actual median and range of time on Blackboard Learning Management System (hours) stratified by module
| Module title | Goals and learning objectives |
Cumulative time Commitment |
|---|---|---|
| Reproducibility crisis |
Goal: To introduce the origins and history of the Reproducibility Crisis.
Describe the origins of the reproducibility crisis. Know what the NIH response to the reproducibility crisis has been. List key stakeholders and describe strategies for addressing the reproducibility crisis. Define reproducibility, replication, and generalizability. |
Expected: 4 h Median time on Blackboard: 1.2 Range: 0–8.2 <5 min: 6% |
| Evaluating rigor of prior research |
Goal: To define the requirements for an NIH scientific premise and provide basic skills to evaluate the rigor of existing research studies and proposals.
Describe the role and importance of rigor and reproducibility in NIH proposal writing and NIH scientific review. Describe the importance of scientific premise in NIH proposal preparation. Critique scientific premise statements. |
Expected: 4 h Median: 0.8 Range: 0–8.9 <5 min: 24% |
| Rigorous experimental design and bias |
Goal: To review the elements of experimental design, tools and standards – including sex as a biological variable (NIH priority); to highlight areas of potential bias.
Discuss the importance of rigorous experimental design and documentation for transparency and replication. Describe when to include sex as a biological variable in research. Define bias and the sources of bias in the conduct of science. Assess bias using the Cochrane Collaboration’s tool for assessing risk of bias in randomized trials. Develop a prospective experimental design that comports with appropriate guidelines. |
Expected: 4 h Median: 0.8 Range: 0–6.1 <5 min: 47% |
| Biological variables, authentication and QC |
Goal: To provide an overview of quality procedures for biomedical research, including authentication procedures. To provide an opportunity to discuss implementation challenges in laboratory settings.
Describe the key elements to include in an authentication plan for an NIH grant application. Describe quality practices important to basic biomedical research. Discuss the implementation of quality practices. |
Expected: 5 h Median: 0.02 Range: 0–2.6 <5 min: 65% |
| Reporting expectations |
Goal: To review reporting guidelines used for manuscript preparation and to provide an overview of image processing and manipulation as it applies to clear and accurate reporting.
Describe how image data may be evaluated to determine whether manipulation has occurred. Describe software tools used to inspect images for manipulation. Using an article of your choosing, evaluate how well authors adhere to transparent reporting publication guidelines. |
Expected: 4.25 h Median: 1.2 Range: 0–5.0 <5 min: 18% |
| Implementing transparency |
Goal: To present a workflow that promotes transparency including detailed record keeping and data management.
Describe the role of lab notebooks in promoting rigor and reproducibility. Describe the roles of the data management plan, metadata, and data dictionary. Describe the challenges and benefits of increased scientific transparency. Critically reflect on practices in your laboratory and consider possible steps toward increased transparency. |
Expected: 3.5 h Median: 0.01 Range: 0–1.8 <5 min: 88% |
| Open science |
Goal: To provide an overview of the principles of open science and practical steps that can be undertaken to promote its implementation.
Define “open science.” Describe the overall goals of open science. Describe the challenges to the implementation of open science. Describe institutional changes that promote rigor and reproducibility. Select an open science objective and identify changes to current practices that promote its achievement. |
Expected: 3.5 h Median: 0.4 Range: 0–5.0 <5 min: 29% |
| Total time on Blackboard Learning Management System across all elements of the class (hours): mean: 12.9, SD: 5.9 | ||
Abbreviations: NIH, National Institutes of Health; QC, quality control.
Including assigned readings, preparatory work, and ongoing work on the project.
Interactive in‐class exercises
| Topic | In‐class activities |
|---|---|
|
Reproducibility Crisis 2 small group (3–5 students) discussions 15 min each 5‐min summary of each group’s discussion |
|
|
Consider what is already being done (or a new idea!) and how “success” of the strategies might be measured. Discuss the implications for implementing (pros/cons) from your stakeholder perspective. List pros/cons from other stakeholder perspectives. | |
|
Evaluating the rigor or previous research, scientific premise 1 small group (3–5 students) discussion 30 min for discussion with 5‐min summary from each group | Each group discusses the high‐level overview of an F30 proposal assigned to the group. Based on readings regarding the importance of scientific premise in NIH review of proposals, what specific prior research studies would your group like to see referenced in support of the scientific premise of this NIH proposal? Has the research your group believes is necessary been done? What is the quality or the previous research which forms as the foundation for the current proposal? Discuss how to determine the rigor of the studies you would like to see before you would highly score the application. |
|
| |
| Rigorous experimental design and bias |
Followed by small group discussion with questions provided by NIH (e.g., can you think of a particular instance in which blinding and randomization could have a dramatic impact on the results?) Cochrane assessment bias tool exercise. Hands on exercise with Experimental Design Assistant Tool |
| Biological variables, authentication, and quality control |
Followed by small group discussion with questions provided by the NIH (e.g., Have you or someone you know only used male mice in an experiment as a way of avoiding the “sex issue?” Do you think this is appropriate? Does it depend on the type of experiment being done?).
|
|
| |
|
Reporting expectations 1 small group (3–5 students) discussion 20 min for discussion with 5‐min summary from each group | Each small group assigned an article. |
|
| |
|
Implementing transparency 2 small group (3–5 students) discussions 15 min each 5‐min summary of each group’s discussion |
Followed by small group discussion with questions provided by NIH (e.g., Do you think the corresponding author should have handled the situation differently?).
|
| Open science |
Standard debate format (see text). Debate 1: Should scientists at our institution be required to use an open science framework for their research? Debate 2: Should federal funders of research in the United States (e.g., NIH, NSF, etc.) participate in Plan S? |
| Reproducibility/ replication projects |
|
Abbreviations: NIH, National Institutes of Health; NSF, National Science Foundation.
Challenges and insights from final team projects (3–5 students per team)
| Title | Challenges | Insights |
|---|---|---|
| TCGABiolinks: An R‐based, Open Source Tool for Genomic Analysis of Published TCGA data |
Updates to TCGABiolinks were not backward compatible. Initial release of software was in 2015. Modifications were required that prevented exact replication. |
Exciting to see how much data set has grown since 2019 publication. Exciting to be able to replicate findings. Relatively stress‐free experience because of excellent documentation. |
| Meta‐analysis of antidepressant efficacy |
Study transparency was overall quite good. Data set was available on‐line and well‐documented. | Challenged by calculation of metrics (e.g., credible interval) |
| Association of electronic cigarette use with subsequent initiation of tobacco cigarettes in US youths |
Figuring out what data were used. Inability to replicate the sample because variables to define inclusion/exclusion criteria were not available in the public data set. Figuring out what weights were used. Lack of detail prevented ability to replicate the recoding. |
Publicly available data sets may lack PHI needed to replicate samples. Independent studies using data from national studies may not publish their own data extract. Replication was impossible. |
| Re‐examination of data: EGFR as receptor of interest on monocytes, causal determination of HCMV on EGFR |
No raw images were included in the omics di repository. Authors made data available, but files were too large to process in R Studio; work arounds identified, but package no longer available with latest version of R. Details provided about wet laboratory procedures certain biological descriptions were ambiguous, but nothing about the data cleaning, missing data, statistical techniques used, and testing of assumptions. No response to emails sent to the authors for more information. |
Data access issues and technical challenges were surprising (backward compatibility). Evidence of image compression artifacts, value inversions, narrow cropping; such issues may be a pervasive issue in biological sciences. Need to include data for all components of a study with user‐friendly documentation. The importance of sharing scripts for data cleaning and statistical practices. |
Abbreviations: HCMV, human cytomegalovirus; PHI, protected health information; TCGA, The Cancer Genome Atlas.
Student focus group and instructor interview findings
| Category | Strengths | Suggestions |
|---|---|---|
|
| ||
| Overall |
Dedicated time to think about and discuss rigor and reproducibility, learning from others’ experiences, opportunity to meet new people; cohort effect a “huge bonus.” Course organization was effective: from overview to different components “and each time coming around with how can we do this better?” “For people who do not know much about reproducibility, in 3 months, I thought it was incredible.” |
Review course to reduce repetition: “Sometimes it got a little repetitive… we started doing similar things as we got closer to the end of the class.” Timing: (1) Offer in the Fall semester as an introduction to research training, (2) stretch the class out over the course of an academic year to integrate with other teaching, (3) run concurrently with a laboratory rotation. Have a teaching assistant to develop summaries of preparation work for each class, update the website, and assist with final projects. |
| Content |
Reviewed general concepts and provided a method of thinking; “a different way… [to] look at things like open science and transparency”; Important principles were covered relevant to future careers; “keep them all.” Addressed NIH expectations; Focused on practical tools/available resources; best practices for maintaining laboratory notebooks was helpful/useful; Cochrane Guidelines session was valuable; “I enjoyed the topics that we were taught,” Provided concrete examples of good and bad science. |
Delve deeper into what makes good research—like how to set up an RCT, how to make figures attractive in a paper or abstract, or graphical abstracts. Focus more on best practices. Include more good and bad examples. Include more on bioinformatics and database research. More analysis of mistakes/misconduct of others. |
| Lectures | Lectures were short and to the point (first lecture was most helpful/effective); defined and clarified terms, explained concepts; “Didn’t really dive too deeply into the weeds”; Image falsification/analysis lecture was particularly interesting; Chat box for questions worked well. | Reduce time in lectures to the bare minimum; make them more interactive (e.g., quiz format); lectures sometimes “blended together”; provide “coming attractions” for next class—stress essential preparation (for in‐class exercises); consistently explain concepts and then show an example; add guest speakers with expertise in the area. |
| Small group exercises | Exercises effectively applied concepts from the lecture; interactive in nature; evaluating and critiquing specific papers was valuable; class presentations allowed for peer teaching; appreciated the opportunity to learn from peers through presentations and discussions; effectively promoted engagement (everyone had a say on a topic). | Discussion time was sometimes too short; devote more time to small group work; first exercise was on an unfamiliar topic (acupuncture); replace with more familiar content; taking a study and formulating a replication plan was too much for a short in‐class exercise; redesign to make it more feasible; group reports could be repetitive when each group was tasked with the same thing. |
| Preparatory assignments |
Good to have a mix of assignment types—engaging. Video assignments were valuable “it’s a nice break and really good to just let it soak in.” Podcasts were a welcomed alternative to articles; “appreciate a more entertainment accessible source material.” |
Reduce/consolidate the number of readings “We just don’t have the time to do it.”; for webinars, to be able to speed up the playback or have transcript; clarify purpose of each reading; provide a distilled summary for prep work; add readings that reflect clinical relevance. Provide in this format: (1) summary document of all key points; (2) essential preparation (discussed in class); (3) required preparation; (4) suggested/ recommended preparation; (5) additional resources. |
| Quizzes | Quick and not a burden; “pretty straight‐forward”; helpful to get a “gist” of the most important take‐aways; seemed helpful to instructors to know what students absorbed. | “Sometimes the quizzes were a bit of a head scratcher”—make them more relevant to key objectives and reinforce main points; build in reminders to ensure completion of quizzes. |
| Final project | A valuable exercise and longitudinal experience; different options were available for type of project meeting different student needs; effectively synthesized learning; led to impressive efforts and presentations by peers—“it seemed like we were experts in what we were doing.” | Clarify the goals of the final project; show an example; provide a more definitive guide on how to do the project; provide a “how to” manual for finding articles with an available dataset; have more frequent meetings with instructors for advice and guidance; add option for a proposal for a replication study instead. |
|
| ||
| Overall |
“I thought it was a phenomenal course.” “This is going to go better than you think, and so prepare for a really good product at the end and do something with it. Like, leverage what the students are producing—capture it in some way and… you can use some of these things on websites. You can use it for recruiting. There’s a lot of product that’s going to come out of this that you want to leverage and make some time to evaluate it and use it.” |
“I think all the [medical] students should get this [course]… you could debate whether the first year is the place to do it or later… every student should be exposed to it.” [Integrate breaking news] “Fortuitously, on the first day of class was the same day that a Nobel Laureate in chemistry withdrew an article from Science because it wasn’t replicable…” “Increase a little bit the amount of primary research material… for example, how journal policies have evolved… a deeper dive into the problem.” |
| Content | “The content was excellent… I wouldn’t get rid of any of the content… it touched on all the points that are relevant… from the ethical to the very technical.” | “One way to grab a medical student or a nursing student’s attention is to give examples that are clinically relevant… have some readings or some examples where the irreproducibility of the research resulted in an adverse clinical outcome.” |
| Lectures | Content was good and covered necessary topics. |
“Make the lectures a bit shorter”… add content and remove slides like Goals & Objectives.” “More information about quantitative aspects… particularly epidemiology.” “Bring in an outside speaker… who’s got more expertise… who’s a real expert in, for example, image manipulation.” “Invited speakers… people that wrote the stuff.” |
| Small group exercises | “Student engagement was very high… the ability to engage all of the students all of the time was a particular strength.” | After small‐group exercises and reporting by small groups: “Maybe something to tie it all together… has your mindset changed?... maybe just [add] like a bit of a summary or final thoughts… go back to the exact same [original] question… was your initial knee‐jerk reaction correct or not? What did you learn that now would make you think something was different?” |
| Preparatory | “One of the great successes is that they [first‐year medical students] didn’t really have to prepare—and there was no penalty. And they were 100% during that time… they get a lot out of it.” | Move some of the content of the lectures to preparatory work. |
| Final project | “The [final] project…that was outstanding…” “I thought it was the best part.” |
Use bioRxiv Final Project: “Give them the project with an eye toward, you’re going to publish this, or at least you are going to blog it… almost all they need to do is narrate their presentation… it would not be too much work if the goal was, okay, now you are going to post it.” |
Abbreviations: NIH, National Institute of Health; RCT, randomized controlled trial.
Self‐assessments of competencies, before and after rigor and reproducibility class
|
Before ( |
Before ( |
After ( | |
|---|---|---|---|
| 1. The origins of the reproducibility crisis. | 1.72 (1.32) | 2.36 (1.29) | 5.45 (1.29) |
| 2. Strategies for addressing the reproducibility crisis. | 2.67 (0.97) | 2.73 (0.90) | 5.82 (1.08) |
| 3. The NIH response to the reproducibility crisis. | 2.44 (1.25) | 2.45 (1.13) | 5.45 (1.51) |
| 4. The role and importance of rigor and reproducibility in NIH proposal writing and scientific review. | 3.78 (1.26) | 3.45 (1.04) | 6.09 (0.94) |
| 5. The importance of scientific premise in NIH proposal preparation. | 3.50 (1.15) | 3.54 (1.04) | 6.00 (0.94) |
| 6. Critically assess sample scientific premise statements. | 3.39 (1.50) | 3.45 (1.75) | 5.36 (1.63) |
| 7. The importance of rigorous experimental design and documentation for transparency. | 4.78 (1.17) | 4.73 (0.65) | 6.45 (5.20) |
| 8. The importance of including sex as a biological variable in research. | 4.50 (1.29) | 4.27 (1.10) | 6.50 (0.71) |
| 9. Bias and the sources of bias in the conduct of science. | 4.56 (1.04) | 4.72 (1.19) | 6.00 (1.00) |
| 10. Assessing bias using the Cochrane Collaboration’s tool for assessing risk of bias in randomized trials. | 1.28 (0.67) | 1.27 (0.65) | 4.64 (1.63) |
| 11. Developing a prospective experimental design that comports with appropriate guidelines. | 3.39 (1.09) | 3.55 (1.04) | 5.45 (1.29) |
| 12. Key elements to include in an authentication plan for an NIH grant application. | 1.72 (0.96) | 1.91 (1.04) | 4.73 (1.68) |
| 13. Quality practices important to basic biomedical research. | 4.06 (1.21) | 4.00 (1.00) | 6.18 (0.98) |
| 14. Implementation of quality practices for basic biological research. | 3.83 (1.42) | 3.73 (1.27) | 5.82 (0.98) |
| 15. Evaluation of image data to determine whether unacceptable manipulation has occurred. | 2.44 (1.62) | 2.82 (1.78) | 5.64 (1.12) |
| 16. Software tools used to inspect images for manipulation. | 1.89 (1.13) | 2.18 (1.25) | 4.82 (1.25) |
| 17. Evaluating adherence to transparent reporting publication guidelines. | 2.39 (1.04) | 2.45 (1.13) | 5.40 (1.17) |
| 18. The role of laboratory notebooks in promoting rigor and reproducibility and transparency. | 4.56 (1.58) | 4.82 (1.60) | 6.45 (0.93) |
| 19. The roles of the data management plan, metadata, and data dictionary in promoting reproducibility and transparency. | 3.56 (1.95) | 3.72 (2.10) | 5.73 (1.42) |
| 20. Challenges and benefits of increased scientific transparency. | 3.94 (1.43) | 4.09 (1.45) | 6.18 (0.75) |
| 21. Critically assessing practices in your laboratory and consider possible steps toward increased transparency. | 3.56 (1.82) | 3.45 (1.69) | 5.82 (0.75) |
| 22. “Open Science” and its overall goals. | 2.94 (1.55) | 2.91 (1.64) | 6.18 (0.98) |
| 23. The challenges to the implementation of Open Science. | 2.39 (1.20) | 2.45 (1.21) | 6.09 (1.04) |
| 24. Identifying changes to current practices that promote Open Science. | 2.67 (1.41) | 2.73 (1.49) | 5.73 (0.79) |
| 25. Institutional changes that promote rigor and reproducibility. | 3.17 (1.20) | 3.55 (0.93) | 5.64 (0.81) |
Abbreviation: NIH, National Institutes of Health.
Students ranked each item on a scale where 1 = know nothing, 2 = very basic understanding, 3 = low/moderate understanding, 4 = moderate understanding, 5 = high/moderate understanding, 6 = strong understanding, and 7 = highly competent.
All students completed the assessment before class, 11 students completed the post‐assessment. All paired t‐tests were <0.05 for the all students and MD/PhD students (n = 7), pre‐post scores were not statistically different for Clinical and Translational Research Pathway students (n = 4) 5, 6, 9, 10, 11, 12, 16, 17, 19, 22, and 25.