| Literature DB >> 33976964 |
Sheeba Samuel1,2, Birgitta König-Ries1,2.
Abstract
Scientific experiments and research practices vary across disciplines. The research practices followed by scientists in each domain play an essential role in the understandability and reproducibility of results. The "Reproducibility Crisis", where researchers find difficulty in reproducing published results, is currently faced by several disciplines. To understand the underlying problem in the context of the reproducibility crisis, it is important to first know the different research practices followed in their domain and the factors that hinder reproducibility. We performed an exploratory study by conducting a survey addressed to researchers representing a range of disciplines to understand scientific experiments and research practices for reproducibility. The survey findings identify a reproducibility crisis and a strong need for sharing data, code, methods, steps, and negative and positive results. Insufficient metadata, lack of publicly available data, and incomplete information in study methods are considered to be the main reasons for poor reproducibility. The survey results also address a wide number of research questions on the reproducibility of scientific results. Based on the results of our explorative study and supported by the existing published literature, we offer general recommendations that could help the scientific community to understand, reproduce, and reuse experimental data and results in the research data lifecycle. ©2021 Samuel et al.Entities:
Keywords: Experiments; FAIR data principles; Reproducibility; Reproducibility crisis; Reproducible research recommendations; Research data lifecycle; Reuse; Understandability
Year: 2021 PMID: 33976964 PMCID: PMC8067906 DOI: 10.7717/peerj.11140
Source DB: PubMed Journal: PeerJ ISSN: 2167-8359 Impact factor: 2.984
The current position of the participants at the time of answering the survey.
| PhD student | 27 |
| PostDoc | 18 |
| Professor | 13 |
| Data manager | 8 |
| Research associate | 7 |
| Student | 5 |
| Junior professor | 4 |
| Lecturer | 1 |
| Technical assistant | 1 |
| Other | 17 |
The primary area of study of the survey participants.
| Computer science | 19 |
| Biology(other) | 17 |
| Environmental sciences | 13 |
| Molecular biology | 6 |
| Neuroscience | 6 |
| Physics | 4 |
| Plant sciences | 3 |
| Health sciences | 3 |
| Cell biology | 2 |
| MicroBiology | 1 |
| Chemistry | 1 |
| Other | 26 |
Summary of survey questions.
| Informed Consent Form (Datenschutzerklärung in German) | Background, purpose, and procedure of study |
| Informed consent | |
| Research context of the participant | Current position |
| Primary area of study | |
| Reproducibility | Reproducibility crisis in your field of research |
| Factors leading to poor reproducibility | |
| Measures taken in different fields to ensure reproducibility of results | Discovery of own project data |
| Discovery of project data for a newcomer | |
| Unable to reproduce published results of others | |
| Contacted for having problems in reproducing results | |
| Repetition of experiments to reproduce results | |
| Important factors to understand a scientific experiment to enable reproducibility | Experimental data |
| Experimental requirements | |
| Experimental settings | |
| Names and contacts of people | |
| Spatial and temporal metadata | |
| Software | |
| Steps and plans | |
| Intermediate and final results | |
| Opinion on sharing other metadata | |
| Experiment Workflow/Research Practices | Kind of data primarily worked with |
| Storage of experimental data | |
| Storage of metadata | |
| Usage of scripts | |
| Knowledge of FAIR principles | |
| Implementation of FAIR principles in research | |
| Opinion on enabling reproducibility in their field |
Figure 1The factors leading to poor reproducibility from the experience of 71 participants who fully responded to this question.
How easy would it be for you vs a newcomer to find all the experimental data related to your own project in order to reproduce the results at a later point in time (e.g. 6 months after the original experiment)?
| Findability of own data at a later point in time | Findability of own data by a newcomer | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| VE | E | NEND | D | VD | VE | E | NEND | D | VD | |
| Input data | 29.6% | 40.7% | 18.5% | 8.6% | 2.5% | 8.3% | 34.5% | 22.6% | 23.8% | 10.7% |
| Metadata about the methods | 19.8% | 39.5% | 32.1% | 7.4% | 1.2% | 1.2% | 22.6% | 40.5% | 27.4% | 8.3% |
| Metadata about the steps | 14.8% | 32.1% | 35.8% | 13.6% | 3.7% | 1.2% | 19.0% | 32.1% | 36.9% | 10.7% |
| Metadata about the setup | 15.6% | 31.2% | 37.7% | 14.3% | 1.3% | 3.6% | 19.0% | 29.8% | 36.9% | 10.7% |
| Results | 42.0% | 37.0% | 18.5% | 1.2% | 1.2% | 8.3% | 40.5% | 27.4% | 13.1% | 10.7% |
Notes.
Very Easy
Easy
Neither easy nor difficult
Difficult
Very Difficult
Figure 2Does your research follow the FAIR (Findable, Accessible, Interoperable, Reusable) principles?