| Literature DB >> 34268467 |
Evanthia Kaimaklioti Samota1,2, Robert P Davey1.
Abstract
We constructed a survey to understand how authors and scientists view the issues around reproducibility, focusing on interactive elements such as interactive figures embedded within online publications, as a solution for enabling the reproducibility of experiments. We report the views of 251 researchers, comprising authors who have published in eLIFE Sciences, and those who work at the Norwich Biosciences Institutes (NBI). The survey also outlines to what extent researchers are occupied with reproducing experiments themselves. Currently, there is an increasing range of tools that attempt to address the production of reproducible research by making code, data, and analyses available to the community for reuse. We wanted to collect information about attitudes around the consumer end of the spectrum, where life scientists interact with research outputs to interpret scientific results. Static plots and figures within articles are a central part of this interpretation, and therefore we asked respondents to consider various features for an interactive figure within a research article that would allow them to better understand and reproduce a published analysis. The majority (91%) of respondents reported that when authors describe their research methodology (methods and analyses) in detail, published research can become more reproducible. The respondents believe that having interactive figures in published papers is a beneficial element to themselves, the papers they read as well as to their readers. Whilst interactive figures are one potential solution for consuming the results of research more effectively to enable reproducibility, we also review the equally pressing technical and cultural demands on researchers that need to be addressed to achieve greater success in reproducibility in the life sciences.Entities:
Keywords: interactive figures; open science; replication of experiments; reproducibility; reproducibility in life sciences; reproducibility metrics; reproducibility of computational experiments; reproducibility survey in life sciences
Year: 2021 PMID: 34268467 PMCID: PMC8276979 DOI: 10.3389/frma.2021.678554
Source DB: PubMed Journal: Front Res Metr Anal ISSN: 2504-0537
Questions used to survey the knowledge of respondents about research reproducibility.
| Survey questions | |
| 1 | How often do you encounter difficulties with working with bioinformatic analysis tools (that are not your own)? (Problems such as: installing, configuring, running the software, working with command-line software)? |
| 2 | How difficult is it to source (or access) the data presented in published papers? |
| 3 | What difficulties have you encountered in accessing the data described in published papers? |
| 4 | How are you currently sourcing the data (if applicable)? Select all that apply to you. |
| 5 | What is your current understanding of the reproducibility of experiments? Please select any that apply. Should you wish to add any additional information, please add it to the “Other” box. |
| 6 | Have you ever tried reproducing any published results? Please select the answer that applies best for you. |
| 7 | In your opinion, what could be done to make published research more reproducible? Other please specify (free text answer). |
| 8 | When thinking about interactive figures, what comes to your mind? (please describe what you understand of what an interactive figure to be, its features, and where you have seen such a feature before if applicable). |
| 9 | An interactive figure is a figure within a paper that is dynamic and becomes “live” when the user interacts with it and where the data displayed changes according to various parameter options. Which of the following features of an interactive figure tool would be good to have? Please rank them in the order of preference, where 1 is the most preferred feature, and 11 the least preferred feature. |
| 10 | What other features an interactive figure could have that were not mentioned in the previous question? |
| 11 | Do you perceive a benefit in being able to publish interactive figures? |
| 12 | Does the provision or option of an interactive figure in the paper affect your decision in choosing the publishing journal or publisher? |
| 13 | Have you heard of the term computationally reproducible data, and do you understand what the term means? If answered yes or unsure, please explain what you understand from the term. |
| 14 | Would you benefit from being able to automatically reproduce computational experiments or other analyses (including statistical tests) described within a paper? |
| 15 | How often do you work with bioinformatic analysis tools (e.g., assemblers, aligners, structure modeling)? |
| 16 | Have you received any of the following training? Training whether formal or informal (training through a colleague etc.). |
| 17 | Which of the following type(s) of data do you work with? |
aQuestions indicated with an asterisk were only available to the eLIFE survey. Answer options to the questions are shown in Supplementary section 1.
FIGURE 1Data types used by NBI and eLIFE respondents. Responses were not mutually exclusive. Data type choices were the same as the article data types available in the eLIFE article categorisation system.
FIGURE 2(A): Difficulty encountered accessing data underlying published research. Whether respondents have attempted to access data underlying previous publications and the level of difficulty typically encountered in doing so. (B): Reasons given for difficulty accessing data. The reasons given by respondents for being unable to access data (restricted to those who have attempted to access data).
Success in reproducing any published results stratified by their knowledge of the term “computationally reproducible data” and training received.
| Number (% of the total sample) | Success in reproducing any published results | |||
|---|---|---|---|---|
| Variable | Successful (% within variable) | Not Successful |
| |
|
| ||||
| Yes | 25 (33.3) | 18 (72) | 7 (28) | 0.048 |
| No | 50 (66.7) | 24 (48) | 26 (33) | |
|
| ||||
| Bioinformatics | 42 (46.7) | 26 (61.9) | 16 (38.1) | 0.73 |
| Not trained in Bioinformatics | 48 (53.3) | 28 (58.3) | 20 (41.7) | |
| Computer Science | 33 (36.7) | 21 (63.6) | 12 (36.4) | 0.59 |
| Not trained in Computer Science | 57 (63.3) | 33 (57.9) | 24 (42.1) | |
| Statistics | 71 (78.9) | 42 (59.2) | 29 (40.8) | 0.75 |
| Not trained in Statistics | 19 (21.1) | 12 (63.2) | 7 (36.8) | |
| No training | 10 (11.1) | 6 (60) | 4 (40) | 0.73 |
| All other training | 80 (88.8) | 48 (60) | 32 (40) | |
n is different for the two variables as not all participants answered all the questions.
Unsuccessful includes answers: “Yes, I have tried reproducing published results, but I have been unsuccessful in producing any results, or the same results”, “No, I have never tried reproducing any published results” and “It is not important to reproduce other people’s published results”.
Statistically significant at the level of p < 0.05.
Chi-square statistic with Yates correction, applied when expected frequencies were lower than five.
FIGURE 3Preferred features for the interactive figure. Responses to question 9: Respondents were asked to rank in order of preference the above features, with one most preferred feature, to 11 the least preferred feature. The average score for each feature was calculated in order of preference as selected by the respondents from both NBI and eLIFE surveys. The lower the average score value (x-axis), the more preferred the feature (y-axis).
FIGURE 4The level of perception of benefit to having the ability to publish papers with interactive figures. The benefit to the author, to the readers of the author’s papers and to the papers the author reads. Answers include the responses from both NBI and eLIFE surveys for question 11.
FIGURE 5Assessment of perceived benefit for automatically reproducing computational experiments or other analyses (including statistical tests). Responses from both NBI and eLIFE for question 14.