| Literature DB >> 35846467 |
A Amina Wilkins1, Paul Whaley2,3, Amanda S Persad1, Ingrid L Druwe1, Janice S Lee1, Michele M Taylor1, Andrew J Shapiro1, Natalie Blanton Southard4, Courtney Lemeris4, Kristina A Thayer1.
Abstract
Background: Environmental health and other researchers can benefit from automated or semi-automated summaries of data within published studies as summarizing study methods and results is time and resource intensive. Automated summaries can be designed to identify and extract details of interest pertaining to the study design, population, testing agent/intervention, or outcome (etc.). Much of the data reported across existing publications lack unified structure, standardization and machine-readable formats or may be presented in complex tables which serve as barriers that impede the development of automated data extraction methodologies.As full automation of data extraction seems unlikely soon, encouraging investigators to submit structured summaries of methods and results in standardized formats with meta-data tagging of content may be of value during the publication process. This would produce machine-readable content to facilitate automated data extraction, establish sharable data repositories, help make research data FAIR, and could improve reporting quality.Entities:
Keywords: Author feedback; Author journal submission requirements; Author opinion; Author willingness; Automated data extraction; Data extraction; Data sharing; Data summary; Data templates; Manuscript submission; Natural language Processing (NLP); Science translation; Standardized data; Structured data; Study evaluation template; Systematic review
Year: 2022 PMID: 35846467 PMCID: PMC9280381 DOI: 10.1016/j.heliyon.2022.e09095
Source DB: PubMed Journal: Heliyon ISSN: 2405-8440
Figure 1Overview of IRIS Program study evaluation approach for animal toxicity studies.
Participant responses describing Pilot's impact on future research activities (from post-extraction survey).
| Impact Query | Impact | Participant Responses |
|---|---|---|
| What fields (if any) are you likely to change about the way you conduct, or report future research based on your experience with the reporting checklist used during this pilot? | Positive (71%) | “Ensuring that my published methods contain all of the elements on the checklist to minimize risk of bias and maximize study quality.” “I will be sure to report all blinding/random selection methodologies in future publications to increase transparency.” “Will think about making endpoint information as clear as possible.” “I am likely to provide more information about chemical purity validation, animals, husbandry practices, and to provide data in a format that can be extracted.” “Be mindful about title/abstract labels, methods details, about chemical purity and origins, animal husbandry details, exposure duration/type/time period, outcomes discussed in the methods are also reflected in the results/discussion or at least summarized in a table” |
| Neutral (29%) | “The chemical and animal assurance. More information on how the animals were assigned to a group/how bias was prevented” “entering the outcome and navigating the interface (e.g., editing and corrections)” | |
| Negative (0%) | • N/A | |
| What fields (if any) are you likely to change about the way you conduct or report research in the future based on your experience with the study quality (internal validity) assessment tool used during this pilot? | Positive (83%) | “Being specific about how data is blinded and randomized (instead of just stating that it was done)” “I will be sure to thoroughly elaborate on all control groups used in experiments, and why those controls were chosen.” “Will provide more details about randomization of animals when publishing or reporting data.” “I am likely to provide more detail about randomization and blinding during analysis” “For biases become more mindful of blinding and random sample assignments, reporting, confounding/variable controls, reporting biases that may occur, exposure methods sensitivity details” |
| Neutral (0%) | N/A | |
| Negative (17%) | “Correcting errors and performing edits to the report could be made easier and straightforward.” |
Reported after participant's 2nd extraction.
Easy and difficult data fields to extract.
| Query: Which data field(s) did you find easy to extract? Which field(s) of data extraction did you find problematic? | Participant Responses | Number of Responses |
|---|---|---|
| Data Summary: Participants provided examples of easy fields to extract which approximately equalled 65% (13 out of 20) of the total number of responses received for this query. | Dose regimen/exposure | 5 |
| Animal husbandry | 3 | |
| Animal group and experiment set up | 3 | |
| Outcome/endpoint | 2 | |
| Data Summary: Participants also provided details regarding fields that were difficult to extract which equalled 35% (7 out of 20) of the total number of responses received for this query. | Behavioral endpoint because there are many different types for each task/difficult to enter data by sex | 1 |
| Outcome | 2 | |
| Endpoint details: not sure how to enter data for system, organ effect and sub-type; not sure what frank effect means | 1 | |
| Reproductive end point entry | 1 | |
| Entering raw/dose response data was challenging [Note: participant had issues with entering | 1 | |
| Dosing regimen: not clear on how to adjust doses; guess it may be done at set up | 1 | |
Reflects one or more responses indicated by a total of 7 participants.
Summary of study participants and selected studies.
| Participant Number | Educational Status | Data Extraction | Study Evaluation | Task Completion Time | Post-Extraction & Post-Pilot Survey | Selected Study and Extracted Endpoints |
|---|---|---|---|---|---|---|
| 1 | Associate dean | X | X | X | X | Neurobehavioral – male rat; open field test for anxiety related behaviors; oral exposure to Bisphenol A (BPA) (Study 1) |
| 2 | Graduate student | X | X | X | X | Neuromuscular – male and female rat; escape latency to visible platform, swimming speed; nicotine exposure via subcutaneous injection (Study 2) |
| 3 | Professor | X | X | X | X | Developmental/Neurobehavioral – adult male and female zebrafish; learning and memory (shock avoidance); exposed to benzo [a]pyrene via water habitat (Study 3) |
| 4 | Professor | X | X | Developmental/Neurobehavioral -weanling, male rat; Morris water maze; spatial learning and memory; lactational exposure from maternal dietary exposure to PCBs (Study 4) | ||
| 5 | Associate dean; Professor | X | X | X | X | Systemic (whole body) – male mouse; body weight; 14-day, oral exposure to valproic acid (Study 5) |
| 6 | Graduate student | X | X | X | X | Systemic (whole body)– male rats; oxidative stress (as measured by malondialdehyde (MDA) and hepatoxicity as measured by histopathology) via subchronic, oral exposure to arsenic (Study 6) |
| 7 | Graduate student | X | X | X | X | Systemic (whole body) – maternal and fetal weight change and placental markers of oxidative stress from maternal dietary TCE exposure and fetal placental exposure (Study 7)) |
| 8 | Graduate student | X | X | X | X | Developmental - Female reproduction and fertility: histopathology; body, organ, and ovary weights; prenatal exposure to benzo [a]pyrene and inorganic lead (Study 8) |
Participant performed two extractions (responses from the first extraction were used in the pilot results).
Willingness to consider structured data entry during publication process.
| Please use the scale below to rate your agreement with the statement (n = 7) | I am open to using structured data entry when submitting articles for publication. | Given the potential application of structured data entry, approaches to implement this during the journal submission process should continue to be explored. | My consideration of using structured data entry depends on the type of article being submitted for publication. |
|---|---|---|---|
| Strongly Agree | 43% (n = 3) | 86% (n = 6) | 43% (n = 3) |
| Moderately Agree | 43% (n = 3) | 14% (n = 1) | 14% (n = 1) |
| Agree | 14% (n = 1) | 0% (n = 0) | 43% (n = 3) |
| Neutral | 0% (n = 0) | 0% (n = 0) | 0% (n = 0) |
| Disagree | 0% (n = 0) | 0% (n = 0) | 0% (n = 0) |
| Moderately Disagree | 0% (n = 0) | 0% (n = 0) | 0% (n = 0) |
| Strongly Disagree | 0% (n = 0) | 0% (n = 0) | 0% (n = 0) |
Participant self-reported time spent performing various pilot tasks.
| Participants' Response to Time Spent Performing These Various Tasks (in minutes) | n | Average (Min) | Range (Min) |
|---|---|---|---|
| Orientation: account creation, becoming acquainted with the HAWC software, watching tutorial videos, etc. | 8 | 65 | 10–120 |
| Data Extraction: summarizing study methods and results into HAWC. | 8 | 83 | 30–120 |
| Receiving Technical Support: assistance required in addition to orientation and team provided written instructions (i.e., help with tool navigation, and rating study and reporting quality, etc. | 7 | 49 | 0–180 |
| Study Evaluation: determining risk of bias | 7 | 56 | 30–90 |
| Reporting Quality: rating how well results were reported and recorded within the study | 7 | 51 | 30–90 |
Summary of Technical Assistance Provided by Team (throughout duration of pilot).
| Summary of Team Provided Technical Assistance | n | Average Time in Minutes (Hours) | Range in Minutes (Hours) |
|---|---|---|---|
| Getting started | 8 | 47 min (0.78 h) | 15–60 min (0.25–1.0 h) |
| Data extraction assistance | 4 | 38 min (0.63 h) | 30–60 min (0.5–1.0 h) |
| Study evaluation assistance | 2 | 30 min (0.50 h) | 30–30 min (0.5–0.5 h) |
| Technical Assistance (provided after the ‘getting started’ phase) | 1 | 60 min (1 h) | 60–60 (1.0–1.0 h) |
Initial survey Likert scale results (from post-extraction survey).
| Question/Response (n = 7) | Easy | Relatively Easy | Neutral | Relatively Difficult | Difficult |
|---|---|---|---|---|---|
| Rate your experience entering study data | 29% (n = 2) | 43% (n = 3) | 14% (n = 1) | 14% (n = 1) | 0% (n = 0) |
| Rate your experience performing a study quality evaluation | 43% (n = 3) | 43% (n = 3) | 14% (n = 1) | 0% (n = 0) | 0% (n = 0) |
| Question/Response (n = 7) | Strongly Agree | Somewhat Agree | Neutral | Somewhat Disagree | Strongly Disagree |
| Completing the reporting quality checklist is likely to impact my future research activities | 57% (n = 4) | 29% (n = 2) | 0% (n = 0) | 0% (n = 0) | 14% (n = 1) |
| Completing the study quality evaluation is likely to impact my future research activities | 57% (n = 4) | 14% (n = 1) | 14% (n = 1) | 0% (n = 0) | 14% (n = 1) |
| I feel comfortable with applying the quality checklist and study quality tool to my other research | 57% (n = 4) | 29% (n = 2) | 14% (n = 1) | 0% (n = 0) | 0% (n = 0) |
Figure 2Participants Aggregated Study Evaluation Summary. ∗ Participant did not extract data from their own study. This heatmap can also be accessed from the following link: https://hawc.epa.gov/summary/visual/assessment/100500015/Table-7-Participant-Aggregated-Study-Evaluation-Su/ and (Supplemental Materials: Supplemental Excel File Tab 1, Figure 5).