| Literature DB >> 32434117 |
Laura Dishaw1, Erin Yost2, Xabier Arzuaga3, April Luke4, Andrew Kraft3, Teneille Walker3, Kris Thayer2.
Abstract
A key aspect of the systematic review process is study evaluation to understand the strengths and weaknesses of individual studies included in the review. The present manuscript describes the process currently being used by the Environmental Protection Agency's (EPA) Integrated Risk Information System (IRIS) Program to evaluate animal toxicity studies, illustrated by application to the recent systematic reviews of two phthalates: diisobutyl phthalate (DIBP) and diethyl phthalate (DEP). The IRIS Program uses a domain-based approach that was developed after careful consideration of tools used by others to evaluate experimental animal studies in toxicology and pre-clinical research. Standard practice is to have studies evaluated by at least two independent reviewers for aspects related to reporting quality, risk of bias/internal validity (e.g., randomization, blinding at outcome assessment, methods used to expose animals and assess outcomes, etc.), and sensitivity to identify factors that may limit the ability of a study to detect a true effect. To promote consistency across raters, prompting considerations and example responses are provided to reviewers, and a pilot phase is conducted. The evaluation process is performed separately for each outcome reported in a study, as the utility of a study may vary for different outcomes. Input from subject matter experts is used to identify chemical- and outcome-specific considerations (e.g., lifestage of exposure and outcome assessment when considering reproductive effects) to guide judgments within particular evaluation domains. For each evaluation domain, reviewers reach a consensus on a rating of Good, Adequate, Deficient, or Critically Deficient. These individual domain ratings are then used to determine the overall confidence in the study (High Confidence, Medium Confidence, Low Confidence, or Deficient). Study evaluation results, including the justifications for reviewer judgements, are documented and made publicly available in EPA's version of Health Assessment Workspace Collaborative (HAWC), a free and open source web-based software application. (The views expressed are those of the authors and do not necessarily represent the views or policies of the US EPA).Entities:
Keywords: Phthalates; Study evaluation; Systematic review
Mesh:
Substances:
Year: 2020 PMID: 32434117 PMCID: PMC8422842 DOI: 10.1016/j.envint.2020.105736
Source DB: PubMed Journal: Environ Int ISSN: 0160-4120 Impact factor: 9.621
Fig. 1.Overview of the study evaluation approach. (a) An overview of the evaluation process. (b) Individual evaluation domains for animal studies. Definitions for the (c) domain and (d) overall confidence ratings.
Study evaluation domains with core and prompting questions.
| Key concept | Domain - Core Question | Prompting questions |
|---|---|---|
| Reporting Quality | Does the study report the following? Critical information necessary to perform study evaluation: Species; test article name; levels and duration of exposure; route (e.g., oral; inhalation); qualitative or quantitative results for at least one outcome of interest Important information for evaluating the study methods: Test animal: strain, sex, source, and general husbandry procedures Exposure methods: source, purity, method of administration Experimental design: frequency of exposure, animal age and lifestage during exposure and at outcome evaluation Outcome evaluation methods: assays or procedures used to measure the outcomes of interest | |
| Risk of Bias | For each study: Did each animal or litter have an equal chance of being assigned to any experimental group (i.e., random allocation)? Is the allocation method described? Aside from randomization, were any steps taken to balance variables across experimental groups during allocation? | |
| For each outcome or grouping of outcomes in a study: Does the study report blinding or other methods/procedures for reducing observational bias? If not, did the study use a design or approach for which such procedures can be inferred? What is the expected impact of failure to implement (or report implementation) of these methods/procedures on results? | ||
| For each study: Are there differences across the treatment groups (e.g., co-exposures, vehicle, diet, palatability, husbandry, health status, etc.) that could bias the results? If differences are identified, to what extent are they expected to impact the results? | ||
| For each study: Are results presented for all outcomes described in the methods? Are all animals accounted for in the results? If there are discrepancies, do authors provide an explanation (e.g., death or unscheduled sacrifice during the study)? If results omissions and/or attrition are identified, what is the expected impact on the interpretation of the results? | ||
| Sensitivity | For each study: Does the study report the source and purity and/or composition (e.g., identity and percent distribution of different isomers) of the chemical? If not, can the purity and/or composition be obtained from the supplier (e.g., as reported on the website) Was independent analytical verification of the test article purity and composition performed? Did the authors take steps to ensure the reported exposure levels were accurate? For inhalation studies: were target concentrations confirmed using reliable analytical measurements in chamber air? For oral studies: if necessary, based on consideration of chemical-specific knowledge (e.g., instability in solution; volatility) and/or exposure design (e.g., the frequency and duration of exposure), were chemical concentrations in the dosing solutions or diet analytically confirmed? Are there concerns about the methods used to administer the chemical (e.g., inhalation chamber type, gavage volume, etc.)? | |
| For each outcome or grouping of outcomes in a study: Does the exposure period include the critical window of sensitivity (if known)? Was the duration and frequency of exposure sensitive for detecting the outcome of interest? | ||
| For each outcome or grouping of outcomes in a study: Are there concerns regarding the specificity and validity of the protocols? Are there serious concerns regarding the sample size? Are there concerns regarding the timing of the outcome assessment? | ||
| For each outcome or grouping of outcomes in a study: Does the level of detail allow for an informed interpretation of the results? Are the data compared or presented in a way that is inappropriate or misleading? | ||
| Overall Confidence | For each outcome or grouping of outcomes in a study: Were concerns (i.e., limitations or uncertainties) related to the reporting quality, risk of bias, or sensitivity identified? If yes, what is their expected impact on the overall reliability and validity of the study results, including (when possible) interpretations of impacts on the magnitude or direction of the reported effects? |
Fig. 2.Summary of study evaluation ratings by domain for (A) DIBP and (B) DEP. The ratings presented in for each study in these heat maps represent the composite ratings across outcomes within each study evaluation domain. There were instances where certain outcomes were rated differently within the same study because of outcome specific considerations (see individual study evaluations in HAWC for details). Links to interactive graphics in HAWC can be found at the following URLs for DIBP (https://hawcprd.epa.gov/summary/visual/100500053/) and DEP (https://hawcprd.epa.gov/summary/visual/100000097/). ++ = Good/High Confidence; + = Adequate/Medium Confidence, − = Deficient/Low Confidence; – = Critically Deficient/Uninformative; NR = Not Reported.