Literature DB >> 29026002

Risk of bias reporting in the recent animal focal cerebral ischaemia literature.

Zsanett Bahor1, Jing Liao1, Malcolm R Macleod2, Alexandra Bannach-Brown1, Sarah K McCann1, Kimberley E Wever3, James Thomas4, Thomas Ottavi5, David W Howells5, Andrew Rice6, Sophia Ananiadou7, Emily Sena1.   

Abstract

BACKGROUND: Findings from in vivo research may be less reliable where studies do not report measures to reduce risks of bias. The experimental stroke community has been at the forefront of implementing changes to improve reporting, but it is not known whether these efforts are associated with continuous improvements. Our aims here were firstly to validate an automated tool to assess risks of bias in published works, and secondly to assess the reporting of measures taken to reduce the risk of bias within recent literature for two experimental models of stroke.
METHODS: We developed and used text analytic approaches to automatically ascertain reporting of measures to reduce risk of bias from full-text articles describing animal experiments inducing middle cerebral artery occlusion (MCAO) or modelling lacunar stroke.
RESULTS: Compared with previous assessments, there were improvements in the reporting of measures taken to reduce risks of bias in the MCAO literature but not in the lacunar stroke literature. Accuracy of automated annotation of risk of bias in the MCAO literature was 86% (randomization), 94% (blinding) and 100% (sample size calculation); and in the lacunar stroke literature accuracy was 67% (randomization), 91% (blinding) and 96% (sample size calculation). DISCUSSION: There remains substantial opportunity for improvement in the reporting of animal research modelling stroke, particularly in the lacunar stroke literature. Further, automated tools perform sufficiently well to identify whether studies report blinded assessment of outcome, but improvements are required in the tools to ascertain whether randomization and a sample size calculation were reported.
© 2017 The Author(s).

Entities:  

Keywords:  animal models; lacunar; middle cerebral artery occlusion; research improvement; risks of bias; text mining

Mesh:

Year:  2017        PMID: 29026002      PMCID: PMC5869854          DOI: 10.1042/CS20160722

Source DB:  PubMed          Journal:  Clin Sci (Lond)        ISSN: 0143-5221            Impact factor:   6.124


Introduction

There has been much recent interest in strategies to improve the usefulness of findings from biomedical research [1]. This has been occasioned by a growing realization that many ‘findings’ in the published literature cannot be replicated, either in single replication attempts [2,3] or in formal replication studies ([4], https://osf.io/e81xl/wiki/home/). There are a number of potential reasons for this, including low positive predictive values of the originator study, publication bias, heterogeneity of treatment effects due to the variable presence of unknown modifying variables, overreliance on tests of statistical rather than biological significance and flexibility in study protocols and statistical analysis plans. Notwithstanding these possibilities, it may be that findings reported from laboratory research are overstated because individual studies are at risk of bias, leading to an overstatement of treatment effects. Across the animal modelling of a range of neurological conditions it is apparent that studies which do not report, for instance, randomization of animals to group or the blinded assessment of outcome, give inflated estimates of treatment effect [5]. Identifying the true prevalence of measures to reduce the risk of bias is not straightforward because it may be that scientists have taken account of these measures but not, for whatever reason, reported this in their manuscript. Recognition of this has led to the development of reporting standards for in vivo research including the ARRIVE guidelines [6] and the Landis checklist [7]. The in vivo stroke research community have been amongst the first to adopt best practice guidelines for the conduct and reporting of animal studies. Following the demonstration that the reporting of risks of bias in in vivo stroke research was low [8], Good Laboratory Practice guidelines for in vivo stroke research were published in 2009 [9] and were adopted as editorial policy by a number of Journals including ‘Stroke’. Subsequently, Minnerup et al. [10] have shown a substantial increase in the prevalence of reported randomization of group allocation, blinded conduct of the experiment, blinded assessment of outcome and sample size calculations in the years from 2010 to 2013. Recent research suggests that improvement may been limited to specific journals reporting predominantly stroke research [11]. In 2014, Pedder et al. [12] reported a systematic review of animal studies in lacunar stroke published between 1992 and 2011, where the reporting of measures to reduce risks of bias was in general better than had been observed in the middle cerebral artery occlusion (MCAO) literature in 2007 [8]. It is not known, however whether the quality of reporting in lacunar stroke has improved since that time Figures 1–3.
Figure 1

Reporting of risks of bias in the recent MCAO literature.

Reporting of (A) blinding, (B) randomization and (C) sample size calculation in the middle cerebral artery occlusion literature in 6-month epochs to October 2016. Values represent mean and 95% Clopper–Pearson exact confidence intervals.

Figure 3

Reporting of risks of bias in the recent literature compared with previous estimates.

(A) Reporting of blinding, randomization and sample size calculation in the MCAO literature collated from systematic reviews and summarized in 2007 in Sena et al. [8] (column 1), from the journal Stroke from 2010 to 2013 reported by Minnerup et al. [10] (column 2), and from 2014 to 2016 reported here (column 3). (B) Reporting of blinding, randomization and sample size calculation in the experimental lacunar stroke literature from the systematic review of 2012 [12] (column 1), and from 2012 to 2017 reported here (column 2).

Reporting of risks of bias in the recent MCAO literature.

Reporting of (A) blinding, (B) randomization and (C) sample size calculation in the middle cerebral artery occlusion literature in 6-month epochs to October 2016. Values represent mean and 95% Clopper–Pearson exact confidence intervals.

Reporting of risks of bias by Journal.

Reporting of (A) blinding, (B) randomization and (C) sample size calculation in the middle cerebral artery occlusion literature in the 2 years to October 2016 by journal of publication for journals with five or more relevant manuscripts, ranked by performance. Values represent mean and 95% Clopper–Pearson exact confidence intervals. The grey shaded bar represents the 95% confidence interval of the overall performance across 918 manuscripts.

Reporting of risks of bias in the recent literature compared with previous estimates.

(A) Reporting of blinding, randomization and sample size calculation in the MCAO literature collated from systematic reviews and summarized in 2007 in Sena et al. [8] (column 1), from the journal Stroke from 2010 to 2013 reported by Minnerup et al. [10] (column 2), and from 2014 to 2016 reported here (column 3). (B) Reporting of blinding, randomization and sample size calculation in the experimental lacunar stroke literature from the systematic review of 2012 [12] (column 1), and from 2012 to 2017 reported here (column 2). Ascertainment of improvement in reporting might serve as an important outcome measure for efforts to improve performance at the level of a journal, funder, or institution, as part of a research improvement activity. However, one problem with this approach is that ascertainment of reporting of measures to reduce risks of bias in published work is a time-consuming process, requiring review of full-text by at least two independent assessors to minimize risk of human error and taking around 1 h for each manuscript. We have therefore set out to develop text mining tools to automatically ascertain the reporting of key measures to reduce risks of bias in published works describing animal research modelling of MCAO and lacunar stroke.

Methods

This development work was not guided by a study protocol. Criteria for successful performance of a text mining system were developed at a consensus meeting organized by the SLIM (Systematic Living Information Machine) consortium in April 2016 (prior to the commencement of this project), which recommended that a tool performing with sensitivity and specificity of at least 0.80 could make a valuable contribution.

Identification of relevant manuscripts

For MCAO studies, we searched PUBMED on 4 October 2016 with the terms OR , limited to animals, with a date of record completion later than 4 October 2014. For lacunar stroke studies, we searched PUBMED on 11 January 2017 using the terms employed in Pedder et al., limited to a deposition date between 6 March 2012 (the date of the search which informed the Pedder review) and 11 January 2017 for the terms ((< Biological Models> OR OR OR OR OR OR OR ) AND ( OR OR OR OR OR OR OR OR OR OR ) AND ( OR OR OR ) AND ( OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR OR

Development of ‘regular expressions’

In the context of a systematic review of studies describing animal models of psychosis, we collected phrases associated with the reporting of randomization, blinded assessment of outcome and sample size calculations and used these to develop target regular expressions (“RegEx”s). These were informally tested through application to a second dataset of 1173 full-text publications describing in vivo research where risks of bias had been previously extracted [13], and no changes were made. We then applied these tools to the converted full-text of stroke publications using the Grepl function in R, resulting in a call of ‘true’ or ‘false’ for each of the measures to reduce risk of bias. We then conducted manual (‘gold standard’) ascertainment of reporting of measures to reduce risk of bias for each publication. To compare the text analytic with human ‘gold standard’, we calculated sensitivity (effectiveness of the tool in identifying positives), specificity (effectiveness of the tool in identifying negatives) and accuracy (number of publications correctly labelled as a proportion of the total number of publications).The R code is available at GitHub (link provided on publication). We present the prevalence of reporting of randomization of animals to experimental groups, blinded assessment of outcome and sample size calculation, along with exact binomial confidence intervals (derived using cii in STATA) for each dataset and, for the MCAO literature, in quartiles of PUBMED accession date and by journal. Least squares linear regression was used to understand the correlation between time of publication and the reporting of measures taken to reduce the risk of bias.

Results

In 2259 studies reporting animal models of psychosis, the RegEx had performed above our pre-defined performance criteria (sensitivity of 0.80 and specificity of 0.91 for randomization, sensitivity of 0.83 and specificity of 0.95 for blinding, and sensitivity of 1.00 and specificity of 0.96 for sample size calculation). In a second dataset of in vivo studies from leading U.K. institutions [13], the tool was somewhat less efficient in calling papers as randomized (sensitivity of 1.00 and specificity of 0.62), but blinding and sample size calculation ascertainment were still above our successful performance criteria level of 0.80 (sensitivity of 0.88 and specificity of 0.98 for blinding, and sensitivity of 0.88 and specificity of 0.97 for sample size calculation). For the MCAO search we identified 1311 publications, of which 1152 met our inclusion criteria. We were able to retrieve full-text pdfs for 918 of these. The RegEx called 670 of 918 MCAO studies (73%) as reporting randomization, 621 (68%) as reporting the blinded assessment of outcome and 21 (2%) as reporting a sample size calculation. Compared with manual assessment, sensitivity was 1.00 (randomization), 0.99 (blinding) and 0.26 (sample size calculation), with specificity of 0.67 (randomization), 0.77 (blinded assessment of outcome) and 0.99 (sample size calculation). Accuracy was 87% (randomization), 90% (blinded assessment of outcome) and 95% (sample size calculation) (Table 1).
Table 1

Summary of performance of RegEx compared with gold standard of manual human ascertainment

Prevalence (RegEx)Prevalence (gold standard)SensitivitySpecificityAccuracy
MCAO (n=918)
Randomization73%60%1.000.6787%
Blinding68%59%0.990.7790%
Sample size calculation2%6%0.260.9995%
Lacunar (n=46)
Randomization67%37%1.000.4867%
Blinding61%53%1.000.8191%
Sample size calculation4%9%0.501.0096%

Accuracy is calculated as the number of publications correctly labelled as a proportion of the total number of publications.

Accuracy is calculated as the number of publications correctly labelled as a proportion of the total number of publications. Reporting of randomization and of sample size calculations was highest in the last quartile and lowest in the first quartile, but using least squares linear regression we found no significant change over time. In journals represented in the dataset by five publications or more, reporting of randomization ranged from 33% to 100% (median 76%), of blinded outcome assessment from 17% to 100% (median 70%), and of sample size calculations from 0% to 17% (median 0%). Overall, reporting of these measures to reduce risk of bias showed substantial improvements from those observed in a selection of stroke systematic reviews reported in 2007. In the 50 publications carried in the journal Stroke, the performance in 2014–2016 was comparable with that reported from 2013 by Minnerup. For the lacunar search we identified 492 publications, of which 61 met our inclusion criteria and of which we were able to retrieve full-text for 46. This lower rate of inclusion (12% vs 87% for MCAO) reflects the reduced sensitivity of the search terms used, reflecting in turn the heterogeneity in the terms used to describe animal experiments modelling lacunar stroke. The RegEx called 31 of 46 lacunar studies (67%) as reporting randomization, 28 (61%) as reporting the blinded assessment of outcome and 2 (4%) as reporting a sample size calculation. Compared with manual ascertainment, sensitivity was 1.00 for each of randomization and blinded assessment of outcome, and 0.50 for sample size calculation, with specificity of 0.48 (randomization), 0.81 (blinded assessment of outcome) and 1.00 (sample size calculation). Accuracy was 67% (randomization), 91% (blinded assessment of outcome) and 96% (sample size calculation). True prevalence of randomization was 37%, of blinding 53% and of sample size calculations 9%. These are essentially unchanged from the Pedder review of the lacunar stroke literature up to 2012, showing no improvements in reporting of measures to reduce risk of bias over time. No journals were represented in the dataset by five publications or more.

Discussion

There have been improvements in reporting of measures to reduce risks of bias in the MCAO literature over the last 10 years, but we did not observe similar improvements in the in vivo lacunar stroke literature. For both MCAO and lacunar stroke the reporting of a sample size calculation remained low, and there remains substantial room for improvement. Within the MCAO literature performance differed between journals, for instance, in journals contributing five or more publications the rate of reporting of randomization ranged from 33% to 100%, of blinding from 17% to 100%, and of sample size calculation from 0% to 17%. There were no substantial differences, in aggregate, in 239 manuscripts published in journals publishing fewer than five included publications compared with 679 manuscripts published in journals publishing five or more included publications (blinding 63% (less than 5) vs 69% (5 or more), randomization 72% vs 73%, sample size calculation 1% vs 3%). While it has been argued that an experiment not guided by a sample size calculation is not in itself at risk of bias, we consider the presentation of a sample size calculation to give much greater confidence in the results presented. Firstly, in the presence of publication bias, underpowered experiments will lead to a literature that overstates observed effects [14]. Secondly, a sample size calculation gives greater confidence that the experiment is presented as originally planned, rather than group size being incrementally increased until an arbitrary level of statistical significance is achieved. Finally, the positive predictive value of an experiment is increased when a study has adequate statistical power [15], and without a sample size calculation this is less likely to be the case. Our study has limitations. Firstly, screening and gold standard risk of bias annotations were performed by a single investigator, and this may have led to errors. Secondly, we were only able to perform data extraction for manuscripts where full-text pdf was available for automated retrieval or pdf could be converted to text, and it may be that the reporting of risks of bias is different in journals for which we do not enjoy an institutional subscription. Moreover, the automated tool did not take into account the existence of any supplementary materials that might have been associated with a publication and may have contained the reporting of methodology pertaining to the measures to reduce risk of bias. Such supplementary materials are not always returned during automatic text retrieval; but in any case reporting guidelines usually recommend that such items be covered in the main text. Finally, the performance of the RegEx for randomization and for sample size calculations did not perform as well for stroke studies as had been the case in other datasets. Problems related mainly to ‘overcalling’ of blinding and randomization (low specificity) and ‘undercalling’ of sample size calculations (low sensitivity). It may be that there are domain specific differences in language used, and these will be incorporated into further iterations of the RegEx used. Future work will include the integration with text mining tools which support screening (such as app.syrf.org.uk) and reduce the amount of annotations required by using topic analysis and active learning [16,17]; and crowdsourced annotations performed in the context of reviews using the SyRF platform will contribute to the pool of annotated data available for tool development. However, for the reporting of the blinded assessment of outcome our RegEx performed at or around the level which had previously been determined as the threshold above which it could be used for the automated ascertainment of the reporting of measures to reduce risk of bias in in vivo research. Such ascertainment, providing as it does almost real-time reporting of the performance of a journal (or an institution, a funder or a nation), might serve as an indicator of changes in the quality of published research, and be used to guide audit for improvement activity, for instance the application of Plan-Do-Study-Act cycles [18] to research activity.
  18 in total

1.  Drug development: Raise standards for preclinical cancer research.

Authors:  C Glenn Begley; Lee M Ellis
Journal:  Nature       Date:  2012-03-28       Impact factor: 49.962

2.  Empirical evidence of bias in the design of experimental stroke studies: a metaepidemiologic approach.

Authors:  Nicolas A Crossley; Emily Sena; Jos Goehler; Jannekke Horn; Bart van der Worp; Philip M W Bath; Malcolm Macleod; Ulrich Dirnagl
Journal:  Stroke       Date:  2008-01-31       Impact factor: 7.914

Review 3.  How can we improve the pre-clinical development of drugs for stroke?

Authors:  Emily Sena; H Bart van der Worp; David Howells; Malcolm Macleod
Journal:  Trends Neurosci       Date:  2007-08-31       Impact factor: 13.837

Review 4.  Quality improvement tools and processes.

Authors:  Catherine Y Lau
Journal:  Neurosurg Clin N Am       Date:  2015-01-31       Impact factor: 2.509

Review 5.  Power failure: why small sample size undermines the reliability of neuroscience.

Authors:  Katherine S Button; John P A Ioannidis; Claire Mokrysz; Brian A Nosek; Jonathan Flint; Emma S J Robinson; Marcus R Munafò
Journal:  Nat Rev Neurosci       Date:  2013-04-10       Impact factor: 34.870

6.  Biomedical research: increasing value, reducing waste.

Authors:  Malcolm R Macleod; Susan Michie; Ian Roberts; Ulrich Dirnagl; Iain Chalmers; John P A Ioannidis; Rustam Al-Shahi Salman; An-Wen Chan; Paul Glasziou
Journal:  Lancet       Date:  2014-01-08       Impact factor: 79.321

Review 7.  Systematic review and meta-analysis of interventions tested in animal models of lacunar stroke.

Authors:  Hugo Pedder; Hanna M Vesterinen; Malcolm R Macleod; Joanna M Wardlaw
Journal:  Stroke       Date:  2014-01-02       Impact factor: 7.914

8.  Risk of Bias in Reports of In Vivo Research: A Focus for Improvement.

Authors:  Malcolm R Macleod; Aaron Lawson McLean; Aikaterini Kyriakopoulou; Stylianos Serghiou; Arno de Wilde; Nicki Sherratt; Theo Hirst; Rachel Hemblade; Zsanett Bahor; Cristina Nunes-Fonseca; Aparna Potluru; Andrew Thomson; Julija Baginskaite; Julija Baginskitae; Kieren Egan; Hanna Vesterinen; Gillian L Currie; Leonid Churilov; David W Howells; Emily S Sena
Journal:  PLoS Biol       Date:  2015-10-13       Impact factor: 8.029

Review 9.  Methodological Rigor in Preclinical Cardiovascular Studies: Targets to Enhance Reproducibility and Promote Research Translation.

Authors:  F Daniel Ramirez; Pouya Motazedian; Richard G Jung; Pietro Di Santo; Zachary D MacDonald; Robert Moreland; Trevor Simard; Aisling A Clancy; Juan J Russo; Vivian A Welch; George A Wells; Benjamin Hibbert
Journal:  Circ Res       Date:  2017-04-03       Impact factor: 17.367

10.  Topic detection using paragraph vectors to support active learning in systematic reviews.

Authors:  Kazuma Hashimoto; Georgios Kontonatsios; Makoto Miwa; Sophia Ananiadou
Journal:  J Biomed Inform       Date:  2016-06-10       Impact factor: 6.317

View more
  13 in total

1.  PR-957 mediates neuroprotection by inhibiting Th17 differentiation and modulating cytokine production in a mouse model of ischaemic stroke.

Authors:  Y Guo; X Chen; D Li; H Liu; Y Ding; R Han; Y Shi; X Ma
Journal:  Clin Exp Immunol       Date:  2018-04-23       Impact factor: 4.330

2.  Translational Block in Stroke: A Constructive and "Out-of-the-Box" Reappraisal.

Authors:  Athanasios Lourbopoulos; Iordanis Mourouzis; Christodoulos Xinaris; Nefeli Zerva; Konstantinos Filippakis; Angelos Pavlopoulos; Constantinos Pantos
Journal:  Front Neurosci       Date:  2021-05-14       Impact factor: 4.677

3.  The study design elements employed by researchers in preclinical animal experiments from two research domains and implications for automation of systematic reviews.

Authors:  Annette M O'Connor; Sarah C Totton; Jonah N Cullen; Mahmood Ramezani; Vijay Kalivarapu; Chaohui Yuan; Stephen B Gilbert
Journal:  PLoS One       Date:  2018-06-28       Impact factor: 3.240

4.  Revision of the ARRIVE guidelines: rationale and scope.

Authors:  Nathalie Percie du Sert; Viki Hurst; Amrita Ahluwalia; Sabina Alam; Douglas G Altman; Marc T Avey; Monya Baker; William Browne; Alejandra Clark; Innes C Cuthill; Ulrich Dirnagl; Michael Emerson; Paul Garner; David W Howells; Natasha A Karp; Catriona J MacCallum; Malcolm Macleod; Ole Petersen; Frances Rawle; Penny Reynolds; Kieron Rooney; Emily S Sena; Shai D Silberberg; Thomas Steckler; Hanno Würbel; Stephen T Holgate
Journal:  BMJ Open Sci       Date:  2018-06-12

5.  A practical guide to preclinical systematic review and meta-analysis.

Authors:  Nadia Soliman; Andrew S C Rice; Jan Vollert
Journal:  Pain       Date:  2020-09-01       Impact factor: 7.926

6.  Weight loss, insulin resistance, and study design confound results in a meta-analysis of animal models of fatty liver.

Authors:  Harriet Hunter; Dana de Gracia Hahn; Amedine Duret; Yu Ri Im; Qinrong Cheah; Jiawen Dong; Madison Fairey; Clarissa Hjalmarsson; Alice Li; Hong Kai Lim; Lorcan McKeown; Claudia-Gabriela Mitrofan; Raunak Rao; Mrudula Utukuri; Ian A Rowe; Jake P Mann
Journal:  Elife       Date:  2020-10-16       Impact factor: 8.140

7.  The potential benefit of leptin therapy against amyotrophic lateral sclerosis (ALS).

Authors:  Agueda Ferrer-Donato; Ana Contreras; Paloma Fernandez; Carmen M Fernandez-Martos
Journal:  Brain Behav       Date:  2021-12-21       Impact factor: 2.708

8.  Small vessels, dementia and chronic diseases - molecular mechanisms and pathophysiology.

Authors:  Karen Horsburgh; Joanna M Wardlaw; Tom van Agtmael; Stuart M Allan; Mike L J Ashford; Philip M Bath; Rosalind Brown; Jason Berwick; M Zameel Cader; Roxana O Carare; John B Davis; Jessica Duncombe; Tracy D Farr; Jill H Fowler; Jozien Goense; Alessandra Granata; Catherine N Hall; Atticus H Hainsworth; Adam Harvey; Cheryl A Hawkes; Anne Joutel; Rajesh N Kalaria; Patrick G Kehoe; Catherine B Lawrence; Andy Lockhart; Seth Love; Malcolm R Macleod; I Mhairi Macrae; Hugh S Markus; Chris McCabe; Barry W McColl; Paul J Meakin; Alyson Miller; Maiken Nedergaard; Michael O'Sullivan; Terry J Quinn; Rikesh Rajani; Lisa M Saksida; Colin Smith; Kenneth J Smith; Rhian M Touyz; Rebecca C Trueman; Tao Wang; Anna Williams; Steven C R Williams; Lorraine M Work
Journal:  Clin Sci (Lond)       Date:  2018-04-30       Impact factor: 6.124

9.  Did a change in Nature journals' editorial policy for life sciences research improve reporting?

Authors: 
Journal:  BMJ Open Sci       Date:  2019-02-26

Review 10.  Technological advances in preclinical meta-research.

Authors:  Alexandra Bannach-Brown; Kaitlyn Hair; Zsanett Bahor; Nadia Soliman; Malcolm Macleod; Jing Liao
Journal:  BMJ Open Sci       Date:  2021-07-25
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.