Literature DB >> 28711679

Reproducibility of studies on text mining for citation screening in systematic reviews: Evaluation and checklist.

Babatunde Kazeem Olorisade1, Pearl Brereton2, Peter Andras3.   

Abstract

CONTEXT: Independent validation of published scientific results through study replication is a pre-condition for accepting the validity of such results. In computation research, full replication is often unrealistic for independent results validation, therefore, study reproduction has been justified as the minimum acceptable standard to evaluate the validity of scientific claims. The application of text mining techniques to citation screening in the context of systematic literature reviews is a relatively young and growing computational field with high relevance for software engineering, medical research and other fields. However, there is little work so far on reproduction studies in the field.
OBJECTIVE: In this paper, we investigate the reproducibility of studies in this area based on information contained in published articles and we propose reporting guidelines that could improve reproducibility.
METHODS: The study was approached in two ways. Initially we attempted to reproduce results from six studies, which were based on the same raw dataset. Then, based on this experience, we identified steps considered essential to successful reproduction of text mining experiments and characterized them to measure how reproducible is a study given the information provided on these steps. 33 articles were systematically assessed for reproducibility using this approach.
RESULTS: Our work revealed that it is currently difficult if not impossible to independently reproduce the results published in any of the studies investigated. The lack of information about the datasets used limits reproducibility of about 80% of the studies assessed. Also, information about the machine learning algorithms is inadequate in about 27% of the papers. On the plus side, the third party software tools used are mostly free and available.
CONCLUSIONS: The reproducibility potential of most of the studies can be significantly improved if more attention is paid to information provided on the datasets used, how they were partitioned and utilized, and how any randomization was controlled. We introduce a checklist of information that needs to be provided in order to ensure that a published study can be reproduced.
Copyright © 2017 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  Citation screening; Reproducibility; Reproducible research; Systematic review; Text mining

Mesh:

Year:  2017        PMID: 28711679     DOI: 10.1016/j.jbi.2017.07.010

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   6.317


  7 in total

1.  Three Dimensions of Reproducibility in Natural Language Processing.

Authors:  K Bretonnel Cohen; Jingbo Xia; Pierre Zweigenbaum; Tiffany J Callahan; Orin Hargraves; Foster Goss; Nancy Ide; Aurélie Névéol; Cyril Grouin; Lawrence E Hunter
Journal:  LREC Int Conf Lang Resour Eval       Date:  2018-05

2.  Data extraction methods for systematic review (semi)automation: A living systematic review.

Authors:  Lena Schmidt; Babatunde K Olorisade; Luke A McGuinness; James Thomas; Julian P T Higgins
Journal:  F1000Res       Date:  2021-05-19

3.  Training sample selection: Impact on screening automation in diagnostic test accuracy reviews.

Authors:  Allard J van Altena; René Spijker; Mariska M G Leeflang; Sílvia Delgado Olabarriaga
Journal:  Res Synth Methods       Date:  2021-08-25       Impact factor: 9.308

Review 4.  Data extraction methods for systematic review (semi)automation: A living review protocol.

Authors:  Lena Schmidt; Babatunde K Olorisade; Luke A McGuinness; James Thomas; Julian P T Higgins
Journal:  F1000Res       Date:  2020-03-25

5.  A question of trust: can we build an evidence base to gain trust in systematic review automation technologies?

Authors:  Annette M O'Connor; Guy Tsafnat; James Thomas; Paul Glasziou; Stephen B Gilbert; Brian Hutton
Journal:  Syst Rev       Date:  2019-06-18

6.  Still moving toward automation of the systematic review process: a summary of discussions at the third meeting of the International Collaboration for Automation of Systematic Reviews (ICASR).

Authors:  Annette M O'Connor; Guy Tsafnat; Stephen B Gilbert; Kristina A Thayer; Ian Shemilt; James Thomas; Paul Glasziou; Mary S Wolfe
Journal:  Syst Rev       Date:  2019-02-20

7.  Performance and usability of machine learning for screening in systematic reviews: a comparative evaluation of three tools.

Authors:  Allison Gates; Samantha Guitard; Jennifer Pillay; Sarah A Elliott; Michele P Dyson; Amanda S Newton; Lisa Hartling
Journal:  Syst Rev       Date:  2019-11-15
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.