| Literature DB >> 31215463 |
Annette M O'Connor1, Guy Tsafnat2, James Thomas3, Paul Glasziou4, Stephen B Gilbert5, Brian Hutton6.
Abstract
BACKGROUND: Although many aspects of systematic reviews use computational tools, systematic reviewers have been reluctant to adopt machine learning tools. DISCUSSION: We discuss that the potential reason for the slow adoption of machine learning tools into systematic reviews is multifactorial. We focus on the current absence of trust in automation and set-up challenges as major barriers to adoption. It is important that reviews produced using automation tools are considered non-inferior or superior to current practice. However, this standard will likely not be sufficient to lead to widespread adoption. As with many technologies, it is important that reviewers see "others" in the review community using automation tools. Adoption will also be slow if the automation tools are not compatible with workflows and tasks currently used to produce reviews. Many automation tools being developed for systematic reviews mimic classification problems. Therefore, the evidence that these automation tools are non-inferior or superior can be presented using methods similar to diagnostic test evaluations, i.e., precision and recall compared to a human reviewer. However, the assessment of automation tools does present unique challenges for investigators and systematic reviewers, including the need to clarify which metrics are of interest to the systematic review community and the unique documentation challenges for reproducible software experiments.Entities:
Keywords: Artificial intelligence; Automation; Data extraction; Machine learning; Screening
Mesh:
Year: 2019 PMID: 31215463 PMCID: PMC6582554 DOI: 10.1186/s13643-019-1062-0
Source DB: PubMed Journal: Syst Rev ISSN: 2046-4053
Levels of automation for human-computer interactions
| Level | Task |
|---|---|
| Level 4 | Tools perform tasks to eliminate the need for human participation in the task altogether, e.g., fully automated article screening decision about relevance made by the automated system. |
| Level 3 | Tools perform a task automatically but unreliably and require human supervision or else provide the option to manually override the tools’ decisions, e.g., duplicate detection algorithms and software, linked publication detection with plagiarism algorithms and software. |
| Level 2 | Tools enable workflow prioritization, e.g., prioritization of relevant abstracts; however, this does not reduce the work time for reviewers on the task but does allow for compression of the calendar time of the entire process. |
| Level 1 | Tools improve the file management process, e.g., citation databases, reference management software, and systematic review management software. |
Proposed additional items for inclusion in a shared dataset for a classification experiment for automation of systematic review processes
| Column | Item |
|---|---|
| 1 | Title of source—publication name, report name, etc. |
| 2 | Indexing data (e.g., PubMed identifier, ISBN, doi) |
| 3 | Author names |
| 4 | Publication venue (e.g., journal name) |
| 5 | Serial data (e.g., volume, issue, and page numbers) |
| 6 | A final classification field. This would be a final category used in the systematic review. For example, if the dataset is designed for screening, this field might refer to inclusion status in the final systematic review (“yes” or “no”), or if the classification task is bias assessment this might refer to bias assessment in the final systematic review (“low”, “high”, “unclear”). |
| 7 | Reviewer 1 classification, i.e., whether Reviewer 1 recommended inclusion of the article in the systematic review |
| Reviewer 1 notes field (free text) whenever notes were provided by the reviewer | |
| 8 | Reviewer 1 notes field supporting text from the manuscript if extracted (optional) |
| 9 | Reviewer 2 classification, i.e., whether Reviewer 2 recommended inclusion of the article in the systematic review |
| 10 | Reviewer 2 notes field (free text) whenever notes were provided by the reviewer |
| Reviewer 2 notes field supporting text from the manuscript if extracted (optional) | |
| 11 | Arbiter notes field (free text) whenever notes were provided by the arbiter |
| 12 | A training field (“yes” or “no”) on whether the entry was used to train human reviewers |
Illustration of the proposed additional metadata documentation for sharing files annotated for systematic reviews