Literature DB >> 21084178

Exploiting the systematic review protocol for classification of medical abstracts.

Oana Frunza1, Diana Inkpen, Stan Matwin, William Klement, Peter O'Blenis.   

Abstract

OBJECTIVE: To determine whether the automatic classification of documents can be useful in systematic reviews on medical topics, and specifically if the performance of the automatic classification can be enhanced by using the particular protocol of questions employed by the human reviewers to create multiple classifiers. METHODS AND MATERIALS: The test collection is the data used in large-scale systematic review on the topic of the dissemination strategy of health care services for elderly people. From a group of 47,274 abstracts marked by human reviewers to be included in or excluded from further screening, we randomly selected 20,000 as a training set, with the remaining 27,274 becoming a separate test set. As a machine learning algorithm we used complement naïve Bayes. We tested both a global classification method, where a single classifier is trained on instances of abstracts and their classification (i.e., included or excluded), and a novel per-question classification method that trains multiple classifiers for each abstract, exploiting the specific protocol (questions) of the systematic review. For the per-question method we tested four ways of combining the results of the classifiers trained for the individual questions. As evaluation measures, we calculated precision and recall for several settings of the two methods. It is most important not to exclude any relevant documents (i.e., to attain high recall for the class of interest) but also desirable to exclude most of the non-relevant documents (i.e., to attain high precision on the class of interest) in order to reduce human workload.
RESULTS: For the global method, the highest recall was 67.8% and the highest precision was 37.9%. For the per-question method, the highest recall was 99.2%, and the highest precision was 63%. The human-machine workflow proposed in this paper achieved a recall value of 99.6%, and a precision value of 17.8%.
CONCLUSION: The per-question method that combines classifiers following the specific protocol of the review leads to better results than the global method in terms of recall. Because neither method is efficient enough to classify abstracts reliably by itself, the technology should be applied in a semi-automatic way, with a human expert still involved. When the workflow includes one human expert and the trained automatic classifier, recall improves to an acceptable level, showing that automatic classification techniques can reduce the human workload in the process of building a systematic review.
Copyright © 2010 Elsevier B.V. All rights reserved.

Entities:  

Mesh:

Year:  2010        PMID: 21084178     DOI: 10.1016/j.artmed.2010.10.005

Source DB:  PubMed          Journal:  Artif Intell Med        ISSN: 0933-3657            Impact factor:   5.326


  11 in total

1.  Direct comparison between support vector machine and multinomial naive Bayes algorithms for medical abstract classification.

Authors:  Stan Matwin; Vera Sazonova
Journal:  J Am Med Inform Assoc       Date:  2012-06-08       Impact factor: 4.497

2.  Validation of a Semiautomated Natural Language Processing-Based Procedure for Meta-Analysis of Cancer Susceptibility Gene Penetrance.

Authors:  Zhengyi Deng; Kanhua Yin; Yujia Bao; Victor Diego Armengol; Cathy Wang; Ankur Tiwari; Regina Barzilay; Giovanni Parmigiani; Danielle Braun; Kevin S Hughes
Journal:  JCO Clin Cancer Inform       Date:  2019-08

3.  A new iterative method to reduce workload in systematic review process.

Authors:  Siddhartha Jonnalagadda; Diana Petitti
Journal:  Int J Comput Biol Drug Des       Date:  2013-02-21

4.  Using Machine Learning and Natural Language Processing to Review and Classify the Medical Literature on Cancer Susceptibility Genes.

Authors:  Yujia Bao; Zhengyi Deng; Yan Wang; Heeyoon Kim; Victor Diego Armengol; Francisco Acevedo; Nofal Ouardaoui; Cathy Wang; Giovanni Parmigiani; Regina Barzilay; Danielle Braun; Kevin S Hughes
Journal:  JCO Clin Cancer Inform       Date:  2019-09

Review 5.  Patient healthcare trajectory. An essential monitoring tool: a systematic review.

Authors:  Jessica Pinaire; Jérôme Azé; Sandra Bringay; Paul Landais
Journal:  Health Inf Sci Syst       Date:  2017-04-12

6.  Screening nonrandomized studies for medical systematic reviews: a comparative study of classifiers.

Authors:  Tanja Bekhuis; Dina Demner-Fushman
Journal:  Artif Intell Med       Date:  2012-06-05       Impact factor: 5.326

Review 7.  Using text mining for study identification in systematic reviews: a systematic review of current approaches.

Authors:  Alison O'Mara-Eves; James Thomas; John McNaught; Makoto Miwa; Sophia Ananiadou
Journal:  Syst Rev       Date:  2015-01-14

8.  Automated confidence ranked classification of randomized controlled trial articles: an aid to evidence-based medicine.

Authors:  Aaron M Cohen; Neil R Smalheiser; Marian S McDonagh; Clement Yu; Clive E Adams; John M Davis; Philip S Yu
Journal:  J Am Med Inform Assoc       Date:  2015-02-05       Impact factor: 4.497

9.  Feature engineering and a proposed decision-support system for systematic reviewers of medical evidence.

Authors:  Tanja Bekhuis; Eugene Tseytlin; Kevin J Mitchell; Dina Demner-Fushman
Journal:  PLoS One       Date:  2014-01-27       Impact factor: 3.240

10.  Prioritising references for systematic reviews with RobotAnalyst: A user study.

Authors:  Piotr Przybyła; Austin J Brockmeier; Georgios Kontonatsios; Marie-Annick Le Pogam; John McNaught; Erik von Elm; Kay Nolan; Sophia Ananiadou
Journal:  Res Synth Methods       Date:  2018-07-30       Impact factor: 5.273

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.