Literature DB >> 31445245

Extractive summarization of clinical trial descriptions.

Christian Gulden1, Melanie Kirchner2, Christina Schüttler3, Marc Hinderer3, Marvin Kampf2, Hans-Ulrich Prokosch3, Dennis Toddenroth3.   

Abstract

PURPOSE: Text summarization of clinical trial descriptions has the potential to reduce the time required to familiarize oneself with the subject of studies by condensing long-form detailed descriptions to concise, meaning-preserving synopses. This work describes the process and quality of automatically generated summaries of clinical trial descriptions using extractive text summarization methods.
METHODS: We generated a novel dataset from the detailed descriptions and brief summaries of trials registered on clinicaltrials.gov. We executed several text summarization algorithms on the detailed descriptions in this corpus and calculated the standard ROUGE metrics using the brief summaries included in the record as a reference. To investigate the correlation of these metrics with human sentiments, four reviewers assessed the content-completeness of the generated summaries and the helpfulness of both the generated and reference summaries via a Likert scale questionnaire.
RESULTS: The filtering stages of the dataset generation process reduce the 277,228 trials registered on clinicaltrials.gov to 101,016 records usable for the summarization task. On average, the summaries in this corpus are 25% the length of the detailed descriptions. Of the evaluated text summarization methods, the TextRank algorithm exhibits the overall best performance with a ROUGE-1 F1 score of 0.3531, ROUGE-2 F1 score of 0.1723, and ROUGE-L F1 score of 0.3003. These scores correlate with the assessment of the helpfulness and content similarity by the human reviewers. Inter-rater agreement for the helpfulness and content similarity was slight and fair respectively (Fleiss' kappa of 0.12 and 0.22).
CONCLUSIONS: Extractive summarization is a viable tool for generating meaning-preserving synopses of detailed clinical trial descriptions. Further, the human evaluation has shown that the ROUGE-L F1 score is useful for rating the general quality of generated summaries of clinical trial descriptions in an automated way.
Copyright © 2019 Elsevier B.V. All rights reserved.

Entities:  

Keywords:  Clinical trials; NLP; Text mining; Text summarization

Mesh:

Year:  2019        PMID: 31445245     DOI: 10.1016/j.ijmedinf.2019.05.019

Source DB:  PubMed          Journal:  Int J Med Inform        ISSN: 1386-5056            Impact factor:   4.046


  4 in total

1.  A systematic review of automatic text summarization for biomedical literature and EHRs.

Authors:  Mengqian Wang; Manhua Wang; Fei Yu; Yue Yang; Jennifer Walker; Javed Mostafa
Journal:  J Am Med Inform Assoc       Date:  2021-09-18       Impact factor: 7.942

2.  An Ensemble Learning Strategy for Eligibility Criteria Text Classification for Clinical Trial Recruitment: Algorithm Development and Validation.

Authors:  Kun Zeng; Zhiwei Pan; Yibin Xu; Yingying Qu
Journal:  JMIR Med Inform       Date:  2020-07-01

3.  Automated classification of clinical trial eligibility criteria text based on ensemble learning and metric learning.

Authors:  Kun Zeng; Yibin Xu; Ge Lin; Likeng Liang; Tianyong Hao
Journal:  BMC Med Inform Decis Mak       Date:  2021-07-30       Impact factor: 2.796

4.  Application of BERT to Enable Gene Classification Based on Clinical Evidence.

Authors:  Yuhan Su; Hongxin Xiang; Haotian Xie; Yong Yu; Shiyan Dong; Zhaogang Yang; Na Zhao
Journal:  Biomed Res Int       Date:  2020-10-07       Impact factor: 3.411

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.