| Literature DB >> 35463810 |
Juan Antonio Lossio-Ventura1, Sebastien Boussard2, Juandiego Morzan3, Tina Hernandez-Boussard1.
Abstract
The adoption of electronic health records has increased the volume of clinical data, which has opened an opportunity for healthcare research. There are several biomedical annotation systems that have been used to facilitate the analysis of clinical data. However, there is a lack of clinical annotation comparisons to select the most suitable tool for a specific clinical task. In this work, we used clinical notes from the MIMIC-III database and evaluated three annotation systems to identify four types of entities: (1) procedure, (2) disorder, (3) drug, and (4) anatomy. Our preliminary results demonstrate that BioPortal performs well when extracting disorder and drug. This can provide clinical researchers with real-clinical insights into patient's health patterns and it may allow to create a first version of an annotated dataset.Entities:
Keywords: clinical research; electronic health records; named-entity recognition; natural language processing
Year: 2020 PMID: 35463810 PMCID: PMC9028678 DOI: 10.1109/bibm47256.2019.8983406
Source DB: PubMed Journal: Proceedings (IEEE Int Conf Bioinformatics Biomed) ISSN: 2156-1125