Literature DB >> 32308910

Transfer Learning from BERT to Support Insertion of New Concepts into SNOMED CT.

Hao Liu1, Yehoshua Perl1, James Geller1.   

Abstract

With advances in Machine Learning (ML), neural network-based methods, such as Convolutional/Recurrent Neural Networks, have been proposed to assist terminology curators in the development and maintenance of terminologies. Bidirectional Encoder Representations from Transformers (BERT), a new language representation model, obtains state-of-the-art results on a wide array of general English NLP tasks. We explore BERT's applicability to medical terminology-related tasks. Utilizing the "next sentence prediction" capability of BERT, we show that the Fine-tuning strategy of Transfer Learning (TL) from the BERTBASE model can address a challenging problem in automatic terminology enrichment - insertion of new concepts. Adding a pre-training strategy enhances the results. We apply our strategies to the two largest hierarchies of SNOMED CT, with one release as training data and the following release as test data. The performance of the combined two proposed TL models achieves an average F1 score of 0.85 and 0.86 for the two hierarchies, respectively. ©2019 AMIA - All rights reserved.

Year:  2020        PMID: 32308910      PMCID: PMC7153142     

Source DB:  PubMed          Journal:  AMIA Annu Symp Proc        ISSN: 1559-4076


  6 in total

Review 1.  Deep learning.

Authors:  Yann LeCun; Yoshua Bengio; Geoffrey Hinton
Journal:  Nature       Date:  2015-05-28       Impact factor: 49.962

2.  Enhancing clinical concept extraction with contextual embeddings.

Authors:  Yuqi Si; Jingqi Wang; Hua Xu; Kirk Roberts
Journal:  J Am Med Inform Assoc       Date:  2019-11-01       Impact factor: 4.497

3.  Using Convolutional Neural Networks to Support Insertion of New Concepts into SNOMED CT.

Authors:  Hao Liu; James Geller; Michael Halper; Yehoshua Perl
Journal:  AMIA Annu Symp Proc       Date:  2018-12-05

4.  Corpus domain effects on distributional semantic modeling of medical terms.

Authors:  Serguei V S Pakhomov; Greg Finley; Reed McEwan; Yan Wang; Genevieve B Melton
Journal:  Bioinformatics       Date:  2016-08-16       Impact factor: 6.937

Review 5.  Abstraction networks for terminologies: Supporting management of "big knowledge".

Authors:  Michael Halper; Huanying Gu; Yehoshua Perl; Christopher Ochs
Journal:  Artif Intell Med       Date:  2015-04-02       Impact factor: 5.326

6.  BioBERT: a pre-trained biomedical language representation model for biomedical text mining.

Authors:  Jinhyuk Lee; Wonjin Yoon; Sungdong Kim; Donghyeon Kim; Sunkyu Kim; Chan Ho So; Jaewoo Kang
Journal:  Bioinformatics       Date:  2020-02-15       Impact factor: 6.937

  6 in total
  3 in total

1.  Deep learning of a bacterial and archaeal universal language of life enables transfer learning and illuminates microbial dark matter.

Authors:  A Hoarfrost; A Aptekmann; G Farfañuk; Y Bromberg
Journal:  Nat Commun       Date:  2022-05-11       Impact factor: 17.694

Review 2.  A review of auditing techniques for the Unified Medical Language System.

Authors:  Ling Zheng; Zhe He; Duo Wei; Vipina Keloth; Jung-Wei Fan; Luke Lindemann; Xinxin Zhu; James J Cimino; Yehoshua Perl
Journal:  J Am Med Inform Assoc       Date:  2020-10-01       Impact factor: 4.497

3.  Automatic Structuring of Ontology Terms Based on Lexical Granularity and Machine Learning: Algorithm Development and Validation.

Authors:  Lingyun Luo; Jingtao Feng; Huijun Yu; Jiaolong Wang
Journal:  JMIR Med Inform       Date:  2020-11-25
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.