| Literature DB >> 30367569 |
Lishuang Li1, Jia Wan2, Jieqiong Zheng2, Jian Wang2.
Abstract
BACKGROUND: Biomedical event extraction is a crucial task in biomedical text mining. As the primary forum for international evaluation of different biomedical event extraction technologies, BioNLP Shared Task represents a trend in biomedical text mining toward fine-grained information extraction (IE). The fourth series of BioNLP Shared Task in 2016 (BioNLP-ST'16) proposed three tasks, in which the Bacteria Biotope event extraction (BB) task has been put forward in the earlier BioNLP-ST. Deep learning methods provide an effective way to automatically extract more complex features and achieve notable results in various natural language processing tasks.Entities:
Keywords: Attention mechanism; Biomedical event extraction; Deep learning; Word representation
Mesh:
Year: 2018 PMID: 30367569 PMCID: PMC6101075 DOI: 10.1186/s12859-018-2275-2
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1Example of entities and Lives in events in the BB task. “Georgia” is the entity of [Geographical]; “populations of gopher tortoises” is the entity of [Habitat]; “M. agassizii” and “M. Testudineum” are the entities of [Bacteria]. There is relation of “Live in” between the bacterium entity and a Habitat or a Geographical entity
Fig. 2The architecture of attention-based BGRU
Fig. 3Hidden activation function of GRU
Comparison with existing systems
| Methods | F-score | Recall | Precision |
|---|---|---|---|
| Baseline | 47.27% | 38.35% | 61.61% |
| TurkuNLP | 52.10% | 44.80% | 62.30% |
| VERSE | 55.80% | 61.50% | 51.00% |
| Li [ | 57.14% | 57.99% | 56.32% |
| Li [ | 58.09% | 56.80% | 59.44% |
| BGRU-Attention | 57.42% | 69.82% | 48.76% |
The results of the attention mechanism
| Input | F-score | Recall | Precision |
|---|---|---|---|
| BGRU | 56.36% | 63.42% | 50.71% |
| BGRU+Attention | 57.42% | 69.82% | 48.76% |
The results of the word embeddings
| Word representation | F-score | Recall | Precision |
|---|---|---|---|
| skip-gram | 55.97% | 62.43% | 50.72% |
| domain-oriented method | 57.42% | 69.82% | 48.76% |