| Literature DB >> 28359255 |
Fei Li1, Meishan Zhang2, Guohong Fu2, Donghong Ji3.
Abstract
BACKGROUND: Extracting biomedical entities and their relations from text has important applications on biomedical research. Previous work primarily utilized feature-based pipeline models to process this task. Many efforts need to be made on feature engineering when feature-based models are employed. Moreover, pipeline models may suffer error propagation and are not able to utilize the interactions between subtasks. Therefore, we propose a neural joint model to extract biomedical entities as well as their relations simultaneously, and it can alleviate the problems above.Entities:
Keywords: Biomedical text; Entity recognition; Joint model; Neural network; Relation extraction
Mesh:
Year: 2017 PMID: 28359255 PMCID: PMC5374588 DOI: 10.1186/s12859-017-1609-9
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1The CNN for extracting character-level representations. A rectangular grid indicates a vector and a square indicates one dimension of this vector, so character embeddings or representations can be denoted as n-dimensional vectors. Shading rectangular grids indicate special padding vectors
Fig. 2The Bi-LSTM-RNN for biomedical entity recognition. Rectangular grids indicate vectors of feature embeddings or representations. At the bottom, three kinds of vectors are concatenated and fed into LSTMs. Dashed arrow lines denote bottom-up computations along the network framework and solid arrow lines denote left-to-right computations along the sentence
Fig. 3The Bi-LSTM-RNN for relation classification. The input sentence is tokenized before it is analyzed by a dependency parser. Tokens are indexed by Arabic numerals. Basic (a.k.a, projective) dependency style is utilized to build a tree. The bold lines in the tree denote the shortest dependency path (SDP) between “gliclazide” and “hepatitis” with their lowest common ancestor “induced”. x indicates the input vector of a LSTM unit as shown in Eq. 6 and i corresponds to the index of a token. In the Bi-LSTM-RNN layer, solid arrow lines denote bottom-up and top-down computations along the SDP in the dependency tree. ↑ h , ↑ h , ↓ h , ↓ h are listed in Eq. 8
Statistics of the ADE and BB data used in our experiments
| ADE | BB | ||
|---|---|---|---|
| Sentences | 6821 | Documents | 161 |
| Entities | 10666 | Entities | 2943 |
| Relations | 6686 | Relations | 864 |
Hyper-parameter settings
| Type | Hyper-parameter |
|---|---|
| Training |
|
| Embedding |
|
|
| |
| CNN |
|
|
| |
| Bi-LSTM-RNN (Entity) |
|
|
| |
| Bi-LSTM-RNN (Relation) |
|
|
|
dim denotes vector dimensions and emb denotes feature embeddings
Result (%) comparisons with other work in the ADE task
| Method | Entity recognition | Relation extraction | ||||
|---|---|---|---|---|---|---|
| P | R | F1 | P | R | F1 | |
| Kang [ | — | — | — | 42.1 | 76.3 | 54.3 |
| Li [ | 79.5 | 79.6 | 79.5 | 64.0 | 62.9 | 63.4 |
| Our model | 82.7 | 86.7 | 84.6 | 67.5 | 75.8 | 71.4 |
Result (%) comparisons with other work in relation extraction of the BB task
| LIMSI | UTS | Our model | |
|---|---|---|---|
| Precision | 19.3 | 33.1 | 49.8 |
| Recall | 19.1 | 13.3 | 19.9 |
| F1 | 19.2 | 19.0 | 28.4 |
| F1(Habitat) | 18.6 | 17.4 | 29.2 |
| F1(Geographical) | 28.3 | 35.0 | 20.5 |
| F1(Intra-sentence) | 28.6 | 23.4 | 35.1 |
The inter-annotator agreement (%) of entity mentions and Lives_In relations [5]
| P | R | F1 | |
|---|---|---|---|
| Entity Mentions | 95.5 | 62.1 | 75.3 |
|
| 95.2 | 31.1 | 46.8 |
Feature contribution experiments for entity recognition
| Features | ADE | BB | ||||
|---|---|---|---|---|---|---|
| P | R | F1 | P | R | F1 | |
| Word | 80.1 | 83.6 | 81.8 | 67.1 | 56.7 | 61.4 |
| +char | 80.2 | 84.0 | 82.1 | 66.4, | 59.4 | 62.7 |
| +pos | 80.5 | 84.7 | 82.5 | 69.4 | 60.8 | 64.8 |
| +label | 82.5 | 85.5 | 84.0 | 66.1 | 59.5 | 62.6 |
| All | 82.4 | 86.4 | 84.3 | 68.0 | 63.4 | 65.6 |
Here “+” means only that feature is added. “char”, “pos” and “label” denote character, POS tag and entity label features, respectively
Feature contribution experiments for relation extraction
| Features | ADE | BB | ||||
|---|---|---|---|---|---|---|
| P | R | F1 | P | R | F1 | |
| Word | 62.7 | 69.9 | 66.1 | 34.5 | 20.4 | 25.6 |
| +dep | 63.3 | 71.0 | 66.9 | 42.0 | 19.9 | 27.0 |
| +entity | 63.4 | 71.2 | 67.1 | 34.1 | 24.7 | 28.6 |
| All | 67.3 | 75.7 | 71.3 | 42.7 | 25.2 | 31.7 |
Here “+” means only that feature is added.“dep” and “entity” denote dependency type and entity representation features, respectively
Performance comparisons of joint and pipeline models
| Task | Method | Entity recognition | Relation extraction | ||||
|---|---|---|---|---|---|---|---|
| P | R | F1 | P | R | F1 | ||
| ADE | Pipeline | 79.6 | 83.5 | 81.5 | 62.5 | 69.9 | 66.0 |
| Joint | 80.1 | 83.6 | 81.8 | 62.7 | 69.9 | 66.1 | |
| BB | Pipeline | 67.2 | 52.0 | 58.6 | 26.6 | 17.7 | 21.2 |
| Joint | 67.1 | 56.7 | 61.4 | 34.5 | 20.4 | 25.6 | |
Error analysis of entity recognition
| Task | Error type | % | |
|---|---|---|---|
| ADE | FP | Incorrect boundaries | 55.3 |
| Incorrect types | 1.3 | ||
| FN | Incorrect boundaries | 42.1 | |
| Incorrect types | 1.3 | ||
| Total | 100 | ||
| BB | FP | Incorrect boundaries | 37.1 |
| Incorrect types | 3.6 | ||
| FN | Incorrect boundaries | 55.7 | |
| Incorrect types | 3.6 | ||
| Total | 100 | ||
Error analysis of relation extraction
| Task | Error type | % | |
|---|---|---|---|
| ADE | FP | Entities incorrectly recognized | 55.7 |
| Entities correct, relations wrong | 3.1 | ||
| FN | Entities not found | 40.7 | |
| Entities found, relations not found | 0.5 | ||
| Total | 100 | ||
| BB | FP | Entities incorrectly recognized | 22.7 |
| Entities correct, relations wrong | 15.2 | ||
| FN | Entities not found | 43.7 | |
| Entities found, relations not found | 18.4 | ||
| Total | 100 | ||
Comparisons with the methods based on co-occurrence entities inside one sentence and gold entity mentions
| Task | Method | Relation Extraction | ||
|---|---|---|---|---|
| P | R | F1 | ||
| ADE | Co-occurrence | 97.3 | 100 | 98.6 |
| Gold mentions | 97.5 | 99.9 | 98.7 | |
| Our model | 67.3 | 75.7 | 71.3 | |
| BB | Co-occurrence | 34.9 | 72.5 | 47.1 |
| Gold mentions | 58.7 | 43.6 | 50.0 | |
| Our model | 42.7 | 25.2 | 31.7 | |