| Literature DB >> 28894463 |
Yuntian Feng1, Hongjun Zhang1, Wenning Hao1, Gang Chen1.
Abstract
We use both reinforcement learning and deep learning to simultaneously extract entities and relations from unstructured texts. For reinforcement learning, we model the task as a two-step decision process. Deep learning is used to automatically capture the most important information from unstructured texts, which represent the state in the decision process. By designing the reward function per step, our proposed method can pass the information of entity extraction to relation extraction and obtain feedback in order to extract entities and relations simultaneously. Firstly, we use bidirectional LSTM to model the context information, which realizes preliminary entity extraction. On the basis of the extraction results, attention based method can represent the sentences that include target entity pair to generate the initial state in the decision process. Then we use Tree-LSTM to represent relation mentions to generate the transition state in the decision process. Finally, we employ Q-Learning algorithm to get control policy π in the two-step decision process. Experiments on ACE2005 demonstrate that our method attains better performance than the state-of-the-art method and gets a 2.4% increase in recall-score.Entities:
Mesh:
Year: 2017 PMID: 28894463 PMCID: PMC5574273 DOI: 10.1155/2017/7643065
Source DB: PubMed Journal: Comput Intell Neurosci
A sentence in ACE2005 dataset.
| Sentence | While either divesting or inviting third parties to take a minority stake in the remaining Entertainment assets. | |
|---|---|---|
| Entity ID = “AFP_ENG_20030319.0879-E24” | Type = “ORG” | third parties |
|
| ||
| Entity ID = “AFP_ENG_20030319.0879-E25” | Type = “ORG” | Entertainment |
|
| ||
| Relation ID = “AFP_ENG_20030319.0879-R2” | Type = “ORG-AFF” | RefID = “AFP_ENG_20030319.0879-E24” |
| RefID = “AFP_ENG_20030319.0879-E25” | ||
Figure 1Two-step decision process.
Figure 2Basic structure of BILSTM.
Figure 3Attention layer.
Figure 4Dependency tree of a relation mention.
Algorithm 1Training procedure for Q-Learning.
Performance for entity extraction task.
| Method | Entity | ||
|---|---|---|---|
| Score |
|
|
|
| LSTM | 81.0 | 78.1 | 79.5 |
| BILSTM | 82.5 | 79.8 | 81.1 |
Performance for relation extraction task.
| Method | Relation | ||
|---|---|---|---|
| Score |
|
|
|
| CNN | 63.1 | 52.9 | 57.6 |
| Tree-LSTM | 63.9 | 54.1 | 58.6 |
| RL | 63.6 | 59.4 | 61.4 |
Performance of two extraction systems.
| Method | Entity | Relation | ||||
|---|---|---|---|---|---|---|
| Score |
|
|
|
|
|
|
| Pipeline | 82.5 | 79.8 | 81.1 | 60.2 | 43.9 | 50.8 |
| Joint | 83.6 | 80.4 | 82.0 | 60.6 | 45.9 | 52.2 |
Figure 5Learning curve of average reward.
Figure 6Learning curves of the performance.
Comparison with state of the art.
| Method | Entity | Relation | ||||
|---|---|---|---|---|---|---|
| Score |
|
|
|
|
|
|
| SPTree | 85.5 | 81.2 | 83.3 | 65.8 | 42.9 | 51.9 |
| Joint | 85.0 |
| 83.7 | 65.9 |
| 53.7 |
Figure 7Objective values in the attention model.
Figure 8Performance of relation mention classification.