| Literature DB >> 31396271 |
Jun Li1, Guimin Huang2,3, Jianheng Chen1, Yabing Wang2.
Abstract
Relation extraction is the underlying critical task of textual understanding. However, the existing methods currently have defects in instance selection and lack background knowledge for entity recognition. In this paper, we propose a knowledge-based attention model, which can make full use of supervised information from a knowledge base, to select an entity. We also design a method of dual convolutional neural networks (CNNs) considering the word embedding of each word is restricted by using a single training tool. The proposed model combines a CNN with an attention mechanism. The model inserts the word embedding and supervised information from the knowledge base into the CNN, performs convolution and pooling, and combines the knowledge base and CNN in the full connection layer. Based on these processes, the model not only obtains better entity representations but also improves the performance of relation extraction with the help of rich background knowledge. The experimental results demonstrate that the proposed model achieves competitive performance.Entities:
Mesh:
Year: 2019 PMID: 31396271 PMCID: PMC6664687 DOI: 10.1155/2019/6789520
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1A knowledge base fragment and the expression of the entities in the texts.
Figure 2Textual relation representation and relative distances.
Figure 3Knowledge-based attention model.
Figure 4Text representation model based on CNN.
Figure 5The architecture of the dual CNN.
Relevant dataset information.
| Datasets | Average sentence length | Number of sentences | Number of sentences in the test set |
|---|---|---|---|
| SST-1 | 18 | 11855 | 2210 |
| SST-2 | 19 | 9613 | 1821 |
| TREC | 10 | 5952 | 500 |
Figure 6Effect of distance decay on the generalization ability of the model. (a) Dataset SST-1. (b) Dataset SST-2. (c) Dataset TREC.
Precision values for the top 100, 200, and 300 extracted relations.
| Accuracy (%) | Top 100 | Top 200 | Top 300 | Average |
|---|---|---|---|---|
| Word embedding | 0.74 | 0.72 | 0.64 | 0.70 |
| Knowledge-based attention model | 0.83 | 0.79 | 0.72 | 0.78 |
Figure 7Aggregate precision/recall curves for a variety of methods.
Figure 8Comparison of the model performances for the datasets.