| Literature DB >> 31798437 |
Jiayuan Mao1, Yuan Yao1, Stefan Heinrich2, Tobias Hinz2, Cornelius Weber2, Stefan Wermter2, Zhiyuan Liu1, Maosong Sun1.
Abstract
The problem of generating structured Knowledge Graphs (KGs) is difficult and open but relevant to a range of tasks related to decision making and information augmentation. A promising approach is to study generating KGs as a relational representation of inputs (e.g., textual paragraphs or natural images), where nodes represent the entities and edges represent the relations. This procedure is naturally a mixture of two phases: extracting primary relations from input, and completing the KG with reasoning. In this paper, we propose a hybrid KG builder that combines these two phases in a unified framework and generates KGs from scratch. Specifically, we employ a neural relation extractor resolving primary relations from input and a differentiable inductive logic programming (ILP) model that iteratively completes the KG. We evaluate our framework in both textual and visual domains and achieve comparable performance on relation extraction datasets based on Wikidata and the Visual Genome. The framework surpasses neural baselines by a noticeable gap in reasoning out dense KGs and overall performs particularly well for rare relations.Entities:
Keywords: inductive logic programming; information extraction; knowledge graphs; relation learning; relation prediction
Year: 2019 PMID: 31798437 PMCID: PMC6861514 DOI: 10.3389/fnbot.2019.00093
Source DB: PubMed Journal: Front Neurorobot ISSN: 1662-5218 Impact factor: 2.650
Figure 1The scene.
Figure 2Resulting semantic graph of the scene in Andersen's The Little Match Girl's dream.
Figure 3An illustration on the comparison between typical neural network-based relation extractors and the proposed hybrid relation extractor. (Left) A general framework for typical neural network-based relation extractors. Dependencies or correlations between relation instances are modeled by viewing other relations as the “context.” Dashed connections denote the cross-pair connections, which may involve an attention mechanism (Sorokin and Gurevych, 2017) or message passing (Xu et al., 2017). (Right) The proposed hybrid relation extractor (HRE) working in an iterative manner. With a pair selector collaborating with a predictor, it naturally resolves the dependencies or correlations between relation instances. The pair prediction module works with a relation induction model based on explanatory logical rules. See section 3 for details.
Figure 4Encoder for textual entity pairs. We use the concatenation of marker embedding and word embedding with an LSTM model (Hochreiter and Schmidhuber, 1997; Greff et al., 2016) to encode the feature.
Statistics of the dataset generated from Wikidata.
| Train | 124,212 | 70,598 | 5.51 | 2.47 |
| Test | 31,054 | 29,148 | 5.56 | 2.33 |
#Avg ent. stands for the average number of entities per paragraph, while #Avg pos. rel. refers to the average number of positive relations per paragraph.
Figure 5Precision-recall on the dense test set.
F1 scores on the dense test set.
| Baseline | 0.75 | 0.56 | 0.640 |
| Ctx Attention | 0.52 | 0.621 | |
| HRE | 0.72 | ||
The best results are highlighted in bold.
Recall at different precision levels on the dense test set.
| R@0.60 | 0.674 | 0.740 | |
| R@0.70 | 0.633 | 0.574 | |
| R@0.80 | 0.488 | 0.444 |
The best results are highlighted in bold.
Figure 6Precision-recall on the entire test set.
F1 scores on the entire test set.
| Baseline | 0.64 | 0.61 | 0.634 |
| Ctx Attention | 0.56 | ||
| HRE | 0.60 | 0.634 | |
The best results are highlighted in bold.
Figure 7Encoder for visual entity pairs. We extend the union box encoder proposed by Lu et al. (2016) and add the entity's features (what is it) and its location (where is it) into the embedding vector.
Experimental results of visual relation extraction on the entire Visual Genome test set.
| UnionBox | 0.279 | 0.350 |
| MsgPass | 0.448 | 0.531 |
| Baseline | 0.489 | 0.570 |
| HRE |
We compare our model with UnionBox (Lu et al., .