| Literature DB >> 36196376 |
Bo Ning1, Deji Zhao1, Xinyi Liu1, Guanyu Li1.
Abstract
Multi-turn dialogue generation is an essential and challenging subtask of text generation in the question answering system. Existing methods focused on extracting latent topic-level relevance or utilizing relevant external background knowledge. However, they are prone to ignore the fact that relying too much on latent aspects will lose subjective key information. Furthermore, there is not so much relevant external knowledge that can be used for referencing or a graph that has complete entity links. Dependency tree is a special structure that can be extracted from sentences, it covers the explicit key information of sentences. Therefore, in this paper, we proposed the EAGS model, which combines the subjective pivotal information from the explicit dependency tree with sentence implicit semantic information. The EAGS model is a knowledge graph enabled multi-turn dialogue generation model, and it doesn't need extra external knowledge, it can not only extract and build a dependency knowledge graph from existing sentences, but also prompt the node representation, which is shared with Bi-GRU each time step word embedding in node semantic level. We store the specific domain subgraphs built by the EAGS, which can be retrieved as external knowledge graph in the future multi-turn dialogue generation task. We design a multi-task training approach to enhance semantics and structure local feature extraction, and balance with the global features. Finally, we conduct experiments on Ubuntu large-scale English multi-turn dialogue community dataset and English Daily dialogue dataset. Experiment results show that our EAGS model performs well on both automatic evaluation and human evaluation compared with the existing baseline models.Entities:
Keywords: Build knowledge graph; Dependency tree; Knowledge graph; Natural language processing; Question answering systems; Text generation
Year: 2022 PMID: 36196376 PMCID: PMC9523637 DOI: 10.1007/s11280-022-01100-8
Source DB: PubMed Journal: World Wide Web ISSN: 1386-145X Impact factor: 3.000
In an example of a multi-turn dialogue on the daily dialogue dataset, we split the multi-turn dialogue into many utterences and current turns
| Contexts | Examples |
|---|---|
| Utterance 1 | The prices are given without engagement. |
| Utterance 2 | Good, if you’ll excuse me. I’ll go over the sheet right now. |
| Utterance 3 | Take your time. |
| Utterance 4 | I can tell you at a glance that your prices are much too high. |
| Utterance 5 | You know that the cost of production |
| has been skyrocketing in recent years. | |
| Current Turn | We only ask that your prices be comparable to others. |
| Response | Well, we can consider making some concessions in our price. |
This is an example in the Daily dialogue dataset. The sentences shown are selected. The sentences in the example use word segmentation and sentence segmentation
Figure 1The architecture of EAGS model. The solid lines represent the direction of vector flow, the dotted lines represent the addition of multi-task losses. The blue circles represent the words in a sentence, the box composed with green circles represents the word distribution of vectors
An example in Daily dialogue dataset
| Contexts | Examples |
|---|---|
| Utterance 5 | You know that the cost of production |
| has been skyrocketing in recent years. | |
| Current Turn | We only ask that your prices be comparable to others. |
Figure 2Examples of converting sentences into graphs
Figure 3The message passing mechanism of GCN usually has many layers in a GCN network. We use the last layer as the final representation of nodes
Performance of different models on Ubuntu dataset and Daily dialogue dataset
| Ubuntu | Daily dialogue | |||||||
|---|---|---|---|---|---|---|---|---|
| Model | PPL | BLEU | Dist-1 | Dist-2 | PPL | BLEU | Dist-1 | Dist-2 |
| Seq2Seq | 133.274 | 0.4813 | 0.885 | 0.996 | 190.147 | 0.5105 | 0.683 | 0.762 |
| HRED | 165.112 | 0.5255 | 1.021 | 2.112 | 178.348 | 0.5428 | 0.745 | 0.984 |
| VHRED | 187.054 | 0.5048 | 1.125 | 2.132 | 175.334 | 0.5610 | 0.758 | 1.032 |
| ReCoSa | 99.358 | 1.1176 | 1.718 | 3.768 | 138.535 | 1.1828 | 1.184 | 2.382 |
| STAR-BTM | 111.242 | 1.2897 | 1.601 | 4.525 | 132.101 | 1.2105 | 1.201 | 2.421 |
| CHG | 108.796 | 1.2814 | 1.5368 | 4.438 | 122.304 | 1.2236 | 1.348 | 2.404 |
| HSAN | 101.601 | 1.3087 | 1.5971 | 4.527 | 120.114 | 1.2560 | 1.427 | 2.488 |
| EAGS | 92.884 | 1.3401 | 1.6613 | 4.558 | 108.561 | 1.3474 | 1.454 | 2.605 |
Comparison of our complete model and simple fusion graph information model on Ubuntu and Daily dialogue dataset
| Ubuntu | Daily dialogue | |||||||
|---|---|---|---|---|---|---|---|---|
| Model | PPL | BLEU | Dist-1 | Dist-2 | PPL | BLEU | Dist-1 | Dist-2 |
| Seq2Seq+G | 141.285 | 0.5204 | 0.941 | 1.128 | 199.325 | 0.5423 | 0.714 | 0.851 |
| HRED+G | 188.483 | 0.5417 | 1.124 | 2.228 | 185.159 | 0.5505 | 0.751 | 1.005 |
| VHRED+G | 188.265 | 0.5715 | 1.230 | 2.304 | 183.257 | 0.5492 | 0.757 | 1.087 |
| ReCoSa+G | 111.247 | 1.2231 | 1.804 | 3.815 | 145.232 | 1.2107 | 1.203 | 2.501 |
| STAR-BTM+G | 124.753 | 1.2874 | 1.719 | 4.579 | 149.528 | 1.2380 | 1.245 | 2.527 |
| CHG+G | 114.651 | 1.3018 | 1.624 | 4.474 | 132.470 | 1.2493 | 1.441 | 2.512 |
| HSAN+G | 105.790 | 1.3242 | 1.632 | 4.508 | 124.081 | 1.2807 | 1.437 | 2.538 |
| EAGS | 92.884 | 1.3401 | 1.6613 | 4.558 | 108.561 | 1.3474 | 1.454 | 2.605 |
Attention ablation experiment on Daily dialogue dataset
| Model | Node Attention | Structure Attention | PPL↑ | BLEU↓ | Dist-1↓ | Dist-2↓ |
|---|---|---|---|---|---|---|
| EAGS | ✓ | ✗ | 7.104 | 0.1011 | 0.015 | 0.091 |
| EAGS | ✗ | ✓ | 9.815 | 0.1208 | 0.021 | 0.104 |
| EAGS | ✓ | ✓ | 0 | 0 | 0 | 0 |
Multi-task ablation experiment on Daily dialogue dataset
| Model | Node Decoder | Structure Decoder | PPL↑ | BLEU↓ | Dist-1↓ | Dist-2↓ |
|---|---|---|---|---|---|---|
| EAGS | ✓ | ✗ | 5.662 | 0.0737 | 0.011 | 0.067 |
| EAGS | ✗ | ✓ | 6.918 | 0.0924 | 0.016 | 0.085 |
| EAGS | ✓ | ✓ | 0 | 0 | 0 | 0 |
Modules ablation experiment on Daily dialogue dataset
| Model | Graph Encoder | Node Encoder | Structure Decoder | Node Decoder | PPL↑ | BLEU↓ | Dist-1↓ | Dist-2↓ |
|---|---|---|---|---|---|---|---|---|
| EAGS | ✗ | ✗ | ✗ | ✗ | 30.024 | 0.1517 | 0.027 | 0.216 |
| EAGS | ✗ | ✓ | ✗ | ✓ | 22.280 | 0.1175 | 0.221 | 0.158 |
| EAGS | ✓ | ✗ | ✓ | ✗ | 24.715 | 0.1451 | 0.021 | 0.184 |
| EAGS | ✓ | ✓ | ✓ | ✓ | 0 | 0 | 0 | 0 |
Human evaluation results of mean score, proportions of three levels (+ 2, + 1, and 0 represent excellent, good and average respectively) on Ubuntu dataset
| Appropriateness | Informativeness | Grammaticality | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Model | Means | + 2 | + 1 | 0 | Means | + 2 | + 1 | 0 | Means | + 2 | + 1 | 0 |
| Seq2Seq | 0.54 | 24.9 | 4.2 | 70.9 | 0.39 | 17.9 | 3.2 | 78.9 | 1.48 | 71.5 | 5.0 | 23.5 |
| ReCoSa | 0.69 | 31.9 | 5.2 | 62.9 | 0.49 | 22.9 | 3.2 | 73.9 | 1.72 | 85.7 | 3.6 | 10.7 |
| STAR-BTM | 0.71 | 31.5 | 8.0 | 60.5 | 0.54 | 23.3 | 7.4 | 69.3 | 1.75 | 84.5 | 6.0 | 9.5 |
| CHG | 0.81 | 37.4 | 6.2 | 56.4 | 0.58 | 26.4 | 5.2 | 68.4 | 1.78 | 85.6 | 6.8 | 7.6 |
| HSAN | 0.83 | 38.6 | 5.8 | 55.6 | 0.66 | 30.9 | 4.2 | 64.9 | 1.72 | 81.6 | 8.8 | 9.6 |
| EAGS | 0.85 | 39.7 | 5.6 | 54.7 | 0.76 | 35.9 | 4.2 | 59.9 | 1.85 | 90.1 | 4.8 | 5.1 |
Human evaluation results of mean score, proportions of three levels (+ 2, + 1, and 0 represent excellent, good and average respectively) on Daily dialogue dataset
| Appropriateness | Informativeness | Grammaticality | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Model | Means | + 2 | + 1 | 0 | Means | + 2 | + 1 | 0 | Means | + 2 | + 1 | 0 |
| Seq2Seq | 0.62 | 29.4 | 3.2 | 67.4 | 0.43 | 20.6 | 1.8 | 77.6 | 1.51 | 74.6 | 1.8 | 23.6 |
| ReCoSa | 0.73 | 35.5 | 2.0 | 62.5 | 0.51 | 22.0 | 7.0 | 71.0 | 1.72 | 83.5 | 5.0 | 11.5 |
| STAR-BTM | 0.74 | 34.1 | 5.8 | 60.1 | 0.56 | 25.3 | 5.4 | 69.3 | 1.75 | 84.4 | 6.2 | 9.4 |
| CHG | 0.78 | 35.1 | 7.8 | 57.1 | 0.64 | 28.4 | 7.2 | 64.4 | 1.71 | 80.2 | 10.6 | 9.2 |
| HSAN | 0.77 | 35.1 | 6.8 | 58.1 | 0.62 | 28.7 | 4.6 | 66.7 | 1.75 | 83.2 | 8.6 | 8.2 |
| EAGS | 0.89 | 41.6 | 5.8 | 52.6 | 0.80 | 38.0 | 4.0 | 58.0 | 1.91 | 93.1 | 4.8 | 2.1 |