| Literature DB >> 35494798 |
Abstract
The whole sentence representation reasoning process simultaneously comprises a sentence representation module and a semantic reasoning module. This paper combines the multi-layer semantic representation network with the deep fusion matching network to solve the limitations of only considering a sentence representation module or a reasoning model. It proposes a joint optimization method based on multi-layer semantics called the Semantic Fusion Deep Matching Network (SCF-DMN) to explore the influence of sentence representation and reasoning models on reasoning performance. Experiments on text entailment recognition tasks show that the joint optimization representation reasoning method performs better than the existing methods. The sentence representation optimization module and the improved optimization reasoning model can promote reasoning performance when used individually. However, the optimization of the reasoning model has a more significant impact on the final reasoning results. Furthermore, after comparing each module's performance, there is a mutual constraint between the sentence representation module and the reasoning model. This condition restricts overall performance, resulting in no linear superposition of reasoning performance. Overall, by comparing the proposed methods with other existed methods that are tested using the same database, the proposed method solves the lack of in-depth interactive information and interpretability in the model design which would be inspirational for future improving and studying of natural language reasoning.Entities:
Keywords: Characterization inference; Deep fusion matching network; Joint-Optimization of Multi-layer semantics; Meta-learning; Natural language reasoning
Year: 2022 PMID: 35494798 PMCID: PMC9044352 DOI: 10.7717/peerj-cs.908
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
Sample data of SNLI dataset.
| Premise sentence | Label | Hypothetical sentence |
|---|---|---|
| Two women are embracing while holding to-go packages | Entailment | Two women are holding packages |
| A man sold donuts to a customer during a world exhibition event held in the city of Angeles | Contradiction | A woman drinks her coffee in a small café |
| A man in a blue shirt is standing in front of a garage-like structure painted with geometric designs | Neutral | A man is repainting a garage |
Sample data of Multi-NLI dataset.
| Categories | Premise sentence | Label | Hypothetical sentence |
|---|---|---|---|
| Novels | The Old One always comforted Ca’daan, except today | Neutral | Ca’daan knew the Old One very well |
| Letters | Your gift is appreciated by every student who will benefit from your generosity | Neutral | Hundreds of students will benefit from your generosity |
| Telephone | Yes, now you know if everybody like in August when everybody’s on vacation or something we can dress a little more casual or | Contradiction | August is a blackout month for vacations in the company |
Figure 1Visualization results of the dependency syntax tree.
Figure 2Network structure without module optimization.
Figure 3Network structure diagram of sentence representation module optimization.
Figure 4Network structure diagram of relation reasoning module optimized separately.
Figure 5Network structure of combined optimization scheme.
Figure 6Network structure of joint optimization scheme.
Accuracy of each optimization scheme on the SNLI dataset.
| Program | Training set (%) | Test set% |
|---|---|---|
| 300D Tree-CNN ( | 83.3 | 82.1 |
| 300D NSE ( | 86.2 | 84.6 |
| 100D LSTMs with attention ( | 85.3 | 83.5 |
| 100D Deep Fusion LSTM ( | 85.2 | 84.6 |
| 300D Matching-LSTM ( | 92.0 | 86.1 |
| 200D Decomposable Attention Models ( | 90.5 | 86.8 |
| 300D Re-read LSTM ( | 90.7 | 87.5 |
| 600D ESIM ( | 92.6 | 88.0 |
| AF-DMN ( | 94.5 | 88.6 |
| Modular optimization (This paper) | 90.4 | 85.0 |
| Sentence representation module optimized (This paper) | 91.7 | 86.1 |
| The relational reasoning module optimized (This paper) | 95.8 | 89.0 |
| Combinatorial optimization (This paper) | 96.4 | 89.4 |
| Joint optimization (This paper) | 96.2 | 88.8 |
Accuracy of each model on the Multi-NLI dataset.
| Program | Matching set% | Unmatched set (%) |
|---|---|---|
| CBOW ( | 64.8 | 64.5 |
| BiLSTM ( | 66.9 | 66.9 |
| ESIM ( | 76.8 | 75.8 |
| AF-DMN ( | 76.9 | 76.3 |
| Modular optimization | 66.9 | 66.9 |
| Sentence representation module optimized | 73.6 | 73.8 |
| The relational reasoning module is optimized | 77.1 | 75.3 |
| Combinatorial optimization | 77.5 | 75.5 |
| Joint optimization | 77.3 | 75.1 |
Comparison of the influence of different optimization methods on sentence representation reasoning method.
| Optimization mode | Accuracy rate (%) | Training time (H) | Hyperparameters |
|---|---|---|---|
| Combinatorial optimization | 89.4 | 32.87 | 49,196,195 |
| Joint optimization | 88.8 | 30.64 | 48,077,358 |
Figure 7Learning curve of joint optimization and combination optimization on the SNLI dataset.
Figure 8Learning curve of joint optimization and single optimization on the SNLI dataset.
Figure 9Learning curve of joint optimization model with different semantic levels on the SNLI dataset.
Figure 10Learning curve of joint optimization model with different matching levels on the SNLI dataset.