| Literature DB >> 35721417 |
Fanqi Meng1,2, Yujie Zheng1, Songbin Bao3, Jingdong Wang1, Shuaisong Yang1.
Abstract
Formulaic language is a general term for ready-made structures in a language. It usually has fixed grammatical structure, stable language expression meaning and specific use context. The use of formulaic language can coordinate sentence generation in the process of writing and communication, and can significantly improve the idiomaticity and logic of machine translation, intelligent question answering and so on. New formulaic language is generated almost every day, and how to accurately identify them is a topic worthy of research. To this end, this article proposes a formulaic language identification model based on GCN fusing associated information. The innovation is that each sentence is constructed into a graph in which the nodes are part-of-speech features and semantic features of the words in the sentence and the edges between nodes are constructed according to mutual information and dependency syntactic relation. On this basis, the graph convolutional neural network is adopted to extract the associated information between words to mine deeper grammatical features. Therefore, it can improve the accuracy of formulaic language identification. The experimental results show that the model in this article is superior to the classical formulaic language identification model in terms of accuracy, recall and F1-score. It lays a foundation for the follow-up research of formulaic language identification tasks. ©2022 Meng et al.Entities:
Keywords: Associated information; Dependency syntactic relation; Formulaic language; Graph convolutional neural network; Mutual information
Year: 2022 PMID: 35721417 PMCID: PMC9202615 DOI: 10.7717/peerj-cs.984
Source DB: PubMed Journal: PeerJ Comput Sci ISSN: 2376-5992
Figure 1The overall structure of the model.
The examples of Part-of-speech analysis result.
| Formulaic language | Part-of-speech tagging | Sentence structure |
|---|---|---|
| X is fundamental to | NN VBZ JJ TO | Subject-Link verb-Predicative Structure |
| X plays a vital role in the metabolism of | NN VBZ DT JJ NN IN DT NN IN | Subject Verb Object Structure |
| Several attempts have been made to | JJ NNS VBP VBN VBN TO | Subject Verb Object Structure |
| In this innovative study, Smith showed that Y | IN DT JJ NN FW FW NN VBD IN NN | Clauses guided by “that” |
| There were some negative comments about Y | EX VBD DT JJ NNS IN NN | There be... |
Figure 2Structure diagram of Bi-LSTM model.
Figure 3Structure diagram of early fusion model.
Figure 4Structure diagram of late fusion model.
Figure 5Examples of dependency syntactic relation.
Figure 6Adjacency matrix constructed based on dependency syntactic parsing.
Figure 7The structure diagram of GCN.
Figure 8The structure diagram of linear-chain CRF.
Comparison of experimental results of ablation experiment.
| Number | Method | Precision | Recall | F1-score | Train_time | Test_time |
|---|---|---|---|---|---|---|
| 1 | Before_Bi-LSTM | 62.12 | 81.46 | 70.49 | 9.40775 | 3.83115 |
| 2 | Before_CNN | 34.23 | 50.33 | 40.75 | 11.24228 | 5.06429 |
| 3 | After_Bi-LSTM | 77.08 | 73.51 | 75.25 | 11.96498 | 5.16604 |
| 4 | After_Bi-LSTM_CNN | 72.46 | 66.23 | 69.20 | 22.86475 | 8.34037 |
| 5 | Bi-LSTM_DS_GCN | 65.22 | 69.54 | 67.31 | 12.20419 | 4.36681 |
| 6 | Bi-LSTM_MI_GCN | 74.19 | 76.16 | 75.16 | 11.19024 | 3.63964 |
| 7 | Bi-LSTM_MI_DS_GCN |
|
|
| 13.55199 | 5.35704 |
Figure 9Comparison of experimental results of ablation experiment.
Figure 10Influence of different network layers on GCN based on dependency parsing.
Figure 11Influence of different network layers on GCN based on mutual information.
Figure 12Comparison of experimental results of three models.
Examples of formulaic language recognition errors.
| Text | Formulaic language | Output |
|---|---|---|
| It is a nontrivial task since we do not have the ground truth labels to decide the adjustment. | It is a nontrivial task since | It is a nontrivial task |
| In section 7 we compare these two variants against the regular hash function as well as to other leading hashing schemes found in the literature. | As well as | As well as to |
| We therefore resort to an approximation of the process which has a negligible impact on the overall precision but greatly reduces run time. | Has a negligible impact on | Resort to; has a negligible impact on |