| Literature DB >> 35830434 |
Abstract
Currently, both manual and automatic evaluation technology can evaluate the translation quality of unreferenced English articles, playing a particular role in detecting translation results. Still, their deficiency is the lack of a close or noticeable relationship between evaluation time and evaluation theory. Thereupon, to realize the automatic Translation Quality Assessment (TQA) of unreferenced English articles, this paper proposes an automatic TQA model based on Sparse AutoEncoder (SAE) under the background of Deep Learning (DL). Meanwhile, the DL-based information extraction method employs AutoEncoder (AE) in the bilingual words' unsupervised learning stage to reconstruct the translation language vector features. Then, it imports the translation information of unreferenced English articles into Bilingual words and optimizes the extraction effect of language vector features. Meantime, the translation language vector feature is introduced into the automatic DL-based TQA. The experimental findings corroborate that when the number of sentences increases, the number of actual translation errors and the evaluation scores of the proposed model increase, but the Bilingual Evaluation Understudy (BLEU) score is not significantly affected. When the number of sentences increases from 1,000 to 6,000, the BLEU increases from 96 to 98, which shows that the proposed model has good performance. Finally, the proposed model can realize the high-precision TQA of unreferenced English articles.Entities:
Mesh:
Year: 2022 PMID: 35830434 PMCID: PMC9278734 DOI: 10.1371/journal.pone.0270308
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Fig 1BPNN structure.
Fig 2AE network structure.
Fig 3Network structure.
Algorithm flow chart.
| Input: | |
|---|---|
| Train: | The training set is used to train the network model and determine its parameters. |
| Valid: | The validation set is used to determine the optimal network model. |
| Test: | The test set tests the classification ability of the network model with different classes. |
| pretrain_lr | Pre-training learning rate. |
| finetune_lr | Fine-tuning learning rate. |
| pretrain_epoches: | Pre-training number of iterations. |
| training_epoches: | The maximum iterations in fine-tuning phase |
| Rho: | Sparsity parameter of SAE |
| Beta: | Control the weight coefficient of the sparsity penalty term |
| Output: | |
| validation_score: | Validation set error rate |
| test performance: | Test set error rate |
| Method: | |
| 1) | A two-layer stacking network model is constructed to determine the size of the block minibatch of the three data sets. |
| 2) | Through the network model constructed in (1), the calculation method of the pre-training loss function of the model is obtained. |
| 3) | For hidden layer |
| For an epoch in pretrain_epoches: | |
| For a batch_index in minibatch: | |
| Call the loss function calculation method obtained in (2) to calculate the loss after each hidden layer is encoded and decoded. Use the BP algorithm to adjust the model parameters. | |
| 4) | Output the average loss of an epoch in each layer. |
| 5) | The model parameters are determined through the network model constructed in (1). The calculation method of the fine-tuning stage of the model is obtained. |
| 6) | Do |
| 7) | Keep iterating to find the best_validation_loss of this_validation_loss in the test set. Update best_validation_loss. |
| 8) | When best_validation_loss is updated, calculate the error rate in the test |
| 9) | Until iteration epoch>training_epoches or done_looping = True |
| 10) | Output validation_score and test performance |
Fig 4Learning diagram.
a: SL A vector reconstruction; b: TL B vector reconstruction.
Fig 5Statement details on news website.
Fig 6Translation details.
Fig 7Evaluation results of the proposed model.
Fig 8Translation quality.
a: declarative sentences, interrogative sentences, and compound sentences; b: special usage sentences.
Fig 9Proposed model evaluation performance affected by the number of sentences.
Fig 10Automatic evaluation results of the TQA model.
a: vocabulary; b: discourse; c: grammar.
Fig 11Comparison of model accuracy under different numbers of sentences.
Fig 12Comparison of model quality evaluation efficiency under different number of sentences.
Fig 13Comparison of experimental results.