| Literature DB >> 36268152 |
Abstract
At present, there is a lack of research on Marx's idea of "combining education and productive labor" and its guiding significance for youth labor education, and no effective teaching model has been formed. In response to this problem, this study proposes a semi-supervised deep learning model based on u-wordMixup (SD-uwM). When there is a shortage of labeled samples, semi-supervised learning uses a large number of unlabeled samples to solve the problem of labeling bottlenecks. However, since the unlabeled samples and labeled samples come from different fields, there may be quality problems in the unlabeled samples, which makes the generalization ability of the model worse., resulting in a decrease in classification accuracy. The model uses the u-wordMixup method to perform data augmentation on unlabeled samples. Under the constraints of supervised cross-entropy and unsupervised consistency loss, it can improve the quality of unlabeled samples and reduce overfitting. The comparative experimental results on the AGNews, THUCNews, and 20Newsgroups data sets show that the proposed method can improve the generalization ability of the model and also effectively improve the time performance. The study found that the SD-uwM model uses the u-wordMixup method to enhance the unlabeled samples and combines the idea of the Mean Teacher model, which can significantly improve the text classification performance. The SD-uwM model can improve the generalization ability and time performance of the model, respectively, 86.4 ± 1.3 and 90.5 ± 1.3. Therefore, the use of SD-uwM in Marx's program is of great practical significance for the guidance process of youth labor education.Entities:
Mesh:
Year: 2022 PMID: 36268152 PMCID: PMC9578841 DOI: 10.1155/2022/2576559
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1SD-uwM model.
Figure 2u-wordMixup method.
Figure 3Comparison of training loss between SD-uwM and Mean Teacher when TextCNN is selected: (a) Nl = 200, Nu = 3000 (THUCNews data set); (b) Nl = 300, Nu = 5000 (AGNews data set).
Classification comparison of four models on three classification results.
| Model | Network structure | Accuracy | ||
|---|---|---|---|---|
| AGNews | THUCNews | 20Newsgroups | ||
| SL | LSTM | 75.4 ± 1.1 | 77.5 ± 1.3 | 71.5 ± 1.3 |
| wM-SL | 80.4 ± 1.3 | 83.2 ± 1.2 | 75.4 ± 1.3 | |
| Mean Teacher | 82.1 ± 1.3 | 86.1 ± 1.3 | 77.5 ± 1.1 | |
| SD-uwM | 90.4 ± 1.2 | 91.4 ± 1.3 | 85.4 ± 1.2 | |
| SL | TextCNN | 76.4 ± 1.2 | 78.4 ± 1.4 | 71.2 ± 1.2 |
| wM-SL | 80.5 ± 1.2 | 84.5 ± 1.3 | 75.3 ± 1.2 | |
| Mean Teacher | 83.6 ± 1.1 | 86.1 ± 1.5 | 78.1 ± 1.3 | |
| SD-uwM | 91.2 ± 1.3 | 92.2 ± 1.3 | 86.2 ± 1.1 | |
Figure 4Comparison of macro-F1 value of each model with iteration times on three data sets using LSTM: (a) AGNews data set; (b) THUCNews data set; (c) 20Newsgroups data set.
Figure 5Comparison of the macro-F1 value of each model with the number of iterations on three data sets using TextCNN: (a) AGNews data set; (b) THUCNews data set; (c) 20 Newsgroups data set.
Time performance comparison of SN-uwM and co-training.
| Dataset | Model | Accuracy (%) | Time (s) | Ratio of train time (SD-uwM/co-training) |
|---|---|---|---|---|
| 20 Newsgroups | SD-uwM | 86.4 ± 1.3 | 0.01 | 1/3000 |
| Nl = 200, Nu = 2000 | Co-training | 83.3 ± 1.2 | 30 | |
| THUCNews | SD-uwM | 90.5 ± 1.3 | 0.02 | 1/2200 |
| Nl = 300, Nu = 4000 | Co-training | 88.4 ± 1.2 | 44 |