| Literature DB >> 33286026 |
Min Zhang1,2, Guohua Geng1, Jing Chen1.
Abstract
Increasingly, popular online museums have significantly changed the way people acquire cultural knowledge. These online museums have been generating abundant amounts of cultural relics data. In recent years, researchers have used deep learning models that can automatically extract complex features and have rich representation capabilities to implement named-entity recognition (NER). However, the lack of labeled data in the field of cultural relics makes it difficult for deep learning models that rely on labeled data to achieve excellent performance. To address this problem, this paper proposes a semi-supervised deep learning model named SCRNER (Semi-supervised model for Cultural Relics' Named Entity Recognition) that utilizes the bidirectional long short-term memory (BiLSTM) and conditional random fields (CRF) model trained by seldom labeled data and abundant unlabeled data to attain an effective performance. To satisfy the semi-supervised sample selection, we propose a repeat-labeled (relabeled) strategy to select samples of high confidence to enlarge the training set iteratively. In addition, we use embeddings from language model (ELMo) representations to dynamically acquire word representations as the input of the model to solve the problem of the blurred boundaries of cultural objects and Chinese characteristics of texts in the field of cultural relics. Experimental results demonstrate that our proposed model, trained on limited labeled data, achieves an effective performance in the task of named entity recognition of cultural relics.Entities:
Keywords: bidirectional long short-term memory network; conditional random fields; cultural relics; embeddings from language models; named-entity recognition; semi-supervised learning
Year: 2020 PMID: 33286026 PMCID: PMC7516692 DOI: 10.3390/e22020252
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Figure 1Overview of the semi-supervised bidirectional long short-term memory (BiLSTM)-conditional random fields (CRF) framework for named-entity recognition (NER).
Figure 2Embedding from Language Model.
Figure 3The architecture of embeddings from language models (ELMO)-based BiLSTM-CRF model.
Figure 4Architecture of the LSTM memory.
Performance comparison of semi-supervised cultural relics named-entity recognition (SCRNER) and semi-supervised baseline methods.
| Model | CRN(%) | CRD(%) | UL(%) | MC(%) | All(%) | |||||
|---|---|---|---|---|---|---|---|---|---|---|
| Acc | F1 | Acc | F1 | Acc | F1 | Acc | F1 | Acc | F1 | |
| Xu et al. [ | 76.7 | 76.2 | 78.5 | 76.8 | 84.8 | 83.1 | 82.5 | 83.4 | 80.6 | 79.9 |
| Liao et.al. [ | 79.2 | 77.1 | 81.9 | 79.5 | 87.5 | 84.9 | 85.2 | 85.6 | 83.5 | 81.8 |
| CSCRNER | 82.6 | 80.3 | 82.0 | 80.4 | 88.3 | 87.1 | 87.4 | 86.8 | 85.1 | 83.7 |
| Luan et al. [ | 83.4 | 81.1 | 83.6 | 81.3 | 88.7 | 87.6 | 88.1 | 88.7 | 86.0 | 84.7 |
| SCRNER | 84.1 | 82.2 | 84.0 | 83.6 | 89.2 | 88.9 | 89.4 | 89.6 | 86.7 | 86.1 |
Note. Entity, CRN, cultural relics’ name; CRD, cultural relics’ dynasty; UL, Unearthed location; MC Museum collection; ALL, the average value of four entities. Model. CSCRNER, a semi-supervised algorithm utilizing the classical self-training based on our framework; SCRNER, the model proposed in this study.
Figure 5Performance comparison of the percentage of initial labeled data. (a) Accuracy curves; (b) F1-score curves.
Figure 6Performance comparison of SCNER and word embeddings. Note: Word, the model that uses word embedding as Word representation; Character, the model that uses character embedding as the word representation; W_C, the model that combines word and character embedding; SCRNER, the model proposed in this study. (a) Accuracy curves; (b) F1-score curves.
Figure 7Performance Comparison of BiLSTM-CRF and SCRNER in four entities. Note: BiLSTM_CRF is the framework proposed by Yang H et al. [11].