| Literature DB >> 30223767 |
Wei Zheng1,2, Hongfei Lin3, Xiaoxia Liu1, Bo Xu4.
Abstract
BACKGROUND: The effective combination of texts and knowledge may improve performances of natural language processing tasks. For the recognition of chemical-induced disease (CID) relations which may span sentence boundaries in an article, although existing CID systems explored the utilization for knowledge bases, the effects of different knowledge on the identification of a special CID haven't been distinguished by these systems. Moreover, systems based on neural network only constructed sentence or mention level models.Entities:
Keywords: Attention mechanism; Chemical-induced diseases; Document level; Knowledge; Neural network; Text mining
Mesh:
Year: 2018 PMID: 30223767 PMCID: PMC6142695 DOI: 10.1186/s12859-018-2316-x
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1The overall architecture of the proposed model
The statistics of the CDR corpus
| Dataset | CID pairs | CD pairs | Inter-sentential CID pairs | Intra-sentential CID pairs |
|---|---|---|---|---|
| Training | 1038 | 5432 | 283 | 755 |
| Development | 1012 | 5263 | 246 | 766 |
| Test | 1066 | 5405 | 303 | 763 |
| Total | 3116 | 16,100 | 832 | 2284 |
The column “CD pairs” represents the total number of candidate instances
Hyperparameters
| Parameter Name | Value |
|---|---|
| 100 | |
| 10 | |
| 200 | |
| The number n1 of sentences in an article | 30 |
| The number n2 of words in a sentence | 120 |
| The window size | 5 |
| The number | 300 |
| Mini-batch | 8 |
| The number of hidden units of two LSTMs | 220,440 |
| The learning rate | 0.001 |
| The dropout rate | 0.5 |
Fig. 2Performance evaluation for the dimension of the word embedding on the test set of the CDR corpus
Fig. 3Performance evaluation for the number f of filters on the test set of the CDR corpus
Performance evaluation for different initialization methods of the knowledge embedding on the test set of the CDR corpus
| Methods | P(%) | R(%) | F(%) |
|---|---|---|---|
| Random | 57.1 | 81.0 | 67.0 |
| TransE | 61.5 | 78.7 | 69.0 |
The post processing step wasn’t applied to the experimental results in this table
Fig. 4Performance evaluation for the dimension of the knowledge embedding on the test set of the CDR corpus
Performance changes with different input representations and post processing on the test set of the CDR corpus
| Feature | P(%) | R(%) | F(%) |
|---|---|---|---|
| (1): word | 52.0 | 64.9 | 57.7 |
| (2): (1) + CTD | 59.5 | 74.8 | 66.3 |
| (3): (2) + POS | 61.5 | 78.7 | 69.0 |
| (4): (3) + PP | 64.4 | 76.7 | 70.0 |
Performance changes with different components of the document level sub-network on the test set of the CDR corpus when knowledge isn’t incorporated
| Architecture | P(%) | R(%) | F(%) |
|---|---|---|---|
| (1): lstm+lstm | 48.0 | 62.8 | 54.4 |
| (2): lstm+cnn | 54.8 | 59.2 | 56.9 |
| (3): lstm+cnnlstm | 56.5 | 61.3 | 58.8 |
| (4): lstm+cnnlstm+topic | 54.3 | 65.9 | 59.5 |
The post processing step wasn’t applied to the experimental results in this table
Performance changes with the different final representations of knowledge on the test set of the CDR corpus
| Methods | P(%) | R(%) | F(%) |
|---|---|---|---|
| (1): Without KB | 54.3 | 65.9 | 59.5 |
| (2): Con | 61.5 | 74.9 | 67.5 |
| (3): Sum | 65.3 | 70.4 | 67.8 |
| (4):ATT_KB_Con | 59.6 | 79.5 | 68.1 |
| (5): ATT_KB_Max | 60.6 | 69.9 | 64.9 |
| (6): ATT_KB_Sum | 61.5 | 78.7 |
|
The post processing step wasn’t applied to the experimental results in this table. The highest F-score is highlighted in bold
Fig. 5Attention value learned by the model with the approach “ATT_KB_Sum” for chemical and disease pairs
Performance comparisons with relevant systems using gold standard entity annotations on the test dataset of the CDR corpus
| Methods | System | Methods | Text and concept level | P(%) | R(%) | F(%) |
|---|---|---|---|---|---|---|
| NN with KB | ATT_KB_sum | LSTM+CNN + CTD | Doc_E | 60.7 | 78.7 |
|
| LSTM+CNN + CTD + pp | 63.6 | 76.8 |
| |||
| Li et al. [ | CNN | Doc_M | 57.8 | 54.2 | 55.9 | |
| CNN + CTD | 60.0 | 81.5 | 69.1 | |||
| Verga et al. [ | Transformer | Doc_E | 55.6 | 70.8 | 62.1 | |
| Transformer+ Extra data | 64.0 | 69.2 | 66.2 | |||
| Tradional ML with KB | Alam et al. [ | SVM + CTD + pp | Doc_E + Sen_M | 43.7 | 80.4 | 56.6 |
| Xu et al. [ | SVM + CTD + SIDER+MEDI | Doc_E + Sen_M | 65.8 | 68.6 | 67.2 | |
| Pons et al. [ | SVM + Graph DB | Doc_E | 73.1 | 67.6 | 70.2 | |
| Peng et al. [ | SVM + CTD + Rules | Doc_E | 68.2 | 66.0 | 67.1 | |
| SVM + CTD + Rules +Extra data | 71.1 | 72.6 |
| |||
| Lowe et al. [ | rules+Ontology+WIKI+PP | Sen_M | 59.3 | 62.3 | 60.8 | |
| NN without KB | Gu et al. [ | CNN + ME+pp | Doc_M + Sen_M | 55.7 | 68.1 |
|
| Zhou et al. [ | LSTM+SVM + pp | Sen_M | 55.6 | 68.4 |
| |
| Gu et al. [ | ME | Doc_M + Sen_M | 62.0 | 55.1 | 58.3 |
The 4-th column denotes the text level and the concept level when candidate instances are constructed. “Doc” denotes the document level, “Sen” denotes the sentence level, “_E” denotes entity-based candidate pairs and “_M” denotes mention-based candidate pairs. In addition, all results listed in this table come from the corresponding improved systems after the CDR challenge. The highest F-scores in each group of methods are highlighted in bold
The recognizing performance of the inter-sentential and intra-sentential CIDs before and after knowledge is introduced into the proposed model
| Knowledge | CIDs | P(%) | R(%) | F(%) | TP | FP | POS |
|---|---|---|---|---|---|---|---|
| Without KB | inter-sentential | 45.0 | 42.9 | 43.9 | 130 | 159 | 303 |
| intra-sentential | 59.5 | 77.9 | 67.4 | 594 | 405 | 763 | |
| With KB | inter-sentential | 52.8 | 60.1 | 56.2 | 182 | 163 | 303 |
| intra-sentential | 68.8 | 83.4 | 75.4 | 636 | 289 | 763 |
The experiments were performed when the new development set is the subset with the index 0 (similarly hereinafter). TP, FP and POS denotes the number of predicted true positive instances, predicted false positive instances and true positive instances of the test dataset, respectively
The recognizing performance on the test dataset of the CDR corpus when different training sets were applied without postprocessing
| Training set | Knowledge | P(%) | R(%) | F(%) |
|---|---|---|---|---|
| Only original training set | Without KB | 45.2 | 68.1 | 54.3 |
| With KB | 55.8 | 81.4 | 66.1 | |
| Train +development set | Without KB | 54.3 | 65.9 | 59.5 |
| With KB | 61.5 | 78.7 | 69.0 |