| Literature DB >> 28415073 |
Jinghang Gu1, Fuqing Sun2, Longhua Qian1, Guodong Zhou1.
Abstract
This article describes our work on the BioCreative-V chemical-disease relation (CDR) extraction task, which employed a maximum entropy (ME) model and a convolutional neural network model for relation extraction at inter- and intra-sentence level, respectively. In our work, relation extraction between entity concepts in documents was simplified to relation extraction between entity mentions. We first constructed pairs of chemical and disease mentions as relation instances for training and testing stages, then we trained and applied the ME model and the convolutional neural network model for inter- and intra-sentence level, respectively. Finally, we merged the classification results from mention level to document level to acquire the final relations between chemical and disease concepts. The evaluation on the BioCreative-V CDR corpus shows the effectiveness of our proposed approach. Database URL: http://www.biocreative.org/resources/corpora/biocreative-v-cdr-corpus/.Entities:
Mesh:
Year: 2017 PMID: 28415073 PMCID: PMC5467558 DOI: 10.1093/database/bax024
Source DB: PubMed Journal: Database (Oxford) ISSN: 1758-0463 Impact factor: 3.451
The CID relation statistics on the corpus
| Task datasets | No. of articles | No. of CID relations |
|---|---|---|
| Training | 500 | 1038 |
| Development | 500 | 1012 |
| Test | 500 | 1066 |
Figure 1.The system workflow diagram.
Figure 2.The architecture of our CNN-based model for the intra-sentence level relation extraction. (a) The overall architecture of our CNN-based model; (b) the CNN model for dependency features.
The contextual features
| No. | Features |
|---|---|
| L1 | Chemical mention |
| L2 | Disease mention |
| L3 | ( |
| L4 | ( |
| L5 | Verbs in between |
Figure 3.The dependency parsing tree of the example sentence.
The dependency paths of the example sentence
| Name | Dependency paths |
|---|---|
| R2C | ROOT↓root↓induced↓nsubj↓dipyridamole |
| R2D | ROOT↓root↓induced↓dobj↓hyperemia |
| C2D | dipyridamole↑nsubj↑induced↓dobj↓hyperemia |
The performance of the CNN-based model on the test dataset at intra-sentence level
| Methods | |||
|---|---|---|---|
| Contextual | 54.8 | 54.9 | 54.8 |
| Dependency | 51.3 | 57.5 | 54.2 |
| Contextual + dependency | 59.7 | 55.0 | 57.2 |
The overall performance on the test dataset
| Methods | |||
|---|---|---|---|
| Inter-sentence level | 51.9 | 7.0 | 11.7 |
| Intra-sentence level | 59.7 | 55.0 | 57.2 |
| Relation merging | 60.9 | 59.5 | 60.2 |
| Post-processing | 55.7 | 68.1 | 61.3 |
Figure 4.The effect of the hyper-parameter w on the development dataset.
Figure 5.The effect of the hyper-parameter v on the development dataset.
Comparisons with the related works
| Methods | System | Description | |||
|---|---|---|---|---|---|
| ML without KB | Ours | CNN | 59.7 | 55.0 | 57.2 |
| CNN+ME | 60.9 | 59.5 | 60.2 | ||
| CNN+ME+PP | 55.7 | 68.1 | 61.3 | ||
| Zhou | CNN | 41.1 | 55.3 | 47.2 | |
| LSTM | 54.9 | 51.4 | 53.1 | ||
| LSTM+SVM | 64.9 | 49.3 | 56.0 | ||
| LSTM+SVM+PP | 55.6 | 68.4 | 61.3 | ||
| Gu | ME | 62.0 | 55.1 | 58.3 | |
| Xu | SVM | 59.6 | 44.0 | 50.7 | |
| ML with KB | Alam | SVM+KBs | 43.7 | 80.4 | 56.6 |
| Xu | SVM+KBs | 65.8 | 68.6 | 67.2 | |
| Pons | SVM+KBs | 73.1 | 67.6 | 70.2 | |
| Peng | SVM+KBs | 68.2 | 66.0 | 67.1 | |
| Extra training data+SVM+KBs | 71.1 | 72.6 | 71.8 | ||
| Rule based | Lowe | Heuristic rules | 59.3 | 62.3 | 60.8 |