| Literature DB >> 29297367 |
Zhehuan Zhao1, Zhihao Yang2, Ling Luo1, Lei Wang3, Yin Zhang4, Hongfei Lin1, Jian Wang1.
Abstract
BACKGROUND: Automatic disease named entity recognition (DNER) is of utmost importance for development of more sophisticated BioNLP tools. However, most conventional CRF based DNER systems rely on well-designed features whose selection is labor intensive and time-consuming. Though most deep learning methods can solve NER problems with little feature engineering, they employ additional CRF layer to capture the correlation information between labels in neighborhoods which makes them much complicated.Entities:
Keywords: Convolutional neural network; Deep learning multiple label strategy; Disease; Named entity recognition
Mesh:
Year: 2017 PMID: 29297367 PMCID: PMC5751782 DOI: 10.1186/s12920-017-0316-8
Source DB: PubMed Journal: BMC Med Genomics ISSN: 1755-8794 Impact factor: 3.063
Fig. 1The processing flow of our method
Fig. 2Generation of the character-level representation using convolutional neural network
Fig. 3The architecture of MCNN
The legal and ill-legal sequences
| Sequence | Legal | Ill-legal |
|---|---|---|
| B, O, … | * | |
| B, B, O, … | * | |
| B, I, O, … | * | |
| B, I, B, I, O,… | * | |
| O, I, O, … | * | |
| O, I, B, O, … | * |
The statistics of CDR and NCBI corpora
| Corpus | Training | Development | Test | |
|---|---|---|---|---|
| CDR | Abstract | 500 | 500 | 500 |
| Mention | 4182 | 4244 | 4424 | |
| NCBI | Abstract | 593 | 100 | 100 |
| Mention | 5145 | 787 | 960 |
The hyper-parameters and corresponding values
| Hyper-parameter | Value |
|---|---|
| Input context window size | 13 (NCBI = CDR) |
| Word-level embedding dimension | 200 (NCBI = CDR) |
| Character-level embedding dimension | 20 (NCBI = CDR) |
| Lexicon feature embedding dimension | 5 (NCBI = CDR) |
| Character-level CNN’s window size | 3 (NCBI = CDR) |
| Character-level CNN’s filters number | 20 (NCBI = CDR) |
| Word-level CNN’s window size | 3 (NCBI = CDR) |
| Word-level CNN’s filters number | 100 (NCBI = CDR) |
| Word-level Convolutional layers size | 3 (NCBI); 4 (CDR) |
Performance comparisons on NCBI and CDR corpora
| Corpus | Method | P | R | F |
|---|---|---|---|---|
| NCBI | BANNER [ | 83.80 | 80.00 | 81.80 |
| Bi-LSTM + WE [ | 84.87 | 74.11 | 79.13 | |
|
|
|
|
| |
| MCNN* | 83.74 | 83.03 | 83.39 | |
| CDR | HITSZ_CDR [ | 88.68 | 85.23 | 86.93 |
| Lee et al.’s [ | 87.34 | 83.75 | 85.51 | |
| CRD-DNER [ | 79.49 | 73.58 | 76.42 | |
|
|
|
|
|
Note. MCNN* is the version of removing the lexicon feature embedding and the post-processing step
The effect analysis of each feature/strategy
| Corpus | Feature/strategy | P | R | F-score | Δ |
|---|---|---|---|---|---|
| NCBI | None | 85.08 | 85.26 | 85.17 | – |
| Character-level | 84.50 | 84.41 | 84.46 | 0.71 | |
| Lexicon | 83.53 | 83.35 | 83.44 | 1.73 | |
| MLS | 83.97 | 84.41 | 84.19 | 0.98 | |
| Post-processing | 84.84 | 84.84 | 84.84 | 0.33 | |
| CDR | None | 88.20 | 87.46 | 87.83 | – |
| Character-level | 87.25 | 87.35 | 87.30 | 0.53 | |
| Lexicon | 85.48 | 84.90 | 85.19 | 2.64 | |
| MLS | 87.18 | 86.17 | 86.67 | 1.16 | |
| Post-processing | 86.67 | 86.69 | 86.68 | 1.15 |
Notes. Δdenotes the corresponding F-score decrease when a strategy or a feature is removed