| Literature DB >> 29017459 |
Wei Zheng1,2, Hongfei Lin3, Ling Luo1, Zhehuan Zhao1, Zhengguang Li1,2, Yijia Zhang1, Zhihao Yang1, Jian Wang1.
Abstract
BACKGROUND: Drug-drug interactions (DDIs) often bring unexpected side effects. The clinical recognition of DDIs is a crucial issue for both patient safety and healthcare cost control. However, although text-mining-based systems explore various methods to classify DDIs, the classification performance with regard to DDIs in long and complex sentences is still unsatisfactory.Entities:
Keywords: Attention; Drug-drug interactions; Long short-term memory; Recurrent neural network; Text mining
Mesh:
Year: 2017 PMID: 29017459 PMCID: PMC5634850 DOI: 10.1186/s12859-017-1855-x
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1The model architecture with input attention. Note: Drug0 and drug1 are the candidate drug pair
Statistics for DDI- 2013 corpus
| Instances | DDI type | DB-2013 | ML-2013 | ||
|---|---|---|---|---|---|
| Training set | Test set | Training set | Test set | ||
| Positive | mechanism | 1257 | 278 | 62 | 24 |
| effect | 1535 | 298 | 152 | 62 | |
| advice | 818 | 214 | 8 | 7 | |
| int | 178 | 94 | 10 | 2 | |
| Total | 3788 | 884 | 232 | 95 | |
| Negative | 22,217 | 4381 | 1555 | 401 | |
| Total | 26,005 | 5265 | 1787 | 434 | |
Hyperparameters
| Parameter | Parameter name | Value |
|---|---|---|
|
| Word emb.Size | 200 |
|
| POS&position emb.Size | 10 |
|
| The length of a pruned sentence | 85 |
| Mini-batch | Minimal batch | 128 |
| LSTM dim. | the number of hidden units | 230 |
Fig. 2Evaluation of the dimensionality of word embedding when our model without the attention mechanism was trained
Fig. 3F-scores of the proposed model as the distance between candidate drugs becomes longer
Performance of different score functions for the DDI classificaton on the Overall-2013 dataset
| Score function | P(%) | R(%) | F(%) |
|---|---|---|---|
| Base_BLSTM | 74.0 | 78.6 | 76.2 |
| dot-score | 78.4 | 76.2 | 77.3 |
| cos-score | 76.3 | 76.5 | 76.4 |
| Tanh-score | 67.9 | 65.9 | 66.9 |
Base_BLSTM is the BLSTM model without an attention mechanism which uses our all preprocessing techniques and all input embeddings including word, POS and position embedding
Fig. 4Input attention visualization. Note: Drug0 and drug1 are the candidate drug pair
Performance changes with different input representations on the overall-2013 dataset
| Input representation | P(%) | R(%) | F(%) |
|---|---|---|---|
| (1): word without attention | 54.7 | 42.8 | 48.0 |
| (2): word + att | 76.5 | 67.5 | 71.7 |
| (3): word + att + pos | 70.9 | 74.7 | 72.7 |
| (4): word + att + position | 79.1 | 73.9 | 76.4 |
| (5): word + att + pos + position | 78.4 | 76.2 | 77.3 |
Every model in this table uses all preprocessing techniques of our approach. Word without attention denotes the model without the attention mechanism which uses only word embedding. Word + att denotes the model which uses the attention mechanism and word embedding
The number of different instances detected by two models for the DDI classificaton on the Overall-2013 dataset
| Model | tp | fp | fn | tp + fn |
|---|---|---|---|---|
| Base_BLSTM | 769 | 270 | 210 | 979 |
| Att-BLSTM | 746 | 205 | 233 | 979 |
tp denotes the number of true-positive instances, fp denotes the number of false-positive instances, and fn denotes the number of false-negative instances
Performance changes with different preprocessing procedures on the overall-2013 dataset
| Processing procedure | P(%) | R(%) | F(%) |
|---|---|---|---|
| (1): only candidate drugs replaced | 75.9 | 68.7 | 71.5 |
| (2): basic processing | 77.5 | 72.3 | 74.8 |
| (3): (2) + Following anaphora | 76.9 | 76.5 | 76.7 |
| (4): (3) + Pruned Sentences | 78.4 | 76.2 | 77.3 |
Every model in this table uses three input embeddings of our approach
Performance comparisons (F-score) with top-ranking systems on the overall-2013 dataset for DDI detection and DDI classification
| Method | Team | CLA | DEC | MEC | EFF | ADV | INT |
|---|---|---|---|---|---|---|---|
| SVM | RAIHANI [ |
| 81.5 |
| 69.6 |
|
|
| Context-Vector [ | 68.4 |
| 66.9 |
| 71.4 | 51.6 | |
| Kim [ | 67.0 | 77.5 | 69.3 | 66.2 | 72.5 | 48.3 | |
| FBK-irst [ | 65.1 | 80.0 | 67.9 | 62.8 | 69.2 | 54.7 | |
| WBI [ | 60.9 | 75.9 | 61.8 | 61.0 | 63.2 | 51.0 | |
| UTurku [ | 59.4 | 69.9 | 58.2 | 60.0 | 63.0 | 50.7 | |
| NN |
|
|
|
| 67.6 |
| 43.1 |
| MCCNN [ | 70.2 | 79.0 | 72.2 | 68.2 | 78.2 |
| |
| Liu CNN [ | 69.8 | – | 70.2 |
| 77.8 | 48.4 | |
| Zhao SCNN [ | 68.6 | 77.2 | – |
| – | – | |
| Ours | Att-BLSTM |
|
|
|
|
|
|
The listed results come from the corresponding papers. The symbol “-” denotes no corresponding values, because the related paper did not provide complete results (similarly hereinafter). “DEC” only indicates DDI detection. “CLA” indicates DDI classification. “MEC”, “EFF”, “ADV” and “INT” denote “mechanism”, “effect”, “advice” and “int” types, respectively. The highest scores are highlighted in bold
Performance comparisons (F-score) with top-ranking systems on the three datasets
| Method | Team | DB-2013 | ML-2013 | Overall-2013 | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| P(%) | R(%) | F(%) | P(%) | R(%) | F(%) | P(%) | R(%) | F(%) | ||
| SVM | RAIHANI [ | – | – |
| – | – | 43.0 |
|
|
|
| Context-Vector [ | – | – | 72.4 | – | – |
| – | – | 68.4 | |
| Kim [ | – | – | 69.8 | – | – | 38.2 | – | – | 67.0 | |
| FBK-irst [ | 66.7 |
| 67.6 | 41.9 | 37.9 | 39.8 | 64.6 | 65.6 | 65.1 | |
| WBI [ | 65.7 | 60.9 | 63.2 | 45.3 |
| 36.5 | 64.2 | 57.9 | 60.9 | |
| UTurku [ |
| 53.5 | 62.0 |
| 16.8 | 26.2 | 73.2 | 49.9 | 59.4 | |
| NN | MCCNN [ | – | – | 70.8 | – | – | – |
|
|
|
| Liu CNN [ |
| 66.7 |
|
|
|
| 75.7 | 64.7 | 69.8 | |
| Zhao SCNN [ | 73.6 |
| 70.2 | 39.4 | 39.1 | 39.2 | 72.5 | 65.1 | 68.6 | |
| Ours | Att-BLSTM |
|
|
|
|
|
|
|
|
|
The highest scores are highlighted in bold
Performance comparisons (F-score) with NN-based systems on the overall-2013 dataset for DDI classification if systems don’t use main processing techniques
| Method | Team | P(%) | R(%) | F(%) |
|---|---|---|---|---|
| NN-based | joint AB-LSTM [ | 71.3 | 66.9 | 69.3 |
| MCCNN [ | – | – | 67.8 | |
| Liu CNN [ | 75.3 | 60.4 | 67.0 | |
| Zhao SCNN [ | 68.5 | 61.0 | 64.5 | |
| Our model | Att-BLSTM |
|
|
|
The listed results come from the corresponding papers. The symbol “-” denotes no corresponding values, because the related paper did not provide complete results. Our model only replaces the candidate drugs (row (1) in Table 6), while other systems use basic text processing and replaced candidate drugs (negative instances aren’t filtered). The highest scores are highlighted in bold
Performance of interaction types on the overall-2013 dataset
| Subtype | P(%) | R(%) | F(%) |
|---|---|---|---|
| EFF | 71.9 | 81.9 | 76.6 |
| MEC | 84.1 | 71.9 | 77.5 |
| ADV | 84.8 | 85.5 | 85.1 |
| INT | 75.0 | 46.9 | 57.7 |