| Literature DB >> 28605776 |
Víctor Suárez-Paniagua1, Isabel Segura-Bedmar1, Paloma Martínez1.
Abstract
Drug-drug interaction (DDI), which is a specific type of adverse drug reaction, occurs when a drug influences the level or activity of another drug. Natural language processing techniques can provide health-care professionals with a novel way of reducing the time spent reviewing the literature for potential DDIs. The current state-of-the-art for the extraction of DDIs is based on feature-engineering algorithms (such as support vector machines), which usually require considerable time and effort. One possible alternative to these approaches includes deep learning. This technique aims to automatically learn the best feature representation from the input data for a given task. The purpose of this paper is to examine whether a convolutional neural network (CNN), which only uses word embeddings as input features, can be applied successfully to classify DDIs from biomedical texts. Proposed herein, is a CNN architecture with only one hidden layer, thus making the model more computationally efficient, and we perform detailed experiments in order to determine the best settings of the model. The goal is to determine the best parameter of this basic CNN that should be considered for future research. The experimental results show that the proposed approach is promising because it attained the second position in the 2013 rankings of the DDI extraction challenge. However, it obtained worse results than previous works using neural networks with more complex architectures.Entities:
Mesh:
Year: 2017 PMID: 28605776 PMCID: PMC5467573 DOI: 10.1093/database/bax019
Source DB: PubMed Journal: Database (Oxford) ISSN: 1758-0463 Impact factor: 3.451
State-of-the-art systems results using the whole DDI Corpus for the classification task
| Systems | Precision | Recall | F1-score |
|---|---|---|---|
| Liu | 75.72% | 64.66% | 69.75% |
| Ebrahimi and Dou [ | 75.31% | 66.19% | 68.64% |
| Zhao | 72.5% | 65.1% | 68.6% |
| Kim | Unknown | Unknown | 67% |
| Chowdhury and Lavelli [ | 64.6% | 65.6% | 65.1% |
Figure 1Some examples of sentences in the DDI corpus [14].
DDI types in the DDI corpus
| DDI types | DDI-DrugBank | DDI-MedLine | Total |
|---|---|---|---|
| Advice | 1035 (22%) | 15 (4.6%) | 1050 (20.9%) |
| Effect | 1855 (39.4%) | 214 (65.4%) | 2069 (41.1%) |
| Int | 272 (5.8%) | 12 (3.7%) | 284 (5.6%) |
| Mechanism | 1539 (32.7%) | 86 (26.3%) | 1625 (32.3%) |
| Total | 4701 | 327 | 5028 |
Figure 2CNN model for DDIExtraction task.
Figure 3Learning curve of a CNN with random initialization. The blue line shows the training-curve variation along the number of epochs, the green represents the validation and the red one the testing curve.
Results obtained for CNN from random initialization on the whole DDI corpus
| Classes | TP | FP | FN | Total | |||
|---|---|---|---|---|---|---|---|
| Advice | 131 | 43 | 90 | 221 | 75.29% | 59.28% | 66.33% |
| Effect | 239 | 220 | 121 | 360 | 52.07% | 66.39% | 58.36% |
| Int | 27 | 3 | 69 | 96 | 90% | 28.12% | 42.86% |
| Mechanism | 176 | 84 | 122 | 298 | 67.69% | 59.06% | 63.08% |
| Overall | 573 | 350 | 402 | 975 | 62.08% | 58.77% | 60.38% |
Results obtained for CNN from random initialization on the DDI-DrugBank dataset
| Classes | TP | FP | FN | Total | |||
|---|---|---|---|---|---|---|---|
| Advice | 130 | 43 | 84 | 214 | 75.14% | 60.75% | 67.18% |
| Effect | 212 | 190 | 86 | 298 | 52.74% | 71.14% | 60.57% |
| Int | 27 | 2 | 67 | 94 | 93.1% | 28.72% | 43.9% |
| Mechanism | 169 | 79 | 109 | 278 | 68.15% | 60.79% | 64.26% |
| Overall | 538 | 314 | 346 | 884 | 63.15% | 60.86% | 61.98% |
Results obtained for CNN from random initialization on the DDI-MedLine dataset
| Classes | TP | FP | FN | Total | |||
|---|---|---|---|---|---|---|---|
| Advice | 1 | 0 | 6 | 7 | 100% | 14.29% | 25% |
| Effect | 27 | 30 | 35 | 62 | 47.37% | 43.55% | 45.38% |
| Int | 0 | 1 | 2 | 2 | 0% | 0% | 0% |
| Mechanism | 7 | 5 | 13 | 20 | 58.33% | 35% | 43.75% |
| Overall | 35 | 36 | 56 | 91 | 49.3% | 38.46% | 43.21% |
Figure 4Distance between entities in sentences describing DDIs.
Results for several filter sizes
| Filter size | |||
|---|---|---|---|
| 2 | 56.89% | 52.1% | 54.39% |
| 4 | 65.65% | 52.92% | 58.6% |
| 6 | 49.23% | 59.55% | |
| (2, 3, 4) | 63.15% | 57.13% | 59.99% |
| (3, 4, 5) | 62.08% | 60.38% | |
| (2, 4, 6) | 73.57% | 52.82% | |
| (2, 3, 4, 5) | 71.31% | 52% | 60.14% |
| 14 | 71.23% | 51.79% | 59.98% |
| (13, 14, 15) | 72.64% | 49.03% | 58.54% |
χ2 and P-value statistics between the different filter sizes
| Filter size | 4 | 6 | (2, 3, 4) | (3, 4, 5) | (2, 4, 6) | (2, 3, 4, 5) | 14 | (13, 14, 15) |
|---|---|---|---|---|---|---|---|---|
| 2 | 13.22* | 50.68* | 155.88* | 1.20 | 25.77* | 119.71* | 44.52* | 5.28* |
| 2.77e−04* | 1.09e−12* | 8.99e−36* | 2.73e−01 | 3.84e−07* | 7.32e−28* | 2.52e−11* | 2.16e−02* | |
| 4 | 20.01* | 118.53* | 21.92* | 3.87* | 97.79* | 17.79* | 0.14 | |
| 7.71e−06* | 1.33e−27* | 2.84e−06* | 4.91e−02* | 4.66e−23* | 2.46e−05* | 7.12e−01 | ||
| 6 | 44.14* | 73.39* | 7.92* | 26.78* | 0.08 | 22.25* | ||
| 3.06e−11* | 1.06e−17* | 4.89e−03* | 2.28e−07* | 7.73e−01 | 2.39e−06* | |||
| (2, 3, 4) | 177.61* | 69.08* | 4.82* | 33.78* | 81.04* | |||
| 1.61e−40* | 9.48e−17* | 2.81e−02* | 6.18e−09* | 2.21e−19* | ||||
| (3, 4, 5) | 33.25* | 146.78* | 67.70* | 11.33* | ||||
| 8.12e−09* | 8.75e−34* | 1.91e−16* | 7.65e−04* | |||||
| (2, 4, 6) | 51.97* | 7.16* | 5.06* | |||||
| 5.63e−13* | 7.45e−03* | 2.45e−02* | ||||||
| (2, 3, 4, 5) | 20.44* | 67.10* | ||||||
| 6.16e−06* | 2.58e−16* | |||||||
| 14 | 31.01* | |||||||
| 2.57e−08* |
Asterisk denotes results statistically significant.
Performance with different word embedding and different position embedding size
| Word embedding | Position embedding | |||
|---|---|---|---|---|
| random | 0 | 62.08% | 58.77% | 60.38% |
| 5 | 69.34% | 55.9% | ||
| 10 | 54.36% | 61.48% | ||
| Wiki_bow_8w_25n | 0 | 60.89% | 54.46% | 57.5% |
| 5 | 59.2% | 60.72% | 59.95% | |
| 10 | 70.64% | 53.54% | 60.91% | |
| Bio_skip_8w_25n | 0 | 62.39% | 57.85% | 60.03% |
| 5 | 67.8% | 53.33% | 59.7% | |
| 10 | 66.92% | 55.18% | 60.48% | |
| Bio_skip_10w_10n | 0 | 70.66% | 49.64% | 58.31% |
| 5 | 61.84% | 56.51% | 59.06% | |
| 10 | 68.77% | 54.87% | 61.04% | |
| Bio_bow_8w_25n | 0 | 64.09% | 54.36% | 58.82% |
| 5 | 69.43% | 54.05% | 60.78% | |
| 10 | 67.27% | 49.95% | 57.33% | |
| Bio_bow_5w_10n | 0 | 58.25% | 59.38% | 58.81% |
| 5 | 60.18% | 60.7% | ||
| 10 | 65.21% | 56.72% | 60.67% |
The prefix Wiki (Wikipedia corpus) or Bio (BioASQ dataset) refers to the corpus used to train the word embedding model. The label bow (CBOW) or skip (skip-gram) refers to the type of architecture used to build the model. The number preceding w and n indicates the size of the context window and the negative sampling, respectively.
χ2 and P-value statistics between the different word embeddings and position embeddings
| Word embedding and position embedding | random | Wiki_bow_8w_25n | Bio_skip_8w_25n | Bio_skip_10w_10n | Bio_bow_8w_25n | Bio_bow_5w_10n | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 5 | 10 | 0 | 5 | 10 | 0 | 5 | 10 | 0 | 5 | 10 | 0 | 5 | 10 | 0 | 5 | 10 | ||
| random | 0 | 78.41* | 11.44* | 0.74 | 11.41* | 0.44 | 10.10* | 0.14 | 20.98* | 133.47* | 0.26 | 0.02 | 127.06* | 71.74* | 10.71* | 5.89* | 132.98* | 8.32* |
| 8.38e−19* | 7.19e−04* | 3.89e−01 | 7.31e−04* | 5.08e−01 | 1.48e−03* | 7.06e−01 | 4.64e−06* | 7.13e−31* | 6.08e−01 | 8.83e−01 | 1.80e−29* | 2.46e−17* | 1.06e−03* | 1.52e−02* | 9.12e−31* | 3.93e−03* | ||
| 5 | 41.42* | 43.25* | 29.50* | 78.75* | 23.52* | 46.94* | 13.26* | 23.89* | 49.09* | 40.16* | 20.05* | 1.48 | 20.31* | 29.08* | 23.86* | 23.58* | ||
| 1.23e−10* | 4.81e−11* | 5.59e−08* | 7.05e−19* | 1.23e−06* | 7.31e−12* | 2.71e−04* | 1.02e−06* | 2.44e−12* | 2.34e−10* | 7.56e−06* | 2.24e−01 | 6.60e−06* | 6.94e−08* | 1.04e−06* | 1.20e−06* | |||
| 10 | 2.96 | 0.21 | 15.52* | 0.26 | 8.78* | 4.07* | 92.25* | 4.61* | 6.76* | 85.78* | 40.50* | 0.39 | 0.03 | 92.07* | 0.03 | |||
| 8.52e−02 | 6.45e−01 | 8.16e−05* | 6.08e−01 | 3.05e−03* | 4.37e−02* | 7.64e−22* | 3.18e−02* | 9.32e−03* | 2.01e−20* | 1.97e−10* | 5.33e−01 | 8.63e−01 | 8.36e−22* | 8.56e−01 | ||||
| Wiki_bow_8w_25n | 0 | 6.20* | 3.09 | 7.15* | 1.72 | 15.94* | 124.17* | 0.10 | 1.13 | 91.84* | 46.65* | 3.90* | 1.57 | 100.12* | 2.78 | |||
| 1.28e−02* | 7.90e−02 | 7.51e−03* | 1.90e−01 | 6.54e−05* | 7.74e−29* | 7.54e−01 | 2.88e−01 | 9.41e−22* | 8.47e−12* | 4.83e−02* | 2.10e−01 | 1.44e−23* | 9.52e−02 | |||||
| 5 | 21.54* | 0.00 | 9.43* | 2.23 | 90.06* | 7.56* | 7.51* | 66.94* | 30.24* | 0.01 | 0.43 | 76.09* | 0.02 | |||||
| 3.46e−06* | 1.00e+00 | 2.13e−03* | 1.36e−01 | 2.31e−21* | 5.96e−03* | 6.13e−03* | 2.80e−16* | 3.81e−08* | 9.41e−01 | 5.12e−01 | 2.71e−18* | 8.76e−01 | ||||||
| 10 | 17.78* | 0.01 | 42.24* | 144.40* | 2.15 | 0.10 | 117.05* | 76.26* | 12.84* | 8.05* | 128.69* | 12.34* | ||||||
| 2.48e−05* | 9.34e−01 | 8.07e−11* | 2.91e−33* | 1.42e−01 | 7.55e−01 | 2.79e−27* | 2.49e−18* | 3.39e−04* | 4.55e−03* | 7.93e−30* | 4.44e−04* | |||||||
| Bio_skip_8w_25n | 0 | 14.78* | 2.16 | 111.21* | 11.72* | 11.36* | 61.25* | 27.80* | 0.00 | 0.49 | 69.84* | 0.05 | ||||||
| 1.21e−04* | 1.41e−01 | 5.33e−26* | 6.18e−04* | 7.51e−04* | 5.02e−15* | 1.35e−07* | 1.00e+00 | 4.82e−01 | 6.43e−17* | 8.18e−01 | ||||||||
| 5 | 24.49* | 139.36* | 1.25 | 0.07 | 101.05* | 63.22* | 9.74* | 5.81* | 102.92* | 8.84* | ||||||||
| 7.47e−07* | 3.67e−32* | 2.63e−01 | 7.87e−01 | 8.98e−24* | 1.85e−15* | 1.81e−03* | 1.59e−02* | 3.49e−24* | 2.95e−03* | |||||||||
| 10 | 70.92* | 26.30* | 20.83* | 52.75* | 19.67* | 1.06 | 3.63 | 55.44* | 2.31 | |||||||||
| 3.72e−17* | 2.93e−07* | 5.03e−06* | 3.79e−13* | 9.19e−06* | 3.03e−01 | 5.66e−02 | 9.63e−14* | 1.28e−01 | ||||||||||
| Bio_skip_10w_10n | 0 | 131.16* | 130.70* | 0.12 | 9.66* | 61.23* | 90.72* | 0.00 | 71.64* | |||||||||
| 2.28e−30* | 2.88e−30* | 7.25e−01 | 1.88e−03* | 5.09e−15* | 1.66e−21* | 1.00e+00 | 2.58e−17* | |||||||||||
| 5 | 0.69 | 93.74* | 53.76* | 5.60* | 2.69 | 100.91* | 4.50* | |||||||||||
| 4.07e−01 | 3.60e−22* | 2.26e−13* | 1.80e−02* | 1.01e−01 | 9.61e−24* | 3.39e−02* | ||||||||||||
| 10 | 95.72* | 55.37* | 7.98* | 4.37* | 95.76* | 7.06* | ||||||||||||
| 1.33e−22* | 9.96e−14* | 4.72e−03* | 3.66e−02* | 1.30e−22* | 7.90e−03* | |||||||||||||
| Bio_bow_8w_25n | 0 | 15.56* | 89.27* | 88.45* | 0.11 | 82.26* | ||||||||||||
| 8.01e−05* | 3.44e−21* | 5.22e−21* | 7.43e−01 | 1.19e−19* | ||||||||||||||
| 5 | 36.01* | 45.52* | 12.88* | 47.91* | ||||||||||||||
| 1.97e−09* | 1.51e−11* | 3.33e−04* | 4.47e−12* | |||||||||||||||
| 10 | 0.71 | 90.31* | 0.15 | |||||||||||||||
| 4.00e−01 | 2.04e−21* | 6.99e−01 | ||||||||||||||||
| Bio_bow_5w_10n | 0 | 96.01* | 0.20 | |||||||||||||||
| 1.15e−22* | 6.56e−01 | |||||||||||||||||
| 5 | 95.07* | |||||||||||||||||
| 1.84e−22* | ||||||||||||||||||
Asterisk denotes results statistically significant.
Results obtained for the best CNN model [random initialization, filter size (2, 4, 6) and position embedding of dimension 5] on the DDI corpus test set
| Classes | TP | FP | FN | Total | |||
|---|---|---|---|---|---|---|---|
| Advice | 145 | 41 | 76 | 221 | 77.96% | 65.61% | 71.25% |
| Effect | 183 | 81 | 177 | 360 | 69.32% | 50.83% | 58.65% |
| Int | 27 | 8 | 69 | 96 | 77.14% | 28.12% | 41.22% |
| Mechanism | 192 | 106 | 106 | 298 | 64.43% | 64.43% | 64.43% |
| Overall | 547 | 236 | 428 | 975 | 69.86% | 56.1% | 62.23% |
Figure 5State-of-the-art F1-scores for DDI classification. The deep blue bars represents the participating system in DDIExtraction 2013. The light blue is the work of Kim et al. [10] and the green ones represent recent systems based on the deep learning techniques for DDI classification, which were subsequently presented. Our best model is represented by the red bar.
Number of instances in each dataset of the DDI corpus after the pre-processing phase
| DDI types | DDI-DrugBank | DDI-MedLine | Total |
|---|---|---|---|
| Advice | 1028 | 14 | 1042 |
| Effect | 1815 | 214 | 2029 |
| Int | 272 | 12 | 284 |
| Mechanism | 1535 | 83 | 1618 |
| Other | 26 486 | 1892 | 28 373 |
| Total | 31 136 | 2215 | 33 351 |
| Train | 25 885 | 1778 | 27 663 |
| Test | 5251 | 437 | 5688 |
The class Other represents the non-interaction between pairs of drug mentions.