| Literature DB >> 30478023 |
Fei Li1,2,3, Weisong Liu1,2,3, Hong Yu1,2,3,4.
Abstract
BACKGROUND: Pharmacovigilance and drug-safety surveillance are crucial for monitoring adverse drug events (ADEs), but the main ADE-reporting systems such as Food and Drug Administration Adverse Event Reporting System face challenges such as underreporting. Therefore, as complementary surveillance, data on ADEs are extracted from electronic health record (EHR) notes via natural language processing (NLP). As NLP develops, many up-to-date machine-learning techniques are introduced in this field, such as deep learning and multi-task learning (MTL). However, only a few studies have focused on employing such techniques to extract ADEs.Entities:
Keywords: adverse drug event; deep learning; multi-task learning; named entity recognition; natural language processing; relation extraction
Year: 2018 PMID: 30478023 PMCID: PMC6288593 DOI: 10.2196/12159
Source DB: PubMed Journal: JMIR Med Inform
Figure 1Study overview. NER: named entity recognition. RE: relation extraction. BiLSTM: bidirectional long short-term-memory. CRF: conditional random field. MTL: multi-task learning. MADE: Medication, Indication, and Adverse Drug Events. HardMTL: multi-task learning model for hard parameter sharing. RegMTL: multi-task learning model for soft parameter sharing based on regularization. LearnMTL: multi-task learning model for soft parameter sharing based on task-relation learning.
Figure 2NER submodel. For simplicity, here we use “Renal Failure” to illustrate the architecture. For “Renal,” the word feature is “Renal,” the capital feature of the initial character is “R,” the POS feature is “JJ,” and the character representation is generated from CNN. NER: named entity recognition. CNN: convolutional neural network. CRF: condition random field. LSTM: long short-term memory. CNN: convolutional neural network. POS: part of speech.
Figure 3RE submodel. The target entities are “renal failure” (e1) and “antibiotics” (e2). Positions represent token distances to the target entities. RE: relation extraction. LSTM: long short-term memory. POS: part of speech.
Figure 4The high-level view of HardMTL. For conciseness, “LSTM” indicates a BiLSTM layer, and the layers above the BiLSTM layer are denoted as D and D. The forward procedures for an NER instance and an RE instance are indicated by blue and green arrow lines, respectively. HardMTL: multi-task learning model for hard parameter sharing. LSTM: long short-term-memory. BiLSTM: bidirectional long short-term-memory. CRF: conditional random field. NER: named entity recognition. RE: relation extraction.
Figure 5The high-level view of RegMTL. LSTM and LSTM indicate the first and second BiLSTM layers of the NER model. LSTM and LSTM indicate the first and second BiLSTM layers of the RE model. NER: named entity recognition. RE: relation extraction. RegMTL: multi-task learning model for soft parameter sharing based on regularization. BiLSTM: bidirectional long short-term-memory. CRF: conditional random field. LSTM: long short-term-memory.
Figure 6The high-level view of LearnMTL. LearnMTL: multi-task learning model for soft parameter sharing based on task-relation learning. CRF: conditional random field. LSTM: long short-term-memory.
Comparison of our model with the existing systems in the Medication, Indication, and Adverse Drug Events dataset. The microaveraged F1s of relation extraction are shown according to the official evaluation report.
| System | Named entity recognition | Relation extraction | F1 |
| Chapman et al [ | CRFa | Random forest | 59.2 |
| Xu et al [ | BiLSTMb-CRF | Support vector machine | 59.9 |
| Dandala et al [ | BiLSTM-CRF | BiLSTM-Attention | 61.7 |
| Our Best (HardMTLc) | BiLSTM-CRF | BiLSTM-Attention | 66.7 |
aCRF: conditional random field
bBiLSTM: bidirectional long short-term memory
cHardMTL: multi-task learning model for hard parameter sharing
Performances (%) of the pipeline and multi-task learning models. The values presented are the means of 5 runs of each model. The microaveraged P, R, and F1s of all entity or relation types are shown.
| Method | Entity recognition | Relation extraction | ||||
| P | R | F1 | P | R | F1 | |
| Pipeline | 85.0 | 83.2 | 84.1 | 69.8 | 62.4 | 65.9 |
| HardMTLa | 85.0 | 84.1 | 84.5 | 70.2 | 63.6 | 66.7 |
| RegMTLb | 84.5 | 84.5 | 84.5 | 66.7 | 63.6 | 65.1 |
| LearnMTLc | 84.5 | 82.8 | 83.6 | 67.2 | 61.5 | 64.2 |
aHardMTL: multi-task learning model for hard parameter sharing
bRegMTL: multi-task learning model for soft parameter sharing based on regularization
cLearnMTL: multi-task learning model for soft parameter sharing based on task relation learning
Performance (%) of each entity type.
| Entity type | P | R | F1 |
| Medication | 91.1 | 92.0 | 91.3 |
| Indication | 65.4 | 64.8 | 64.8 |
| Frequency | 87.1 | 86.5 | 86.3 |
| Severity | 84.6 | 84.7 | 84.7 |
| Dosage | 87.9 | 86.4 | 88.0 |
| Duration | 75.3 | 76.6 | 77.6 |
| Route | 91.6 | 91.9 | 91.9 |
| Adverse drug events | 59.5 | 57.6 | 55.4 |
| SSLIFa | 83.9 | 84.8 | 84.9 |
aSSLIF: any sign, symptom, and disease that is not an ADE or Indication
Performance (%) of each relation type.
| Relation type | P | R | F1 |
| Severity-Adverse drug events | 55.0 | 54.4 | 54.1 |
| Route-Medication | 81.0 | 82.5 | 82.1 |
| Medication-Indication | 53.9 | 52.5 | 52.9 |
| Dosage-Medication | 80.9 | 79.8 | 81.0 |
| Duration-Medication | 60.3 | 63.7 | 59.5 |
| Frequency-Medication | 77.7 | 78.6 | 78.4 |
| Medication-Adverse drug events | 50.4 | 47.6 | 45.5 |
Results (%) of comparisons between our pipeline model and the MedEx system.
| Entity type | MedEx system | Pipeline model | ||||
| P | R | F1 | P | R | F1 | |
| Route-Medication | 71.9 | 47.9 | 57.5 | 81.0 | 82.5 | 82.1 |
| Dosage-Medication | 29.7 | 3.5 | 6.2 | 80.9 | 79.8 | 81.0 |
| Duration-Medication | 25.5 | 15.6 | 19.4 | 60.3 | 63.7 | 59.5 |
| Frequency-Medication | 52.5 | 36.2 | 42.8 | 77.7 | 78.6 | 78.4 |