| Literature DB >> 35761319 |
Chunming Yang1,2, Dan Xiao3, Yuanyuan Luo3, Bo Li3, Xujian Zhao3, Hui Zhang4.
Abstract
BACKGROUND: Building a large-scale medical knowledge graphs needs to automatically extract the relations between entities from electronic medical records (EMRs) . The main challenges are the scarcity of available labeled corpus and the identification of complexity semantic relations in text of Chinese EMRs. A hybrid method based on semi-supervised learning is proposed to extract the medical entity relations from small-scale complex Chinese EMRs.Entities:
Keywords: Bootstrapping; Medical knowledge graphs; Relation extraction; Residual network; Semi-supervised learning
Mesh:
Year: 2022 PMID: 35761319 PMCID: PMC9235238 DOI: 10.1186/s12911-022-01908-4
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 3.298
Our relation annotation standard of the Chinese EMRs relation corpus
| Entity pair category | Number of entity pairs | Relation category | Number of relations | Relation description |
|---|---|---|---|---|
| Disease-position | 16538 | DAP | 304 | Disease is applied to the position in the body |
| Symptom-position | 19280 | SAP | 518 | Symptom is applied to the position in body |
| SNAP | 893 | Symptom is not applied to the position in the body | ||
| Test-disease | 5743 | TeRD | 342 | Test reveals the disease |
| Test-position | 30673 | TeAP | 1194 | Test is applied to the position in body |
| TeCP | 572 | The results of the test contains the position in the body | ||
| Test-symptom | 13617 | TeRS | 190 | Test reveals the symptom |
| TeAS | 110 | Test is applied to the symptom | ||
| Treatment-disease | 5629 | TrAD | 679 | Treatment is applied to the disease |
| TrRD | 227 | Treatment (mainly surgery) reveals the disease | ||
| Treatment-position | 8871 | TrAP | 128 | Treatment is applied to the position in body |
Classical relation extraction methods
| Classes | Principle | Classic Methods |
|---|---|---|
| Manual | Rule-based methods | Dependency parse trees[ |
| Machine learning | Based on kernel functions | Convolution kernel[ |
| Based on feature vectors | SVM[ | |
| Deep learning | Based on convolutional neural networks | CNN[ |
| Based on recurrent neural network | RNN[ | |
| Graph-based neural networks | GNN[ |
Fig. 1The architecture of the ResGRU-Att model
Fig. 2Example of position embedding
Fig. 3Residual convolution block
Fig. 4GRU unit
Bootstrapping algorithm
| Train a single relation extraction model on L | |
| Run the relation extraction model on U | |
| Find (at most) N instances in U that the probability predicted by the relation extraction model is greater than | |
| Add them into L | |
| |
Fig. 5Training process of relation extraction model based on Bootstrapping algorithm
Fig. 6An example of Chinese EMRs relation extraction
Experimental parameters settings
| Parameters | value |
|---|---|
| Batch size | 64 |
| Dimension of character embedding | 300 |
| Dimension of position embedding | 25 |
| GRU hidden units | 512 |
| GRU hidden layer | 3 |
| Window size | 3,5,7 |
| Number of filters | 128 |
| Learning rate | 0.015 |
| Optimizer | Adam |
| Dropout | 0.5 |
Fig. 7Comparison of F1-score of ResNet and CNN with different depths
Comparison of F1-score of all models on different scale datasets
| Models | |||||
|---|---|---|---|---|---|
| CNN | 67.99 | 69.67 | 72.8 | 74.86 | 76.71 |
| CNN-Att | 70.49 | 73.57 | 75 | 78.44 | 81.06 |
| PCNN | 74.46 | 75.24 | 77.8 | 78.26 | 79.57 |
| ResNet | 78.27 | 79.75 | 81.52 | 83.34 | 86.13 |
| BiLSTM-Att | 76.12 | 80.98 | 82.71 | 84.96 | 85.21 |
| BiGRU-Att | 77.96 | 81.18 | 83.9 | 85.11 | 85.94 |
| ResGRU | 80 | 84.08 | 86 | 86.74 | 87.09 |
| ResGRU-Att |
Bold indicates the best value for this column
Time comparison of all models on different scale datasets
| Models | |||||
|---|---|---|---|---|---|
| CNN | 13min43s | 25min12s | 40min5s | 48min35s | 57min3s |
| CNN-Att | 17min18s | 31min20s | 47min35s | 59min48s | 1h20min15s |
| PCNN | 13min20s | 24min41s | 41min48s | 47min25s | 56min25s |
| ResNet | 50min35s | 1h40min | 2h39min11s | 3h18min35s | 4h25min24s |
| BiLSTM-Att | 1h30min9s | 2h38min41s | 4h18min46s | 5h6min29s | 6h13min27s |
| BiGRU-Att | 1h17min25s | 2h30min22s | 3h54min35s | 4h55min4s | 5h48min16s |
| ResGRU | 1h50min46s | 2h48min19s | 5h1min8s | 6h40min39s | 8h14min8s |
| ResGRU-Att | 2h4min5s | 3h9s | 5h24min34s | 7h12min24s | 8h51min21s |
Fig. 8Comparison of precision and recall for the ResGRU-Att model on various relation categories
Comparison of performance for different models on overall relation categories
| Models | Precision | Recall | F1-score |
|---|---|---|---|
| CNN | 79.39 | 74.21 | 76.71 |
| CNN-Att | 85.46 | 77.09 | 81.06 |
| PCNN | 83.56 | 75.94 | 79.57 |
| ResNet | 88.44 | 83.94 | 86.13 |
| BiLSTM-Att | 85.4 | 85.02 | 85.21 |
| BiGRU-Att | 87.75 | 84.20 | 85.94 |
| ResGRU | 86.47 | 87.72 | 87.09 |
| ResGRU-Att |
Comparison of F1-score for different models on various relation categories
| Models | DAP | SAP | SNAP | TeRD | TeAP | TeCP | TeRS | TeAS | TrAD | TrRD | TrAP |
|---|---|---|---|---|---|---|---|---|---|---|---|
| CNN | 64.86 | 78.04 | 75.95 | 79.51 | 85 | 74.29 | 80.4 | 66.76 | 83.9 | 72.74 | 82.36 |
| CNN-Att | 70.33 | 82.89 | 84.79 | 77.16 | 84.98 | 80.1 | 84.89 | 75.67 | 83.73 | 83.71 | 83.42 |
| PCNN | 69.53 | 84.74 | 80.06 | 75.69 | 82.28 | 76.36 | 79.95 | 75.43 | 84.74 | 83.44 | 83.05 |
| ResNet | 75.98 | 91.11 | 86.95 | 84.18 | 89.89 | 78.59 | 85.86 | 86 | 86.24 | ||
| BiLSTM-Att | 74.96 | 92.83 | 92.48 | 84.84 | 85.27 | 73.57 | 85.6 | 87.47 | 92.48 | 80.34 | 87.43 |
| BiGRU-Att | 75.34 | 90.62 | 90 | 87.43 | 83.18 | 85.27 | 81.84 | 92.08 | 82.24 | 87.44 | |
| ResGRU | 78.53 | 92 | 85 | 83.39 | 84.46 | 80.13 | 91.83 | 87.11 | 90.17 | ||
| ResGRU-Att | 92.96 | 88.43 | 86.54 | 93.01 | 87.58 |
Bold indicates the best value for this column