| Literature DB >> 35917162 |
Wentai Tang1, Jian Wang1, Hongfei Lin1, Di Zhao1, Bo Xu1, Yijia Zhang1, Zhihao Yang1.
Abstract
BACKGROUND: The ever-increasing volume of medical literature necessitates the classification of medical literature. Medical relation extraction is a typical method of classifying a large volume of medical literature. With the development of arithmetic power, medical relation extraction models have evolved from rule-based models to neural network models. The single neural network model discards the shallow syntactic information while discarding the traditional rules. Therefore, we propose a syntactic information-based classification model that complements and equalizes syntactic information to enhance the model.Entities:
Keywords: classification; extraction; interaction; literature; medical literature; medical relation extraction; medical text; neural networks; pruning method; semantic; syntactic; syntactic features; text
Year: 2022 PMID: 35917162 PMCID: PMC9382554 DOI: 10.2196/37817
Source DB: PubMed Journal: JMIR Med Inform
Figure 1Interaction features by introducing shallow syntactic information and equalization. (A) Dependency tree without processing; (B) dependency tree after syntactic structure fusion; and (C) dependency tree after the pruning process. The weight of each arc in the forest is indicated by its number. Some edges were omitted for the sake of clarity.
Figure 2Diagrammatic representation of the syntactic enhancement graph convolutional network model showing an instance and its syntactic information processing flow. The syntactic structure tree can be obtained from the encoder, and a matrix-tree can transform the syntactic dependency tree in the feature processor.
Results of the cross-sentence task.
| Model | Binary-class, accuracy | Multi-class, accuracy | ||||
|
| Ternary | Binary | Ternary | Binary | ||
|
| Single | Cross | Single | Cross | Cross | Cross |
| Feature-Based | 74.7 | 77.7 | 73.9 | 75.2 | —a | — |
| Graph LSTMb | 77.9 | 80.7 | 75.6 | 76.7 | — | — |
| DAGc LSTM | 77.9 | 80.7 | 74.3 | 76.5 | — | — |
| GS LSTMd | 80.3 | 83.2 | 83.5 | 83.6 | 71.7 | 71.7 |
| GCNe + Pruned | 85.8 | 85.8 | 83.8 | 83.7 | 78.1 | 73.6 |
| AGGCNf | 87.1 | 87.0 | 85.2 | 85.6 | 80.2 | 77.4 |
| LFGCNg | 87.3 | 86.5 | 86.7 | 85.7 | 79.9 | 77.6 |
| AGGCN + BERTh | 87.2 | 87.1 | 86.1 | 84.9 | 80.5 | 78.1 |
| LFGCN + BERT | 87.3 | 86.5 | 86.5 | 86.7 | 80.3 | 78.0 |
| SEGCNi | 88.5 | 88.2 | 87.2 | 87.5 | 81.7 | 80.2 |
| SEGCN + BERT | 88.7 | 88.4 | 86.8 | 87.7 | 81.9 | 80.4 |
aNot determined.
bLSTM: long short-term memory.
cDAG: directed acyclic graph.
dGS LSTM: graph-structured long short-term memory.
eGCN: graph convolutional network.
fAGGCN: attention-guided graph convolutional network.
gLFGCN: Lévy Flights graph convolutional network.
hBERT: Bidirectional Encoder Representations from Transformers.
iSEGCN: syntactic edge-enhanced graph convolutional network.
Results of the sentence-level task.
| Type and model | Multi-class (BioCreative ViCPR data set), F1 score | Binary-class (Phenotype-Gene Relationship data set), F1 score | |
|
| |||
|
| Random-DDCNNa | 45.4 | —b |
|
| Att-GRUc | 49.5 | — |
|
| Bran | 50.8 | — |
|
| BioBERTd | — | 67.2 |
|
| |||
|
| BO-LSTMe | — | 52.3 |
|
| GCNf | 52.2 | 81.3 |
|
| Edgewise-GRNg | 53.4 | 83.6 |
|
| kBest-GRN | 52.4 | 85.7 |
|
| ForestFT-DDCNN | 55.7 | 89.3 |
| AGGCNh | 56.7 | 88.5 | |
| LFGCNi | 64.0 | 89.6 | |
| LFGCN+BERT | 64.2 | 89.8 | |
|
| |||
|
| SEGCNj | 65.4 | 91.3 |
|
| SEGCN+BERT | 65.6 | 91.5 |
aDDCNN: Dilated and Depthwise separable convolutional neural network.
bNot determined.
cAtt-GRU: attention-based multilayer gated recurrent unit.
dBioBERT: Bidirectional Encoder Representations from Transformers for Biomedical Text Mining.
eBO-LSTM: biological ontology–based long short-term memory.
fGCN: graph convolutional network.
gGRN: graph recurrent network.
hAGGCN: attention-guided graph convolutional network.
iLFGCN: Lévy Flights graph convolutional network.
jSEGCN: syntactic enhancement graph convolutional network.
An ablation study using the Phenotype-Gene Relationship data set.
| Model | F1 score |
| SEGCNa (All) | 91.5 |
| SEGCN (- BERT Pretraining) | 91.3 |
| SEGCN (- Matrix-tree pruning) | 90.0 |
| SEGCN (- Feature capture) | 89.1 |
| Baseline (- All) | 88.5 |
aSEGCN: syntactic enhancement graph convolutional network.
Figure 3Performance against sentence length and Bidirectional Encoder Representations from Transformers (BERT) pretraining. (A) F1 scores at different sentence lengths. Results of the ForestFT– Dilated and Depthwise separable convolutional neural network are based on Jin et al [10]. (B) F1 scores against sentence length after BERT pretraining. AGGCN: attention-guided graph convolutional network; LFGCN: Lévy Flights graph convolutional network.
Figure 4The heat maps of an example sentence in the syntactic enhancement graph convolutional network model.