| Literature DB >> 32938366 |
Jian Wang1, Mengying Li1, Qishuai Diao1, Hongfei Lin1, Zhihao Yang1, YiJia Zhang2.
Abstract
BACKGROUND: Biomedical document triage is the foundation of biomedical information extraction, which is important to precision medicine. Recently, some neural networks-based methods have been proposed to classify biomedical documents automatically. In the biomedical domain, documents are often very long and often contain very complicated sentences. However, the current methods still find it difficult to capture important features across sentences.Entities:
Keywords: Biomedical document triage; Biomedical literature; Capsule network; Hierarchical attention mechanism
Mesh:
Year: 2020 PMID: 32938366 PMCID: PMC7495737 DOI: 10.1186/s12859-020-03673-5
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1Negative examples and positive examples of JON-format from PM datset. a negative examples of JON-format from PM datset. b positive examples of JON-format from PM dataset
Fig. 2Negative examples and positive examples of XML-like format from IAS dataset. a negative examples of XML-like format from IAS dataset. b Positive examples of XML-like format from IAS dataset
Fig. 3The schematic overview of our model
Fig. 4The self-attention mechanism calculation process. We can get three vectors that are a query vector, a key vector and a value vector for each word by multiplying the embedding word vector by the three matrices trained during the training. The size of these new vectors is 3, while the dimensions of the embedding word and encoder are 4. We evaluate the dependence between the words with the dot product operation
Fig. 5The hierarchical attention mechanism architecture. We connect the word-level and sentence-level features by attention mechanism to get a feature vector with more information and the Softmax function is used to normalize the result
Fig. 6The capsule network model architecture
Dataset statistics
| Corpus | Positive | Negative | Total |
|---|---|---|---|
| PM Training set | 1729 | 2353 | 4028 |
| PM Test set | 704 | 723 | 1427 |
| IAS Training set | 3536 | 1959 | 5495 |
| IAS Test set | 338 | 339 | 667 |
| ACT Training set | 1140 | 1140 | 2280 |
| ACT Development set | 682 | 3318 | 4000 |
| ACT Test set | 910 | 5090 | 6000 |
The results of baseline methods
| Methods | P | R | F1 |
|---|---|---|---|
| CNN | 0.581 | 0.774 | 0.664 |
| Capsule network | 0.629 | 0.755 | 0.686 |
| Self-Attention | 0.584 | 0.895 | 0.707 |
The results of hierarchical attention
| Methods | P | R | F1 |
|---|---|---|---|
| CNN | 0.581 | 0.774 | 0.664 |
| CNN+self-attention | 0.584 | 0.895 | 0.707 |
| CNN+hierarchical attention | 0.623 | 0.840 | 0.715 |
The results of capsule network
| Methods | P | R | F1 |
|---|---|---|---|
| CNN | 0.581 | 0.774 | 0.664 |
| Capsule network | 0.629 | 0.755 | 0.686 |
| CNN+hierarchical attention | 0.623 | 0.840 | 0.715 |
| CapsNet+hierarchical attention | 0.624 | 0.895 | 0.723 |
Performance compared with other methods on PM corpus
| Methods | P | R | F1 |
|---|---|---|---|
| PrecMed Baseline [ | 0.610 | 0.636 | 0.622 |
| Team 418 | 0.629 | 0.766 | 0.691 |
| Team 421 | 0.570 | 0.874 | 0.690 |
| PrecMed-best [ | 0.603 | 0.821 | 0.695 |
| Ensemble model | 0.629 | 0.815 | 0.710 |
| CapsNet+hierarchical attention | 0.624 | 0.895 | 0.723 |
Performance compared with other methods on ACT corpus
| Methods | P | R | F1 |
|---|---|---|---|
| Team 65 | - | - | 0.598 |
| Team 70 | - | - | 0.549 |
| Team 73 | - | - | 0.614 |
| Team 81 | - | - | 0.311 |
| Team 88 | - | - | 0.344 |
| Team 89 | - | - | 0.608 |
| Team 90 | - | - | 0.596 |
| Team 92 | - | - | 0.572 |
| Team 100 | - | - | 0.594 |
| Team 104 | - | - | 0.539 |
| Our method | 0.570 | 0.676 | 0.618 |
Performance compared with other methods on IAS corpus
| Methods | P | R | F1 |
|---|---|---|---|
| Team 4 | 0.712 | 0.792 | 0.750 |
| Team 6 | 0.708 | 0.860 | 0.777 |
| Team 7 | 0.684 | 0.858 | 0.761 |
| Team 11 | 0.676 | 0.781 | 0.725 |
| Team 14 | 0.746 | 0.470 | 0.757 |
| Team 19 | 0.645 | 0.565 | 0.602 |
| Team 27 | 0.607 | 0.852 | 0.709 |
| Team 28 | 0.750 | 0.810 | 0.779 |
| Team 30 | 0.686 | 0.789 | 0.643 |
| Team 31 | 0.667 | 0.594 | 0.629 |
| Team 37 | 0.575 | 0.946 | 0.715 |
| Team 41 | 0.619 | 0.890 | 0.730 |
| Team 44 | 0.688 | 0.868 | 0.764 |
| Team 48 | 0.588 | 0.863 | 0.700 |
| Team 49 | 0.526 | 0.985 | 0.685 |
| Team 51 | 0.717 | 0.828 | 0.769 |
| Team 52 | 0.692 | 0.834 | 0.757 |
| Team 57 | 0.703 | 0.875 | 0.780 |
| Team 58 | 0.667 | 0.730 | 0.697 |
| Our method | 0.704 | 0.879 | 0.782 |
Fig. 7The visualization of the multi-head attention mechanism. The blue lines represent the dependence relationship between the words in the single-head self-attention mechanism and the purple lines represent the dependence relationship between the words in the multi-head self-attention mechanism
The confusion matrix on PM corpus
| Actual | Predict | |
|---|---|---|
| True | False | |
| True | 865 | 142 |
| False | 522 | 98 |
Fig. 8The false positive examples and false negative examples on PM corpus. a the false positive examples on PM corpus. b the false negative examples on PM corpus