| Literature DB >> 18547432 |
Steven Bethard1, Zhiyong Lu, James H Martin, Lawrence Hunter.
Abstract
BACKGROUND: Automatic semantic role labeling (SRL) is a natural language processing (NLP) technique that maps sentences to semantic representations. This technique has been widely studied in the recent years, but mostly with data in newswire domains. Here, we report on a SRL model for identifying the semantic roles of biomedical predicates describing protein transport in GeneRIFs - manually curated sentences focusing on gene functions. To avoid the computational cost of syntactic parsing, and because the boundaries of our protein transport roles often did not match up with syntactic phrase boundaries, we approached this problem with a word-chunking paradigm and trained support vector machine classifiers to classify words as being at the beginning, inside or outside of a protein transport role.Entities:
Mesh:
Year: 2008 PMID: 18547432 PMCID: PMC2474622 DOI: 10.1186/1471-2105-9-277
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Model performance by feature set.
| Labeled | Unlabeled | Labeled | ||||
| Precision | Recall | F-measure | Precision | Recall | Accuracy | |
| Word-Chunking | 79.7 | 64.4 | 71.3 | 80.1 | 64.7 | 99.6 |
| Phrase-Chunking | 81.0 | 71.9 | 76.2 | 81.9 | 72.7 | 98.9 |
| Protein-Transport | 87.6 | 79.0 | 83.1 | 87.9 | 79.2 | 99.7 |
| Protein-Transport (ABNER) | 87.0 | 74.5 | 80.3 | 87.3 | 74.8 | 99.7 |
This table shows precision, recall, F-measure and labeled accuracy statistics for semantic role labeling models trained on various feature sets. "ABNER" indicates the results when using the ABNER-identified proteins instead of the manually annotated ones.
Model performance by role type.
| Precision | Recall | F-measure | % of Roles | |
| AGENT | 100.0 | 33.3 | 50.0 | 0.8 |
| PATIENT | 86.5 | 74.4 | 80.0 | 51.7 |
| ORIGIN | 82.9 | 75.6 | 79.1 | 11.7 |
| DESTINATION | 90.3 | 87.7 | 89.0 | 35.8 |
This table shows precision, recall, F-measure split up by the type of role being identified. The final column indicates the percent of the total roles each role type accounted for. Performance numbers were calculated from the output of the Protein-Transport model when using manually annotated protein boundaries.
Model performance by seen vs. unseen.
| Precision | Recall | F-measure | % | |
| Seen roles | 97.7 | 88.9 | 93.1 | 60.1 |
| Unseen roles | 71.6 | 63.6 | 67.4 | 39.2 |
| Seen AGENT roles | 100.0 | 100.0 | 100.0 | 33.3 |
| Unseen AGENT roles | 100.0 | 0.0 | 0.0 | 66.7 |
| Seen PATIENT roles | 100.0 | 86.7 | 92.9 | 37.7 |
| Unseen PATIENT roles | 78.3 | 66.9 | 72.2 | 62.3 |
| Seen ORIGIN roles | 97.1 | 84.6 | 90.4 | 86.7 |
| Unseen ORIGIN roles | 14.3 | 16.7 | 15.4 | 13.3 |
| Seen DESTINATION roles | 96.5 | 91.6 | 94.0 | 86.2 |
| Unseen DESTINATION roles | 57.1 | 63.2 | 60.0 | 13.8 |
This table shows precision, recall, F-measure comparing role phrases which appeared in both the training and testing data ("seen" roles) to role phrases which appeared only in the testing data ("unseen" roles). Both overall results and results by role type are shown. The final column indicates the percent of the particular category of roles that were seen (or unseen). Performance numbers were calculated from the output of the Protein-Transport model when using manually annotated protein boundaries.
Model performance by predicate type.
| Feature Set | Precision | Recall | F-measure | % of Roles | |
| Nominal predicates | Phrase-Chunking | 78.8 | 63.0 | 70.7 | 79.2 |
| Verbal predicates | Phrase-Chunking | 64.3 | 52.9 | 58.1 | 20.8 |
| Nominal predicates | Protein-Transport | 86.2 | 75.8 | 80.6 | 79.2 |
| Verbal predicates | Protein-Transport | 88.5 | 67.6 | 76.7 | 20.8 |
| Wall Street Journal predicates | Phrase-Chunking | 74.0 | 62.7 | 67.9 | 34.3 |
| GeneRIF-only predicates | Phrase-Chunking | 77.3 | 60.7 | 68.0 | 65.7 |
| Wall Street Journal predicates | Protein-Transport | 87.0 | 79.7 | 83.2 | 34.3 |
| GeneRIF-only predicates | Protein-Transport | 86.3 | 72.1 | 78.6 | 65.7 |
This table shows precision, recall, F-measure split up by the predicate part of speech and domain. A predicate was considered to be in the Wall Street Journal domain if the word appeared anywhere in the Wall Street Journal section of the Penn TreeBank, and was considered a GeneRIF-only predicate otherwise. The second column indicates which model (which feature set) the performance numbers are for. The final column indicates the percent of roles accounted for by each part of speech and each predicate domain.
Figure 1Example syntactic tree. This figure shows the syntactic tree for the phrase This protein, overexpressed in prostate cancer, shuttles between the cytoplasm and the nucleus.
Corpus statistics.
| All | Train | Test | |
| GeneRIFs | 837 | 637 | 200 |
| Words | 21620 | 16446 | 5174 |
| Unique words | 3841 | 3249 | 1459 |
| Predicates | 911 | 693 | 218 |
| Unique predicates | 86 | 72 | 44 |
| Unique predicate lemmas | 34 | 28 | 25 |
| Roles | 1544 | 1159 | 385 |
| AGENT roles | 17 | 14 | 3 |
| PATIENT roles | 822 | 623 | 199 |
| ORIGIN roles | 173 | 128 | 45 |
| DESTINATION roles | 532 | 394 | 138 |
This table shows some basic statistics for the semantic roles annotated over the GeneRIFs in the protein transport corpus.
Newswire semantic role chunk labels.
| Sales | B_ARG0 |
| declined | O |
| 10 | B_ARG2 |
| % | I_ARG2 |
| to | O |
| $ | B_ARG4 |
| 251.2 | I_ARG4 |
| million | I_ARG4 |
| from | O |
| $ | B_ARG3 |
| 278.7 | I_ARG3 |
| million | I_ARG3 |
| . | O |
This table shows the semantic role chunk labels for the Wall Street Journal sentence Sales declined 10% to $251.2 million from $258.7 million.
GeneRIF semantic role chunk labels.
| to | TO | O |
| induce | VB | O |
| the | DT | O |
| nuclear | JJ | B_DESTINATION |
| translocation | NN | O |
| of | IN | O |
| NF-kappaB | NN | B_PATIENT |
| transcription | NN | I_PATIENT |
| factor | NN | I_PATIENT |
This table shows the semantic role chunk labels for the GeneRIF phrase to induce the nuclear translocation of NF-kappaB transcription factor. The table also includes Penn TreeBank style part of speech tags for each word, and identifies what a one-word feature window for the classification of the word transcription looks like.
Features for Example 13.
| Word | Predicate | POS | Phrase | Clause |
| by | import | IN | B-PP | * |
| increasing | import | VBG | B-VP | (* |
| RING-dependent | import | JJ | B-NP | * |
| BRCA1 | import | NNP | I-NP | * |
| nuclear | import | JJ | I-NP | * |
| import | import | NN | I-NP | * |
| and | import | CC | O | * |
| inhibiting | import | VBG | B-VP | * |
| BRCA1 | import | JJ | B-NP | * |
| nuclear | import | JJ | I-NP | * |
| export | import | NN | I-NP | *) |
This table shows the basic word-level features that were used to characterize each word in Example 13.