| Literature DB >> 19073593 |
Yusuke Miyao1, Kenji Sagae, Rune Saetre, Takuya Matsuzaki, Jun'ichi Tsujii.
Abstract
MOTIVATION: While text mining technologies for biomedical research have gained popularity as a way to take advantage of the explosive growth of information in text form in biomedical papers, selecting appropriate natural language processing (NLP) tools is still difficult for researchers who are not familiar with recent advances in NLP. This article provides a comparative evaluation of several state-of-the-art natural language parsers, focusing on the task of extracting protein-protein interaction (PPI) from biomedical papers. We measure how each parser, and its output representation, contributes to accuracy improvement when the parser is used as a component in a PPI system.Entities:
Mesh:
Substances:
Year: 2008 PMID: 19073593 PMCID: PMC2639072 DOI: 10.1093/bioinformatics/btn631
Source DB: PubMed Journal: Bioinformatics ISSN: 1367-4803 Impact factor: 6.937
Fig. 1.CoNLL-X dependency tree (CoNLL).
Fig. 2.Penn Treebank-style phrase structure tree (PTB).
Fig. 3.Head dependencies (HD).
Fig. 4.Stanford dependencies (SD).
Fig. 5.Predicate argument structure (PAS).
Fig. 6.Sentences including protein names.
Fig. 7.Dependency path.
Fig. 8.Tree representation of a dependency path.
Parser output representations
| CoNLL | Dependency trees used in CoNLL 2006 and 2007 |
| PTB | Penn Treebank-style phrase structure trees |
| HD | Dependency trees of lexical heads |
| SD | Stanford dependency graphs |
| PAS | Predicate argument structures |
Fig. 9.Conversion of parser output representations.
Parsing time and accuracy (precision/recall/f -score) on the PPI task
| Time (s) | CoNLL | PTB | HD | SD | PAS | |
|---|---|---|---|---|---|---|
| MST | 425 | 49.1/65.6/55.9 | N/A | N/A | N/A | N/A |
| KSDEP | 111 | 51.6/67.5/58.3 | N/A | N/A | N/A | N/A |
| NO-RERANK | 1372 | 53.9/60.3/56.8 | 51.3/54.9/52.8 | 53.1/60.2/56.3 | 54.6/58.1/56.2 | N/A |
| RERANK | 2125 | 52.8/61.5/56.6 | 48.3/58.0/52.6 | 52.1/60.3/55.7 | 53.0/61.1/56.7 | N/A |
| BERKELEY | 1198 | 52.7/60.3/56.0 | 48.0/59.9/53.1 | 54.9/54.6/54.6 | 50.5/63.2/55.9 | N/A |
| STANFORD | 1645 | 49.3/62.8/55.1 | 44.5/64.7/52.5 | 49.0/62.0/54.5 | 54.6/57.5/55.8 | N/A |
| ENJU | 727 | 54.4/59.7/56.7 | 48.3/60.6/53.6 | 56.7/55.6/56.0 | 54.4/59.3/56.6 | 52.0/63.8/57.2 |
| ENJU-GENIA | 821 | 56.4/57.4/56.7 | 46.5/63.9/53.7 | 53.4/60.2/56.4 | 55.2/58.3/56.5 | 57.5/59.8/58.4 |
| Baseline | 48.2/54.9/51.1 | |||||
Results of parser/representation ensemble (f -score)
| RERANK CoNLL | HD | SD | ENJU CoNLL | HD | SD | PAS | ||
|---|---|---|---|---|---|---|---|---|
| KSDEP | CoNLL | 58.5 (+0.2) | 57.1 (-1.2) | 58.4 (+0.1) | 58.5 (+0.2) | 58.0 (-0.3) | 59.1 (+0.8) | 59.0 (+0.7) |
| RERANK | CoNLL | 56.7 (+0.1) | 57.1 (+0.4) | 58.3 (+1.6) | 57.3 (+0.7) | 58.7 (+2.1) | 59.5 (+2.3) | |
| HD | 56.8 (+0.1) | 57.2 (+0.5) | 56.5 (+0.5) | 56.8 (+0.2) | 57.6 (+0.4) | |||
| SD | 58.3 (+1.6) | 58.3 (+1.6) | 56.9 (+0.2) | 58.6 (+1.4) | ||||
| ENJU | CoNLL | 57.0 (+0.3) | 57.2 (+0.5) | 58.4 (+1.2) | ||||
| HD | 57.1 (+0.5) | 58.1 (+0.9) | ||||||
| SD | 58.3 (+1.1) |
Fig. 10.Parser training set size (number of sentences) versus parse accuracy and PPI extraction accuracy (f-score).
Fig. 11.Parser accuracy (f-score) versus PPI extraction accuracy (f-score).
Comparison with previous results on PPI extraction (f -score)
| Bag-of-words features | 51.1 |
| Yakushiji | 33.4 |
| Mitsumori | 47.7 |
| Giuliano | 52.4 |
| Sætre | 52.0 |
| Airola | 56.4 |
| This article | 59.5 |