| Literature DB >> 34920707 |
Mayla R Boguslav1, Negacy D Hailu2, Michael Bada2, William A Baumgartner2, Lawrence E Hunter2.
Abstract
BACKGROUND: Automated assignment of specific ontology concepts to mentions in text is a critical task in biomedical natural language processing, and the subject of many open shared tasks. Although the current state of the art involves the use of neural network language models as a post-processing step, the very large number of ontology classes to be recognized and the limited amount of gold-standard training data has impeded the creation of end-to-end systems based entirely on machine learning. Recently, Hailu et al. recast the concept recognition problem as a type of machine translation and demonstrated that sequence-to-sequence machine learning models have the potential to outperform multi-class classification approaches.Entities:
Keywords: Computational resources; Concept recognition; Machine translation; Named entity normalization; Named entity recognition
Year: 2021 PMID: 34920707 PMCID: PMC8678974 DOI: 10.1186/s12859-021-04141-4
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1Example of the full translation pipeline. Each step is seen as a translation problem. The input is text and the final output is the ontology class identifiers for each detected text mention
Statistics for the concept annotations in the training (67-document) and evaluation (30-document) data sets for all ontologies
| Ontology | # training set annotations | avg/median # training set annotations per article | # evaluation set annotations | Avg/median # evaluation set annotations per article |
|---|---|---|---|---|
| ChEBI | 4548 | 68/45 | 2200 | 73/45 |
| ChEBI_EXT | 11,915 | 178/142 | 5248 | 175/142 |
| CL | 4043 | 60/32 | 1749 | 58/32 |
| CL_EXT | 6276 | 94/64 | 2872 | 96/64 |
| GO_BP | 9280 | 139/108 | 3681 | 123/108 |
| GO_BP_EXT | 13,954 | 208/158 | 5847 | 195/158 |
| GO_CC | 4075 | 61/33 | 1184 | 39/33 |
| GO_CC_EXT | 8495 | 127/91 | 3217 | 107/91 |
| GO_MF | 375 | 6/2 | 94 | 3/2 |
| GO_MF_EXT | 4070 | 61/34 | 1822 | 61/34 |
| MOP | 240 | 4/2 | 101 | 3/2 |
| MOP_EXT | 386 | 6/2 | 111 | 4/2 |
| NCBITaxon | 7362 | 110/90 | 3101 | 103/90 |
| NCBITaxon_EXT | 7592 | 113/97 | 3219 | 107/97 |
| PR | 17,038 | 254/198 | 6409 | 214/198 |
| PR_EXT | 19,862 | 296/246 | 7932 | 264/246 |
| SO | 8797 | 131/118 | 3446 | 115/118 |
| SO_EXT | 24,955 | 372/341 | 9136 | 305/341 |
| UBERON | 12,269 | 183/130 | 6551 | 218/130 |
| UBERON_EXT | 14,910 | 223/165 | 7416 | 247/165 |
Avg average
Statistics for the concept annotation classes used in the training (67-document) and evaluation (30-document) data sets and for those added as additional training data for concept normalization for all ontologies
| Ontology | # training set annotation classes | Avg/median # training set annotation classes per article | # classes added to training set | # evaluation set annotation classes | Avg/median # evaluation set annotation classes per article |
|---|---|---|---|---|---|
| ChEBI | 1463 | 22/18 | 58,214 | 627 | 21/20 |
| ChEBI_EXT | 2852 | 43/38 | 58,439 | 1167 | 39/39 |
| CL | 581 | 9/7 | 2163 | 253 | 8/9 |
| CL_EXT | 651 | 10/8 | 2168 | 286 | 10/10 |
| GO_BP | 1586 | 24/21 | 29,213 | 682 | 23/23 |
| GO_BP_EXT | 2511 | 37/33 | 29,301 | 1090 | 36/37 |
| GO_CC | 677 | 10/9 | 4052 | 212 | 7/6 |
| GO_CC_EXT | 896 | 13/12 | 4086 | 296 | 10/9 |
| GO_MF | 49 | 1/1 | 10951 | 19 | 1/1 |
| GO_MF_EXT | 738 | 11/11 | 10,031 | 377 | 13/12 |
| MOP | 85 | 1/1 | 3574 | 32 | 1/1 |
| MOP_EXT | 108 | 2/1 | 3578 | 40 | 1/1 |
| NCBITaxon | 690 | 10/9 | 1,175,661 | 315 | 11/9 |
| NCBITaxon_EXT | 757 | 11/10 | 1,175,682 | 346 | 12/10 |
| PR | 1278 | 19/18 | 213,371 | 466 | 16/16 |
| PR_EXT | 1534 | 23/22 | 213,531 | 588 | 20/19 |
| SO | 1216 | 18/18 | 2256 | 544 | 18/19 |
| SO_EXT | 3172 | 47/47 | 2405 | 1409 | 47/48 |
| UBERON | 2048 | 31/24 | 14,057 | 1040 | 35/31 |
| UBERON_EXT | 2409 | 36/29 | 14,113 | 1217 | 41/38 |
Avg average
BIO(−) labeling for the discontinuous and overlapping ontology class mentions in the phrase “red and white blood cells” (from PMCID:15314655)
| Red | And | White | Blood | Cells | |
|---|---|---|---|---|---|
| Labels for the annotation of | B | O− | O− | I | I |
| Labels for the annotation of | O | O | B | I | I |
| Final labeling | B | O− | B | I | I |
The O− would simply be O in the canonical BIO labeling
Quantification of discontinuous and overlapping words in all concept mentions
| Ontology | # words in all concept mentions | % words in discontinuous mentions (%) | % words between text spans of discontinuous mentions (%) | % words overlapping multiple mentions (%) |
|---|---|---|---|---|
| ChEBI | 5985 | 0.3 | 0.6 | 0.1 |
| CL | 6576 | 4.3 | 4.3 | 2.6 |
| GO_BP | 12,956 | 5.2 | 7.0 | 1.6 |
| GO_CC | 5864 | 1.5 | 2.1 | 0.5 |
| GO_MF | 376 | 0 | 0 | 0 |
| MOP | 257 | 0 | 0 | 0 |
| NCBITaxon | 7696 | 0.03 | 0.03 | 0.03 |
| PR | 23,261 | 0.5 | 0.2 | 0.9 |
| SO | 10,348 | 1.2 | 1.8 | 0.5 |
| UBERON | 15,681 | 2.0 | 2.3 | 0.8 |
All numbers are based on the number of words, not concepts
Full end-to-end system evaluation on the core set comparing F1 score
| Ontology | CRF | BiLSTM | BiLSTM-CRF | Char-Embeddings | BiLSTM-ELMo | BioBERT | UZH@CRAFT-ST |
|---|---|---|---|---|---|---|---|
| ChEBI | 0.7882 | 0.6394 | 0.5027 | 0.5942 | 0.0550 | 0.7700 | |
| CL | 0.6779 | 0.5134 | 0.3859 | 0.5611 | 0.0526 | 0.6657 | |
| GO_BP | 0.7505 | 0.5137 | 0.3642 | 0.6182 | 0.0720 | 0.7405 | |
| GO_CC | 0.7225 | 0.1689 | 0.3049 | 0.3244 | 0.0506 | 0.7645 | |
| GO_MF | 0.9778 | 0.9770 | 0.9778 | 0.8906 | 0.3704 | 0.9783 | |
| MOP | 0.8129 | 0.7721 | 0.7158 | 0.5985 | 0.0930 | 0.8705 | |
| NCBITaxon | 0.9026 | 0.7736 | 0.8391 | 0.8518 | 0.0948 | 0.8910 | |
| PR | 0.4040 | 0.3136 | 0.2827 | 0.2732 | 0.0516 | 0.5295 | |
| SO | 0.8987 | 0.4106 | 0.4096 | 0.7815 | 0.0813 | 0.9027 | |
| UBERON | 0.7474 | 0.6812 | 0.5029 | 0.6901 | 0.0793 | 0.7488 |
For all results shown here, the span detection algorithm is listed, and the concept normalization algorithm is OpenNMT. UZH@CRAFT-ST is the best performing system from Furrer et al. [4] in the CRAFT-ST, shown as a comparison to our methods. The best-performing algorithm is bolded with an asterisk*
Full end-to-end system evaluation on the core + extensions set comparing F1 score for the top two algorithms found in the core set
| Ontology | CRF | BioBERT | UZH@CRAFT-ST |
|---|---|---|---|
| ChEBI_EXT | 0.7891 | 0.8039 | |
| CL_EXT | 0.7381 | 0.7484 | |
| GO_BP_EXT | 0.7279 | 0.7353 | |
| GO_CC_EXT | 0.8738 | 0.8936 | |
| GO_MF_EXT | 0.6413 | 0.6255 | |
| MOP_EXT | 0.8000 | 0.8437 | |
| NCBITaxon_EXT | 0.8710 | 0.8624 | |
| PR_EXT | 0.4397 | 0.5188 | |
| SO_EXT | 0.7682 | 0.7829 | |
| UBERON_EXT | 0.7558 | 0.7711 |
For all results shown here, the span detection algorithm is listed, and the concept normalization algorithm is OpenNMT. UZH@CRAFT-ST is the best performing system from Furrer et al. [4] in the CRAFT-ST, shown as a comparison to our methods. The best-performing algorithm is bolded with an asterisk*
Hardware, memory, and time used for training for all evaluated algorithms
| Algorithm | Hardware | Training memory (GBs) | Training time (h) |
|---|---|---|---|
| CRF | CPUs | 2–13 | 1–4 |
| BiLSTM* | GPUs/CPUs** | 17 | 29 |
| BiLSTM-CRF | CPUs | 7 | 15 |
| Char-Embeddings | CPUs | 30 | 84 |
| BiLSTM-ELMo* | GPUs | 42 | 700–1000 |
| BioBERT | GPUs/CPUs** | 5 | 20 |
| UZH@CRAFT-ST BioBERT* [ | GPUS | 120*** | 200 |
| OpenNMT* | CPUs | 620 | 515 |
| ConceptMapper [ | CPUs | N/A | N/A |
A given training time specifies the total hours if training for all ontology annotation sets were run consecutively, but these can be parallelized by ontology
ConceptMapper runs on CPUs but has no training, as it is a dictionary-based lookup tool, hence the specifications as N/A
*Parallelized per ontology due to time constraints
**Runs significantly faster on GPUs
***Total free RAM available
Span detection F1 score results for all algorithms tested against the core evaluation annotation set of the 30 held-out articles
| Ontology | CRF | BiLSTM | BiLSTM-CRF | Char-Embeddings | BiLSTM-ELMo | BioBERT |
|---|---|---|---|---|---|---|
| ChEBI | 0.7234 | 0.6545 | 0.5000 | 0.5280 | 0.0620 | |
| CL | 0.8333 | 0.5882 | 0.3774 | 0.8000 | 0.0000 | |
| GO_BP | 0.5498 | 0.3661 | 0.6346 | 0.0685 | 0.8646 | |
| GO_CC | 0.9412 | 0.1379 | 0.2689 | 0.2581 | 0.1000 | |
| GO_MF | > | > | > | 0.8421 | 0.0000 | > |
| MOP | > | > | > | > | 0.0000 | > |
| NCBITaxon | 0.8551 | 0.9440 | 0.9569 | 0.0711 | 0.9453 | |
| PR | 0.4351 | 0.2979 | 0.2151 | 0.0995 | 0.0339 | |
| SO | 0.4935 | 0.4897 | 0.8203 | 0.1059 | 0.9081 | |
| UBERON | 0.7913 | 0.7206 | 0.4758 | 0.7440 | 0.0854 |
The best-performing algorithm per ontology is bolded with an asterisk*
Span detection F1 score results for all algorithms tested against the core + extensions evaluation annotation set of the 30 held-out articles
| Ontology | CRF | BioBERT |
|---|---|---|
| ChEBI_EXT | 0.8802 | |
| CL_EXT | 0.8000 | |
| GO_BP_EXT | 0.8516 | |
| GO_CC_EXT | 0.8667 | |
| GO_MF_EXT | 0.9211 | |
| MOP_EXT | > | > |
| NCBITaxon_EXT | 0.9919 | |
| PR_EXT | 0.5598 | |
| SO_EXT | 0.8054 | |
| UBERON_EXT | 0.8418 |
The best-performing algorithm per ontology is bolded with an asterisk*
F1 score results for detection of discontinuous spans for all algorithms tested against the core evaluation annotation set of the 30 held-out articles
| Ontology | Support | CRF | BiLSTM | BiLSTM-CRF | Char-Embeddings | BiLSTM-ELMo | BioBERT |
|---|---|---|---|---|---|---|---|
| ChEBI | 14 | 0 | 0 | 0 | 0 | 0 | 0 |
| CL | 175 | 0.1176 | 0.1158 | 0.1171 | 0 | 0.0107 | 0.1818 |
| GO_BP | 272 | 0.0952 | 0 | 0.014 | 0 | 0.007 | 0.2742 |
| GO_CC | 14 | 0.1053 | 0 | 0 | 0.0526 | 0 | 0.3 |
| PR | 44 | 0 | 0 | 0 | 0 | 0 | 0.08 |
| SO | 45 | 0.0408 | 0 | 0 | 0 | 0 | 0.3 |
| UBERON | 118 | 0.04 | 0 | 0 | 0 | 0 | 0.0915 |
Note that there are no discontinuous spans in the GO_MF, MOP, and NCBITaxon sets
F1 score results for detection of discontinuous spans for all algorithms tested against the core + extensions evaluation annotation set of the 30 held-out documents
| Ontology | Support | CRF | BioBERT |
|---|---|---|---|
| ChEBI_EXT | 19 | 0 | 0 |
| CL_EXT | 175 | 0.1164 | 0.1608 |
| GO_BP_EXT | 287 | 0.085 | 0.2651 |
| GO_CC_EXT | 30 | 0.1579 | 0.3721 |
| GO_MF_EXT | 20 | 0 | 0 |
| PR_EXT | 44 | 0 | 0.1224 |
| SO_EXT | 72 | 0.1505 | 0.2979 |
| UBERON_EXT | 133 | 0.0485 | 0.2128 |
Note that there are no discontinuous spans in the MOP_EXT and NCBITaxon_EXT sets
CRF tuning parameters and resulting tuning F1 scores
| Ontology | L1 | L2 | Time (h) | F1 score (macro) | F1 score (micro) |
|---|---|---|---|---|---|
| ChEBI | 0.0862 | 0.000186 | 3 | 0.61 | 0.99 |
| CL | 0.00477 | 0.0473 | 3 | 0.73 | > 0.99 |
| GO_BP | 0.0862 | 0.000186 | 3.17 | 0.63 | 0.99 |
| GO_CC | 0.269 | 0.00892 | 3 | 0.55 | 0.99 |
| GO_MF | 0.215 | 0.00392 | 3 | 0.66 | 0.99 |
| MOP | 0.0862 | 0.000186 | 3 | 0.65 | > 0.99 |
| NCBITaxon | 0.0862 | 0.000186 | 3 | 0.61 | > 0.99 |
| PR | 0.00477 | 0.0473 | 3.25 | 0.54 | 0.97 |
| SO | 0.315 | 0.00578 | 3.22 | 0.66 | > 0.99 |
| UBERON | 0.221 | 0.003005 | 3.3 | 0.63 | 0.99 |
The overall memory usage for all tuning was 6 GB
BiLSTM tuning parameters and resulting tuning F1 scores that are used for the BiLSTM-CRF and Char-Embeddings models also
| Ontology | Batch size | # Epochs | # Neurons | Time (h) | Memory (GBs) | F1 score (macro) | F1 score (micro) |
|---|---|---|---|---|---|---|---|
| ChEBI | 53 | 10 | 12 | 99 | 6.5 | 0.67 | > 0.99 |
| CL | 36 | 10 | 12 | 92 | 6.5 | 0.74 | > 0.99 |
| GO_BP | 36 | 10 | 12 | 99 | 6.5 | 0.68 | > 0.99 |
| GO_CC | 106 | 100 | 12 | 97 | 6.5 | 0.66 | > 0.99 |
| GO_MF | 106 | 10 | 3 | 108 | 8.4 | 0.99 | > 0.99 |
| MOP | 106 | 10 | 12 | 99 | 6.4 | 0.61 | > 0.99 |
| NCBITaxon | 106 | 100 | 12 | 95 | 6.5 | 0.96 | > 0.99 |
| PR | 36 | 10 | 12 | 95 | 6.5 | 0.71 | > 0.99 |
| SO | 36 | 100 | 12 | 98 | 6.5 | 0.66 | > 0.99 |
| UBERON | 18 | 10 | 12 | 97 | 6.5 | 0.68 | > 0.99 |
BiLSTM-ELMo parameters and resulting tuning F1 scores
| Ontology | Batch size | # Epochs | # Neurons | F1 score (macro) | F1 score (micro) |
|---|---|---|---|---|---|
| ChEBI | 18 | 10 | 3 | 0.65 | > 0.99 |
| CL | 18 | 10 | 12 | 0.72 | > 0.99 |
| GO_BP | 18 | 10 | 12 | 0.66 | > 0.99 |
| GO_CC | 18 | 10 | 3 | 0.65 | > 0.99 |
| GO_MF | 18 | 100 | 3 | 0.66 | > 0.99 |
| MOP | 18 | 100 | 12 | 0.61 | > 0.99 |
| NCBITaxon | 18 | 100 | 12 | 0.96 | > 0.99 |
| PR | 18 | 10 | 12 | 0.71 | > 0.99 |
| SO | 18 | 100 | 12 | 0.65 | > 0.99 |
| UBERON | 18 | 10 | 12 | 0.68 | > 0.99 |
Due to limited resources, the batch size is 18 for all ontologies
Concept normalization exact match results on the core evaluation annotation set of the 30 held-out documents compared to the baseline ConceptMapper approach
| Ontology | % OpenNMT class ID (%) | % ConceptMapper class ID (%) | % ConceptMapper FN Class ID (%) | % OpenNMT character (%) | % ConceptMapper character (%) |
|---|---|---|---|---|---|
| ChEBI | 55 | 41 | 58 | ||
| CL | 52 | 12 | 77 | ||
| GO_BP | 29 | 59 | 36 | ||
| GO_CC | 54 | 44 | 55 | ||
| GO_MF | 0 | 100 | 0 | ||
| MOP | 65 | 34 | 66 | ||
| NCBITaxon | 86 | 13 | 87 | ||
| PR | 10 | 26 | 57 | ||
| SO | 75 | 21 | 78 | ||
| UBERON | 64 | 34 | 65 |
We report both the percent exact match at the class ID level and the character level. We also report the percentage of false negatives (FN) for ConceptMapper (i.e., no class ID prediction for a given text mention). Note that for each ontology the better performance between OpenNMT and ConceptMapper is bolded with an asterisk* for both class ID and character levels
Exact match results for the unseen and seen text mentions (relative to the training data) for the core evaluation annotation set of the 30 held-out documents
| Ontology | Total/unique # unseen mentions | % Unseen OpenNMT class ID (%) | % Seen OpenNMT class ID (%) | % Unseen OpenNMT character (%) | % Seen OpenNMT character (%) |
|---|---|---|---|---|---|
| ChEBI | 345/148 | 17 | 94 | 69 | 99 |
| CL | 774/208 | 39 | 98 | 92 | > 99 |
| GO_BP | 727/367 | 17 | 98 | 65 | 99 |
| GO_CC | 301/85 | 29 | 99 | 67 | > 99 |
| GO_MF | 3/3 | 33 | > 99 | 70 | > 99 |
| MOP | 18/7 | 83 | 98 | 96 | > 99 |
| NCBITaxon | 81/52 | 0 | 89 | 72 | 98 |
| PR | 2926/388 | 0 | 19 | 71 | 80 |
| SO | 181/105 | 62 | 99 | 89 | > 99 |
| UBERON | 1584/514 | 20 | 97 | 80 | 99 |
Reporting the total number of mentions and the number of unique mentions along with the percent exact match on the class ID level and character level for both unseen and seen text mentions
Concept normalization exact match results on the core + extensions evaluation annotation set of the 30 held-out documents compared to the baseline ConceptMapper approach
| Ontology | % OpenNMT class ID (%) | % ConceptMapper class ID (%) | % ConceptMapper FN class ID (%) | % OpenNMT character (%) | % ConceptMapper character (%) |
|---|---|---|---|---|---|
| ChEBI_EXT | 64 | 26 | 66 | ||
| CL_EXT | 67 | 11 | 84 | ||
| GO_BP_EXT | 34 | 44 | 38 | ||
| GO_CC_EXT | 80 | 18 | 84 | ||
| GO_MF_EXT | 60 | 30 | 64 | ||
| MOP_EXT | 64 | 35 | 44 | ||
| NCBITaxon_EXT | 83 | 13 | 87 | ||
| PR_EXT | 9 | 28 | 21 | ||
| SO_EXT | 19 | 40 | 22 | ||
| UBERON_EXT | 68 | 29 | 75 |
We report both the percent exact match on the class ID level and the character level. We also report the percentage of false negatives (FN) for ConceptMapper (i.e. no class ID prediction for a given text mention). Note that the best performance between OpenNMT and ConceptMapper is bolded with an asterisk* for both class ID and character level
Exact match results for the unseen and seen text mentions (relative to the training data) for the core + extensions evaluation annotation set of the 30 held-out documents
| Ontology | Total/unique # unseen Mentions | % Unseen OpenNMT class ID (%) | % Seen OpenNMT class ID (%) | % Unseen OpenNMT character (%) | % Seen OpenNMT character (%) |
|---|---|---|---|---|---|
| ChEBI_EXT | 476/188 | 32 | 92 | 67 | 85 |
| CL_EXT | 775/209 | 36 | 99 | 77 | 97 |
| GO_BP_EXT | 861/431 | 26 | 89 | 57 | 78 |
| GO_CC_EXT | 339/113 | 39 | 99 | 57 | 98 |
| GO_MF_EXT | 515/146 | 31 | 83 | 45 | 78 |
| MOP_EXT | 21/10 | 67 | 98 | 85 | > 99 |
| NCBITaxon_EXT | 123/79 | 1 | 86 | 68 | 94 |
| PR_EXT | 3114/429 | 0 | 25 | 66 | 75 |
| SO_EXT | 318/183 | 51 | 94 | 69 | 92 |
| UBERON_EXT | 1609/532 | 23 | 96 | 79 | 95 |
Reporting the total number of mentions and the number of unique mentions along with the percent exact match on the class ID level and character level for both unseen and seen text mentions
Percentage of predicted non-existent class IDs out of the total number of predicted mismatch class IDs for the core set for the training, validation and evaluation sets
| Ontology | % Non-existent class IDs in training (%) | % Non-existent class IDs in validation (%) | % Non-existent class IDs in evaluation (%) |
|---|---|---|---|
| ChEBI | 3 | 4 | 11 |
| CL | 0 | 0 | 0 |
| GO_BP | 2 | 2 | 2 |
| GO_CC | 2 | 4 | 1 |
| GO_MF | 1 | 1 | 50 |
| MOP | 0 | 2 | 0 |
| NCBITaxon | 7 | 7 | 11 |
| PR | 10 | 10 | 2 |
| SO | 0 | 1 | 0 |
| UBERON | 0 | 1 | 2 |
Percentage of predicted non-existent class IDs out of the total number of predicted mismatch class IDs for the core + extensions set for the training, validation and evaluation sets
| Ontology | % Non-existent class IDs in training (%) | % Non-existent class IDs in validation (%) | % Non-existent class IDs in evaluation (%) |
|---|---|---|---|
| ChEBI_EXT | 8 | 8 | 17 |
| CL_EXT | 1 | 1 | 9 |
| GO_BP_EXT | 2 | 2 | 3 |
| GO_CC_EXT | 4 | 4 | 21 |
| GO_MF_EXT | 2 | 2 | 4 |
| MOP_EXT | 0 | 6 | 0 |
| NCBITaxon_EXT | 9 | 9 | 25 |
| PR_EXT | 7 | 7 | 5 |
| SO_EXT | 1 | 1 | 2 |
| UBERON_EXT | 1 | 1 | 1 |
Exact match results for the concept normalization experiments on the core evaluation annotation set of 30 held-out documents (class ID level)
| Ontology | Token-ids (%) | Type-ids (%) | Shuffled-ids (%) | Random-ids (%) | Alphabetical-ids (%) |
|---|---|---|---|---|---|
| ChEBI | 65 | 0 | 0 | 78 | |
| CL | 69 | 70 | 56 | ||
| GO_BP | 79 | 64 | 27 | 52 | |
| GO_CC | 81 | 80 | 81 | 76 | |
| GO_MF | 51 | ||||
| MOP | 95 | 92 | 80 | ||
| NCBITaxon | 0 | 0 | 0 | 0 | |
| PR | 3 | 0 | 0 | 8 | |
| SO | 96 | 96 | |||
| UBERON | 74 | 0 | 0 | 74 |
We report the exact match percentage at the class ID level. The highest percentage is bolded and with an asterisk*
Exact match results for the concept normalization experiments on the core evaluation annotation set of 30 held-out documents (character level)
| Ontology | Token-ids (%) | Type-ids (%) | Shuffled-ids (%) | Random-ids (%) | Alphabetical-ids (%) |
|---|---|---|---|---|---|
| ChEBI | 89 | 60 | 58 | ||
| CL | 86 | 80 | 65 | ||
| GO_BP | 91 | 85 | 56 | 84 | |
| GO_CC | 91 | 89 | 82 | ||
| GO_MF | 94 | ||||
| MOP | 99 | > | 99 | 99 | 86 |
| NCBITaxon | 74 | 73 | 68 | 74 | |
| PR | 75 | 40 | 30 | 74 | |
| SO | 98 | 96 | |||
| UBERON | 93 | 69 | 54 | 88 |
We report the exact match percentage at the character level. The highest percentage is bolded and with an asterisk*