| Literature DB >> 29335022 |
Daniel Faria1, Catia Pesquita2, Isabela Mott2, Catarina Martins3, Francisco M Couto2, Isabel F Cruz4.
Abstract
BACKGROUND: Biomedical ontologies pose several challenges to ontology matching due both to the complexity of the biomedical domain and to the characteristics of the ontologies themselves. The biomedical tracks in the Ontology Matching Evaluation Initiative (OAEI) have spurred the development of matching systems able to tackle these challenges, and benchmarked their general performance. In this study, we dissect the strategies employed by matching systems to tackle the challenges of matching biomedical ontologies and gauge the impact of the challenges themselves on matching performance, using the AgreementMakerLight (AML) system as the platform for this study.Entities:
Keywords: Biomedical ontologies; Ontology matching
Mesh:
Year: 2018 PMID: 29335022 PMCID: PMC5769431 DOI: 10.1186/s13326-017-0170-9
Source DB: PubMed Journal: J Biomed Semantics
Overview of the ontology matching systems that participated in OAEI biomedical tracks
| System | Size | Lexicon | Relations | Repair | Background knowledge | OAEI Bio tracks |
|---|---|---|---|---|---|---|
| AgrMaker [ | + | weights |
| - | Bio; Man; Med | A |
| AML [ | +++ | WN; weights | all | Logic | Bio; Auto; M/E | all |
| Anchor-Flood [ | + | WN | - | - | Man; Exp | A |
| Aroma [ | +++ | - | - | - | - | A, LB |
| ASMOV [ | + | WN | - | - | Bio; Man; Exp | A |
| AUTOMSv2[ | ++ | WN; weights | - | - | Man; Exp | LB- |
| BLOOMS [ | + | WN |
| - | Bio; Man; Med | A |
| COMMAND [ | + | - | - | Logic | - | A |
| CroMatcher [ | + | WN | - | - | Man; Exp | A |
| DKP-AOM [ | + | WN | - | - | Man; Exp | A, LB- |
| DSSim [ | + | WN | - | - | Man; Exp | A |
| FCA-Map [ | ++ | UMLS | - | Logic | Man; Exp | all- |
| GMap [ | + | external? | - | Logic | Man; Exp | A |
| GOMMA [ | +++ | - | - | - | Bio; Auto; Med | A, LB |
| kosimap [ | + | - | - | - | - | A |
| LogMap [ | +++ | WN; UMLS | - | Logic | Bio; Auto; M/E | all |
| LP HOM [ | + | - | - | - | A | |
| Lyam++[ | ++ | BabelNet | - | - | Man; Exp | all- |
| MapPSO [ | + | - | - | - | - | A |
| OACAS [ | + | - | - | - | - | A |
| PhenomeNET [ | ++ | (AML) |
| - | Bio; Man; Med | DP |
| SAMBO [ | + | WN |
| - | Bio; Man; Exp | A |
| ServOMap [ | +++ | WN | - | Logic | Man; Exp | A, LB |
| TaxoMap [ | + | - | - | - | - | A |
| TOAST [ | + | - | - | - | - | A |
| WikiMatch [ | ++ | Wikipedia | - | - | Man; Exp | A, LB- |
| YAM++[ | +++ | WN | - | Rules | Man; Exp | A, LB |
Size details the systems’ capability of handling medium-sized (+), large (++) or very large (+++) ontologies; Lexicon lists their use of lexical tools such as WordNet (WN) or the UMLS SPECIALIST Lexicon (UMLS), as well as whether they use synonym weights; Relations lists the types of relations they contemplate in addition to subclass relations; Repair details whether they perform alignment repair based on logic or rules; Background Knowledge describes whether they use biomedical ontologies as background knowledge (Bio), whether the process of background knowledge selection is manual (Man) or automatic (Auto), and whether background knowledge is used as a mediator (Med) or for lexical expansion (Exp); OAEI Bio Tracks lists the tracks in which the system successfully competed in out of Anatomy (A), Large Biomedical Ontologies (LB), and Disease & Phenotype (DP), with - indicating that the system did not complete the largest LB tasks
Fig. 1Flowchart representation of AML’s pipeline. The pipeline is divided into three stages: ontology loading, where input or background knowledge (BK) ontologies are parsed and loaded into AML’s data structures; ontology matching, where matching algorithms generate mapping candidates which are combined into a preliminary alignment; and filtering, where problem-causing mappings are removed from the preliminary alignment to produce a final alignment
Fig. 2Example of a logical definition shared by a Human Phenotype Ontology and a Mammalian Phenotype Ontology class, which enables their mapping
Fig. 3Run time versus number of lexical entries, in a log-log scale, for the hash-based searching LexicalMatcher and its functional equivalent but pairwise comparing QuadraticLexicalMatcher. Power law regression curve and equation are shown for each algorithm. The LexicalMatcher was executed using a single CPU thread, whereas the QuadraticLexicalMatcher ran asynchronously on 4 CPU threads
Run time comparison between the hash-based LexicalMatcher (in milliseconds) and its functional equivalent QuadraticLexicalMatcher that performs pairwise comparisons (in seconds), on all biomedical OAEI 2016 tasks
| Task |
|
| ||
|---|---|---|---|---|
| Hash-based searches | Time (ms) | Pairwise comparisons | Time (s) | |
| Anatomy | 3,072 | 38 | 15,587,328 | 9 |
| FMA-NCI small | 11,515 | 65 | 224,715,225 | 84 |
| FMA-SNOMED small | 26,295 | 311 | 959,057,535 | 323 |
| DOID-ORDO | 40,683 | 191 | 2,892,154,470 | 1,071 |
| HP-MP | 59,240 | 358 | 3,696,339,040 | 1,672 |
| SNOMED-NCI small | 76,134 | 372 | 5,803,999,356 | 2,579 |
| FMA-SNOMED whole | 184,484 | 761 | 41,136,795,772 | 14,984 |
| SNOMED-NCI whole | 184,484 | 801 | 35,189,031,612 | 15,640 |
| FMA-NCI whole | 190,743 | 721 | 42,532,446,369 | 15,509 |
Fig. 4Run time versus number of classes, in a log-log scale, for the StringMatcher applied locally, in the neighborhood of the mappings found by the other matching algorithms in AML’s pipeline. Both power law regression curve and equation, and bounding n.log(n) curve and equation are shown. The algorithm ran asynchronously on 4 CPU threads
Evaluation of AML’s full matching pipeline with the StringMatcher run locally and run globally
| Task | Local | Global | ||||||
|---|---|---|---|---|---|---|---|---|
| Prc | Rec | F-m | Time (s) | Prc | Rec | F-m | Time (s) | |
| Anatomy | 95.0% | 93.6% | 94.3% | 0.74 | 94.1% | 93.5% | 93.8% | 141 |
| FMA-NCI small | 95.8% | 91.0% | 93.3% | 0.36 | 95.1% | 91.5% | 93.2% | 164 |
| FMA-SNOMED small | 92.3% | 76.2% | 83.5% | 0.85 | 88.2% | 79.7% | 83.8% | 796 |
| SNOMED-NCI small | 91.4% | 73.6% | 81.6% | 9.7 | 86.1% | 75.5% | 80.5% | 5010 |
Precision (Prc), Recall (Rec), and F-measure (F-m) of the alignment produced by each strategy, and execution time of the StringMatcher algorithm
Comparison between the LexicalMatcher using all class names and synonyms, and using only primary names
| Task | All names & synonyms | Primary name | ||||
|---|---|---|---|---|---|---|
| Precision | Recall | F-measure | Precision | Recall | F-measure | |
| Anatomy | 96.1% | 69.7% | 80.8% | 99.7% | 61.5% | 76.1% |
| FMA-NCI small | 97.5% | 77.5% | 86.4% | 99.5% | 51.7% | 68.1% |
| FMA-SNOMED small | 98.8% | 19.9% | 33.2% | 99.2% | 15.1% | 26.2% |
| SNOMED-NCI small | 96.5% | 52.6% | 68.1% | 98.9% | 40.6% | 57.6% |
| FMA-NCI whole | 69.5% | 77.5% | 73.3% | 96.7% | 51.7% | 67.4% |
| FMA-SNOMED whole | 94.6% | 19.9% | 32.9% | 96.6% | 15.1% | 26.1% |
| SNOMED-NCI whole | 86.7% | 52.6% | 65.5% | 96.8% | 40.6% | 57.2% |
Comparison between AML’s full matching pipeline with and without the use of Lexicon weights to score the mappings
| Task | No | |||||
|---|---|---|---|---|---|---|
| Precision | Recall | F-measure | Precision | Recall | F-measure | |
| Anatomy | 95.0% | 93.6% | 94.3% | 81.3% | 95.8% | 88.0% |
| FMA-NCI small | 95.8% | 91.0% | 93.3% | 94.6% | 91.1% | 92.8% |
| FMA-SNOMED small | 92.3% | 76.2% | 83.5% | 90.9% | 77.3% | 83.5% |
| SNOMED-NCI small | 91.4% | 73.6% | 81.6% | 89.9% | 74.1% | 81.2% |
| FMA-NCI whole | 83.8% | 87.2% | 85.5% | 78.3% | 86.4% | 82.2% |
| FMA-SNOMED whole | 88.0% | 69.0% | 77.4% | 84.4% | 70.3% | 76.7% |
| SNOMED-NCI whole | 89.7% | 67.1% | 76.8% | 86.4% | 67.1% | 75.6% |
Comparison between the combination of LexicalMatcher and ThesaurusMatcher, and the LexicalMatcher alone
| Task | Lexical+Thesaurus | Lexical only | ||||
|---|---|---|---|---|---|---|
| Precision | Recall | F-measure | Precision | Recall | F-measure | |
| Anatomy | 95.7% | 71.4% | 81.8% | 96.1% | 69.7% | 80.8% |
| FMA-NCI small | 95.8% | 83.7% | 89.3% | 96.8% | 81.8% | 88.7% |
| FMA-SNOMED small | 96.9% | 62.9% | 76.3% | 97.8% | 62.2% | 76.0% |
| SNOMED-NCI small | 95.5% | 59.9% | 73.6% | 95.9% | 59.1% | 73.1% |
| FMA-NCI whole | 63.0% | 82.9% | 71.6% | 65.4% | 81.8% | 72.7% |
| FMA-SNOMED whole | 90.1% | 62.8% | 74.0% | 92.8% | 62.2% | 74.5% |
| SNOMED-NCI whole | 81.3% | 59.7% | 68.8% | 82.0% | 59.1% | 68.7% |
Evaluation of AML’s matching pipeline in the Anatomy and FMA-NCI small tasks with different combinations of background knowledge information source (lexical vs. cross-references) and usage strategies (mediator vs. lexical expansion)
| Task | BK info | BK usage | Precision | Recall | F-measure |
|---|---|---|---|---|---|
| Anatomy | Lexical | Mediator | 94.8% | 91.0% | 92.9% |
| (1365-1352) | Expansion | 94.6% | 90.2% | 92.4% | |
| Cross-refs | Mediator | 93.4% | 92.5% | 93.0% | |
| (1389-1401) | Expansion | 95.0% | 93.6% | 94.3% | |
| FMA-NCI small | Lexical | Mediator | 93.9% | 91.7% | 92.8% |
| (1624-1598) | Expansion | 93.6% | 91.5% | 92.5% | |
| Cross-refs | Mediator | 95.8% | 91.0% | 93.3% | |
| (1541-1559) | Expansion | 94.2% | 92.0% | 93.1% |
The number of classes of the two input ontologies covered by each information source is shown within parentheses below the source in each dataset