| Literature DB >> 33742785 |
Karamarie Fecho1, James Balhoff1, Chris Bizon1, William E Byrd2, Sui Hang3, David Koslicki4, Stefano E Rensi5, Patrick L Schmitt1, Mathias J Wawer6, Mark Williams7, Stanley C Ahalt1.
Abstract
"Knowledge graphs" (KGs) have become a common approach for representing biomedical knowledge. In a KG, multiple biomedical data sets can be linked together as a graph representation, with nodes representing entities, such as "chemical substance" or "genes," and edges representing predicates, such as "causes" or "treats." Reasoning and inference algorithms can then be applied to the KG and used to generate new knowledge. We developed three KG-based question-answering systems as part of the Biomedical Data Translator program. These systems are typically tested and evaluated using traditional software engineering tools and approaches. In this study, we explored a team-based approach to test and evaluate the prototype "Translator Reasoners" through the application of Medical College Admission Test (MCAT) questions. Specifically, we describe three "hackathons," in which the developers of each of the three systems worked together with a moderator to determine whether the applications could be used to solve MCAT questions. The results demonstrate progressive improvement in system performance, with 0% (0/5) correct answers during the first hackathon, 75% (3/4) correct during the second hackathon, and 100% (5/5) correct during the final hackathon. We discuss the technical and sociologic lessons learned and conclude that MCAT questions can be applied successfully in the context of moderated hackathons to test and evaluate prototype KG-based question-answering systems, identify gaps in current capabilities, and improve performance. Finally, we highlight several published clinical and translational science applications of the Translator Reasoners.Entities:
Mesh:
Year: 2021 PMID: 33742785 PMCID: PMC8504839 DOI: 10.1111/cts.13021
Source DB: PubMed Journal: Clin Transl Sci ISSN: 1752-8054 Impact factor: 4.689
Translator performance and lessons learned when applying Translator Reasoners to answer MCAT questions over three 4‐h moderated hackathons
| Hackathon date | Success rate | Lessons learned |
|---|---|---|
| January 2019 | 0/5 questions (0%) |
Missing/incomplete data sources Errors with existing data sources Inadequate specificity with existing data sources Entity identifier mismatches “One‐hop” graph queries insufficient |
| July 2019 | 3/4 questions (75%) |
Missing/incomplete relationships between entities Limited or absent annotation for certain data sources Lack of relative/contextual relationships “Opposites” under‐represented or absent in data sources “Synonymization” or equivalence of text terms challenging Lack of differentiation or unclear differentiation between data types (e.g., disease vs. phenotype, protein vs. gene) Multiple implementation strategies (e.g., direct match, process of elimination, and inference) improves success rate |
| September 2019 | 5/5 questions (100%) |
“Two‐hop” graph queries and other more complex queries more effective than “one‐hop” queries Query directionality and choice of predicate important Missing or incomplete predicates Terminology challenges with pluralities Exact matches to correct answers uncommon Generalization and inference required for terms that lack specificity Careful review of supporting evidence improves success rate Biomedical input facilitates developer identification of correct answer |
Abbreviation: MCAT, Medical College Admission Test.
The goal was to tackle five questions for this hackathon session, but only four questions were attempted due to time constraints.
The correct answer to one of the five questions was confirmed during a subsequent November 2019 meeting with the moderator and the lead developer of one of the Translator Reasoners who was unable to attend the September 2019 hackathon.