| Literature DB >> 33784373 |
Janisha A Patel1, Dennis A Dean2, Charles Hadley King1,3, Nan Xiao2, Soner Koc2, Ekaterina Minina4, Anton Golikov4, Phillip Brooks2, Robel Kahsay1, Rahi Navelkar1, Manisha Ray2, Dave Roberson2, Chris Armstrong1, Raja Mazumder1,3, Jonathon Keeney1.
Abstract
Developments in high-throughput sequencing (HTS) result in an exponential increase in the amount of data generated by sequencing experiments, an increase in the complexity of bioinformatics analysis reporting and an increase in the types of data generated. These increases in volume, diversity and complexity of the data generated and their analysis expose the necessity of a structured and standardized reporting template. BioCompute Objects (BCOs) provide the requisite support for communication of HTS data analysis that includes support for workflow, as well as data, curation, accessibility and reproducibility of communication. BCOs standardize how researchers report provenance and the established verification and validation protocols used in workflows while also being robust enough to convey content integration or curation in knowledge bases. BCOs that encapsulate tools, platforms, datasets and workflows are FAIR (findable, accessible, interoperable and reusable) compliant. Providing operational workflow and data information facilitates interoperability between platforms and incorporation of future dataset within an HTS analysis for use within industrial, academic and regulatory settings. Cloud-based platforms, including High-performance Integrated Virtual Environment (HIVE), Cancer Genomics Cloud (CGC) and Galaxy, support BCO generation for users. Given the 100K+ userbase between these platforms, BioCompute can be leveraged for workflow documentation. In this paper, we report the availability of platform-dependent and platform-independent BCO tools: HIVE BCO App, CGC BCO App, Galaxy BCO API Extension and BCO Portal. Community engagement was utilized to evaluate tool efficacy. We demonstrate that these tools further advance BCO creation from text editing approaches used in earlier releases of the standard. Moreover, we demonstrate that integrating BCO generation within existing analysis platforms greatly streamlines BCO creation while capturing granular workflow details. We also demonstrate that the BCO tools described in the paper provide an approach to solve the long-standing challenge of standardizing workflow descriptions that are both human and machine readable while accommodating manual and automated curation with evidence tagging. Database URL: https://www.biocomputeobject.org/resources. © Oxford University Press 2021.Entities:
Year: 2021 PMID: 33784373 PMCID: PMC8009203 DOI: 10.1093/database/baab008
Source DB: PubMed Journal: Database (Oxford) ISSN: 1758-0463 Impact factor: 3.451
Figure 1.HIVE BCO App deployment diagram.
Figure 2.BCO portal deployment diagram.
Figure 3.User BCO training process.
Training module user survey
| Question | Answer |
|---|---|
| How clear were the evaluation unit objectives? | 1 (unclear) → 5 (extremely clear) |
| Did the structure and sequence of the lectures make sense? | 1 (not at all) → 5 (very much so) |
| Did the unit expose you to new knowledge, tools and practices? | Definitely, yes |
| Of the new knowledge, tools and practices this module taught, how comfortable do you feel using GitHub as a means of code/project management and collaboration? | 1 (not at all comfortable) → 5 (extremely comfortable) |
| Of the new knowledge, tools and practices this module taught, how comfortable do you feel with running computational analysis on HIVE? | 1 (not at all comfortable) → 5 (extremely comfortable) |
| Of the new knowledge, tools and practices this module taught, how comfortable do you feel with running computational analysis on CGC? | 1 (not at all comfortable) → 5 (extremely comfortable) |
| Of the new knowledge, tools and practices this module taught, how comfortable do you feel with generating BioCompute Objects in CGC? | 1 (not at all comfortable) → 5 (extremely comfortable) |
| Of the new knowledge, tools and practices this module taught, how comfortable do you feel with creating BioCompute Objects in the BCO Editor? | 1 (not at all comfortable) → 5 (extremely comfortable) |
| GUI: In your opinion, how user friendly is HIVE’s interface? | 1 (not at all) → 5 (extremely) |
| GUI: In your opinion, how user friendly is CGC’s interface? | 1 (not at all) → 5 (extremely) |
| GUI: In your opinion, how user friendly is BCO Editor’s interface? | 1 (not at all) → 5 (extremely) |
BCOs recording HIVE platform testing protocol
| Tool(s) | BCO(s) | Computation | Inputs |
|---|---|---|---|
| HIVE-Hexagon | BioCompute Portal: | HIVE Object ID: 3249 |
|
| GitHub: | HIVE Object ID: | ||
| HIVE-Heptagon | BioCompute Portal: | HIVE Object ID: 3260 |
|
| GitHub: | HIVE Object ID: 3249 | ||
| CensuScope | BioCompute Portal: | HIVE Object ID: 3242 |
|
| GitHub: | HIVE Object ID: 2218 |