| Literature DB >> 26287255 |
Yushu Yao1, Terence Sun2, Tony Wang3, Oliver Ruebel4, Trent Northen5, Benjamin P Bowen6.
Abstract
Even with the widespread use of liquid chromatography mass spectrometry (LC/MS) based metabolomics, there are still a number of challenges facing this promising technique. Many, diverse experimental workflows exist; yet there is a lack of infrastructure and systems for tracking and sharing of information. Here, we describe the Metabolite Atlas framework and interface that provides highly-efficient, web-based access to raw mass spectrometry data in concert with assertions about chemicals detected to help address some of these challenges. This integration, by design, enables experimentalists to explore their raw data, specify and refine features annotations such that they can be leveraged for future experiments. Fast queries of the data through the web using SciDB, a parallelized database for high performance computing, make this process operate quickly. By using scripting containers, such as IPython or Jupyter, to analyze the data, scientists can utilize a wide variety of freely available graphing, statistics, and information management resources. In addition, the interfaces facilitate integration with systems biology tools to ultimately link metabolomics data with biological models.Entities:
Keywords: IPython; LC/MS; MS/MS; Python; SciDB; biology; data analysis; metabolite atlas; metabolomics
Year: 2015 PMID: 26287255 PMCID: PMC4588804 DOI: 10.3390/metabo5030431
Source DB: PubMed Journal: Metabolites ISSN: 2218-1989
Integrated metabolite atlas API for simultaneously querying raw data along with compound specifications.
| Method | URL | Options | Description |
|---|---|---|---|
| GET | /run/ | {“L”:<level>, “P”:polarity, “arrayname”:<myArray>, “fileidlist”:<myList>, “max_mz”:<mzMax>, “min_mz”:<mzMin>, “min_rt”:<rtMin>, “max_rt”:<rtMax>, “nsteps”:<2000>, “queryType”:”XICofFile_mf”} JSON | Gets chromatograms for a given mz and rt specification for one or more files. |
| GET | /api/dict/<dict_id>/ | Gets details about a specified compound dictionary | |
| PUT | /api/dict/<dict_id>/ | {“<field_name>”: <field_val> ...} JSON | Completely replaces the compound dictionary fields with the JSON object |
| GET | /api/compound/<compound_id>/ | Gets details about a specified compound | |
| PATCH | /api/compound/<compound_id>/ | {“<field_name>”: <field_val> ..., “removed_fields”: [...list of removed field names...]} JSON | Updates the compound fields with the specified values |
| PUT | /api/compound/<compound_id>/ | {“<field_name>”: <field_val>...} JSON | Completely replaces the compound fields with the JSON object |
Figure 1Overview of Metabolite Atlas implementation scheme. (A) raw data extraction of chromatograms and spectra from a large number of LC/MS runs is facilitated by high-performance computing applications such as SciDB; (B) the specification, update, and management of metadata about experiments, samples, and of compounds in a Metabolite Atlas is facilitated by standard database applications such as MongoDB. The integrated analysis of these components via web-based interfaces makes the analysis and sharing of experimental observations in the context of raw data possible.
Figure 2User interface for adjusting the retention time bounds. Integrated access to raw LC/MS data and a Metabolite Atlas is used to adjust retention time bounds. As improved retention and m/z bounds are specified the parameters for each compound are automatically updated in a Metabolite Atlas.
Figure 3Authenticated users can acquire data from Metabolite Atlas using IPython and Jupyter notebooks. These notebooks provide a user friendly interface to the Python programming language which contains extensive libraries for data processing including peak fitting as shown here. These notebooks can be easily shared via the nbviewer service [32]. Typical notebooks contain code for analysis, results, and text explaining the purpose of the code.
Figure 4After optimizing the bounds for an Atlas, a user can acquire peak areas from Metabolite Atlas and perform statistical analysis for the compounds detected in their experiment. Python’s scientific libraries for statistical analysis can easily be implemented to perform common analysis such as hierarchical clustering and statistical confidence testing. Development of peak-shape modeling tools will be an important next step to deal with low-intensity peaks and missing values.