Literature DB >> 33244585

Probabilistic framework for integration of mass spectrum and retention time information in small molecule identification.

Eric Bach1, Simon Rogers2, John Williamson2, Juho Rousu1.   

Abstract

MOTIVATION: Identification of small molecules in a biological sample remains a major bottleneck in molecular biology, despite a decade of rapid development of computational approaches for predicting molecular structures using mass spectrometry (MS) data. Recently, there has been increasing interest in utilizing other information sources, such as liquid chromatography (LC) retention time (RT), to improve identifications solely based on MS information, such as precursor mass-per-charge and tandem mass spectrometry (MS2).
RESULTS: We put forward a probabilistic modelling framework to integrate MS and RT data of multiple features in an LC-MS experiment. We model the MS measurements and all pairwise retention order information as a Markov random field and use efficient approximate inference for scoring and ranking potential molecular structures. Our experiments show improved identification accuracy by combining MS2 data and retention orders using our approach, thereby outperforming state-of-the-art methods. Furthermore, we demonstrate the benefit of our model when only a subset of LC-MS features has MS2 measurements available besides MS1.
AVAILABILITY AND IMPLEMENTATION: Software and data are freely available at https://github.com/aalto-ics-kepaco/msms_rt_score_integration. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
© The Author(s) 2020. Published by Oxford University Press.

Entities:  

Year:  2021        PMID: 33244585      PMCID: PMC8289373          DOI: 10.1093/bioinformatics/btaa998

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


1 Introduction

The identification of small molecules, such as metabolites or drugs, in biological samples is a challenging task posing a bottleneck in various research fields, such as biomedicine, biotechnology, environmental chemistry and drug discovery. In untargeted metabolomics studies, the samples typically contain thousands of different molecules, the vast majority of which remain unidentified (Aksenov ; da Silva ). Liquid chromatography (LC) coupled with tandem mass spectrometry (MS2) is arguably the most important measurement platform in metabolomics (Blaženović ), due to its suitability to high-throughput screening, its high sensitivity and applicability to a wide range of molecules. Briefly explained, LC separates molecules by their differential physicochemical interaction between the stationary and mobile phase, which results in retention time (RT) differences and MS separates molecular ions by their mass per charge (MS1). Subsequently, MS2 can be used to fragment molecules in a narrow mass window and to record the fragment intensities (MS2-spectrum). In an untargeted metabolomics experiment, large sets of MS features (MS1 and RT, plus optionally MS2), are observed, corresponding to the different molecules in the sample. Metabolite identification concerns then the structural annotation of the observed MS features. In recent years, numerous powerful approaches (Nguyen ; Schymanski et al., 2017) to predict molecular structure annotations for MS2 spectra have been developed (Allen ; Brouard ; Dührkop , 2019; Nguyen et al., 2018b, 2019; Ruttkies et al., 2016, 2019). Typically, these methods output a ranked list of molecular structure candidates, that can be shown to human experts, or further post-processed, e.g. by using additional information available for the analysed sample. Sources of additional information include, e.g. RT (Bach ; Ruttkies ; Samaraweera et al., 2018), collision cross-section (Plante et al., 2019) or prior knowledge on the data generating process, such as the source organism’s metabolic characteristics (Rutz et al., 2019). RT, i.e. the time that a molecule takes to elute from the LC column, is readily available in all LC-MS pipelines, and is frequently used in aiding annotation (Stanstrup et al., 2015). A basic technique is to use the difference between the observed and predicted RT (Domingo-Almenara ; Samaraweera ) to prune the list of candidate molecular structures. A major challenge for utilizing RT information, however, is that the RT of the same molecule can vary significantly across different LC systems and configurations, necessitating system specific candidate RT reference databases and RT predictors. Different approaches have been proposed to tackle this challenge, such as using physicochemical properties (e.g. partition coefficient, LogP) as RT proxies (Hu ; Ruttkies ), RT mapping across LC systems (Stanstrup ) or predicting retention orders, which are largely preserved within a family of LC systems (e.g. reversed phase) (Bach ; Liu ). Using LogP as an RT proxy is simple to implement, but only models the hydrophobic separation effects of the LC system. RT mapping, on the other hand, is limited to pairs of LC systems in which the same molecules have been measured. Retention order prediction can overcome those drawbacks, by learning the LC system’s separation directly from RT data of multiple systems (Bach ). This study proposes a probabilistic framework to integrate MS1 or MS2-based annotations with predicted retention order for improved small molecule identification given a set of MS features measured within one LC-MS run by building on the work by Bach and Del Carratore . The latter proposed a probabilistic approach for integrating different types of additional information to MS1 data, including RT information. We too define a probabilistic approach, but differ in how RT is handled. Where Del Carratore use absolute RT information, we follow Bach and use pairwise retention order predictions for molecules eluting within the same LC-MS run. In contrast to the work done by Bach , our model makes use of pairwise retention order information between all MS features rather than only the ones adjacent in terms of their RTs, resulting in more accurate annotations. Furthermore, our model allows to rank all candidate lists, instead of just returning the most likely candidate assignment for each MS feature, as done by Bach . Our framework models the score integration as an inference problem on a graphical model, where the edges correspond to retention order predictions, the nodes correspond to MS features and the node labels correspond to candidate molecular structures, scored by a MS2 based predictor, such as CSI: FingerID (Dührkop ), MetFrag (Ruttkies ) or IOKR (Brouard ), or in the absence of MS2 information, MS1 precursor mass deviation. This graph is fully connected, which makes exact inference an NP-hard problem. To solve this challenge, we resort to efficient approximate inference, in particular spanning tree approximations (Marchand ; Pletscher et al., 2009; Su and Rousu, 2015; Wainwright et al., 2005).

2 Materials and methods

2.1 Overall workflow

We assume data arising from a typical LC-MS-based experimental workflow (including chromatographic peak picking, and alignment): MS features consisting MS1 measurement and the associated RT. A subset of these will include an MS2 spectrum. In the following, we present our score-integration model in the most general form in which it is provided with MS features and a set of possible candidate molecular structures. The candidate list can be generated, e.g. by querying molecular structures from a structure database, such as ChemSpider (Pence and Williams, 2010), that have the same mass as the observed MS feature. In addition, we assume that to each candidate structure a score is assigned by either an MS2-based predictor, or, if no MS2-spectrum is available, a score based on the mass deviation of the candidates from the MS mass. For all molecular candidate pairs, associated with the different MS features, the retention order is predicted. Here, we use the Ranking Support Vector Machine (RankSVM)-based predictor by Bach . The candidate structure scores and predicted retention orders are integrated through a probabilistic graphical model (described in the following). This allows us to rank the molecular candidate structures by their inferred marginal probabilities, given both the MS and RT information. More formally, the output of an LC-MS experiment is given as a tuple set , with being the spectrum of feature i (either an MS2 or a spectrum containing only a single peak at the mass of the precursor ion if no MS2 information is available), being its RT, and being the associated molecular candidates. Here, represents a molecular candidate structure and n is the number of molecular candidates for the ith MS feature. Figure 1 shows an overview of our workflow.
Fig. 1.

Workflow of our framework and its main components. (a) Data acquisition in an LC-MS experiment resulting in a set of (MS2, RT)-tuples of unknown molecules. (b) Illustration of the underlying graphical model. (c) Ensemble of spanning trees to approximate the MRF and their integration using averaged marginals. (d) Output to the user: ranked molecular candidate lists based on the approximated marginals. (e) Incorporation of the predicted retention orders for a particular assignment for z via the edge potential function. (f) Illustration of the RankSVM model

Workflow of our framework and its main components. (a) Data acquisition in an LC-MS experiment resulting in a set of (MS2, RT)-tuples of unknown molecules. (b) Illustration of the underlying graphical model. (c) Ensemble of spanning trees to approximate the MRF and their integration using averaged marginals. (d) Output to the user: ranked molecular candidate lists based on the approximated marginals. (e) Incorporation of the predicted retention orders for a particular assignment for z via the edge potential function. (f) Illustration of the RankSVM model

2.2 Probabilistic model

Let be an undirected graph, in which each node, represents one observed MS feature, and with an edge for all MS feature pairs . The edge-set E does not contain any parallel edges. The number of MS features is denoted with N, i.e. . We associate each node i in the vertex set with a discrete random variable z that takes values from the space . Intuitively, z defines which candidate has been assigned to the ith MS feature. The full vector corresponds to the molecular structure assignment to each MS feature in the LC-MS experiment, and it takes values from the set . In this work, we consider to be fixed and finite for a given a set of MS features, due to our definition of the molecular candidates sets, which assumes that we can restrict the putative annotation for a given MS feature.

Markov random field

The probability distribution of z is given as a pairwise Markov Random Field (MRF) (MacKay, 2005): composed of node ψ and edge ψ potential functions, and omits higher-order cliques (hence the term pairwise). Above, is the potential function of node i measuring how well the i’s candidates matches the measured MS information, and encodes the consistency of the observed retention orders for MS feature i and j and the predicted retention order of their candidates z and z and is the partition function (MacKay, 2005).

Node potential function ψ

For each candidate , we predict a matching score expressing how well it matches the observed MS1 or MS2 spectrum x. For that, we assume a pre-trained model, such as CSI: FingerID (Dührkop ), MetFrag (Ruttkies ) or IOKR (Brouard ). We use the latter two in our experiments as representative MS2 scoring methods (Section 3.3). MetFrag performs an in silico fragmentation of m, compares these fragments peaks with the observed ones in x and outputs a matching score. IOKR, on the other hand, can be used to directly predict a matching score f(x, m) for any (MS2 feature, molecular structure)-tuple. All matching scores θ are normalized to the range . Finally, we express the potential of a molecular candidate m given the spectrum x as follows: where c > 0 is a constant used to avoid zero potentials. In our experiments, we select c such that it is 10 times smaller than the minimum of all non-zero scores across all candidate sets.

Edge potential function ψ

For each candidate pair associated with the MS pair (i, j), we compute how well the candidates’ predicted retention order is aligned with the observed one defined by the RTs t and t. To this end, we apply the framework for retention order prediction developed by Bach . The edge potential is defined as follows: where is the RankSVM’s parameter vector, and are the feature vectors of the candidates’ molecular structures, and is a monotonic function mapping the predicted preference value difference to a value between zero and one. In our experiments, we consider two mapping functions: The different functions can be interpreted as follows. The sigmoid makes full use of the information from the RankSVM margin, i.e. the score of each candidate pair depends on the preference score difference. In this work, we consider k as a hyper-parameter of our method that needs to be estimated from data (Section 3.5). The step-function, on the other hand, only differentiates between aligned and not aligned pairs.

Weighting of information sources

To control the contribution of each information source, i.e. MS information and retention orders, we introduce a modification on the potential functions: with . A D value close to one, e.g. will result in a score mainly based on the observed retention orders. In our experiments, we explain how this hyper-parameter can be estimated in practice (Section 3.5).

2.3 Ranking candidates through approximated marginals

We rank the molecular candidates using the marginals of the MRF (1). The marginal for the candidate r of MS feature i is given as: In practice, the calculation of (2) is intractable due to the size of the domain of z, which grows exponentially with the number of MS features, thus we will resort to approximate inference methods.

Tree approximation of G

To enable feasible inference of (2), we approximate the MRF (1) using spanning trees of the original graphical model G (Marchand ; Pletscher ; Su and Rousu, 2015; Wainwright ). In the following let T be a spanning tree of G with the same nodes, but an edge-set , with , that ensures T being a cycle-free single connected component. The probability distribution of the graphical model induced by T is given as: As the graphical model associated with (3) is a tree, we can exactly infer its marginals through the sum-product algorithm (MacKay, 2005), The sum-product algorithm is a message-passing algorithm using dynamic programming that has linear time complexity in the number of MS features. See, e.g. MacKay (2005) for further details on the algorithm. The output of the sum-product algorithm are the unnormalized marginals for all and . We calculate the normalized marginals as follows (MacKay, 2005):

Random spanning trees sampling

We compare two approaches to retrieve spanning trees from G. The first approach is to randomly sample spanning trees from G (c.f. Pletscher ; Su and Rousu, 2015; Wainwright ). We sample the trees by applying the minimum weighted spanning tree algorithm to a random adjacency matrix. If for an MS feature pair (i, j) both RTs are equal, i.e. t=t, than their corresponding edge is not sampled. This is justified by the observation, that MS features with a RT difference equal zero, do not impose constraints on the retention order of their corresponding candidates. We will refer to a sampled spanning tree as T. The second approach was implicitly used by Bach and corresponds to a linear Markov chain where edges connect adjacent MS features ordered by increasing RT, which can be seen as a degenerate spanning tree. In the remaining text, we refer to this tree as .

Averaged marginal over a random spanning tree ensemble

Using tree-like graphical models for the inference is motivated by the exact and fast inference it enables us to do. However, a single tree, such as or a sampled T, will most likely be only a rough approximation of the original probability distribution (1). Therefore, the use of random spanning tree ensembles has been proposed. In particular, Wainwright show that an expectation over trees can be used to obtain an upper bound on the maximum a posteriori (MAP) estimate of the original graph, and showed that this approximation can be tight if the underlying trees agree about the MAP configuration. More recently, Marchand demonstrated generalization bounds for joint learning and inference using tree ensembles. More applied work in using tree-based approximation can be found in Pletscher , who use majority voting and Su and Rousu (2015), who empirically study several aggregation schemes in multilabel classification. Motivated by the mentioned literature, we opted to average the marginals of a random spanning tree ensemble T, where for each tree T, we independently retrieve the marginals using the sum-product:

Max-marginals

The exact inference on trees allows us to use the max-marginal, as an alternative to the sum-marginal shown in Equation (2). The max-marginal is closely related to the MAP estimate. For a single tree T it is given as: The interpretation of the two marginals (sum and max) differs slightly. Whereas the sum-marginal expresses the sum of the probabilities of all configurations with , the max-marginal is the maximum probability that a configuration with the constraint can reach. In our experiments, we compare the performance of both marginal types (Section 4.1). The max-product algorithm is used to calculate the unnormalized max-marginals , which is a modification of the sum-product algorithm, in which summations are replaced by maximization. The normalized marginal can be calculated as (MacKay, 2005): By plugging Equation (5) into (4) instead of the sum-marginal, we obtain the averaged max-marginal .

Run-time complexity

Calculating the marginals for an individual tree and all MS features i has run-time complexity . Here, N is the total number of features and the maximum number of molecular candidates for a feature.

3 Material and experiments

3.1 Evaluation datasets

To evaluate our score-integration approach, we use two publicly available datasets. These are described in this section and summarized in Table 1.
Table 1.

Summary of the datasets used for the evaluation of our score-integration framework

DatasetIonizationMass spectra info.Molecular candidatesaChromatography
MS1 info.#MS2Tot. #Cand.Median #Cand.ColumnEluent
CASMI 2016NegativePrecursor m/z8174 589420Phenomenex Kinetex EVO C18H2O → MeOH (both 0.1% formic acid)
CASMI 2016PositivePrecursor m/z127183 633919Phenomenex Kinetex EVO C18H2O → MeOH (both 0.1% formic acid)
EA (Massbank)NegativePrecursor m/z15475 107119.5Waters XBridge C18H2O → MeOH (both 0.1% formic acid)
EA (Massbank)PositivePrecursor m/z319215 893246Waters XBridge C18H2O → MeOH (both 0.1% formic acid)

Extracted from ChemSpider. CASMI: ±5 ppm window around monoisotopic exact mass of correct candidate. EA: MF of correct candidate.

Summary of the datasets used for the evaluation of our score-integration framework Extracted from ChemSpider. CASMI: ±5 ppm window around monoisotopic exact mass of correct candidate. EA: MF of correct candidate.

CASMI 2016

The Critical Assessment of Small Molecule Identification (CASMI) challenge is a contest organized for the computational spectrometry community (Schymanski ). For its implementation in 2016, a dataset of 208 (MS2, retention-time)-tuples was released. The dataset contains 81 negative and 127 positive ionization mode MS2 spectra. The molecular candidate structure sets were extracted from ChemSpider, using a ± 5 ppm window around the monoisotopic exact mass of the correct candidate, by the challenge organizers.

EA (Massbank)

Massbank (Horai ) is a publicly available repository for MS data. For the development of MetFrag 2.2, Ruttkies extracted 473 (MS2, retention-time)-tuples of 359 unique molecular structures from Massbank (EA dataset). The dataset is split into 154 negative and 319 positive ionization mode MS2 spectra. We used the molecular candidates provided by Ruttkies extracted from ChemSpider using the molecular formula (MF) of the correct candidate. For each dataset and ionization mode, we repeatedly subsample training and test (MS2, RT)-tuple sets: CASMI (negative) 50-times ; CASMI (positive) 50-times ; EA (negative) 50-times ; and EA (positive) 100-times . No molecular structure, determined by its InChI representation, appears simultaneously in test and training. The training set is used for the hyper-parameter selection (Section 3.5) and the test sets are used to assess the average identification performance of our score-integration framework (Section 3.4).

3.2 Training setup for the retention order predictor

To calculate the edge potentials of our MRF model (1), we use the RankSVM retention order prediction approach by Bach . The RankSVM model is trained using seven publicly available RT datasets. Six where published by Stanstrup along with their RT mapping tool PredRet: UFZ_Phenomenex, FEM_long, FEM_orbitrap_plasma, FEM_orbitrap_urine, FEM_short and Eawag_XBridgeC18. The seventh dataset contains examples for which RTs were published as part of the training dataset for the CASMI 2016 challenge (Schymanski ). The joint dataset covered four different chromatographic columns all using H2O → MeOH (with 0.1% formic acid as additive) as eluent. In total, the dataset contained 1248 (molecule, RT)-tuples of 890 unique molecular structures, after the same pre-processing as in Bach was applied. We represent the molecular structures using Substructure counting fingerprints calculated with rcdk and CDK 2.2 (Willighagen et al., 2017). We use the MinMax-kernel (Ralaivola et al., 2005) to calculate the similarity between the fingerprints. For our experiments, we build an individual RankSVM model for each (MS, RT)-tuple subsample (Section 3.1), ensuring no molecular structure in the subsample is used for the RankSVM training.

3.3 MS2-based match scores from MetFrag and IOKR

We apply MetFrag (Ruttkies ) and IOKR (Brouard ) as representative methods to obtain MS2 matching scores for the molecular structures in the candidate list of each MS2 spectrum.

MetFrag

We use the latest MetFrag version 2.4.5 (http://msbi.ipb-halle.de/ cruttkie/metfrag/MetFrag2.4.5-CL.jar) and utilize it as described in Ruttkies . The MS2 matching scores are calculated using the FragmenterScore feature of MetFrag.

IOKR

Two IOKR models are trained, for negative and positive mode MS2 spectra, respectively. The training (MS2, molecular structure)-tuples are extracted from GNPS (Wang et al., 2016), Massbank and the CASMI 2016 training data. We remove training molecular structures that appear in our evaluation datasets (Section 3.1). This results in 3255 negative and 6773 positive mode training examples. We use a uniform combination of 16 MS2 spectra and fragmentation tree (FT) kernels as input kernel (Supplementary Section S4). On the output side, we use the same molecular fingerprint definitions as Dührkop as feature representation and a Gaussian kernel those distances are derived from the Tanimoto kernel (Brouard ) as output kernel. For all MS2 spectra in our evaluation datasets, we calculate the FTs using SIRIUS 4.0.1 (Dührkop ) and keep the highest scoring tree for each spectra to calculate the MS2 and FT kernels used by the IOKR.

3.4 Performance evaluation

In our experiments, we use the top-k accuracy to determine the metabolite identification performance, i.e. the percentage of correctly ranked molecular candidates at rank k or less. Different approaches can be used to determine the rank of the correct structure. We follow the protocol used by Schymanski . If multiple stereo-isomers were present in the candidate list, only the one with the highest MS2-score was retained. The correct molecular structure was found by comparing the InChIs containing no stereo information. The top-k accuracies are calculated the test sets.

3.5 Hyper-parameter estimation

The training set of each individual subsample is used to determine optimal weighting D between MS and retention order information. For that, we run the score-integration framework for a different D values, and calculate the area under-the-ranking curve up to rank 20: , where is the number of correct structures up to rank i and N is the number of MS features. Subsequently, we select the retention order weight with the highest (Supplementary Section S2). The optimal sigmoid parameter k is estimated using Platt’s method (Lin ; Platt, 2000) calibrated using RankSVM’s training data (Section 3.2).

3.6 Experiments

Full MS2 information available

We compare our approach for combining MS2 and RT information for metabolite identification against the baseline, which only uses MS2 information for the candidate ranking. This allowed us to quantify the performance gain by using RTs. Furthermore, we applied two recently published methods for the integration of MS2 and RT scores and compared them to our approach. The first one is MetFrag 2.2 (Ruttkies ), which exploits the RT information by establishing a linear relationship between the candidates’ predicted LogP values with the observed RTs. Each molecular candidate receives an additional score by comparing its predicted RT against that observed for the corresponding MS2 spectra. We use the CDK (Willighagen ) XLogP predictions, which are automatically calculated by the MetFrag software. The weight of the RT feature RetentionTimeScore is determined as described in Section 3.5. Our second comparison is the approach by Bach , which uses predicted retention order and dynamic programming over a chain-graph connecting adjacent MS features. Bach focussed in extracting the most likely assignment z [Equation (3)] given the chain-graph using dynamic programming. Here, we use our generalized framework to also compute the marginals of all candidates given the chain-graph Tchain. The approach by Bach implicitly used a hinge-sigmoid to compute edge potentials: . Its parameter k is determined as described in the Supplementary Section S2. We refer to this method as Chain-graph.

Missing MS2

In a second experiment, we simulated the common application scenario, in which during an LC-MS experiment, a set of MS features (MS and RT) have been measured, but MS2 spectra have only been acquired for a subset of the features. There can be multiple reasons for this, such as limited measuring time when using, e.g. data-dependent acquisition (Xiao et al., 2012) protocols, bad fragmentation quality or inability to deconvolute all spectra when using data-independent acquisition. In this case, besides the RT, only MS1 related information is available for some features, which includes the mass of the ion (precursor m/z) and its isotope pattern. We use our proposed score-integration framework to perform structural identification when the proportion of MS features that have an MS2 spectrum varies. We vary the percentage of available MS2-spectra from 0% to 100% and investigate how the joint use of MS1, MS2 and RT information can improve over the baseline solely relying on only MS information. For the candidates r of an MS feature i, simulated to be without MS2 spectrum, we assign the following candidate score (Del Carratore ): where is the neutral exact mass of the measured ion calculated from the precursor m/z using the ground truth adduct, here either (positive) or (negative), and is the exact mass of candidate r associated with MS feature i, and variance of Gaussian noise model. ppm expresses the MS-device accuracy, which we set to ppm=5.

4 Results

4.1 Parameters of our framework

This section investigates the influence of different settings for framework, such as number of random spanning trees or the marginal type.

Number of random spanning-trees and marginal type

Figure 2 shows the top-k accuracy as a function of the number of random spanning trees L averaged across the datasets and ionizations. The identification performance increases for larger L, whereby the improvement per tree decreases. For the top-1 performance remains similar for trees. However, for top-20, we observe improvements till L = 128. Figure 2 also shows that the max-marginal approach performs slightly better than the sum-margin. An explanation could be that max-marginal is more robust against candidate configurations z with very low probability. The sum-marginal averages over such cases, whereas the max-marginal only includes the one with maximum probability.
Fig. 2.

Top-k accuracies, averaged across all datasets and ionizations, plotted against the number of random spanning trees (L) used for the approximation. The baseline using only MS2 information is plotted in black. The sigmoid function is used in the score integration. The differences between the sum- and max-marginals’ average performance for the L values is significant (P < 0.001 for Top-1 and 20, Two-sided Wilcoxon Signed-Rank test

Top-k accuracies, averaged across all datasets and ionizations, plotted against the number of random spanning trees (L) used for the approximation. The baseline using only MS2 information is plotted in black. The sigmoid function is used in the score integration. The differences between the sum- and max-marginals’ average performance for the L values is significant (P < 0.001 for Top-1 and 20, Two-sided Wilcoxon Signed-Rank test

Comparison of the edge potential functions

The average metabolite identification performance does not differ much between two edge potential functions (see Supplementary Table S1). This is interesting specifically for the Step-function, which uses the predicted retention orders in a binary fashion only. However, the Sigmoid function still can significantly outperform the Step-function for top-1 and top-5 accuracy.

4.2 Performance of our score integration framework

Here, we compare our score-integration framework with other methods and evaluate it under different data setups. We use L = 128 with max-marginals and the Sigmoid as edge potential function for the experiments.

Comparison to other approaches

In Table 2, we compare the performance of our score-integration framework with other approaches from the literature that utilize RT information for metabolite identification. It can be seen that our framework performs well across all datasets and ionization modes and we reach significant improvements over the baseline (Only MS2). Especially for the positive mode spectra, our method seems to have an advantage, as both competing approaches, cannot consistently improve the identification by including RT information. The least performance improvement of our approach can be observed for the negative CASMI dataset, which might be due to the small training set. The other approaches, MetFrag 2.2 and Chain-graph, can consistently (top-1, 5, 10 and 20) improve the results only on particular (dataset, ionization mode) combinations. However, they almost always increase top-1 performance. The results in Table 3 show that our framework significantly outperforms MetFrag 2.2 and Chain-graph in terms of identification performance.
Table 2.

Identification accuracies (top-k) for the different datasets and ionization modes

NegativePositive
DatasetMethodTop-1Top-5Top-10Top-20Top-1Top-5Top-10Top-20
CASMI 2016MS2 + RT (our) 15.2 (***)47.2 (***)57.0 (**)70.1 (***) 14.0 (***) 40.7 (***) 52.2 (***) 62.8 (***)
MS2 + RT (Chain-graph)13.2 (***) 49.4 (***) 61.0 (***)69.4 (***)11.936.550.2 (***)60.7 (***)
MS2 + RT (MetFrag 2.2)14.0 (***)42.055.5 71.2 (***)13.7 (***)36.246.257.5
Only MS211.144.255.368.011.837.347.058.3
EA MassbankMS2 + RT (our)28.7 (***) 61.9 (***) 73.8 (***)83.6 (***) 27.3 (***) 61.6 (***) 72.9 (***) 80.7 (***)
MS2 + RT (Chain-graph)27.2 (***)59.5 (***)72.4 (***)81.8 (***)23.9 (***)59.270.179.1 (***)
MS2 + RT (MetFrag 2.2) 30.2 (***)59.2 (***)73.6 (***) 84.4 (***)24.0 (***)59.069.577.1
Only MS222.857.669.578.521.259.069.777.6

Note: Compares our score-integration framework (MS2 + RT (our)), against the baseline (Only MS2), MetFrag 2.2 with predicted RT and the Chain-graph model. The best performance for each dataset and ionization is indicated by bold-font. The stars (*) represent the significant improvement over the baseline calculated using a one-sided Wilcoxon signed-rank test on the sample top-k accuracies (P < 0.05 (*), P < 0.01 (**) and P < 0.001 (***)).

Table 3.

Pairwise test for significant improvement of the MS2 + RT score-integration methods: Our, MetFrag 2.2 and Chain-graph

MethodTop-1Top-20
(MS2 + RT)Chain-graphMetFrag 2.2Chain-graphMetFrag 2.2
Our 8.7·1024 6.1·108 2.1·1013 1.5·1014
Chain-graphn.s. 8.1·1004
MetFrag 2.2 4.3·1008 n.s.

Note: We show the P-values for testing the improvement of the row over the column method using a one-sided Wilcoxon signed-rank test. The test is performed over all top-k accuracy samples (datasets and ionization). MetFrag 2.2 and Chain-graph could not significantly outperform our framework. P-values are marked with ‘n.s.’.

Identification accuracies (top-k) for the different datasets and ionization modes Note: Compares our score-integration framework (MS2 + RT (our)), against the baseline (Only MS2), MetFrag 2.2 with predicted RT and the Chain-graph model. The best performance for each dataset and ionization is indicated by bold-font. The stars (*) represent the significant improvement over the baseline calculated using a one-sided Wilcoxon signed-rank test on the sample top-k accuracies (P < 0.05 (*), P < 0.01 (**) and P < 0.001 (***)). Pairwise test for significant improvement of the MS2 + RT score-integration methods: Our, MetFrag 2.2 and Chain-graph Note: We show the P-values for testing the improvement of the row over the column method using a one-sided Wilcoxon signed-rank test. The test is performed over all top-k accuracy samples (datasets and ionization). MetFrag 2.2 and Chain-graph could not significantly outperform our framework. P-values are marked with ‘n.s.’.

Influence of MS2 scoring method

Table 4 shows the performance using of our score-integration framework for two difference MS2 scoring methods, MetFrag and IOKR (Section 3.3). Retention order information (MS2 + RT) can improve the identification performance in both cases; however the improvement with IOKR scores is lower.
Table 4.

Top-k accuracies averaged across all datasets for two MS2-scorers

MS2-scorersMethodTop-1Top-5Top-10Top-20
MetFragMS2 + RT21.352.964.074.3
Only MS216.749.560.470.6
IOKRMS2 + RT26.752.162.570.3
Only MS225.149.560.367.6
Top-k accuracies averaged across all datasets for two MS2-scorers

Influence of the candidate set

Here, we study the effect of two commonly used ways of defining the candidate lists of molecular structures: first approach (‘All’) includes all molecules with a matching exact mass to the list, and the second approach (‘Correct MF’) only includes molecular structures matching the pre-determined MF [e.g. SIRIUS (Dührkop ) uses this approach]. To determine the effect of the candidate set definition on our framework, we modify the CASMI dataset, such that for a spectrum i only candidates are used that have the same MFs as the correct structures. This leads to significantly reduced candidate sets: For the positive mode spectra, the median number of candidates decreases from 919 to 238 and for the negative ones from 420 to 58. For the Massbank data, we cannot do this modification, as the candidates are already restricted to structures with the correct MF. Table 5 shows that the baseline performance using MetFrag MS2 scores (Only MS2) improves after filtering of the candidates. Further improvement can be reached by using retention order information (MS2 + RT) even though the absolute improvement is slightly lower than without candidate filtering. For IOKR, we notice that RT information significantly improves the top-k accuracies when all matching exact mass candidates are used, whereas when the candidate sets only contain molecules with the correct, i.e. ground truth, MF, using RT information can only improve top-10 and top-20 accuracies.
Table 5.

Top-k accuracies averaged on the CASMI data (pos. & neg.) using either MetFrag or IOKR as MS2-scorer for two different candidate sets: ‘All’ molecules queried using a mass window; only those with ‘correct molecular formula’

MetFragIOKR
Candidate SetMethodTop-1Top-5Top-10Top-20Top-1Top-5Top-10Top-20
AllMS2 + RT14.6 (***)44.0 (***)54.6 (***)66.5 (***)26.0 (***)48.0 (***)60.0 (***)69.1 (***)
Only MS211.440.751.263.224.446.058.465.5
Correct MFMS2 + RT17.7 (***)48.4 (***)59.8 (***)71.0 (***)30.652.366.2 (***)75.1 (*)
Only MS213.146.056.968.730.653.965.374.8

Note: The stars (*) represent the significant improvement over the Only MS2 (see Table 2 for details on the significance test).

Top-k accuracies averaged on the CASMI data (pos. & neg.) using either MetFrag or IOKR as MS2-scorer for two different candidate sets: ‘All’ molecules queried using a mass window; only those with ‘correct molecular formula’ Note: The stars (*) represent the significant improvement over the Only MS2 (see Table 2 for details on the significance test).

4.3 Missing MS2

In Figure 3, we show the identification accuracy using our score-integration framework compared to the baseline (Only MS) when only some percentage of the MS features has an MS2 spectrum. The features without spectra only use the precursor mass as MS information (Section 3.6). We vary the percentage from 0% to 100% with 25%-point steps. The retention order weight D was optimized using the 100% setting. At 0%, the score-integration framework only uses the mass of the candidates and their predicted retention order for the ranking. In the absence of MS2 information, we observe a high performance gain for top-20. The more MS2 information we add, the smaller the gain in top-20 accuracy using the retention orders. The fact that RT is a weaker information than MS2 could explain this observation. The more MS2 are available, the less additional information RT can add. For the top-1, there is constant improvement for all MS2%’s.
Fig. 3.

Top-k accuracies and improvements averaged across all datasets. Plots for different percentages of available MS2 spectra: 0% (only MS1 and RT) to 100% of MS2 spectra (previous experiments)

Top-k accuracies and improvements averaged across all datasets. Plots for different percentages of available MS2 spectra: 0% (only MS1 and RT) to 100% of MS2 spectra (previous experiments)

5 Discussion

In this article, we have put forward a rigorous probabilistic framework for the integration of MS-based candidate structure and retention order predictions. Our framework allows the use of any of the popular models, such as CSI: FingerID, IOKR or Metfrag for scoring candidate structures on MS data. Our method takes into account the retention orders of all candidate structure pairs in distinct candidate lists through an approximated fully connected MRF model. It generally achieves higher quality structural annotations of observed MS features than using a single Markov chain as implied in the Bach model. It also improves on the method of Ruttkies , which uses predicted RTs, in three out of four experiments. For the latter approach, we believe using the RankSVM scores instead of the predicted LogP values could improve the performance. Both measures are proxies for retention behaviour and our results show that the RankSVM predicts the retention order more accurately than the LogP values (see Supplementary Table S2). We also demonstrate that our framework improves the identifications, if only a subset of the MS features come with an MS2 spectrum. The framework is computationally efficient, e.g. ranking the candidates for a set of N = 75 MS features takes <4 min (see Section S.5), and can be trained using modest-sized datasets. The amount of improvement using RT information was shown to depend on the dataset and MS2 scorer (here MetFrag or IOKR). This indicates that RT information rather fine tunes the ranking given by the MS2 scorer, e.g. by better tie-breaking. The underlying factors could be ambiguities in the candidate sets that can be only be resolved by RT or molecular features that cannot be predicted by MS. Stereochemistry is an obvious factor, but annotations of stereochemistry are not always provided for the RT databases limiting the use of this information for training better retention order prediction models. Thus improved modelling of stereochemistry features is an important open problem (Witting and Böcker, 2020). A further research direction could be to include the LC system’s configuration, e.g. column or eluent, into the retention order prediction. As LC systems can be configured to separate certain molecular classes, this could provide additional information to certain molecular candidates. Also, using the LC peak shape to train a model directly predicting the retention order probabilities could be more accurate, e.g. by incorporating RT variance. However, such data are currently not part of the public RT databases. Click here for additional data file.
  28 in total

1.  SIRIUS 4: a rapid tool for turning tandem mass spectra into metabolite structure information.

Authors:  Kai Dührkop; Markus Fleischauer; Marcus Ludwig; Alexander A Aksenov; Alexey V Melnik; Marvin Meusel; Pieter C Dorrestein; Juho Rousu; Sebastian Böcker
Journal:  Nat Methods       Date:  2019-03-18       Impact factor: 28.547

2.  Improving MetFrag with statistical learning of fragment annotations.

Authors:  Christoph Ruttkies; Steffen Neumann; Stefan Posch
Journal:  BMC Bioinformatics       Date:  2019-07-05       Impact factor: 3.169

3.  Integrated Probabilistic Annotation: A Bayesian-Based Annotation Method for Metabolomic Profiles Integrating Biochemical Connections, Isotope Patterns, and Adduct Relationships.

Authors:  Francesco Del Carratore; Kamila Schmidt; Maria Vinaixa; Katherine A Hollywood; Caitlin Greenland-Bews; Eriko Takano; Simon Rogers; Rainer Breitling
Journal:  Anal Chem       Date:  2019-09-30       Impact factor: 6.986

4.  Fast metabolite identification with Input Output Kernel Regression.

Authors:  Céline Brouard; Huibin Shen; Kai Dührkop; Florence d'Alché-Buc; Sebastian Böcker; Juho Rousu
Journal:  Bioinformatics       Date:  2016-06-15       Impact factor: 6.937

5.  The Chemistry Development Kit (CDK) v2.0: atom typing, depiction, molecular formulas, and substructure searching.

Authors:  Egon L Willighagen; John W Mayfield; Jonathan Alvarsson; Arvid Berg; Lars Carlsson; Nina Jeliazkova; Stefan Kuhn; Tomáš Pluskal; Miquel Rojas-Chertó; Ola Spjuth; Gilleain Torrance; Chris T Evelo; Rajarshi Guha; Christoph Steinbeck
Journal:  J Cheminform       Date:  2017-06-06       Impact factor: 5.514

6.  Critical Assessment of Small Molecule Identification 2016: automated methods.

Authors:  Emma L Schymanski; Christoph Ruttkies; Martin Krauss; Céline Brouard; Tobias Kind; Kai Dührkop; Felicity Allen; Arpana Vaniya; Dries Verdegem; Sebastian Böcker; Juho Rousu; Huibin Shen; Hiroshi Tsugawa; Tanvir Sajed; Oliver Fiehn; Bart Ghesquière; Steffen Neumann
Journal:  J Cheminform       Date:  2017-03-27       Impact factor: 5.514

7.  SIMPLE: Sparse Interaction Model over Peaks of moLEcules for fast, interpretable metabolite identification from tandem mass spectra.

Authors:  Dai Hai Nguyen; Canh Hao Nguyen; Hiroshi Mamitsuka
Journal:  Bioinformatics       Date:  2018-07-01       Impact factor: 6.937

8.  Taxonomically Informed Scoring Enhances Confidence in Natural Products Annotation.

Authors:  Adriano Rutz; Miwa Dounoue-Kubo; Simon Ollivier; Jonathan Bisson; Mohsen Bagheri; Tongchai Saesong; Samad Nejad Ebrahimi; Kornkanok Ingkaninan; Jean-Luc Wolfender; Pierre-Marie Allard
Journal:  Front Plant Sci       Date:  2019-10-25       Impact factor: 5.753

Review 9.  Software Tools and Approaches for Compound Identification of LC-MS/MS Data in Metabolomics.

Authors:  Ivana Blaženović; Tobias Kind; Jian Ji; Oliver Fiehn
Journal:  Metabolites       Date:  2018-05-10

10.  Improved Small Molecule Identification through Learning Combinations of Kernel Regression Models.

Authors:  Céline Brouard; Antoine Bassé; Florence d'Alché-Buc; Juho Rousu
Journal:  Metabolites       Date:  2019-08-01
View more
  2 in total

1.  Probabilistic metabolite annotation using retention time prediction and meta-learned projections.

Authors:  Constantino A García; Alberto Gil-de-la-Fuente; Coral Barbas; Abraham Otero
Journal:  J Cheminform       Date:  2022-06-07       Impact factor: 8.489

Review 2.  Strategies for structure elucidation of small molecules based on LC-MS/MS data from complex biological samples.

Authors:  Zhitao Tian; Fangzhou Liu; Dongqin Li; Alisdair R Fernie; Wei Chen
Journal:  Comput Struct Biotechnol J       Date:  2022-09-07       Impact factor: 6.155

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.