Literature DB >> 29950009

SIMPLE: Sparse Interaction Model over Peaks of moLEcules for fast, interpretable metabolite identification from tandem mass spectra.

Dai Hai Nguyen1, Canh Hao Nguyen1, Hiroshi Mamitsuka1,2.   

Abstract

Motivation: Recent success in metabolite identification from tandem mass spectra has been led by machine learning, which has two stages: mapping mass spectra to molecular fingerprint vectors and then retrieving candidate molecules from the database. In the first stage, i.e. fingerprint prediction, spectrum peaks are features and considering their interactions would be reasonable for more accurate identification of unknown metabolites. Existing approaches of fingerprint prediction are based on only individual peaks in the spectra, without explicitly considering the peak interactions. Also the current cutting-edge method is based on kernels, which are computationally heavy and difficult to interpret.
Results: We propose two learning models that allow to incorporate peak interactions for fingerprint prediction. First, we extend the state-of-the-art kernel learning method by developing kernels for peak interactions to combine with kernels for peaks through multiple kernel learning (MKL). Second, we formulate a sparse interaction model for metabolite peaks, which we call SIMPLE, which is computationally light and interpretable for fingerprint prediction. The formulation of SIMPLE is convex and guarantees global optimization, for which we develop an alternating direction method of multipliers (ADMM) algorithm. Experiments using the MassBank dataset show that both models achieved comparative prediction accuracy with the current top-performance kernel method. Furthermore SIMPLE clearly revealed individual peaks and peak interactions which contribute to enhancing the performance of fingerprint prediction. Availability and implementation: The code will be accessed through http://mamitsukalab.org/tools/SIMPLE/.

Entities:  

Mesh:

Year:  2018        PMID: 29950009      PMCID: PMC6022642          DOI: 10.1093/bioinformatics/bty252

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.937


1 Introduction

Metabolites are small molecules which are used in, or created by, chemical reactions occurring in living organisms (Wishart, 2007). They play a lot of important functions such as energy transport, signaling, building block of cells and inhibition/catalysis. Understanding biochemical characteristics of metabolites is an essential and important part of metabolomics to enlarge the knowledge of biological systems. It is also key to development of many applications and areas such as biotechnology, biomedicine or pharmaceutical sciences. Mass Spectrometry is a common technique in analytical chemistry (de Hoffmann and Stroobant, 2007) for metabolite identification. A mass spectrometer analyses a chemical sample by fragmenting it and measuring the mass-to-charge ratios (m/z) of its fragments. The resulting mass spectrum (MS) or tandem mass spectrum (MS/MS) is represented by a graph with m/z on the x-axis and the relative abundance of ions with m/z values on the y-axis (Fig. 1). Another common representation of mass spectrum is a list of peaks, each defined by its m/z and intensity value (top-right corner of Fig. 1).
Fig. 1.

Example MS spectrum from Human Metabolome Database (Wishart ) for 1-Methylhistidine (HMBD00001), with its corresponding chemical structure (top-left) and peak list (top-right)

Example MS spectrum from Human Metabolome Database (Wishart ) for 1-Methylhistidine (HMBD00001), with its corresponding chemical structure (top-left) and peak list (top-right) Many methods have been proposed for metabolite identification. A traditional approach is to compare a query MS or MS/MS spectrum of unknown compound against a database of a number of reference MS or MS/MS spectra (Scheubert ). The candidate molecules from the database are ranked based on the similarity, and the query spectrum and the best matched candidates are returned. However, the main disadvantage of this approach is that the reference database often contains only a small fraction of molecules in reality, leading to unreliable matching results if the molecule of the query spectrum is not in the database. Consequently, to mitigate the insufficiency of the reference databases, alternative approaches for metabolite identification are devised to deal with unavailable measured reference spectra. In general, we can divide computational methods for metabolite identification into the following three categories: i) searching in spectral libraries; ii) in silico fragmentation; iii) machine learning (ML). We focus on the ML approach, where the common scheme is, given a set of MS/MS spectra, to learn a map from a MS/MS spectrum to a molecule (Fig. 2). This has two steps: i) fingerprint prediction: predict a fingerprint with supervised ML, and ii) candidate retrieval: use the predicted fingerprint to query the database. Our particular focus is the first step, which predicts the fingerprint that is a binary (or rarely real-valued) vector, indicating the presence (or absence) of certain substructure. This step has been tackled by many ML methods, including linear discriminative analysis (LDA) (Imre ) and decision tree (Hummel ). A notable method is FingerID (Heinonen ), which used support vector machine (SVM) with kernels for pairs of mass spectra, including integral mass kernel and probability product kernel (PPK) (Jebara ).
Fig. 2.

A general scheme to identify unknown metabolites based on molecular fingerprint vectors. There are two main stages: 1) fingerprint prediction: learning a mapping from a molecule to the corresponding binary molecular fingerprint vector by classification methods, given a set of MS/MS spectra and fingerprints; 2) candidate retrieval: using the predicted fingerprints to retrieve candidate molecules from the databases of known metabolites. Note that the step of constructing an affinity matrix is optional and is used in L-SIMPLE only

A general scheme to identify unknown metabolites based on molecular fingerprint vectors. There are two main stages: 1) fingerprint prediction: learning a mapping from a molecule to the corresponding binary molecular fingerprint vector by classification methods, given a set of MS/MS spectra and fingerprints; 2) candidate retrieval: using the predicted fingerprints to retrieve candidate molecules from the databases of known metabolites. Note that the step of constructing an affinity matrix is optional and is used in L-SIMPLE only We point out three drawbacks of the existing ML approaches: i) all methods are based on the information from individual peaks in spectra, without explicitly considering peak interactions, which leads to the limitation of predictive performance. ii) There is an interesting (not ML-based) software, which outputs, given a spectrum, a so-called fragmentation tree (FT) (Böcker and Rasche, 2008; Rasche ). In a FT, possible fragments corresponding to peaks in spectrum are shown as node labels, where parent–child relationships in this tree are inclusive relations of fragments in chemical structure. FT is indeed interesting, but this software of converting a given spectrum into a FT is very slow. CSI: FingerID (Dührkop ; Shen ) used both mass spectra and FTs as input, thinking that structural information of chemical compounds can be captured by FTs. Indeed using FTs might be similar to considering peak interactions. However if FTs are used as input features, spectrum must be converted into a FT not only in training and but also in prediction, which needs a heavy computational load. iii) The size/length of a fingerprint is not short, while the number of peaks is usually small, meaning that most elements of a fingerprint vector are zero. So the training data is very sparse, while sparse models have not been considered yet. In addition, sparse models are advantageous in that their results are easily interpretable. Motivated by these drawbacks, we address the following two problems: i) incorporation of peak interactions into the learning model to improve the predictive performance; ii) introduction of sparsity into the models for interpretation. For the above i), we propose a kernel for peak interactions and combine this kernel with other kernels through multiple kernel learning (MKL). For the above both i) and ii), we propose a sparse, interpretable model, which we call Sparse Interaction Model over Peaks of moLEcules (SIMPLE). Additionally, we also propose a FT-induced Laplacian regularization to make SIMPLE more robust. We note that in SIMPLE, FTs are used for regularization only and not for input, by which we do not need FTs for prediction, meaning that computationally SIMPLE is much lighter for prediction than (Shen ). We formulate the model as a convex optimization problem, for which we develop an alternating direction method of multipliers (ADMM) algorithm. We evaluated our two proposed models by using real data obtained from the MassBank dataset (Horai ), which was used in (Shen ). We found that incorporation of peak interactions can significantly improve the prediction accuracy of not using interactions, resulting in comparable performance with the current top method. Furthermore, SIMPLE could show the interpretability of results, i.e. peaks and peak interactions which contribute to high predictive performance.

2 Related work

In Figure 2, the standard data preprocessing converts spectra into high-dimensional feature vectors by dividing m/z range into bins and taking accumulated intensity within each bin as a feature value. However the width of bins is hard to determine. In fact, wide bins can cause noise, and narrow bins can induce alignment errors due to mass error. One idea to overcome this issue is using a kernel, say probability product kernel (PPK; Jebara ), which can be computed directly from mass spectra. PPK treats each peak in a spectrum as a two-dimensional Gaussian distribution and a spectrum as a uniform mixture of these distributions. Then, kernel between two spectra is computed by all-against-all matching between the component Gaussians. The detail is in Section 3.1.1. Also we briefly explain the current cutting-edge ML method by MKL, CSI: FingerID (Dührkop ; Shen ), where the input is both MS/MS spectra and FTs. Motivation behind this method is to use structural information, which might be captured by FTs. So they use numerous kernels for FTs to capture any information of FTs, ranging from simple ones, such as node kernels: node binary (NB), node intensity (NI) and edge kernels: loss binary (LB), loss count (LC), loss intensity (LI) to more complex one like common path counting (CPC) which is one of path based kernels. Subsequently, these kernels are combined by MKL methods, such as centered alignment (Gönen and Alpaydin, 2011), quadratic combination and norm regularized combination (Kloft ). The combined kernel is then used in learning the final model for fingerprint prediction. We again emphasize that these methods consider mainly only peaks in MS/MS spectra without explicitly taking peak interactions into account. Also kernels using FTs have the limitation of processing MS/MS spectra, particularly for prediction, due to the need of computationally heavy conversion of spectra into FTs. More importantly, kernels are difficult to interpret and deal with sparse data, such as MS/MS spectra, where each spectrum (fingerprint vector) has only a few number of peaks (nonzeros).

3 Materials and methods

3.1 Kernel method

We develop kernel for peak interactions and combine this kernel with PPK (kernel for peaks) (Heinonen ) through the framework of MKL.

Preliminary: kernel for peaks

MS/MS spectra have peak information, i.e. m/z and intensity. The problem is that peaks are not correctly aligned each other, due to measurement errors in mass spectrometry devices (alignment error). To overcome this problem, PPK (Jebara ) was developed to calculate the similarity between two spectra. The idea is that a peak can be considered as a two-dimensional Gaussian distribution and the spectrum is a uniform mixture of peaks (or distributions). The kernel between two spectra is computed by all-against-all matching between the Gaussians. More specifically, given a mass spectrum S, being a list of peaks, i.e. . The kth peak of the spectrum S is represented by a Gaussian distribution p centered around the peak measurement and with covariance shared with all peaks: , where and are the variances for the mass and intensity, respectively. Hence, the spectrum S can be represented as a uniform mixture of the peaks contained in it, i.e. . Likewise, another spectrum is also represented by a mixture of distributions q, and . With definitions above, the kernel for peaks, , between two spectra S and is given by: where is the PPK between two component Gaussian distributions p and q, computed by (3). where and denote the mass and intensity values of the peaks corresponding to p.

Kernel for peak interactions

Being rather straight-forward, kernel for peak interactions between two spectra S and can be defined as follows: where K is the function defined by (3). Note that also can overcome the alignment error problem by taking advantage of . Intuitively, can be considered as a probabilistic version of the number of common peaks. Similarly can be considered as a probabilistic version of the number of common edges or interactions between two spectra.

Combining kernels for peaks and peak interactions

We combine kernels for peaks and peak interactions by using a regular approach: Multiple Kernel Learning (MKL; Gönen and Alpaydin, 2011), which can combine kernels from different sources. We use three approaches: the first is the uniform combination of the kernels (UNIMKL: the weights for kernels are equal), which can produce good results for prediction in many practical applications. The second and third are two different MKL algorithms: centered alignment (ALIGN) and alignment maximization algorithm (ALIGNF) (Cortes ). These algorithms have the same setting. Given a set of kernels , computed from n data points and representing different sources of information. The vector corresponds to the output or labels of data points. The aim of MKL is to seek a combination of these kernels, as defined by: A popular approach to MKL, which is called two-stage MKL, separate kernel learning from prediction learning. In the first stage, the combined kernel is learned through the use of an objective function, such as centered alignment that measures the similarity of two kernels over the training dataset. For the purposes of MKL, alignment is often measured between the combined kernel in (5) and the target kernel . Subsequently, the learned combined kernel is used in the second stage with kernel based classifiers such as relevance vector machine (RVM) used in (Burden and Winkler, 2015) or SVM used in our experiments. Formally, the centered kernel of a given kernel is defined by: where I is the identity matrix and e is the vector with all ones. The centered alignment between two kernels K and is defined by: Cortes proposed two MKL algorithms: i) a simple centered alignment algorithm (ALIGN), which assigns alignment scores computed in (7) to combination weights in (5), i.e. . and ii) the alignment maximum algorithm (ALIGNF) seeks the weights α by maximizing the alignment scores between the combined kernels and target kernel , resulting in the objective function: In our experiments, we conducted comparative experiments by using these algorithm for combining kernels for peaks and interactions.

3.2 Sparse Interaction Model over Peaks of moLEcules (SIMPLE)

Kernel-based methods are difficult to deal with sparse data and lack of interpretation. Thus we present a more interpretable, fast, sparse learning: Sparse Interaction Model over Peaks of moLEcules (SIMPLE), to incorporate peak interactions explicitly. We first preprocess each MS/MS spectrum to generate a feature vector: we divide m/z range into bins and taking accumulated peak intensities in a bin as a feature value to obtain high dimensional vector for each spectrum. These feature vectors are normalized such that all feature values are in [0, 1].

Prediction model

Given a MS/MS spectrum, represented by a feature vector, , we formulate the model for individual peaks and peak interactions as follows: where and . The prediction function consists of a bias b and two terms: main effect term parameterized by the weight vector w and interaction term parameterized by the weight matrix W. Their roles are different. While the former capture information about the peaks, the latter captures information about peak interactions. Since our task is classification which predicts the presence or absence of properties in fingerprint vector, the output of the model can be computed by . In order to predict a binary vector of fingerprint, we predict one response variable (hereafter a property or a task) at each time with a separate classifier. Let consider a property and give a set of n training input/output pairs , where represent the ith spectrum and indicates the presence (+1) or absence (-1) of the property. We can learn the parameters by minimizing the following optimization function: where is the hinge loss function and computed by , in which the operator denotes . and are the regularization terms for vector w and matrix W, respectively. Our purpose is to seek a model which is accurate as well as interpretable. Different from SVM using -norm regularizer, , our model uses sparsity-induced regularizer (Tibshirani, 1994), to yield sparse solution and interpretation. As for interaction term, it is natural to impose low-rankness on matrix W due to the existence of groups of peaks interacting with each other. Thus, we propose to use trace norm, similar to (Blondel ), i.e. . Furthermore, due to the symmetry of matrix W, we also impose the positive semidefiniteness on matrix W, i.e. . Therefore, putting all above things together, the objective function of SIMPLE becomes: where α and β are hyperparameters to control the sparsity of w and low-rankness of W, respectively. Note that (15) is the convex function which guarantees to find the global optimal solution.

Fragmentation trees (FTs) as prior information

As the number of interactions is large, it would be a good idea to regularize (14) with background knowledge. We propose to regularize the interaction matrix W by FTs of the spectra. However, there are two questions to deal with: i) How to represent the structural information from FTs; ii) how to incorporate them into the convex objective function (14) while preserving its convexity to guarantee the global optimal solutions. To answer the first question, our idea is to construct an affinity (adjacency) matrix A for all peaks in the spectra. Concretely, each spectrum is converted into a FT by algorithms in (Rasche ). Our assumption is that, the frequency of an interaction present in the fragmentation trees reflects how strongly the corresponding features are interacting with each other. From that, we calculate the co-occurrence of all peak pairs in the trees to construct the affinity matrix (Fig. 3).
Fig. 3.

Illustration of constructing affinity matrix A from the set of fragmentation trees. The constructed matrix A is used as prior information for regularizing interaction matrix W

Illustration of constructing affinity matrix A from the set of fragmentation trees. The constructed matrix A is used as prior information for regularizing interaction matrix W To answer the second question, we impose the constraint of being positive semidefinite on interaction matrix W. By this, W can be decomposed into the low rank matrix, i.e. where . Thus, the interaction term in the prediction function (12) can be rewritten as following: where v and v are the ith and jth bins of spectrum x and correspond to representation of ith and jth features in the space . We assume that if ith and jth peaks are strongly interacting with each other, then their representation in the vector space should be close, i.e. should be small. From this observation, we include the following term, Laplacian smoothness, into the objective function (14): where , D is the degree matrix of A, i.e. Thus, we can formulate the following optimization problem, which we call L-SIMPLE: where , γ is an additional hyperparameter to control Laplacian smoothness of the objective function (18). It is noteworthy that, our derived formulation is still convex due to convexity of regularization terms, thus, the solution of (18) is guaranteed to be the global optimal. Below we will present an optimization method for efficiently solving (18).

Optimization: ADMM algorithm

Equation (18) is convex and guaranteed that the algorithm converges to the global optimal solution. However, it is challenging to solve this problem directly, because terms are nondifferentiable. Our optimization method is based on alternating direction method of multipliers (ADMM). Due to the nondifferentiability of the hinge loss, we introduce the auxiliary variable C, where and , then (18) can be reformulated into the following equivalent constrained problem: where . The augmented Lagrange function of (19) is defined by: where is a dual variable corresponding to the constraint . Note that the constraint is not included in (20) because this property will be imposed automatically on W after each iteration of updating W, as explained in the Appendix section. We solve the problem of finding the saddle point of the augmented Lagrangian function (20) through an iterative algorithm between the primal and the dual optimization as follows: where the first three steps update the primal variables based on the current estimate of the dual variable and the final step updates the dual variable based on the current estimate of the primal variables. Note that the efficiency of ADMM for solving (20) depends on whether the subproblems in (21) can be solved quickly. Algorithm 1 summarizes the ADMM steps for solving the optimization problem (20). To avoid confusion, the detailed derivation of the update rules for subproblems of (21) are in Appendix.

3.3 Model summary

We here summarize the advantageous features of (L-)SIMPLE: Peak interactions: SIMPLE has, in its formulation, the term for peak interactions explicitly, which has not been considered by existing methods, particularly kernel-based methods. Sparse interpretability: MS spectra are with many zeros and sparse data. We formulate SIMPLE as a sparse model, by which peaks or peak interactions which contribute to improve the predictive performance can be checked and found easily. Convex formulation: our formulation keeps SIMPLE (L-SIMPLE) a convex model, which guarantees to find the global optimum. We have developed an alternating direction method of multipliers (ADMM) algorithm, to realize the detection of the global optimum. No fragmentation trees in prediction: L-SIMPLE uses fragmentation trees for regularization in training, and they are not inputs, meaning that we do not need fragmentation trees in prediction, which can avoid heavy computational cost of generating fragmentation trees from spectrum.

4 Experimental evaluation

Our focus is to incorporate peak interactions of mass spectra into fingerprint prediction, and to build sparse models for model interpretability, and so we conducted experiments to answer the following two questions: (Q1): Can peak interactions due to sets of correlated peaks in spectra be used to predict fingerprint vectors more accurately? (Q2): For the purpose of interpretation, how to identify a smaller subset of predictors (i.e. peaks or peak interactions) that exhibit the strongest effects on fingerprints?

4.1 Data, preprocessing and evaluation measures

MassBank (Horai ) was used. We used the same dataset with 402 compounds as (Shen ), and followed the same preprocessing steps as in (Shen ): for each compound, peaks recorded from different energies were merged for each MS/MS spectra. Then, spectra were normalized such that the sum of intensities is up to 100%. In terms of the output for the learning models, molecular fingerprints, which are binary vectors of 528 bits in total, were generated using OpenBabel (O’Boyle ). Fingerprints have a high class imbalance, i.e. mostly +1 or reverse. Thus we only trained models for predicting fingerprints in which the majority class occupies less than 90% of instances. We used micro-average accuracy, F1 score to evaluate the performance of different methods, computed by taking the average of accuracies and F1 scores over all tasks. ADMM algorithm for optimizing the objective function (19). 1: Inputs: A set of MS/MS spectra and associated output . Laplacian L (only for L-SIMPLE). 2: Outputs: weight vector w and interaction matrix W. 3: Initialize: 4: while not converged do 5:   1. Fix W, C, u and update b, w 6:   Precompute 7:   while not converged do 8:   end while 9:   2: Fix b, w, C, u and update W 10:   precompute 11:   while not converged do 12:   end while 13:   3: Update C 14: 15:   4: Update the dual variable u 16: 17: end while

4.2 Benefit of incorporating interaction

We show the benefit of incorporating peak interactions using kernel methods described in Section 3.1. The MKL algorithms, i.e. UNIMKL, ALIGN and ALIGNF, were used to combine two kernels, and (we call their results ComUNIMKL, ComALIGN, ComALIGNF, respectively) and compared with PPK, which has only kernel for peaks. The combined kernels were coupled with SVM to predict the fingerprint properties. Each property was separately trained by a classifier. Five-fold cross-validation was conducted to seek suitable margin parameter C where . Table 1 shows the results, in which the micro-average accuracy and F1 score of combined kernels were higher than PPK. This shows that incorporating peak interactions can improve the performance on predicting molecular fingerprint properties. Additionally, the performance of ALIGNF algorithm was slightly better than the rest.
Table 1.

Micro-average performance of kernels: PPK (Heinonen ) is used to compute

MethodAcc (%)F1-score (%)
PPK75.74 (±8.13)60.59 (±13.75)
ComUNIMKL78.41 (±6.82)65.05 (±12.16)
ComALIGN78.57 (±6.24)65.34 (±11.99)
ComALIGNF79.03 (±7.89)65.67 (±13.02)

Note: ComUNIMKL, ALIGN, ALIGNF are combinations of and by algorithms UNIMKL, ALIGN, ALIGNF, respectively.

Micro-average performance of kernels: PPK (Heinonen ) is used to compute Note: ComUNIMKL, ALIGN, ALIGNF are combinations of and by algorithms UNIMKL, ALIGN, ALIGNF, respectively.

4.3 Benefit of (L-)SIMPLE, sparse interaction models

Since kernel methods in Section 3.1 are unable to provide sparse solutions for interpretability, we examined (L-)SIMPLE to gain insight into the models to predict fingerprints. (L-)SIMPLE needs to construct feature vectors from MS/MS spectra: we divided m/z range into 500 bins and took accumulated peak intensities in a bin as a feature value to obtain high dimensional vector for each spectrum. These feature vectors were normalized such that all feature values are in [0, 1]. (L-)SIMPLE was trained by ADMM (see algorithm 1) with all variables initialized at zero. The algorithm was iterated until the relative difference in training errors fell below 0.0001 or the number of iterations reaches 100. Five-fold cross-validation was used to evaluate the generalization of the learning machines. Specifically, parameters α, β and γ, for controlling sparsity, low-rankness and Laplacian smoothness were chosen from the lists: {1, 2, 3}, and , respectively. Also, to evaluate the effects of adding the Laplacian smoothness term into the objective function, we compared the accuracy and F1 score over tasks by performing five-fold cross-validation. Table 2 shows the accuracy and F1 score obtained for the first ten tasks (Note that we used only tasks in which the majority class occupies less than 90% of instances). The micro-average over all tasks is also displayed at the bottom. We can see that Laplacian regularization worked to make SIMPLE more robust, resulting in better predictive performance of L-SIMPLE.
Table 2.

Performance comparison between SIMPLE and L-SIMPLE

Task Id/NameSIMPLE
L-SIMPLE
Acc (%)F1 (%)Acc (%)F1 (%)
3 (Aldehyde)71.1669.2473.1470.75
27 (Hydroxy)91.2495.1990.2994.82
29 (Primary alcohol)79.3653.7479.8553.98
30 (Secondary alcohol)80.3554.1881.8456.17
37 (Ether)80.3571.3980.6172.03
38 (Dialkyl etherEther)82.3570.9882.671.16
45 (Aryl)83.8381.6783.3482.16
50 (Carboxylic acid)69.3862.0069.6562.15
56 (Primary Carbon)73.1240.4273.8844.46
57 (Secondary Carbon)71.8867.1572.3967.28
60 (Alkene)81.3523.4984.0827.75
Avg ± Std78.33 ± 6.0566.69 ± 13.0378.86 ± 5.8767.59 ± 12.35
Performance comparison between SIMPLE and L-SIMPLE We further compared (L-)SIMPLE with various kernel, namely PPK, NB, NI, LB, LC, LI, CPC and their combination, which we call ComFT, all from (Shen ). While these kernels (except for PPK) are all computed from the fragmentation trees, in which the cost for converting MS/MS spectra to these trees is heavy and time-consuming, our method uses peaks from spectra only and is efficiently computable in prediction. Table 3 shows the results of accuracy, F1 scores and computation time for prediction by all compared methods. The prediction time (in milliseconds) was averaged over all spectrum in the dataset. The first four methods including PPK achieved around the accuracy of 75%, which is clearly worse than the other methods, which achieved around 78–80% and are very comparable each other. In fact ComFT, the current cutting-edge MKL-based method, performed best in both accuracy and F1 score, while the second best was not clear (NB by accuracy and L-SIMPLE by F1 score). On the other had, about computation time, FT-based methods, i.e. from LB to ComFT, were clearly slower than the others, because they need convert spectra into FTs. In fact ComFT needed more than 1500 ms, which is more than 300 times slower than that of (L-)SIMPLE, which just spent only less than 5 ms. This is a sizable difference when we have to process a huge amount of spectra produced by the current high-throughput MS/MS. We stress that the performance difference between (L-)SIMPLE and ComFT was very slight and statistically insignificant, while (L-)SIMPLE was exceedingly faster than ComFT.
Table 3.

Micro-average performance and computation time (for prediction) of kernel-based methods in Shen and proposed methods in this paper

MethodAccF1 scoreRun. time
(%)(%)(ms)
PPK (Peaks)75.74 (±6.72)60.59 (±14.54)52.37
LB (Loss binary)76.63 (±7.03)61.64 (±15.48)1501.02
LC (Loss count)75.33 (±5.4)61.25 (±13.99)1501.02
LI (Loss intensity)74.54 (±8.49)58.46 (±16.01)1501.02
NB (Node binary)79.11 (±5.02)67.34 (±11.75)1501.09
NI (Node intensity)78.41 (±4.99)66.87 (±12.11)1501.01
CPC (Common path count)79.02 (±7.4)67.55 (±12.93)1501.11
ComFT (combining all above)80.98 (±6.05)69.04 (±11.98)1559.20
ComALIGNF (Proposed: MKL)79.03 (±7.89)65.67 (±13.02)471.71
SIMPLE (Proposed)78.33 (±6.05)66.70 (±13.03)4.57
L-SIMPLE (Proposed)78.86 (±5.87)67.59 (±12.35)4.32
Micro-average performance and computation time (for prediction) of kernel-based methods in Shen and proposed methods in this paper

4.4 Model interpretation

One advantage of sparse learning models over kernel based methods is interpretation. For illustration purposes, Figure 4 shows the weights of main effect terms (w) and the interaction weight matrix (W) obtained by L-SIMPLE for nine fingerprint properties (or tasks): 29, 37, 56, 70, 139, 192, 236, 356 and 370 (these are randomly selected for investigation). As observed, the weight of main effect and interaction terms were different between properties, suggesting that different properties are strongly affected by different subsets of a few peaks in spectra.
Fig. 4.

(a) Weight vectors (w) of the main effect terms and (b) smooth heat map of weight matrices (W) of the interaction terms learned by L-SIMPLE for properties or tasks: 29 (Primary alcohol), 37 (Ether), 56 (Primary Carbon), 70 (Alkylarylether), 139 (Thioenol), 192 (Carbonic acid monoester), 236 (Heteroaromatic), 356 (1,5-Tautomerizable) and 366 (Actinide)

(a) Weight vectors (w) of the main effect terms and (b) smooth heat map of weight matrices (W) of the interaction terms learned by L-SIMPLE for properties or tasks: 29 (Primary alcohol), 37 (Ether), 56 (Primary Carbon), 70 (Alkylarylether), 139 (Thioenol), 192 (Carbonic acid monoester), 236 (Heteroaromatic), 356 (1,5-Tautomerizable) and 366 (Actinide) Table 4 shows case studies to illustrate the effects of peak interactions. We consider four interaction pairs frequently present in these tasks. w1 and w2 denote the weights for peaks and W12 denotes the weight for their interactions. We can raise the following three interesting findings:
Table 4.

Case studies of weight vector w and interaction matrix W learned by L-SIMPLE over a set of randomly selected tasks

Tasks/Name(42, 85)
(42, 163)
(85, 227)
(130, 201)
w1w2W12w1w2W12w1w2W12w1w2W12
29 (Primary Alcohol)0.00.00.0016−0.05450.00.0442−0.4085−0.05450.0765−0.40850.00.0218
37 (Ether)0.00.00.01200.02600.00.04710.00.02600.12640.00.00.3389
56 (Primary Carbon)0.00.00.0271−0.56570.00.0047−0.9833−0.56570.0271−0.98330.00.0104
70 (Alkylarylether)0.00.00.01590.00.00.0551−0.12380.00.0972−0.12380.00.0265
139 (Thioenol)0.15750.39390.0−0.2308−0.40940.0−0.3589−0.23080.0−0.3589−0.11260.0
192 (Carbonic acid monoester)−0.15420.00.00.00.00.0−0.69450.00.0−0.69450.00.0
236 (Heteroaromatic)0.00.00.01070.00.24990.0537−0.30690.00.1062−0.30690.14010.0201
356 (1,5-Tautomerizable)0.00.00.01700.00.00.03010.55390.00.06070.55390.00.0107
366 (Actinide)0.00.00.0245−0.18910.60650.0282−0.5373−0.18910.0399−0.53730.00.0153

Note: w1 and w2 denote weights corresponding two mass positions and W denotes the weight of their interactions. Four pairs of mass positions which are frequently present in these tasks, including (42, 85), (42, 163), (85, 227) and (130, 201) are shown.

Case studies of weight vector w and interaction matrix W learned by L-SIMPLE over a set of randomly selected tasks Note: w1 and w2 denote weights corresponding two mass positions and W denotes the weight of their interactions. Four pairs of mass positions which are frequently present in these tasks, including (42, 85), (42, 163), (85, 227) and (130, 201) are shown. Either w1 or w2 (or both) can be zero but the interaction weights are often nonzero: For example, W12 of interaction (42, 85) are mostly nonzero, while w1 and w2 are both zero with respect to tasks 29, 37, 56, 70. This means that individual peaks are not good predictive descriptors of properties, while their interactions are. Thus this result clearly shows the importance of considering peak interactions in fingerprint prediction. Despite of negative impacts of individual peaks, their interactions can be positive: For example, interaction (85, 227) with respect to tasks 29, 56 and 366. This means that while individual peaks indicate the absence of a property, the interactions of two peaks mean the presence of the property. Interaction of peaks can be zero, indicating these peaks are independent of each other even if their corresponding weights are nonzeros: For example, interaction (42, 85) with respect to task 139. They are just part of numerous findings, but even these examples show a fact that even individual peaks are not good indicators of some fingerprints, their interactions with others may significantly contribute to the prediction. This again confirms the importance of peak interactions for fingerprint prediction.

5 Discussion and conclusion

The goal of this work is to propose machine learning models which are able to incorporate peak interactions for fingerprint prediction. Our experiments showed that peak interactions are definitely useful to improve fingerprint prediction, along with discriminative information about peaks. Our first model is based on kernel learning, defining two kernels, one for peaks and the other for peak interactions, which are combined through MKL. Again we note that Shen used fragmentation trees for prior structural information of spectral, and converting spectra into fragmentation trees is definitely computationally expensive. On the other hand, our model of kernel learning uses only peaks in the spectrum as input for prediction, indicating that our model is much more efficient. Kernel learning does not have to construct feature vectors for spectra, and instead this is done implicitly by kernels defined, which can avoid any error caused when generating feature vectors. However, a big issue of kernel learning is interpretability. That is, it is difficult for kernel learning to figure out which subset of peaks or peak interactions exhibit the strongest effects on fingerprint prediction, despite that clearly each property depends on a very few number of mass positions in each given spectrum. Our sparse interaction models, (L-)SIMPLE, have a number of advantages, which are summarized in Section 3.3: (L-)SIMPLE is formulated as a sparse, convex optimization model, which can capture peak interactions and also give interpretable solutions. We emphasize that next generation fingerprint prediction needs a ML model, which should learn, from huge but sparse spectra, peaks as well as peak interactions comprehensibly and predict fingerprints against again huge spectra highly efficiently. (L-)SIMPLE would be a reasonable solution for this situation.

Funding

This work was partially supported by MEXT KAKENHI Grant Number 16H02868, Grant Number JPMJAC1503 ACCEL JST, FiDiPro Tekes (currently Business Finland) and AIPSE Academy of Finland. Conflict of Interest: none declared.
  14 in total

1.  MassBank: a public repository for sharing mass spectral data for life sciences.

Authors:  Hisayuki Horai; Masanori Arita; Shigehiko Kanaya; Yoshito Nihei; Tasuku Ikeda; Kazuhiro Suwa; Yuya Ojima; Kenichi Tanaka; Satoshi Tanaka; Ken Aoshima; Yoshiya Oda; Yuji Kakazu; Miyako Kusano; Takayuki Tohge; Fumio Matsuda; Yuji Sawada; Masami Yokota Hirai; Hiroki Nakanishi; Kazutaka Ikeda; Naoshige Akimoto; Takashi Maoka; Hiroki Takahashi; Takeshi Ara; Nozomu Sakurai; Hideyuki Suzuki; Daisuke Shibata; Steffen Neumann; Takashi Iida; Ken Tanaka; Kimito Funatsu; Fumito Matsuura; Tomoyoshi Soga; Ryo Taguchi; Kazuki Saito; Takaaki Nishioka
Journal:  J Mass Spectrom       Date:  2010-07       Impact factor: 1.982

2.  Searching molecular structure databases with tandem mass spectra using CSI:FingerID.

Authors:  Kai Dührkop; Huibin Shen; Marvin Meusel; Juho Rousu; Sebastian Böcker
Journal:  Proc Natl Acad Sci U S A       Date:  2015-09-21       Impact factor: 11.205

Review 3.  Current progress in computational metabolomics.

Authors:  David S Wishart
Journal:  Brief Bioinform       Date:  2007-07-11       Impact factor: 11.622

4.  Towards de novo identification of metabolites by analyzing tandem mass spectra.

Authors:  Sebastian Böcker; Florian Rasche
Journal:  Bioinformatics       Date:  2008-08-15       Impact factor: 6.937

5.  Relevance Vector Machines: Sparse Classification Methods for QSAR.

Authors:  Frank R Burden; David A Winkler
Journal:  J Chem Inf Model       Date:  2015-07-21       Impact factor: 4.956

6.  Decision tree supported substructure prediction of metabolites from GC-MS profiles.

Authors:  Jan Hummel; Nadine Strehmel; Joachim Selbig; Dirk Walther; Joachim Kopka
Journal:  Metabolomics       Date:  2010-02-16       Impact factor: 4.290

7.  SCALABLE FUSED LASSO SVM FOR CONNECTOME-BASED DISEASE PREDICTION.

Authors:  Takanori Watanabe; Clayton D Scott; Daniel Kessler; Michael Angstadt; Chandra S Sripada
Journal:  Proc IEEE Int Conf Acoust Speech Signal Process       Date:  2014-05

8.  HMDB 3.0--The Human Metabolome Database in 2013.

Authors:  David S Wishart; Timothy Jewison; An Chi Guo; Michael Wilson; Craig Knox; Yifeng Liu; Yannick Djoumbou; Rupasri Mandal; Farid Aziat; Edison Dong; Souhaila Bouatra; Igor Sinelnikov; David Arndt; Jianguo Xia; Philip Liu; Faizath Yallou; Trent Bjorndahl; Rolando Perez-Pineiro; Roman Eisner; Felicity Allen; Vanessa Neveu; Russ Greiner; Augustin Scalbert
Journal:  Nucleic Acids Res       Date:  2012-11-17       Impact factor: 16.971

9.  Open Babel: An open chemical toolbox.

Authors:  Noel M O'Boyle; Michael Banck; Craig A James; Chris Morley; Tim Vandermeersch; Geoffrey R Hutchison
Journal:  J Cheminform       Date:  2011-10-07       Impact factor: 5.514

10.  Computational mass spectrometry for small molecules.

Authors:  Kerstin Scheubert; Franziska Hufsky; Sebastian Böcker
Journal:  J Cheminform       Date:  2013-03-01       Impact factor: 5.514

View more
  7 in total

1.  Recent advances and prospects of computational methods for metabolite identification: a review with emphasis on machine learning approaches.

Authors:  Dai Hai Nguyen; Canh Hao Nguyen; Hiroshi Mamitsuka
Journal:  Brief Bioinform       Date:  2019-11-27       Impact factor: 11.622

2.  MetFID: artificial neural network-based compound fingerprint prediction for metabolite annotation.

Authors:  Ziling Fan; Amber Alley; Kian Ghaffari; Habtom W Ressom
Journal:  Metabolomics       Date:  2020-09-30       Impact factor: 4.747

3.  ADAPTIVE: leArning DAta-dePendenT, concIse molecular VEctors for fast, accurate metabolite identification from tandem mass spectra.

Authors:  Dai Hai Nguyen; Canh Hao Nguyen; Hiroshi Mamitsuka
Journal:  Bioinformatics       Date:  2019-07-15       Impact factor: 6.937

4.  Machine learning for identification of silylated derivatives from mass spectra.

Authors:  Milka Ljoncheva; Tomaž Stepišnik; Tina Kosjek; Sašo Džeroski
Journal:  J Cheminform       Date:  2022-09-15       Impact factor: 8.489

5.  Convolutional Neural Network-Based Compound Fingerprint Prediction for Metabolite Annotation.

Authors:  Shijinqiu Gao; Hoi Yan Katharine Chau; Kuijun Wang; Hongyu Ao; Rency S Varghese; Habtom W Ressom
Journal:  Metabolites       Date:  2022-06-29

Review 6.  Strategies for structure elucidation of small molecules based on LC-MS/MS data from complex biological samples.

Authors:  Zhitao Tian; Fangzhou Liu; Dongqin Li; Alisdair R Fernie; Wei Chen
Journal:  Comput Struct Biotechnol J       Date:  2022-09-07       Impact factor: 6.155

7.  Probabilistic framework for integration of mass spectrum and retention time information in small molecule identification.

Authors:  Eric Bach; Simon Rogers; John Williamson; Juho Rousu
Journal:  Bioinformatics       Date:  2021-07-19       Impact factor: 6.937

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.