Literature DB >> 35758801

MLGL-MP: a Multi-Label Graph Learning framework enhanced by pathway interdependence for Metabolic Pathway prediction.

Bing-Xue Du1, Peng-Cheng Zhao1, Bei Zhu1, Siu-Ming Yiu2, Arnold K Nyamabo3, Hui Yu3, Jian-Yu Shi1.   

Abstract

MOTIVATION: During lead compound optimization, it is crucial to identify pathways where a drug-like compound is metabolized. Recently, machine learning-based methods have achieved inspiring progress to predict potential metabolic pathways for drug-like compounds. However, they neglect the knowledge that metabolic pathways are dependent on each other. Moreover, they are inadequate to elucidate why compounds participate in specific pathways.
RESULTS: To address these issues, we propose a novel Multi-Label Graph Learning framework of Metabolic Pathway prediction boosted by pathway interdependence, called MLGL-MP, which contains a compound encoder, a pathway encoder and a multi-label predictor. The compound encoder learns compound embedding representations by graph neural networks. After constructing a pathway dependence graph by re-trained word embeddings and pathway co-occurrences, the pathway encoder learns pathway embeddings by graph convolutional networks. Moreover, after adapting the compound embedding space into the pathway embedding space, the multi-label predictor measures the proximity of two spaces to discriminate which pathways a compound participates in. The comparison with state-of-the-art methods on KEGG pathways demonstrates the superiority of our MLGL-MP. Also, the ablation studies reveal how its three components contribute to the model, including the pathway dependence, the adapter between compound embeddings and pathway embeddings, as well as the pre-training strategy. Furthermore, a case study illustrates the interpretability of MLGL-MP by indicating crucial substructures in a compound, which are significantly associated with the attending metabolic pathways. It is anticipated that this work can boost metabolic pathway predictions in drug discovery.
AVAILABILITY AND IMPLEMENTATION: The code and data underlying this article are freely available at https://github.com/dubingxue/MLGL-MP.
© The Author(s) 2022. Published by Oxford University Press.

Entities:  

Mesh:

Year:  2022        PMID: 35758801      PMCID: PMC9235472          DOI: 10.1093/bioinformatics/btac222

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.931


1 Introduction

Enzymes catalyze drug or drug-like compounds into their metabolites, which differ significantly from these compounds themselves (Zhang and Tang, 2018). As a complex biotransformation, a compound metabolic pathway contains a set of interlocking enzymatic reactions (Jia ). In drug discovery, it matters to identify what metabolic pathways a compound attends in the stage of lead compound optimization (Baranwal ; Cho ; Sankar ). More importantly, there is a crucial need in drug design to understand why compounds attend specific metabolic pathways. However, since a compound (e.g. beta-Alanine) would attend one or more pathways (Fig. 1), biological assays are always costly and time-consuming to identify pathways among a vast set of pathway combinations. In recent years, computational methods, especially machine learning-based methods, are promising to predict possible metabolic pathways rapidly for given compounds (Baranwal ; Zhu ).
Fig. 1.

Illustration of compounds and their metabolic pathways. Labels l∼l represent different types of metabolic pathways. For example, beta-Alanine is metabolized by five pathways, labeled as l, l, l, l and l. Among them, pathway l is shared by Pentulose-5-phosphate and beta-Alanine while pathway l is shared by Phosphoenolpyruvate acid and beta-Alanine. Especially, pathway l is commonly shared by beta-Alanine, Pentulose-5-phosphate and Phosphoenolpyruvate acid. The list of metabolic pathway names can be found in Table 1

Illustration of compounds and their metabolic pathways. Labels l∼l represent different types of metabolic pathways. For example, beta-Alanine is metabolized by five pathways, labeled as l, l, l, l and l. Among them, pathway l is shared by Pentulose-5-phosphate and beta-Alanine while pathway l is shared by Phosphoenolpyruvate acid and beta-Alanine. Especially, pathway l is commonly shared by beta-Alanine, Pentulose-5-phosphate and Phosphoenolpyruvate acid. The list of metabolic pathway names can be found in Table 1
Table 1.

Statistics of metabolic pathway dataset

TypeMetabolic pathway typesInvolving compounds
l0 Carbohydrate metabolism1126
l1 Energy metabolism750
l2 Lipid metabolism1066
l3 Nucleotide metabolism342
l4 Amino acid metabolism1440
l5 Metabolism of other amino acids597
l6 Glycan biosynthesis and metabolism325
l7 Metabolism of cofactors and vitamins948
l8 Metabolism of terpenoids and polyketides1483
l9 Biosynthesis of other secondary metabolites1906
l10 Xenobiotics biodegradation and metabolism1452
Former machine learning-based methods can be roughly categorized into network-based and classification-based. Network-based methods generally construct certain interaction networks and leverage network propagation algorithms to infer potential pathways for compounds. For example, Hu constructed a network of chemical–chemical interactions (CCIs) to predict the association of a query compound to 11 kinds of metabolic pathways. As its extension, Gao integrated three networks, involving CCIs, protein–protein interactions (PPIs) and chemical–protein interactions (CPIs) to predict metabolic pathways. Very recently, Zhu proposed a heterogeneous network involving chemicals, enzymes, CCI, CPI and PPI information, where nodes are chemicals or enzymes, and edges are interactions between nodes. However, the major limitation of network-based methods cannot process the compounds which are isolated in the network. Classification-based methods leverage the technique of multi-label learning to infer potential metabolic pathways. Usually, there are three kinds of multi-label learning strategies. The first technique converts a multi-label classification into one or more binary classifications. For example, taking SVMs the binary classifier, Fang and Chen (2017) set compound-pathway pairs as samples where positive samples are validated pairs and negative samples are unknown pairs. Guo first construct seven compound association networks from KEGG and STITCH. Then, they apply Random Walk with Restart to generate network embeddings, which further were fused as the final compound features. Last, they train a set of pathway-specific binary SVMs. As an extension of Guo , Jia recently use random forest (RF) as basic classifiers for accommodating more sub-types of pathways. However, these methods ignore the relationship between labels and aggregate the imbalance between positive and negative samples by generating many negative samples. The second sort uses a random k-label sets (RAKEL) (Tsoumakas ) to treat each kind of label combinations as a new label. Thus, it turns the multi-label classification into a multi-class classification. Compared to the previous binary classification, it provides the co-occurrence of multiple labels. For example, iMPT-FRAKEL (Jia ) encodes compounds by fingerprints and leverages a random k-label sets algorithm to tackle the multi-label classification by SVMs and RF. However, these methods also face the imbalance issue where new combined labels usually account for few samples. The last one directly performs a direct multi-label classification by deep learning. For example, Baranwal and Yang utilized graph neural networks (GNNs) [i.e. graph convolutional network (GCN) and GATs, respectively] to extract compound features based on 2D compound graphs and concatenated them with additional features derived from fingerprints (e.g. MACCS) and molecular properties (e.g. the number of aromatic rings, molecular weight and log P) as the final features. However, these methods ignore the dependence between pathways. In summary, although existing methods have achieved inspiring performance in metabolic pathway prediction, they neglect the dependence between pathways (i.e. pathway crosstalk in terms of biology). For example, both Lipid Metabolism (LM) and Carbohydrate Metabolism (CM) are affected by Metabolism of Cofactors and Vitamins (CVM). Deficiencies of vitamin B1, folic acid and vitamins B6 and B12 (in CVM) lead to a significant increment of lipid deposits in the aorta (in LM) (Kalyesubula ; McNeil ). In addition, The B group vitamins help convert carbohydrates into energy in CM (Calderón-Ospina and Nava-Mesa, 2020). Moreover, such a pathway interdependence is asymmetric due to upstream/downstream relationships between pathways (Yan ). Therefore, the characterization of asymmetric interdependences among pathways would enhance the prediction task in the context that a compound participates in uncertain numbers of metabolic pathways. Furthermore, existing computational methods are inadequate to interpret why a compound attends a specific metabolic pathway (Baranwal ; Yang ). In fact, metabolic pathways are usually related to the presence of certain chemical substructures. For example, amino and carboxylic substructures play an important role in binding to enzymes in Amino Acid Metabolism (Lopez and Mohiuddin, 2021). Therefore, the capture of crucial substructures (possibly revealing functional groups), will help reveal the mechanism of a compound metabolized by enzymes. To address the above issues (pathway interdependence and interpretability), we develop a Multi-Label Graph Learning framework enhanced by pathway interdependence for Metabolic Pathway prediction (MLGL-MP). This end-to-end framework contains a compound encoder, a pathway encoder and a multi-label predictor. The compound encoder learns compound embedding representations based on molecular graphs, while the pathway encoder learns pathway interdependence embeddings. The multi-label prediction discriminates which pathways a compound attends based on two kinds of embeddings. Overall, the main contributions of our MLGL-MP are as follows. It provides an interpretable manner to indicate crucial compound substructures which are significantly associated with metabolic pathways. By capturing the pathway interdependence, it significantly improves the characterization of the relevance between compounds and their metabolic pathways. It proposes a direct multi-pathway prediction approach by measuring the proximity between compounds and metabolic pathways in a common embedding space.

2 Materials and methods

2.1 Problem formulation and model construction

Given m compounds and a listof metabolic pathways . Suppose that a compound is assigned with a set of attending pathways . The task is to predict the pathway set of a newly coming compound , where . The prediction can be modeled as a problem of multi-label learning, which learns a function mapping . For this task, we design a multi-label graph learning framework, which contains a compound encoder, a pathway encoder and a multi-label predictor (shown in Fig. 2).
Fig. 2.

The overall framework of MLGL-MP for multi-label metabolic pathway prediction. It is an end-to-end learning model, which contains a compound encoder, a pathway encoder and a multi-label predictor. The compound encoder generates compound embeddings based on molecular graphs by the composite of a GAT and a GCN. The pathway encoder generates pathway embeddings by two-layer GCNs on a pathway dependence graph, where nodes are pathways, node features are obtained by a pre-training strategy and edges are asymmetric pathway dependences. The multi-label predictor directly discriminates the metabolic pathways of given compounds by the proximity of pathway embeddings and compound embeddings

The overall framework of MLGL-MP for multi-label metabolic pathway prediction. It is an end-to-end learning model, which contains a compound encoder, a pathway encoder and a multi-label predictor. The compound encoder generates compound embeddings based on molecular graphs by the composite of a GAT and a GCN. The pathway encoder generates pathway embeddings by two-layer GCNs on a pathway dependence graph, where nodes are pathways, node features are obtained by a pre-training strategy and edges are asymmetric pathway dependences. The multi-label predictor directly discriminates the metabolic pathways of given compounds by the proximity of pathway embeddings and compound embeddings

2.2 Compound encoder

The compound encoder, adopting a two-layer GNN architecture, learns compound embedding representations by utilizing molecule graphs. Its first layer is a GAT (Veličković ), which capture the importance of chemical bonds to pathways. The second layer is a GCN (Kipf and Welling, 2017), which further extracts atom features by aggregating neighboring atom features and is followed by a global pooling layer (Nguyen ) to generate compound embeddings. According to chemical structure, each compound is represented as a molecule graph , where V is the set of N atoms and E is a set of chemical bonds. Let (N=|V|) be its adjacency matrix, in which a= 1 indicates the occurring bond between atom i and atom j, and a = 0 indicates no bond. Here, each node v (atom) is initially represented by a q-dimensional binary feature vector . As suggested in Nguyen , the initial node features typically include the atom symbol, the number of adjacent atoms, the number of adjacent hydrogens, the implicit value of the atom and the atom occurrence in an aromatic structure. First of all, for each atom in the molecular graph of compound c, the GAT layer updates its features by aggregating the features of its neighboring atoms. The aggregation is implemented by a shared self-attention operation a: , which defines the importance of its neighboring atom to atom as follows: Here, the learnable weight matrix accounts for a linear transformation from input features into higher-level features, where s is the dimension of updated atom features. Then, the updated features of atom v can be defined by where = softmax (is the normalized attention coefficient and is a non-linear activation function (i.e. ReLU). The normalization of attention coefficients makes themselves comparable among different nodes. Furthermore, a multihead attention is adopted to enhance the expression of the attention layer. Specifically, K independent attention mechanisms are performed in parallel and their features are concatenated as the updated features of atom as follows: where is the dimension of updated atom features, K is the number of attention heads, and is the concatenation operation of vectors. It is remarkable that the average of accounts for the importance of the chemical bond between atom v and atom v. After that, the GCN layer following the GAT layer further updates atom features by emphasizing the topology of molecule graph. According to the propagation rules of GCN, the updated features of atoms of compound c is determined by the following matrix form: where is the GAT-based feature matrix stacked by , is the GCN-updating feature matrix stacked by the updated atom feature vector , is the weight matrix in the GCN layer, , is the identity matrix, D is the degree matrix, in which diagonal elements are the degrees of each vertex and . Once the atom embedding vectors of compound c are obtained, a readout operation finally turns them into the embedding vector of the compound. In the readout, both a global Max-pooling and a global Mean-pooling are performed in parallel. Their resulting embeddings, and , are concatenated as the final embedding of compound c.

2.3 Pathway encoder

The metabolic pathway encoder, containing a two-layer GCN architecture, learns pathway embedding representations by constructing a pathway dependence graph, which characterizes asymmetric pathway interdependence. In such a graph, nodes are pathways and directed edges are their asymmetric dependences. A compound is assigned with a set of attending pathways . Inspired by Chen , the construction of the graph includes two phases, node initialization and edge building. The node initialization assigns initial node features. However, they cannot be directly learned in an end-to-end manner due to the small number of pathways (i.e. 11 pathways in the benchmark dataset). To address this issue, a pre-training strategy is adopted to obtain initial pathway embeddings. Considering that pathway names are semantic, we apply GloVe, a word representation tool, to implement the pre-training (Pennington ). For the pathway set, GloVe generates initial pathway embeddings by learning word embeddings of semantic pathway names based on the Common Crawl dataset (https://nlp.stanford.edu/projects/glove/). The edge building contains three steps to capture the pathway interdependence. First, the pathway co-occurrence is calculated based on the dataset in hand. Define the co-occurrence matrix as , where u is the pairwise co-occurrence counts between pathway t and t, u = u and C denotes the number of metabolic pathway types. Then, a probability matrix P is calculated by the co-occurrence matrix . Define p = p (t|t) as the probability of t occurring when given t occurring, and as the total occurrence count of t. The probability matrix p can be calculated by = u/N . Note that P is usually not equal to P. As Chen suggested, a binarization of P is performed to address the possible long-tail distribution of co-occurrence patterns, where a few {P}are significantly greater than others. Formally, the conditional probability matrix P is turned to a binary matrix by a hard threshold as: This binarization removes trivial edges, which could be noise. Last, to enhance the node distinguishability during the aggregation of neighboring nodes, an extra re-weighting scheme is utilized (Li ) as follows: where 1 and is the re-weighted dependence matrix. Once the pathway interdependence graph is constructed, two layers of GCNs are used to learn pathway embeddings. Denote as the initial feature vector of a given pathway and as the initial pathway feature matrix stacked by . Passing through a GCN layer, is updated by the following propagation rule: where , is its degree matrix where , is the embedding matrix in the layer, is a trade-off coefficient, which determines how neighboring nodes are emphasized in the convolutional aggregation. In addition, is the identity matrix, denotes an LeakyReLU activation function and is a transformation weight matrix to be learned. Specifically, the input pathway feature matrix , while the output of the second layer is just the pathway embedding feature matrix, denoted as , where B is the dimension of pathway embedding features. The dimension can be same as that of compound embedding features for the further purpose of measuring the proximity between pathways and compounds.

2.4 Multi-label predictor

After obtaining compound embeddings and pathway embeddings, we can directly perform discriminate the pathway set, in which a compound could attend. Inspired by Chen , we measure the proximities between a given compound and a list of pathways as follows: where is the embedding vector of a compound c and is the pathway embedding feature matrix, of which each row denotes the embedding of pathway j. The proximity is the predicting score of the given compound attending in the i-th pathway among the pathway list . However, such a direct proximity measure would be senseless since the compound embedding space and the pathway embedding space are of different vector spaces. To address this issue, we design an adapter to map the compound embedding space into the pathway embedding space. The adapter can be implemented by a dense neural network (DNN) containing an input layer, a hidden layer, and an output layer. Thus, the final compound representation feature is defined as . Last, the multi-label classification loss (He ) is used when training MLGL-MP. It is defined as follows: where } are a true label indicating whether or not the compound participates in pathway i, and is the corresponding confidence score output by MLGL-MP.

2.5 Evaluation metrics

To evaluate the performance of multi-label learning models, we follow the conventional settings in Baranwal and Yang , which use Accuracy, Precision, Recall and F1_score as the performance metrics. The greater value these metrics are, the better performance the model achieves. Furthermore, we use four additional indicators designated for multi-label learning (Paniri ; Zhang ), including Hamming Loss (HL), Ranking Loss (RL), Coverage and One Error (OE). HL provides an assessment how many times a pair of sample label is misclassified. RL provides an assessment about the fraction of reversely ordered label pairs. Coverage provides an assessment how far the list of ranked labels goes down to cover all the truth labels of samples on average. OE provides an assessment about the fraction of samples whose top-ranked label is not in the set of proper labels. For HL, RL, Coverage and OE, the smaller the values, the better the performance.

3 Experiments

3.1 Dataset

The experimental dataset was taken from Baranwal , which includes 6669 compounds and their metabolic pathway entries. The dataset was originally collected from KEGG Pathway (https://www.genome.jp/kegg/pathway.html), which contains 11 types of metabolic pathways, including (i) Carbohydrate metabolism; (ii) Energy metabolism; (iii) Lipid metabolism; (iv) Nucleotide metabolism; (v) Amino acid metabolism; (vi) Metabolism of other amino acids; (vii) Glycan biosynthesis and metabolism; (viii) Metabolism of cofactors and vitamins; (ix) Metabolism of terpenoids and polyketides; (x) Biosynthesis of other secondary metabolites; and (xi) Xenobiotics biodegradation and metabolism. As the compound encoder of our MLGL-MP requires, we removed 21 compounds that cannot be converted to molecular graphs. We finally constructed a dataset containing 6648 compounds and their 11435 metabolic pathway entries. Among them, 4898 compounds attend only one metabolic pathway while the remaining 1750 compounds attend multiple metabolic pathways. Specifically, 38 out of 1750 compounds attend all the 11 metabolic pathways. The metabolic pathway dataset is summarized in Table 1 and in Fig. 3. More details can be found at https://github.com/dubingxue/MLGL-MP.
Fig. 3.

Compound distribution on multiple metabolic pathways

Compound distribution on multiple metabolic pathways Statistics of metabolic pathway dataset

3.2 Parameter setting

In the compound encoder, each node of input compound graph was initially represented by a 78-dimensional (78-d) binary atom feature vector, as suggested in Nguyen . In brief, the initial representation of a compound contains five groups of atom indicators, including the atom symbol (44-d), the number of adjacent atoms (11-d), the number of adjacent hydrogens (11-d), the implicit value of the atom (11-d) and the atom occurrence in an aromatic structure (1-d). More details can be found in DeepChem (Ramsundar, 2018). Moreover, its GAT layer adopted 10 heads of attention layers, of which the outputting atom dimension was also set as 78 to capture the importance of chemical bonds. Similarly, its GCN layer kept the same dimension (780) of outputting atom features as that in the GAT layer. Finally, through the Global Max-pooling and the Global Mean-pooling on atoms, each compound was represented a 1560-dimensional embedding vector. For the pathway encoder, the pre-training of 11 metabolic pathways was implemented by the GloVe algorithm (Pennington ), which represent their names in 300-dimensional embedding vectors based on the Common Crawl dataset containing 840 billion tokens and 2.2 million vocab. Both the two layers of GCNs in the pathway encoder were represented 1024-dimensional node embeddings. Finally, each pathway was represented as a 1024-dimensional vector. In the multi-label predictor, the adapter was implemented by a DNN, of which the input layer, the hidden layer and the output layer contain 1560, 1500 and 1024 neurons, respectively. After setting up the architecture of MLGL-MP, we investigated how its hyperparameters, including the threshold in Equation (5) and the coefficient in Equation (7) influence the metabolic pathway prediction. For the pathway dependence matrix, we tuned the value of for the list of {0.1, 0.2, 0.3…,0.9}. We discarded two non-convergence cases where indicates no edge removed and generates a zero-dependence matrix. Furthermore, we set in a set of {0.1, 0.2, 0.3…,0.9, 1.0}. Similarly, we discarded the case of , which makes the dependence matrix as an identity matrix. The Accuracy metric was adopted to evaluate the investigation generated by the grid research on and . The results show that the pair of = 0.5 and =0.3 accounts for the best performance of MLGL-MP (Fig. 4). Also, we tuned the learning rate from the list {0.1, 0.01, 0.001, 0.0005, 0.0001}, where 0.0005 accounts for the best performance.
Fig. 4.

Grid search expanded by in terms of accuracy

Grid search expanded by in terms of accuracy In addition to these fine-tuned hyper-parameters, we empirically set the number of epochs as 200, set the batch size as 256 and selected Adam as the optimizer. Similarly, ReLU and LeakyReLU with default parameters were used as the activation functions in the compound encoder and the pathway encoder, respectively. All the remaining experiments were run under the optimal values of these parameters.

3.3 Comparisons with baselines

We assessed the performance of MLGL-MP by the comparison with two state-of-the-art shallow learning methods, (i.e. RF and XGBoost), as well as two state-of-the-art deep learning methods, including a GCN-based method (Baranwal ) and a GAT-based method (Yang ). They are briefly summarized as follows. RF-based model: Since RF was originally designed for multi-class classification but not multi-label classification, we implemented RF in the exact same way as that in Baranwal . Eleven RF classifiers were trained separately to recognize each pathway type with the parameter setting where the maximum depth of the tree is 60 and the number of decision trees is 300. XGBoost-based model: The implementation of XGBoost is similar to that of RF. Eleven XGBoost classifiers were trained separately with a parameter setting where the maximum depth of the tree is 30 and the number of decision trees is 300. The GCN-based model (Baranwal ): it proposes a compound subgraph representation learning based on GCNs and combined the learned subgraph embeddings as local features with global features (diverse molecular properties, MACCS fingerprints, adjacency matrix, etc.) to a feedforward neural network. We used the same parameters as those in the original paper. The GAT-based model (Yang ): it adopts GATs to obtain compound subgraph representation as local features and used the same global features same to the GCN-based model (Baranwal ). We used the default values of parameters as those in the original paper. For a fair comparison, we utilized 10-fold cross-validation (10-CV) for all the methods and measured their performance by two groups of metrics (Table 2). The first group contains the average Accuracy, Precision, Recall and F1_score. The greater, the better. The second group contains HL, Coverage, OE and RL, which are designated metrics for multi-label learning. The smaller, the better. The results show that the deep learning-based methods (the GCN-based, the GAT-based and our MLGL-MP) outperform the shallow learning-based methods (the RF-based and the XGBoost-based) overall. Moreover, it reveals that MLGL-MP achieves the best performance with significant improvements over all the metrics, compared with the GCN-based model and the GAT-based model. Therefore, the comparison demonstrates the superiority of our MLGL-MP.
Table 2.

Performance evaluation on the KEGG dataset of multi-label metabolic pathway prediction

MethodAccuracy (%)Precision (%)Recall (%)F1_score (%)HLCoverageOERL
RF97.59 ± 0.1983.58 ± 0.8483.54 ± 0.7983.56 ± 0.810.024 ± 0.0021.809 ± 0.0690.156 ± 0.0080.167 ± 0.008
XGBoost98.04 ± 0.1889.66 ± 0.5890.49 ± 0.8590.07 ± 0.640.020 ± 0.0021.447 ± 0.0870.099 ± 0.0050.100 ± 0.008
GCN-based97.53 ± 0.4191.37 ± 1.2093.22 ± 1.6092.28 ± 1.300.025 ± 0.0041.033 ± 1.1400.100 ± 0.1530.040 ± 0.082
GAT-based97.57 ± 0.1892.71 ± 0.6492.04 ± 0.8792.53 ± 0.390.024 ± 0.0020.830 ± 0.3180.064 ± 0.0820.024 ± 0.028
MLGL-MP 98.64 ± 0.47 95.26 ± 2.25 94.21 ± 1.94 94.73 ± 1.89 0.014 ± 0.005 0.559 ± 0.113 0.050 ± 0.019 0.011 ± 0.003
Performance evaluation on the KEGG dataset of multi-label metabolic pathway prediction

3.4 Ablation studies

In this section, we investigated why MLGL-MP can achieve inspiring prediction by ablation studies. We made three variants of MLGL-MP, of which the first removes the pathway encoder (denoted as w/o PE), the second one lacks the adapter in the multi-label predictor (denoted as w/o AP), the third (denoted as MLGL-MP-r) alters the pre-trained node feature vectors in the pathway dependence graph to randomly initialized Gaussian vectors (Fig. 5).
Fig. 5.

Ablation comparison. Compared with MLGL-MP, w/o PE removes the pathway encoder, w/o AP lacks the adapter in the multi-label predictor, MLGL-MP-r denotes alters the pre-trained node feature vectors in the pathway dependence graph to randomly initialized Gaussian vectors. The left panel indicates the performance by Accuracy, Precision, Recall and F1-score. The middle one indicates the performance with regard to Coverage. The right panel indicates the comparison in terms of Hamming Loss (HL), One Error (OE) and Ranking Loss (RL)

Ablation comparison. Compared with MLGL-MP, w/o PE removes the pathway encoder, w/o AP lacks the adapter in the multi-label predictor, MLGL-MP-r denotes alters the pre-trained node feature vectors in the pathway dependence graph to randomly initialized Gaussian vectors. The left panel indicates the performance by Accuracy, Precision, Recall and F1-score. The middle one indicates the performance with regard to Coverage. The right panel indicates the comparison in terms of Hamming Loss (HL), One Error (OE) and Ranking Loss (RL) MLGL-MP significantly outperforms w/o PE over all the evaluation metrics. In detail, compared with w/o PE, MLGL-MP improves the Accuracy by 2.70%, the Precision by 5.88%, the Recall by 15.86% and the F1-score by 11.27% due to its pathway encoder. Meanwhile, MLGL-MP reduces the HL by 0.027, the Coverage by 0.212, the OE by 0.076 and the RL by over 0.018. The result indicates that the designated pathway embeddings improve metabolic pathway prediction greatly because it captures the pathway interdependence. Moreover, the comparison with w/o AP shows a similar improvement over all the metrics. In detail, MLGL-MP improves the accuracy by 1.01%, the precision by 4.43%, the recall by 3.16% and the F1-score by 3.84%. Again, it shows a better performance of multi-label learning with reducing the HL by 0.009, the Coverage by 0.046, the OE by 0.019 and the RL by 0.003. This result shows that the adapter in the predictor improves the prediction significantly by aligning compound embeddings with pathway embeddings. In addition, the comparison shows that the version with pathway pre-training (MLGL-MP) is better than that with pathway random initialization (MLGL-MP-r) over all the evaluation metrics. Thus, the pre-training strategy can improve the prediction. In general, the pathway encoder, the adapter and the pre-training strategy play indispensable roles in predicting multi-label metabolic pathways.

3.5 Case study: interpretability of MLGL-MP

Although deep learning is known as a black-box model, it is essential to understand how the model makes a prediction and whether the model can guide lead compound optimization in drug discovery. MLGL-MP leverages the GAT layer in its compound encoder to access why a compound participates in a specific pathway. Since the attention weights learned in the GAT layer can reflect the importance of chemical bonds in compounds, we can reveal the association between compounds’ substructures and their metabolic pathway. For example, Energy Metabolism and Amino Acid Metabolism are two important pathways in organisms (Rui, 2014; Vettore ). The former maintains the regular activity of metabolic enzymes (Foo ; Motohashi and Akaike, 2019) while the latter are an essential process in cells (Lopez and Mohiuddin, 2021). Thus, we selected them as examples to illustrate the interpretability of MLGL-MP (Fig. 6).
Fig. 6.

Visualization of compound substructure importance. (A) Energy metabolism; (B) Amino acid metabolisms.

Visualization of compound substructure importance. (A) Energy metabolism; (B) Amino acid metabolisms. Overall, the visualized attention weights show that most carbon (C)-based chemical bonds constructing compound backbones usually have small attention values. More importantly, the visualization reveals that crucial substructures having high attentions are pathway-specific. We went deeper into the case of Energy Metabolism (Fig. 6A), where Sulfur (S) and Phosphorus (P)-based chemical bonds in Energy Metabolism have higher attention values and are highlighted in red. The result is consistent with the early knowledge that Sulfur metabolism and Oxidative phosphorylation occur in the pathway of Energy Metabolism (Gibson and Skett, 2013). Meanwhile, recent works also provided more pieces of evidence (Foo ; Motohashi and Akaike, 2019). For example, the unique role of sulfur in organisms is mainly related to redox reactions and its functions include cell protection and energy metabolism (Motohashi and Akaike, 2019). Oxidative phosphorylation has garnered increasing interest in energy metabolism as a new target space (e.g. the mycobacterial druggable target) (Foo ). Moreover, the case of Amino Acid Metabolism (Fig. 6B) shows that the substructures of amino and carboxylic have greater attention weights. The result is validated by the work (Lopez and Mohiuddin, 2021), which indicates both amino (−NH2) and carboxylic acid (−COOH) functional groups play an important role in Amino Acid Metabolism. In summary, MLGL-MP is an interpretable model, which can indicate compound substructures significantly associated with metabolic pathways. It would help reveal why a compound participates in a specific pathway.

4 Conclusion

In this paper, we have proposed an MLGL-MP, which contains a compound encoder, a pathway encoder and a multi-label predictor. This end-to-end framework can address two existing issues, including inadequate characterization of pathway dependences and interpretable prediction. The comparison with popular shallow learning models and deep learning models demonstrates the superiority of MLGL-MP. Moreover, the ablation studies as well as the case study validate its contributions. First, it provides an interpretable manner to indicate crucial compound substructures which are significantly associated with metabolic pathways by molecular graph attention embedding. Secondly, by capturing the pathway interdependence, it significantly improves the characterization of the relevance between compounds and their metabolic pathways. Thirdly, by measuring the proximity between compounds and metabolic pathways in a common embedding space, it proposes a direct multipathway prediction approach without extra label strategy. In summary, we believe that our study provides new insights into label dependence representation learning for other multi-label classification problems (e.g. drug toxicity prediction) in drug discovery. Moreover, though the GAT can interpret the importance of drug substructures to metabolic pathways in some sense, other parts (i.e. the pathway encoder and the adapter) in the model are of the black box. In the coming future, it is anticipated that interpretable techniques derived from image processing (e.g. visualization of hidden layers, nearest neighbors and GAN) can be utilized to achieve better interpretability in predicting metabolic pathways for compounds.

Funding

This work has been supported by the National Nature Science Foundation of China [Grant number 61872297) and Shaanxi Provincial Key Research & Development Program, China (Grant number 2020KW-063). Conflict of Interest: We declare that we have no conflict of interest.
  16 in total

1.  Prediction of novel synthetic pathways for the production of desired chemicals.

Authors:  Ayoun Cho; Hongseok Yun; Jin Hwan Park; Sang Yup Lee; Sunwon Park
Journal:  BMC Syst Biol       Date:  2010-03-28

2.  A Network Integration Method for Deciphering the Types of Metabolic Pathway of Chemicals with Heterogeneous Information.

Authors:  Zi-Han Guo; Lei Chen; Xian Zhao
Journal:  Comb Chem High Throughput Screen       Date:  2018       Impact factor: 1.339

3.  A Binary Classifier for Prediction of the Types of Metabolic Pathway of Chemicals.

Authors:  Yemin Fang; Lei Chen
Journal:  Comb Chem High Throughput Screen       Date:  2017       Impact factor: 1.339

4.  A deep learning architecture for metabolic pathway prediction.

Authors:  Mayank Baranwal; Abram Magner; Paolo Elvati; Jacob Saldinger; Angela Violi; Alfred O Hero
Journal:  Bioinformatics       Date:  2020-04-15       Impact factor: 6.937

Review 5.  Energy metabolism in the liver.

Authors:  Liangyou Rui
Journal:  Compr Physiol       Date:  2014-01       Impact factor: 9.090

6.  Predicting metabolic pathways of small molecules and enzymes based on interaction information of chemicals and proteins.

Authors:  Yu-Fei Gao; Lei Chen; Yu-Dong Cai; Kai-Yan Feng; Tao Huang; Yang Jiang
Journal:  PLoS One       Date:  2012-09-21       Impact factor: 3.240

Review 7.  New aspects of amino acid metabolism in cancer.

Authors:  Lisa Vettore; Rebecca L Westbrook; Daniel A Tennant
Journal:  Br J Cancer       Date:  2019-12-10       Impact factor: 7.640

8.  Analysing the meta-interaction between pathways by gene set topological impact analysis.

Authors:  Shen Yan; Xu Chi; Xiao Chang; Mengliang Tian
Journal:  BMC Genomics       Date:  2020-10-27       Impact factor: 3.969

9.  High-dose vitamin B1 therapy prevents the development of experimental fatty liver driven by overnutrition.

Authors:  Mugagga Kalyesubula; Ramgopal Mopuri; Jimmy Asiku; Alexander Rosov; Sara Yosefi; Nir Edery; Samuel Bocobza; Uzi Moallem; Hay Dvir
Journal:  Dis Model Mech       Date:  2021-03-18       Impact factor: 5.758

Review 10.  Drug metabolism in drug discovery and development.

Authors:  Zhoupeng Zhang; Wei Tang
Journal:  Acta Pharm Sin B       Date:  2018-04-12       Impact factor: 11.413

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.