Literature DB >> 35758803

SPARSE: a sparse hypergraph neural network for learning multiple types of latent combinations to accurately predict drug-drug interactions.

Duc Anh Nguyen1, Canh Hao Nguyen1, Peter Petschner1,2, Hiroshi Mamitsuka1,3.   

Abstract

MOTIVATION: Predicting side effects of drug-drug interactions (DDIs) is an important task in pharmacology. The state-of-the-art methods for DDI prediction use hypergraph neural networks to learn latent representations of drugs and side effects to express high-order relationships among two interacting drugs and a side effect. The idea of these methods is that each side effect is caused by a unique combination of latent features of the corresponding interacting drugs. However, in reality, a side effect might have multiple, different mechanisms that cannot be represented by a single combination of latent features of drugs. Moreover, DDI data are sparse, suggesting that using a sparsity regularization would help to learn better latent representations to improve prediction performances.
RESULTS: We propose SPARSE, which encodes the DDI hypergraph and drug features to latent spaces to learn multiple types of combinations of latent features of drugs and side effects, controlling the model sparsity by a sparse prior. Our extensive experiments using both synthetic and three real-world DDI datasets showed the clear predictive performance advantage of SPARSE over cutting-edge competing methods. Also, latent feature analysis over unknown top predictions by SPARSE demonstrated the interpretability advantage contributed by the model sparsity.
AVAILABILITY AND IMPLEMENTATION: Code and data can be accessed at https://github.com/anhnda/SPARSE. SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
© The Author(s) 2022. Published by Oxford University Press.

Entities:  

Mesh:

Year:  2022        PMID: 35758803      PMCID: PMC9235485          DOI: 10.1093/bioinformatics/btac250

Source DB:  PubMed          Journal:  Bioinformatics        ISSN: 1367-4803            Impact factor:   6.931


1 Introduction

A drug-drug interaction (DDI) is a reaction between two drugs, whereby the effects of one drug are modified by the concomitant use of the second drug. A DDI might cause side effects, which are unwanted effects and are responsible for significant patient morbidity and mortality (Magro ). Hence, predicting side effects of a DDI, i.e. DDI prediction, is a very important task to guarantee drug safety. Using machine learning has emerged as a prominent approach for DDI prediction, making the prediction fast and highly accurate (Xu ; Zitnik ). The traditional machine learning methods such as support-vector machines (Kastrin ), logistic regression (Mei and Zhang, 2021) or feedforward neural networks (Wang ) use predefined drug features to predict side effects as labels. However, DDI data have more information. Particularly, DDIs can be represented by a graph, called a DDI graph, where nodes are drugs and edges are interacting drugs. The DDI graph can be learned with graph neural networks (Zitnik ). Nonetheless, DDI graphs are only limited to pairwise relationships of drug pairs while there still exist many side effects, which can be represented by other relationships, such as co-occurrence. Then, a state-of-the-art generalization of a DDI graph can be a DDI hypergraph, which can capture higher-order relationships, where drugs and side effects are both nodes, and each hyperedge is a triple of a side effect with two interacting drugs. On the DDI hypergraph, hypergraph neural networks can be applied to learn the representations of drugs and side effects altogether. In DDIs, two drugs with totally different properties can still interact with each other, hence the traditional hypergraph neural networks using similarity assumption on node representations are not suitable (Feng ). Instead, CentSmoothie, a current cutting-edge hypergraph neural network for DDIs (Nguyen ), assumes that each side effect is caused by a unique combination of latent features of the corresponding interacting drugs. However, in real life, each side effect might have many different mechanisms (Suleyman ) that cannot be reflected in a single combination of drug latent features. Hence, it is necessary to learn different types of combinations of drug latent features for each side effect. This is the first problem (P1), which we address in this article. To solve P1, we borrow one idea of stochastic block models (SBMs) on hypergraphs such that each node (e.g. drug or side effect) has one or several latent features (Anandkumar ; Pal and Zhu, 2021) and there exist interactions (associations) of latent features. This method can learn different types of combinations of drug latent features for each side effect, at once. In addition, to improve the quality of learned latent features, input node features also can be used (Zhang ). However, transformations from input node features and node relationships in the hypergraphs to latent features might be complex and, especially, non-linear. This is the second problem (P2), which has not been addressed in existing SBMs and we address in this article. Moreover, DDI data are sparse (e.g. in the largest DDI dataset, 97.6% of all triples of drug–drug-side effects are not a DDI), suggesting that the model for learning DDIs also should be sparse. However, recent work on DDIs has not used this sparsity of the data (Nguyen ; Zitnik ), which might potentially impair model performance. This is the third problem (P3), which we address in this article. We propose SPARSE, a new model for DDI prediction, to solve the above three problems. For P1, we assume that there exist drug and side effect latent features with latent interactions so that each side effect latent feature interacts with several pairs of drug latent features. For P2, we encode drug features and the DDI hypergraph altogether in the latent representations using a suitable hypergraph neural network. For P3, we guide the model to preserve the sparsity of the data using a suitable sparsity control. Figure 1 schematically illustrates these ideas of our model. That is, the model consists of two parts: (i) an encoder and (ii) a decoder. The encoder encodes the input of the DDI hypergraph (e.g. three hyperedges in Fig. 1) with drug features into latent spaces of drug and side effect latent representations, and interactions of latent features. The decoder reconstructs from the latent spaces the DDI hypergraph with new DDI predictions (e.g. the dotted hyperedge in Fig. 1). Finally, a sparsity prior (horseshoe priors in our model) is used to control the sparsity of the latent interactions.
Fig. 1.

A schematic illustration of the procedure in the proposed model, SPARSE

A schematic illustration of the procedure in the proposed model, SPARSE Our extensive experiments first validated the advantage of SPARSE in terms of prediction performance by using both synthetic and real-world datasets. Throughout all experiments on prediction performance, SPARSE achieved better prediction performances than competing methods, such as CentSmoothie and SBM. For example, in the experiment of using the largest real DDI dataset, called TWOSIDES, SPARSE achieved area under the ROC curve (AUC) of 0.9524 and (area under the precision-recall curve (AUPR) of 0.882, while CentSmoothie achieved AUC of 0.9348 and AUPR of 0.8749 and SBM achieved AUC of 0.9337 and AUPR of 0.8583. Similarly when using JADERDDI, another DDI dataset, SPARSE achieved AUC of 0.9698 and AUPR of 0.7348, while CentSmoothie was AUC of 0.9684 and AUPR of 0.6044 and SBM was AUC of 0.9428 and AUPR of 0.5963. We then examined the top prediction obtained by SPARSE, which is trained by using the whole TWOSIDES. That is, we checked the number of overlaps between the top 400 predictions by one method and DDIs in drugs.com (Drugs.com, 2021; Thelwall ), which is a commonly used online web checker for DDI. We found 98 DDIs in drugs.com out of the top 400 predictions, while by using the same procedure, CentSmoothie found only 71 DDIs out of the top 400 predictions, implying that SPARSE can find new DDIs more than competing methods. Finally, we validated the prediction results by characterizing the top predictions obtained by SPARSE. In more detail, we checked the biological properties, such as target proteins, of the top 10 triples of drug–drug-side effect, predicted by SPARSE, by using latent features connected to these top 10 predictions. We then found that top predictions can be associated with some biological mechanisms and particularly with responsible proteins/pathways. These results indicate that our model, SPARSE, can provide high predictive performances as well as latent biological knowledge beneficial to understand the background behind predicted DDIs.

2 Related work

Machine-learning models for DDI prediction can be divided into non-graph-based and graph-based ones. For non-graph-based models, the inputs are the predefined feature vectors of pairs of drugs, the outputs are the corresponding side effects, and the models are multi-label classifiers, e.g. support-vector machines (Kastrin ) or a multilayer feedforward neural network (Wang ). Instead of only using predefined drug feature vectors, graph-based methods for DDI use graph neural networks to learn new latent representations of drugs from molecular graphs or DDI graphs. In molecular graphs, each drug is considered as a graph that nodes are atoms and edges are connections of atoms (Harada ; Xu ). In DDI graphs, DDIs are considered as pairwise relationships and formulated in the form of a graph where nodes are drugs and edges are drug interactions with side effects as labels (Zitnik ). The latter one has shown to be more effective for DDI prediction since it can use both pharmacological information and biological information rather than only molecular graphs (Zitnik ). However, one drawback of using graph neural networks on DDI graphs is that it does not use multiple relationships (labels) at the same time. Side effects themselves have relationships with each other, e.g. co-occurrences. Existing work often fixes them as one-hot vectors to indicate the presence of the side effects. This representation considers side effects independently, potentially making the models under-utilize the side effect relationships. Hypergraph neural networks on DDI overcome the above drawback by learning representations of drug and side effect nodes altogether in latent spaces (Nguyen ). DDI is considered as high-order relationships of drug–drug-side effects in the form of a hypergraph where nodes are both drugs and side effects, and each hyperedge is a triple of two interacting drugs and a side effect caused by the drugs. There are two types of hypergraph neural networks models on the DDI: similarity based and non-similarity based. The similarity-based models, e.g. traditional spectral-based hypergraph neural networks, assume that interacting drugs should have similar representations (Fan ; Feng ). However, in DDI, two interacting drugs are not necessarily similar. For non-similarity models, the current state-of-the-art method is CentSmoothie (Nguyen ) that assumes that the representation of a side effect can be represented by a combination of latent features of two drugs causing the side effect. However, CentSmoothie cannot deal with multiple combinations of latent features at the same time. In order to deal with multiple combinations of latent features, one possible approach is to use the idea of SBMs, which can be applied to hypergraphs, with each node belonging to several latent features (groups) and associations of latent features (groups) (Anandkumar ). However this has not been applied to DDI hypergraphs, and more importantly, SBM is based on linear assumption, while DDI can be generated through more complex relations to be represented by non-linearity. Many studies have shown the benefits of sparsity regularization, which is a commonly used method to achieve sparsity of models, especially on noisy and sparse data (Carvalho ; Tibshirani, 1996). In a Bayesian viewpoint, sparsity regularization can be understood as a result of using sparse prior distributions. A state-of-the-art method for sparsity regularization is to use horseshoe priors (Carvalho ; Piironen and Vehtari, 2017). It shows an advantage in comparison with traditional Laplace prior (Lasso regularization) (Tibshirani, 1996) in that the horseshoe prior allows to shrink in both directions: no shrinkage for important features and complete shrinkage for non-important (noise) features. A comparable shrinkage prior with the horseshoe prior is the spike-and-slab prior (Hoeting ). However, the spike-and-slab prior is a discrete prior that requires the Markov chain Monte Carlo sampling for optimization, which is not effective for large-scale datasets like DDI.

3 Materials and methods

3.1 Background

We recall definitions for horseshoe priors and n-mode tensor product for 3D tensors, which will be used later.

Horseshoe priors

We summarize the horseshoe prior (Carvalho ), a state-of-the-art prior for sparsity control, for a non-negative 3D tensor: . The idea of the horseshoe prior is that each follows a normal distribution with the same zero mean and a different variance. Each variance has two parts: one is a global parameter sharing among all variances to decide the sparsity of B and one is a local parameter to decide the magnitude of each variance by using a heavy tail distribution with the half-Cauchy distribution. In more detail: where τ is a global parameter for sparsity, and is a half-Cauchy distribution defined by: for . Both the horseshoe prior and Laplace prior (for Lasso regularization) are shrinkage priors such that by using priors, values of features tend to be shrunk (Piironen and Vehtari, 2017). Let be the optimal values without priors, then the optimal values having priors has the form: , where is a shrinkage factor depending on the priors. With Laplace prior (Lasso regularization), the density of tends to be a constant near 1 and disappears near 0, meaning that it always shrinks all features, containing important ones. In contrast, the density of with the horseshoe prior has two peaks at 0 and 1, meaning that the horseshoe prior allows two kinds of shrinkage: no shrinkage to maintain important features and complete shrinkage to remove unimportant features.

N-mode tensor product

The n-mode tensor product can be understood as a generalization of the matrix dot product in high-dimension that the product is processed at the nth dimension. Considering in the 3D space with a tensor: and a matrix , the n-mode product of B and H is denoted by and is defined for each of n = 1, 2 and 3, as follows:

3.2 Problem formulation: DDI prediction

We formulate the DDI prediction problem as follows. Input: Given a DDI hypergraph: , where V is a set of drug nodes, V is a set of side effect nodes (given , two triples (u, v, t) and (v, u, t) are the same). The drug node features are and the side effect node features are one-hot vectors: Output: For , calculate a prediction score for interaction m(e).

3.3 Proposed model

We propose SPARSE: a sparse model for learning multiple types of latent combinations of side effects and drugs to predict DDIs. Our model follows an auto-encoder framework with two parts: an encoder and a decoder. The encoder encodes the DDI hypergraph with drug node features to latent spaces with latent representations of drugs and side effects (H), and interactions of latent features (B). The decoder aims to reconstruct the DDI hypergraph with new predicted hyperedges from H and B. In the following parts, we first present our latent interaction assumption with sparsity for the interactions of drugs and side effects, and then we describe the encoder and decoder.

Latent interaction assumption

To model DDIs, we suppose that there exist latent spaces with drug latent features and side effect latent features where DDIs occur. The latent interaction assumption is that two interacting drugs cause a side effect if there exist a pair of drug latent features of the two drugs that interact with a latent feature of the side effect. In detail, the formulation for the latent interaction assumption can be described as follows. Let and be the sets of indices of latent features of drugs and side effect with K and K be the numbers of latent features. Let be a 3D tensor representing interactions of latent features of drugs and side effects. The set of interacting latent features is: . Considering a triple of two drugs and one side effect . Let be the vectors representing the presence of latent features of the two drugs and the side effect, respectively. Let and be the sets of latent features of u, v and t, respectively. Under the latent interaction assumption, u interacts with v to cause t if: or with tensor product formulation: In practice, we can change the value 0 on the right side of Equation (7) to a positive threshold. Equation (6) will be used to generate synthetic data in the experimental section. Equation (7) will be used in the decoder of the model. Sparsity property We first define formulations for sparsity measures of the DDI data and the latent interactions using the percentages of non-interactions. Let s be the sparsity of the hypergraph G: The sparsity of the latent interactions s is defined as the percentage of the number of non-interacting triples of the latent features per the total number of all triples of the latent features. DDI data are sparse as per statistics in Table 1. It is shown that 97.6% and 99.87% of all triples are non-interacting in TWOSIDES and JADERDDI, respectively.
Table 1.

Statistics of three real datasets

DatasetNo. of DrugsNo. of Side effectsNo. of Drug–drug pairsNo. of Drug–drug-side effects (DDIs)Avg. no. of side effects/No. ofdrug–drug pairsSparsity (%)
TWOSIDES55796449,6773 606 04672.5897.6
CADDDI58796921,918373 97617.0699.77
JADERDDI54592236,929222 0816.0199.83
Statistics of three real datasets The motivation for us to use sparse models is that sparse models, according to statistical learning theory, are usually more reliable models if they could fit the training data well (Hastie, 2015). As our sparse models have sparse interactions among latent features, we will prove that they tend to generate sparse data and are suitable for DDI data. We show a relationship between sparsity of the models and sparsity of data generated by the models, which are the ones that best fit the models, as follows. Assume that the DDI data are generated from the true generation model according to formula (7). Assuming that each drug and side effect has exactly n and n non-zero latent features, respectively. Then, there exists a relationship between the sparsity of the model and the expected sparsity of the generated data as follows: Proof: For a pair of drug u, v to cause side effect t, then This means that there is at least one non-zero entry of B corresponding to latent features of u, v and t. Since there are exactly possible entries of B corresponding latent features of u, v and t, then the probability of a uniform sampling of entries of B to corresponding to these latent features is . This is the probability of having an interaction among the features (that generates a side effect data point). Since entries of B are assumed to be randomly sampled according to a uniform distribution, the number of interactions when B have non-zero entries follows a binomial distribution . With the assumption that the hypergraph is generated from this generative process, the expected number of non-zero data points (the number of hyperedges) becomes . The expected sparsity of the hypergraph becomes . This result leads to . It shows a relationship between the sparsity of the model (s) and the expected sparsity of the data generated by the model (). It shows that the model can be sparse but cannot be as sparse as we want. It can be a hint on setting sparsity of the model in learning processes.

Encoder

For the encoder, we use a hypergraph neural network with message passing (Yadati, 2020) to encode the input hypergraph and node features into latent spaces with node latent representations H and latent interactions B (for simplicity, B can be considered as a free parameter to learn). where and are hypergraph neural networks based on message passing (Yadati, 2020) with parameters to learn w0, w1, (node representations of drugs) and (node representations of side effects). The formulation of each message passing layer has the following form: where is the representation of node at layer (l), σ is an activation function, is an aggregation function (e.g. an average function), and is a message passing function at layer (l) to pass information from neighbor nodes in hyperedge e to a: where is a two-layer feedforward neural network, and are the node types.

Decoder

The reconstruction of the hypergraph is from the latent interaction assumption. The likelihood to reconstruct each triple follows a Gaussian distribution: where , i(e) = 0 if , and is the mean value for the latent interaction of e: Equation (17) is also the score for the interactions of triples (u, v, t) used for prediction. The likelihood for the decoder is:

Objective function

The objective function for our method is to maximize a posterior of the model. The objective function consists of two parts: one for log-likelihood of the model and one for the prior for sparsity control. Let be the horseshoe prior parameter for B and τ be the hyperparameter for the global sparsity of the horseshoe prior. We have the following objective function: where is the log-likelihood of Equation (18) with H in Equation (11) and B in Equation (12), and is the logarithm of the horseshoe prior: We then use stochastic gradient descent libraries in the PyTorch framework for optimizing Equation (19). We also consider two other variants of : for not using any sparsity prior and for using Laplace prior (Lasso regularization), to examine the effect of using the horseshoe prior.

4 Experimental results

We validated SPARSE in two scenarios: synthetic data and real data. On the synthetic data, assuming that the data are generated from the latent interactions, we examined if SPARSE can recover the latent interactions under changing hyperparameters of data: the number of latent features, sparsity and amount of noise. On real data, we checked the prediction performance of SPARSE in comparison with state-of-the-art DDI prediction methods by using three real-world DDI datasets. Additionally, we evaluated if the top unknown predictions by SPARSE can be related to biological phenomena like functions and mechanisms. For all experiments, we used 20-fold cross-validation by dividing hyperedges into 20-folds, keeping the same number of hyperedges (side effects) in each fold. We reported the mean and standard deviation of the two commonly used measures AUC and AUPR. Also, all reported results were the highest performances through grid searches of hyperparameters. There were three hyperparameters for grid searches for SPARSE: (i) latent feature sizes. The tested values were 30, 40, 50 and 60. We set the same size for all layers. (ii) Global sparsity τ. The tested values were 0.01, 0.02, 0.03, 0.05 and 0.1 and (iii) the numbers of neural layers. The tested values were 1, 2 and 3. The hyperparameter values obtained were 50 for the latent feature size, for TWOSIDES and for CADDDI and JADERDDI, and the number of neural layers was 2. All experiments were run in a computer with Intel Core I7-9700 CPU, 8 GB GeForce RTX 2080 GPU and 32 GB RAM.

4.1 Synthetic data

Data generation

The generation process for synthetic data consists of two steps: (i) generating latent interactions and (ii) generating triples of interacting drug–drug-side effects from the latent interactions, as follows. Generating latent interactions. Given sets of indices of drug latent features: and side effect latent features: . Initialize a set of latent interactions . For each : Sample the number of drug latent feature pairs: , where M is the maximum number of pairs. Sample n pairs . For each pair (i, j): . Generating drug interactions: Generate drug and side effect latent features. Assume that there are V drugs and V side effects. For each drug : Sample the number of drug latent features: , where N1 is the maximum number of drug latent features. Sample . For drug feature vectors F: . . For each side effect , sample the number of side effect latent feature and Sample . Generating true triples . Initialize . For , if then (u, v, t) is a true triple: . Adding noise: For each , replace e by a random sample with probability r. The final set of triples of drug–drug-side effects is E. Finally, we have a synthetic dataset with triples of drug–drug-side effects E and drug feature vectors F.

Experiments

The synthetic data has five hyperparameters: the number of drugs, the number of side effects, the number of latent interactions, data sparsity and the amount of noise (noise rate). We evaluated our methods by changing one hyperparameter, fixing the other four. The hyperparameters changed are (i) number of latent features, (ii) data sparsity and (iii) noise rate.

1) Changing the number of latent features

Setting: , noise rate r = 0.01. We changed . For each (K, K), we selected N1, N2 and M such that the sparsity of the generated data is kept at 0.98. Compared methods: We compared four methods: (no sparsity control), CentSmoothie (Nguyen ), a similarity-based hypergraph neural network, HPNN (Feng ) and SBM on hypergraph (Anandkumar ). Results: Figure 2a shows the results, where achieved the highest performances among the compared methods in all cases. We had the following two findings:
Fig. 2.

Performances on synthetic data, when changing (a) #latent features, (b) sparsity and (c) amount of noise

Performances on synthetic data, when changing (a) #latent features, (b) sparsity and (c) amount of noise For the small number of latent features, the performance of CentSmoothie was close to (both AUC and AUPR were around 0.99 under ). However, by increasing the number of latent features, the performance gap between and CentSmoothie also increased (gaps in AUC and AUPR were around 0.01 and 0.03, respectively, when ). This result implies that CentSmoothie was unable to distinguish latent interactions clearly for a large number of latent interactions, while worked better for capturing multiple latent interactions. The performances of SBM were lower than both CentSmoothie and , since SBM did not use the node features, which decreased the performance. HPNN, a similarity-based hypergraph neural network, had the lowest performance since the two drugs of a DDI do not necessarily have similarity in the data generated from latent interactions. Overall, these results indicated that can recover the latent interactions better than the other methods.

2) Changing data sparsity

Setting: and . We changed M in , resulting in data sparsity in , respectively. Compared methods: Since in the previous experiment, outperformed the compared methods already, we compared SPARSE and two variants and (please see the end of Section 3.3.4) to check the effect of sparse priors. Results: Figure 2b shows the results, where SPARSE achieved the highest performance, followed by and . In particular, the performance advantage by SPARSE using sparsity control was clearer with higher sparsity. These results indicate that the horseshoe prior is suitable for learning sparse data.

3) Changing the amount of noise

Setting: , M = 1 (keeping the data sparsity of 0.98). We changed noise r in . Compared methods: We again compared SPARSE with two variants and to examine the effectiveness of the sparse priors to deal with noise. Results: Figure 2c shows the results, where again SPARSE achieved the highest performances among the three methods for all different amounts of noise. When there are no noises, the performances of the three methods were very close to each other. However, as the amount of noise is increased, the advantage of SPARSE over the other two methods became clearer. For example, when the amount of noise is 20%, the gap between SPARSE and reached around 0.07, and the gap between SPARSE and was around 0.1. These results suggest that the horseshoe prior could deal with noise better than the Laplace prior and the case with no sparsity prior.

4.2 Real data

Data description

We used three real-world datasets for DDI, namely TWOSIDES (Tatonetti ), CADDDI and JADERDDI. To our knowledge, TWOSIDES is the largest benchmark dataset for DDI. The other two datasets, i.e. CADDDI and JADERDDI, were generated from Canada Vigilance Adverse Reaction Reports and Japanese Adverse Drug Event Reports, respectively, in the same manner as the way that TWOSIDES was generated from the adverse events reported to US Food and Drug Administration (Nguyen ). For all datasets, we only chose small molecular drugs, which can be found in DrugBank. Also, we focused drugs appearing in more than five interactions (hyperedges) in each dataset. For each drug, we used a feature (binary) vector, with the size of 2329, consisting of 881 substructures and 1448 interacting proteins. Table 1 shows a summary statistics of the three real benchmark datasets, TWOSIDES, CADDDI and JADERDDI.

Predictive performance experiments

Compared methods: For our method, we used SPARSE and two variants and . We further used five methods as competing methods against SPARSE. These competing methods were CentSmoothie (Nguyen ), the traditional similarity-based hypergraph neural network (HPNN) (Feng ), two DDI graph-based graph neural networks: Decagon (Zitnik ) and SpecConv (Kipf and Welling, 2016) and, a molecular graph-based graph neural network, MRGNN (Xu ). Decagon and CentSmoothie provide available codes, and we ran them with the recommended settings. For MLNN, MGRNN, SpecConv, HPNN and SBM, we implemented them and did a grid search for finding the best hyperparameter values. Results—Cross-validation predictive performance: Table 2 shows AUC and AUPR results of all competing methods. From this table, SPARSE and two variants ( and ) achieved the highest performances, followed by CentSmoothie, SBM and HPNN. On the other hand, the performances of SpecConv, Decagon and MRGNN were significantly lower. Amazingly, (SPARSE without any sparsity prior) achieved still better performance over CentSmoothie, particularly in AUPR. There was only one case (CADDDI), where the AUC of SPARSE was slightly smaller than that of CentSmoothie. We then ran t-test over the prediction results of these two methods, to examine the significance of the difference between CentSmoothie and SPARSE. The resultant P-value of t-test was 0.057, indicating that the performance advantage of CentSmoothie over SPARSE was NOT significant, under the regular significance level of 0.05. Also, it has to be noted that AUPR is more useful than AUC for imbalanced data (Saito and Rehmsmeier, 2015), which can be often seen practically. We emphasize that DDI is a typical example of this situation. In fact, the AUPR performance gap between and CentSmoothie reached around 1%, 5% and 12% in TWOSIDES, CADDDI and JADERDDI, respectively. The performance gap in JADERDDI is especially sizable. This might be caused by the high sparsity of JADERDDI (see Table 1).
Table 2.

Comparison of performances of the methods on the real DDI datasets

MethodTWOSIDES
CADDDI
JADERDDI
AUCAUPRAUCAUPRAUCAUPR
MRGNN0.8452 ± 0.00360.8029 ± 0.00390.9226 ± 0.00150.7113 ± 0.00310.9049 ± 0.00090.3698 ± 0.0019
Decagon0.8639 ± 0.00290.8094 ± 0.00240.9132 ± 0.00140.6338 ± 0.00290.9099 ± 0.00120.4710 ± 0.0027
SpecConv0.8785 ± 0.00250.8256 ± 0.00220.8971 ± 0.00550.6640 ± 0.00140.8862 ± 0.00250.5162 ± 0.0047
HPNN0.9044 ± 0.00030.8410 ± 0.00070.9495 ± 0.00040.7020 ± 0.00180.9127 ± 0.00040.5198 ± 0.0016
SBM0.9337 ± 0.00020.8583 ± 0.00040.9588 ± 0.00060.8170 ± 0.00080.9428 ± 0.00060.5963 ± 0.0018
CentSmoothie0.9348 ± 0.00020.8749 ± 0.00130.9846 ± 0.00010.8230 ± 0.00190.9684 ± 0.00040.6044 ± 0.0025
SPARSEO 0.9511 ± 0.00020.8811 ± 0.00010.9824 ± 0.00090.8773 ± 0.00140.9692 ± 0.00070.7230 ± 0.0008
SPARSEL 0.9517 ± 0.00010.8815 ± 0.0002 0.9859 ±0.00070.8797 ± 0.00100.9694 ± 0.00110.7276 ± 0.0017
SPARSE 0.9524 ±0.0001 0.8820 ±0.00020.9837 ± 0.0010 0.8843 ±0.0012 0.9698 ±0.0008 0.7348 ±0.0018
Comparison of performances of the methods on the real DDI datasets These results suggest that the latent interaction assumption in SPARSE is more reasonable and suitable for DDI prediction than CentSmoothie and the other competing methods. Among SPARSE, and , SPARSE achieved the highest performance. Note that the performance gap between SPARSE and in AUPR became clearer for more sparse data: e.g. only around 0.1% for TWOSIDES, while the gap reached around 1% for CADDDI and JADERDDI. Hence, we can see that with more sparse data, the horseshoe prior had advantage over Laplace prior and also the case with no sparsity prior. Results—Unknown DDI prediction performance: We evaluated the predictive ability of unknown DDIs. That is, we first trained a model by using the whole TWOSIDES data (the largest dataset), then predicted the scores of unknown triples (drug–drug-side effect), and finally sorted the predicted triples in the descending order of the scores. We focused on the top 400 predictions of each method and checked the overlap with the DDIs stored in drugs.com (Drugs.com, 2021; Thelwall ), a commonly used web checker for DDIs. Table 3 shows the number of overlaps between the DDIs in drugs.com and the top 400 predictions. SPARSE found 98 overlapped DDIs with drugs.com, this number being the highest and followed by CentSmoothie with 71 and HPNN with 48.
Table 3.

Number of overlaps with DDIs in drugs.com for the top 400 predictions

Method#Overlaps
SPARSE98
CentSmoothie71
HPNN48
Number of overlaps with DDIs in drugs.com for the top 400 predictions

Case studies: interpretation of top 10 unknown predictions

SPARSE is an SBM with latent features for drugs, side effects, and interactions. In particular, the model has connections between latent drug features and latent interactions. Thus from the trained model, we can extract the drug features, which are most associated with each drug latent feature and further extract the drug features most associated with each latent interaction through the corresponding latent drug feature. This means that we can retrieve drug features of a DDI if we can connect the DDI with the latent interactions. Algorithm 1 shows the pseudocode of this procedure (with T = 20 in our cases). SPARSE is a sparse model, which allows only a limited number of latent interactions and eventually allows to extract only a limited number of drug features. This is a sizable advantage of SPARSE for understanding the biological/chemical background behind predicted DDIs.For case studies, we extracted drug features (such as protein/pathway names) of the top unknown DDI predictions by using SPARSE, which was trained by the entire TWOSIDES. Table 4 shows the top 10 predictions (out of the 400 predictions in the experiment of the previous section) with the observable features associated with latent drug features [fifth column from the right-hand side. In this column, ‘Not clear’ means that to our current understanding of the potential DDI mechanisms, we could not explain the corresponding low-level (molecular level) background, although our algorithm could find associated drug features], the target protein of the corresponding drug using DrugBank (sixth column) and the corresponding reference to each DDI (seventh column). The top predictions are likely to be similar to each other, since the similar triples are likely to have similar scores. In fact the top predictions in Table 4 have large overlaps, but from the table, we could find the following four points:
Table 4.

Top 10 new (unknown) predictions with potentially associated latent features of proteins and extracted proteins

No.Drug ADrug BSide effectObservable features associated with latent featuresExtracted proteins of drugs from DrugBankReferences
1CiprofloxacinMefenamic acidAbdominal distensionCytochrome enzymes Venkataraman et al. (2014)
2NaratriptanOxycodoneAbnormal ECGSerotonin transporters and receptors Baldo (2018) and Ritter et al. (2019)
3NaratriptanTramadolAbnormal ECGSerotonin transporters and receptors Baldo (2018)
4NaratriptanSertralineAbnormal ECGSerotonin transporters and receptors5-Hydroxytryptamine receptor 1B (and 1D) and sodium-dependent serotonin transporter Ritter et al. (2019)
5NaratriptanParoxetineAbnormal ECGSerotonin transporters and receptors5-Hydroxytryptamine receptor 1B (and 1D) and sodium-dependent serotonin transporter Ritter et al. (2019)
6TrihexyphenidylThiothixeneAbnormal EEGDopamine receptors Ritter et al. (2019)
7CarisoprodolOrphenadrineAbnormal visionNot clear Downs et al. (2019)
8BuspironeOrphenadrineAbnormal visionNot clear Ritter et al. (2019)
9OxycodoneOrphenadrineAbnormal visionNot clear Ritter et al. (2019)
10CarisoprodolZaleplonAbnormal visionNot clear Fagiolini et al. (2004)
Extracting potentially associated drug features Input: Learned parameters , drug features matrix , a predicted triple (u, v, t), hyperparameter T Outout: Associated drug features for the triple //Extract drug features for each latent feature fordo end for //Calculate non-zeros latent interactions. is the pairwise dot product, ⊗ is the outer product. //Extract potentially associated drug features for the triple fordo end for Return Re Top 10 new (unknown) predictions with potentially associated latent features of proteins and extracted proteins The fourth and fifth predictions show the cases, where SPARSE could specify target proteins precisely, confirming the high credibility of these predictions and more importantly, approving the high ability of SPARSE for detecting unknown DDIs. The first, second, third and sixth predictions show the cases, where SPARSE could identify possible interacting protein groups (fourth column), not necessarily directly associated with the drugs, indicating that SPARSE allows suggesting novel interactions as well as potential target proteins. The validity of the seventh, eighth, ninth and tenth predictions might be understood by high-level views, like the connection between vision and dizziness/sedation. This result implies that SPARSE can predict probable interactions, which however cannot be straightforwardly inferred from low-level data. Entirely, we could find relevant references for all top 10 predictions (Baldo, 2018; Fagiolini ; Rho ; Venkataraman ), giving plausibility of these prediction and at the same time an additional layer of evidence for the usefulness of SPARSE in practical settings. To facilitate medical research and confirmation of our findings by subsequent clinical or preclinical studies, we provide the potential mechanisms as a Supplementary Material for our top predictions. Also, we discuss below the main biological mechanism for a predicted top 10 interaction: Naratriptan, Sertraline and abnormal ECG: Sertraline belongs to the selective serotonin reuptake inhibitor class antidepressants. Members of this class inhibit the reuptake of the neurotransmitter serotonin into cells (Ritter ). Through this inhibition, sertraline increases serotonin levels outside of the cells and allows serotonin to remain longer at its site of action. Naratriptan is known to cause heart-related side effects through serotonin receptor agonism at serotonin type 1 receptors (Dodick ; Ritter ). Therefore, the predicted side effect can be a direct consequence of sertraline increasing the level of endogenous serotonin and naratriptan acting at serotonin receptors in the heart, with the resulting changes visible in electrocardiogram recordings.

5 Conclusion and discussion

We have proposed SPARSE to learn the latent representations of drugs, side effects and interactions, through hypergraph neural networks. SPARSE addresses three important issues of state-of-the-art DDI prediction, which have not been addressed by any other methods. Extensive empirical validation using both synthetic and real data showed that SPARSE outperformed all current, cutting-edge methods for DDI prediction, verifying the effectiveness of multiple types of latent interaction assumptions and the sparsity control setting of SPARSE. Possible future work is to generalize SPARSE for higher-order drug interactions with multiple drugs. Another interesting direction might be to apply SPARSE to other sparse, high-dimensional data in bioinformatics.

Funding

This work was supported in part by Otsuka Toshimi Scholarship Foundation (to D.A.N.); MEXT KAKENHI [grant number 22K12150] (to C.H.N.); the Japan Society for the Promotion of Science (Postdoctoral Fellowships for Research in Japan, standard program) [P20809] (to P.P.); and MEXT KAKENHI [grant numbers: 19H04169, 20F20809, 21H05027 and 22H03645] and the AIPSE program of the Academy of Finland (to H.M.). Conflict of Interest: none declared. Click here for additional data file.
  16 in total

1.  Data-driven prediction of drug effects and interactions.

Authors:  Nicholas P Tatonetti; Patrick P Ye; Roxana Daneshjou; Russ B Altman
Journal:  Sci Transl Med       Date:  2012-03-14       Impact factor: 17.956

2.  Barbiturate-like actions of the propanediol dicarbamates felbamate and meprobamate.

Authors:  J M Rho; S D Donevan; M A Rogawski
Journal:  J Pharmacol Exp Ther       Date:  1997-03       Impact factor: 4.030

3.  Trihexyphenidyl rescues the deficit in dopamine neurotransmission in a mouse model of DYT1 dystonia.

Authors:  Anthony M Downs; Xueliang Fan; Christine Donsante; H A Jinnah; Ellen J Hess
Journal:  Neurobiol Dis       Date:  2019-01-30       Impact factor: 5.996

Review 4.  Cardiovascular tolerability and safety of triptans: a review of clinical data.

Authors:  David W Dodick; Vincent T Martin; Timothy Smith; Stephen Silberstein
Journal:  Headache       Date:  2004-05       Impact factor: 5.887

5.  Specific GABAA circuits for visual cortical plasticity.

Authors:  Michela Fagiolini; Jean-Marc Fritschy; Karin Löw; Hanns Möhler; Uwe Rudolph; Takao K Hensch
Journal:  Science       Date:  2004-03-12       Impact factor: 47.728

6.  Heterogeneous Hypergraph Variational Autoencoder for Link Prediction.

Authors:  Haoyi Fan; Fengbin Zhang; Yuxuan Wei; Zuoyong Li; Changqing Zou; Yue Gao; Qionghai Dai
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2022-07-01       Impact factor: 6.226

7.  The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets.

Authors:  Takaya Saito; Marc Rehmsmeier
Journal:  PLoS One       Date:  2015-03-04       Impact factor: 3.240

8.  Predicting potential drug-drug interactions on topological and semantic similarity features using statistical learning.

Authors:  Andrej Kastrin; Polonca Ferk; Brane Leskošek
Journal:  PLoS One       Date:  2018-05-08       Impact factor: 3.240

9.  Detecting Potential Adverse Drug Reactions Using a Deep Neural Network Model.

Authors:  Chi-Shiang Wang; Pei-Ju Lin; Ching-Lan Cheng; Shu-Hua Tai; Yea-Huei Kao Yang; Jung-Hsien Chiang
Journal:  J Med Internet Res       Date:  2019-02-06       Impact factor: 5.428

10.  Dual graph convolutional neural network for predicting chemical networks.

Authors:  Shonosuke Harada; Hirotaka Akita; Masashi Tsubaki; Yukino Baba; Ichigaku Takigawa; Yoshihiro Yamanishi; Hisashi Kashima
Journal:  BMC Bioinformatics       Date:  2020-04-23       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.