Literature DB >> 35922768

Multi-scaled self-attention for drug-target interaction prediction based on multi-granularity representation.

Yuni Zeng1, Xiangru Chen2, Dezhong Peng2,3,4, Lijun Zhang5,6, Haixiao Huang7.   

Abstract

BACKGROUND: Drug-target interaction (DTI) prediction plays a crucial role in drug discovery. Although the advanced deep learning has shown promising results in predicting DTIs, it still needs improvements in two aspects: (1) encoding method, in which the existing encoding method, character encoding, overlooks chemical textual information of atoms with multiple characters and chemical functional groups; as well as (2) the architecture of deep model, which should focus on multiple chemical patterns in drug and target representations.
RESULTS: In this paper, we propose a multi-granularity multi-scaled self-attention (SAN) model by alleviating the above problems. Specifically, in process of encoding, we investigate a segmentation method for drug and protein sequences and then label the segmented groups as the multi-granularity representations. Moreover, in order to enhance the various local patterns in these multi-granularity representations, a multi-scaled SAN is built and exploited to generate deep representations of drugs and targets. Finally, our proposed model predicts DTIs based on the fusion of these deep representations. Our proposed model is evaluated on two benchmark datasets, KIBA and Davis. The experimental results reveal that our proposed model yields better prediction accuracy than strong baseline models.
CONCLUSION: Our proposed multi-granularity encoding method and multi-scaled SAN model improve DTI prediction by encoding the chemical textual information of drugs and targets and extracting their various local patterns, respectively.
© 2022. The Author(s).

Entities:  

Keywords:  Deep learning; Drug–target interaction; Representations learning; Self-attention networks

Mesh:

Substances:

Year:  2022        PMID: 35922768      PMCID: PMC9347097          DOI: 10.1186/s12859-022-04857-x

Source DB:  PubMed          Journal:  BMC Bioinformatics        ISSN: 1471-2105            Impact factor:   3.307


Background

Drug–target interaction (DTI) indicates the binding of drug compounds to their targets. The targets refer to the proteins or some bio-molecules to which the drug directly binds, and which are responsible for the therapeutic efficacy of the drug in vivo [1]. The drugs exert their clinical effects in treating diseases by changing the structure of the targets or regulating their metabolism. Therefore, accurate identification of DTI is one crucial step of drug discovery and development [1-3]. For example, in process of drug repositioning [4] task, DTI prediction is regarded as the foundation to find new targets of existing drugs. Nowadays, due to the high-cost and time-consuming traditional biological experiments, effective computational methods are urgently needed [5-7]. In response to this demand, many DTI prediction methods have been proposed in recent years. These methods mainly includes two parts: encoding methods and DTI prediction methods. As for the encoding methods, most studies for DTI prediction label their inputs by a character-based dictionary. For example, in DeepDTA [6], with a dictionary like {‘C’:1,‘H’:2,‘N’:3,,‘=’:63}, the drug simplified molecular input line entry system (SMILES) sequence ‘CN=C=O’ was labelled as [1 3 63 1 63 5]. It labelled each character of drug SMILES by its corresponding integer in the character-based dictionary. In addition, in other chemical compounds related fields, some works applied tokenization methods to extract substrings from drug sequences as their functional groups at the chemical level. Study [8] tokenized the names of chemical compounds by the open parser for systematic IUPAC nomenclature (OPSIN) tokenizer [9] and byte-pair-encoding (BPE) [10] in predicting chemical compounds task. Based on BPE, study [11] introduced a tokenization algorithm named SMILES pair encoding (SPE) to label the SMILES by the learned chemical groups. It has been applied to generative and predictive tasks and molecular tasks. Study [12] proposed a ChemBoost approach to predict protein-ligand binding affinity scores based on substrings extracted by Word2vec [13] and BPE. In these studies, tokenizer methods in the fields of natural language processing (NLP) were used for drug SMILES segmentation, and then the segmented SMILES were applied to compound-related tasks. For DTI prediction methods,many efforts have been conducted to predict drug–target binding affinity scores in recent years. The traditional approach to DTI prediction mainly based on similarity [14, 15]. Study [16] used the 2D compound similarity of drugs and Smith-Waterman similarity of targets as the inputs. Then, the Kronecker regularized least squares (KronRLS) algorithm was employed to predict the binding affinity values of drug-tart pairs. Study [17] also utilized a number of similarity-based information and features to predict DTI by a gradient boosting machine. DTINet [18] was based on the assumption that similar drugs may share similar targets. Taking a series of similar matrices as input, it was designed to find an optimal projection from drug space onto target space by the random walk with restart (RWR) algorithm. With the significant success of deep learning in computer version, speech recognition and NLP, deep learning models are widely used in DTI prediction. DeepDTA [6] employed two convolutional neural network (CNN) models to extract features for deep representations of drugs and targets. Then, an fully connected network was utilized to predict the interaction of drug and protein representations. OnionNet [19] also utilized CNNs for drug and protein representations and so as to predict the binding affinity values. GANsDTA [20] used the generative adversarial networks (GANs) to learn deep representations for drugs and targets, and then predicted the binding affinity scores of drug–target pairs. DeepCDA [21] also was proposed for binding affinity score prediction. It employed two CNNs to extract feature of drug and target. Then, long-short-term memory (LSTM) layers and a two-side attention mechanism were used in interaction learning to predict DTIs. Moreover, self-attention networks (SANs) also were applied to generate deep representations of drugs and targets [22-24]. Especially, study [23] proved that SANs have the ability to capture the long-distance relation between atoms in drug and target sequences. Despite these efforts, the existing methods have several areas for improvement:In order to address the above problems, we introduce a new multi-scaled SAN model for drug–target binding affinity prediction based on multi-granularity representations in this work. Taking protein sequences and drug SMILES sequences as inputs, we first introduce a multi-granularity encoding method for them. The multi-granularity encoding is built upon the BPE algorithm which is a widely used tokenization algorithm in field of NLP. BPE calculates the frequency of occurrence of each consecutive byte pair, and then forms a vocabulary from high-frequency byte pairs. The multi-granularity representations are labelled by the vocabulary and then transmitted as inputs to our proposed multi-scale SAN model. By assigning different window sizes to heads in SAN, the multi-scaled SAN is exploited to learn the multi-scaled local patterns and generate deep representations of drugs and targets. Finally, the prediction is made on fused deep representations. The existing encoding method labels molecular input character by character and it cannot encode fundamental chemical groups: (1) atoms with multiple characters in compounds, like ‘Br’, ‘Cl’, and (2) chemical functional groups, like ‘CC’, ‘OH’. These chemical groups are the determining part of chemical compounds and protein sequences. Therefore, the existing encoding method leads to the loss of essential chemical information. The existing deep models do not fully model different chemical correlations between atoms and atoms, atoms and chemical groups, chemical groups and chemical groups. Although CNNs can capture local features of these correlations, they failed to model long-distant atoms [23]. Besides, SANS focus on the overall input sentence, but they may overlook fine-grained information in drug and target sequences [25]. Thus, the existing deep model for DTI prediction need to improve. To the end, we evaluate the effectiveness of our proposed model on benchmark datasets (Davis [26] and KIBA [27]). Experimental results demonstrate that our multi-granularity multi-scaled model yields better accuracy over baselines and existing DTI deep models. Moreover, the experiment analyses reveal that both the multi-granularity encoding and multi-scaled features extracted by our multi-scaled SANs are beneficial to DTI prediction.

Methods

In this work, we propose a multi-granularity multi-scaled method for DTI prediction, as shown in Fig. 1. The proposed method includes four components: multi-granularity encoding, drug representation learning, protein representation learning, and the interaction learning part. Firstly, we introduce a multi-granularity encoding method for drug and protein input sequences. In this process, the input sequences are encoded by a multi-granularity vocabulary, which are generated by a segmentation method. Then, taken the multi-granularity representations as inputs, a multi-scaled SAN is proposed to extract and fuse multi-scaled local features. Finally, the prediction is made on fused deep drug representations and deep protein representations by fully connected feed-forward networks.
Fig. 1

Our proposed multi-granularity multi-scaled SAN model for DTI prediction

Our proposed multi-granularity multi-scaled SAN model for DTI prediction

Multi-granularity encoding

The current labeling method is not sufficient to encode chemical sequences since it ignores the chemical textual information from chemical groups in drugs and proteins, for example, chemical functional groups ‘[C@@H]’,‘Br’. Thus, the intuitive way for representing a chemical sequence is to find out the substrings in the sequence by a computational method. Here, the substring is the chemical functional groups or atoms with multiple characters. BPE [10] is a data compression method that can obtain high-frequency substrings to segment the sequence. In the field of NLP, BPE is widely used in different text tasks and as the first step to understand text sentences. BPE initializes the symbol vocabulary with the character vocabulary, and then it iteratively counts the frequency of adjacent character pairs in the corpus and merges the pair with the highest frequency to a new symbol. Finally, the vocabulary update is stopped when the number of merge operations reaches a threshold. In this work, we utilize the BPE algorithm to generate vocabularies for encoding molecular inputs (SMILES or proteins). First, the segmentation datasets of drugs and targets are built and used to train BPE. Then, the BPE model trained by drug data would generate a vocabulary with a threshold for drugs, and and for targets. T determines the size of the generated vocabulary which consists of the segmented inputs by BPE. For example, taken the ‘COC1=C(C=C2C(= C1)N=CN=C2NC3=C(C(=CC=C3)Cl)F)CN4CCCC[C@@H]4C(=O)N’ as the input, the segmented outputs of BPE is shown in Table 1 with different T.
Table 1

Examples of segmented outputs of ’COC1=C(C=C2C(=C1)N=CN=C2NC3 =C(C(=CC=C3)Cl)F)CN4CCCC[C@@H]4C(=O)N’ with different T

TSegmented SMILES Sequence (Vocabulary)
1kCOC1=C(C=C, 2C(=C1), N=C, N=C2, N, C3=C, (C(=CC=C,
3)C, l), F)C, N, 4, CCCC, [C@@H]4, C(=O)N
5kCO, C1=C, (C=C, 2, C(=C1), N=C, N=C, 2, N, C3=C, (,
C(=CC=C, 3)C, l), F, )C, N, 4, CCCC, [C@@H], 4, C(=O), N
25kC, O, C1=C, (C=C, 2, C(=C, 1), N=C, N=C, 2, N, C3,
=C, (, C(, =CC=C, 3, )C, l, ), F, )C, N, 4, CCCC,
[C@@H], 4, C(=O), N
Examples of segmented outputs of ’COC1=C(C=C2C(=C1)N=CN=C2NC3 =C(C(=CC=C3)Cl)F)CN4CCCC[C@@H]4C(=O)N’ with different T Finally, a multi-granularity dictionary is constructed by assigning each group in the vocabulary a corresponding integer like the character-level dictionary in study [6]. Thus, an input sequence is labelled as multi-granularity representation where and the length of X is varied, which depends on the length of the input sequence.

Multi-scaled self-attention model for drug–target binding affinity prediction

Our multi-scaled SAN is built upon Transformer block [28] which has shown excellent capability on sequence processing tasks. Given a drug multi-granularity representation and protein multi-granularity representation , we first adopt an input embedding module to integrate multiple embeddings. Then, for drug embedding and protein embedding , two multi-scaled SAN blocks are exploited to capture the local patterns features of drugs and proteins, respectively. Finally, an interaction block is proposed to fuse and extract interaction features from deep drug representations and deep protein representations . The final prediction is the output of the interaction block.

Input embedding

Given a multi-granularity drug input asand a multi-granularity protein input aswe define a hyper-parameter l to restrict the max input length. Specially, restricts drug input and restricts target input . If the length of X is shorter than l, the lack value is setting as 0. According to Transformer [28] and MT-DTI [23], the input of multi-scaled SAN is the sum of token embedding of the input sequence and position embedding of the input sequence, that is calculated as:Here, the token embedding has a trainable weight . The is the vocabulary size of drugs and is the embedding length of drugs. The position embedding has a trainable weight . As for protein embedding,where is the token embedding of , is the position embedding of and is the embedding size of protein sequence.

Multi-scaled self-attention block

Multi-head SAN is the main component of Transformer [28]. It performs multiple self-attention modules on input expressions, then jointly pay attention to the information of different expression at different position. In this work, in order to generate a more informative deep representations of drugs and proteins, we adopt multi-scaled SAN to their embedings, which assign different window size to heads in multi-head SAN, that is formulated as,where MSSAN() denotes a multi-scaled self-attention block, as shown in Fig. 2. and are the hyper-parameters notating the number of multi-scaled SAN blocks.
Fig. 2

Our proposed multi-scaled SAN block

Our proposed multi-scaled SAN block Especially, suppose the input to multi-scaled SAN blocks is E. Our model first transforms input sequence into N subspace with different linear projections,where is the index and , the denotes the dimensionality of the head subspace. Then, we utilize a mask matrix for the head to achieve multi-scaled SAN. The output of head on multi-scaled SAN is calculated as,where is determined by a hyper-parameter named window size ,Then, the h heads are concatenated,where is a concatenation function. Next, a residual connection [29] and the layer normalization (LN()) [30] are employed,Thus, the output of a multi-scaled SAN block is formulated,where FFN(Z, 1) denotes one fully connected feed-forward layer (FCN) with ReLU activation [31] and Z as input. The hidden size of the FCN is .

Interaction block

The interaction block in this work is to combine deep drug and protein representations and predicts the binding affinity scores of drug–target pairs. Mathematically, firstly,Next, 4 layers of FCN are employed to capture the interaction information from R.where is the predicted binding affinity value of the drug–target pair.

Data and experiments

Datasets

Benchmark datasets for DIT prediction

We evaluated our proposed model on Davis [26] and KIBA [27] datasets because they are widely used in existing drug–target interaction studies. Specially, in order to ensure the uniqueness of drug input sequence, we only use Isomeric SMILES strings in this paper. The number of proteins, compounds and interactions of the Davis and KIBA dataset are summarised in Table 2. In particular, the Davis dataset contains the 442 kinase proteins, their relevant inhibitors (68 ligands) and their respective dissociation constant () value. The binding affinity scores of drug–target pairs were transformed into log space , as [6, 17], as follows,The used KIBA dataset comprised 229 proteins, 2111 drugs and their KIBA scores. Here, the KIBA scores measure the kinase inhibitor bioactivities as the binding affinity values in following experiments.
Table 2

The summary of the benchmark datasets

ProteinsCompoundsInteractionsTraining DataTest Data
Davis4426830056250465010
KIBA22921111182549854519709
The summary of the benchmark datasets

Segmentation dataset

We collect drug SMILES sequences from the National Center for Biotechnology Information (NCBI) 1 and protein sequences from The Universal Protein Resource2. Finally, 147546 SMILES sequences and 114500 protein sequences are collected as segmentation data to train the segmentation methods.

Experiment setup and metric

Table 3 summaries other hyper-parameter settings. We use five-time leave-one-out cross-validation to train our model and list the average results on test data. All models were trained on 1 NVIDIA 3080 GPU.
Table 3

Summary of parameter settings

ParameterKIBADavis
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$l_d$$\end{document}ld8036
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$l_p$$\end{document}lp800900
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$m^h$$\end{document}mh0,1,2,30,1,2,3
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$e_d$$\end{document}ed12864
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_d$$\end{document}Ld22
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$L_p$$\end{document}Lp21
Hidden size in FFN1024,1024,512,1
Epoch300300
Dropout0.10.1
OptimizerAdamAdam
Learning rate0.00010.0001
Summary of parameter settings To measure the performance of our model, three metrics are included: mean squared error (MSE), Concordance Index (CI) and the metric. MSE is the loss of the optimizer in the deep model.where the is the predicted binding affinity value, y is the ground-truth and n is the number of drug–target pairs. CI is the probability of the predicted scores of two randomly chosen drug–target pairs in the correct order, aswhere is the predicted value with larger affinity , is the prediction score for smaller affinity and N is a normalization constant. Moreover, the f(x) is a step function [16],Then metric [32, 33] is another widely used metric in this filed. Mathematically,where and are the squared correlation coefficient values between the observed and predicted values with and without intercept, respectively. Especially, the value of an acceptable model should be larger than 0.5.

Experiments 1: Effects of the segmentation method

In this paper, the BPE algorithm is utilized as the segmentation method to learn the substrings in drug SMILES and protein sequences. As seen in Table 1, the threshold T determines the degree of segmentation. The larger T in BPE indicates the more fine-grained and longer segmentation outputs. We first investigated the effects of T to DTI prediction on KIBA and Davis dataset. We extract various multi-granularity representations by setting different T, and then build DeepDTA [6] models with these representations as inputs. As plotted in Figs. 3 and 4, the prediction results on KIBA and Davis dataset are demonstrated, respectively.
Fig. 3

Results of DeepDTA [6] model on the KIBA dataset with different multi-granularity representations as inputs. These multi-granularity representations are encoded by BPE algorithm with different threshold T. Here, is the threshold T for drug segmentation and is the threshold T for protein segmentation

Fig. 4

Results of DeepDTA [6] model on the Davis dataset with different multi-granularity representations as inputs. These multi-granularity representations are encoded by BPE algorithm with different threshold T. Here, is the threshold T for drug segmentation and is the threshold T for protein segmentation

Results of DeepDTA [6] model on the KIBA dataset with different multi-granularity representations as inputs. These multi-granularity representations are encoded by BPE algorithm with different threshold T. Here, is the threshold T for drug segmentation and is the threshold T for protein segmentation Results of DeepDTA [6] model on the Davis dataset with different multi-granularity representations as inputs. These multi-granularity representations are encoded by BPE algorithm with different threshold T. Here, is the threshold T for drug segmentation and is the threshold T for protein segmentation Discussion: For both KIBA and Davis dataset, the and is superior to other settings. It is clear that when and , the prediction quality goes up as T increases. Conversely when and , the increase of T seems to cause performance degradation. One possible reason is that the segmented SMILE with and the segmented protein sequences with include more chemical textual information for predicting DTI. As the result, and in following experiments.

Experiments 2: Encoding methods for DTI prediction

The starting point of our approach is an observation in encoding methods. Considering the improvements of existing character-based encoding methods, we adopt segmentation method to learn the chemical groups in drug and target sequences. Thus, in this subsection, we evaluate whether deep representations learned from multi-granularity representations contains more drug–target interaction information than deep representations learned from character encoded representations, We also implemented DeepDTA [6], as baseline, with multi-granularity representations and character encoded representations as inputs. Table 4 lists the average results of the drug–target binding affinity prediction on KIBA and Davis dataset.
Table 4

Results of DeepDTA (CNN model) on KIBA and Davis dataset with character-based and multi-granularity encoding. Especially, the character-based encoding methods is original labelling method in DeepDTA [6]

Encoding MethodCIMSE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r^2_m$$\end{document}rm2
KIBACharacter Encoding0.863 (0.002)0.1940.673 (0.009)
Multi-Granularity0.875 (0.001)0.1850.696 (0.017)
DavisCharacter Encoding0.878 (0.004)0.2610.630 (0.017)
Multi-Granularity0.884 (0.005)0.2500.655 (0.019)

Bold values indicate the best results on the datasets

Discussion: As seen, the multi-granularity encoding method improves the prediction quality in both two datasets, reconfirming the necessity of encoding the chemical groups in drug and protein sequences. Results of DeepDTA (CNN model) on KIBA and Davis dataset with character-based and multi-granularity encoding. Especially, the character-based encoding methods is original labelling method in DeepDTA [6] Bold values indicate the best results on the datasets

Experiments 3: Multi-scaled SAN for DTI prediction

In this section, we conducted experiments about deep models based on multi-granularity encoding. Table 5 gives the average test results on the drug–target binding affinity prediction tasks. One intuition of our work is to capture the local patterns in multi-granularity representations by multi-scaled SANs. To evaluate it, we implemented models with CNNs from DeepDTA [6], SANs from Transformer [28] which also employed in MT-DTI [23] and our multi-scaled SAN.
Table 5

Results of deep models on KIBA and Davis dataset with multi-granularity representations as inputs

Deep modelsCIMSE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r^2_m$$\end{document}rm2
KIBACNNs0.863 (0.002)0.1940.673 (0.009)
SANs0.875 (0.003)0.1790.691 (0.019)
Multi-scaled SANs0.890 (0.002)0.1550.742 (0.010)
DavisCNNs0.878 (0.004)0.2610.630 (0.017)
SANs0.888 (0.004)0.2320.689 (0.007)
Multi-scaled SANs0.890 (0.005)0.2330.681(0.014)

Bold values indicate the best results on the datasets

Discussion: As shown in Table 5, the multi-scaled SAN outperforms the SANs model, indicating that the local pattern information can raise the ability of SANs to capture the drug–target interaction information. Moreover, as all known, CNNs have the ability to capture the local features. According to Table 5, the multi-scaled model achieved higher results than CNNs model, revealing extracting local features by the dynamic weights of multi-scaled SANs is superior to fixed weight from CNNs. Results of deep models on KIBA and Davis dataset with multi-granularity representations as inputs Bold values indicate the best results on the datasets

Experiments 4: Comparison to existing approaches

In the end, we compare our multi-granularity multi-scaled SANs model to traditional methods, like KronRLS [16], SimBoost [17], and other recent deep sequence representation methods, like DeepDTA [6], MT-DTI [23], GANsDTA [20], CrossAttentionDTI [24]. Table 6 lists the results of these models on drug–target binding affinity prediction task.
Table 6

Results on KIBA and Davis of our proposed multi-granularity multi-scaled SANs model, transitional methods and existing deep sequence representation methods

MethodDrugProteinInteractionCIMSE\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r^2_m$$\end{document}rm2
KIBAKronRLS [16]Pubchem SimS-W0.782 (0.001)0.4110.342 (0.001)
SimBoost [17]Pubchem SimS-W0.836 (0.001)0.2220.629 (0.007)
DeepDTA [6]CNNsCNNsConcatenation0.863 (0.002)0.1940.673 (0.009)
MT-DTI [23]SANsSANsConcatenation0.882 (0.002)0.1520.738 (0.006)
GANsDTA [20]GANsGANsConcatenation0.866 (−)0.2240.675 (−)
CrossAttentionDTI [24]Cross SANsCross SANsConcatenation0.874 (0.001)0.175
OursMSSANMSSANConcatenation0.890 (0.002)0.155 0.742(0.010)
DavisKronRLS [16]Pubchem SimS-W0.871 (0.001)0.3790.407 (0.005)
SimBoost [17]Pubchem SimS-W0.872 (0.001)0.2820.644 (0.006)
DeepDTA [6]CNNsCNNsConcatenation0.878 (0.004)0.2610.630 (0.017)
MT-DTI [23]SANsSANsConcatenation0.887 (0.003)0.2450.665 (0.014)
GANsDTA [20]GANsGANsConcatenation0.881 (−)0.2760.653 (−)
CrossAttentionDTI [24]Cross SANsCross SANsConcatenation0.876 (0.006)0.244
OursMSSANMSSANConcatenation 0.890 (0.005)0.233 0.681 (0.014)

Bold values indicate the best results on the datasets

Results on KIBA and Davis of our proposed multi-granularity multi-scaled SANs model, transitional methods and existing deep sequence representation methods Bold values indicate the best results on the datasets Discussion: As seen, these sequence-based deep models improve prediction quality than transitional methods, reconfirming the effectiveness of modeling sequence information. Besides, our proposed model improves CI to 0.890 on both KIBA and Davis dataset, and improve to 0.742 and 0.681 on KIBA and Davis dataset, respectively. Thus, our model outperforms the recent sequence-based works, indicating the superiority of the proposed approaches.

Discussion

DTI prediction is to identify the interactions between drugs and targets, which is a substantial task in the drug discovery field. Many studies proposed computation methods to reduce dependence on time, cost and traditional biological experiments. Based on these related works, we proposed a deep model for DTI prediction based on the multi-granularity encoding and the multi-scaled SAN model in this work. The main contribution of this paper can be summarized as follows.Via in-depth analyses, our work may contribute to subsequent researches on this topic: (1) the multiple encoding methods of SMILES sequence and protein sequence in DTI prediction as well as other bioinformatics tasks, (2) the learning method for local patterns in sequence, and (3) the representation learning of drug and target sequences. In order to encode fundamental chemical groups, a multi-granularity encoding method is introduced to label the molecular inputs of drugs and targets as the corresponding multi-granularity representations (Section Method). In order to model the multiple kinds of chemical correlations, a multi-scaled SAN model is proposed to learn the local patterns in drugs and targets by the dynamic weights (Section Method). Our proposed method achieves higher results on KIBA and DAVIS datasets, compared to traditional methods and recent deep sequence representation methods (Section Experiments).

Conclusion

In this paper, we investigate and propose effective approaches to improve drug–target binding affinity prediction from both encoding method and model architecture perspectives. As for the encoding method, we employ the BPE algorithm and segmentation dataset to train a multi-granularity encoding method for drug SMILES and protein sequences. It contributes to encode atoms with multiple characters and chemical functional groups. Secondly, we build a multi-scaled SAN model for their multi-granularity representations by assigning various window size to heads in original SANs. Experimental results demonstrate that the proposed approach not only is of benefit to predict DTI but also makes DTIs prediction surpass baselines on various metrics. Our proposed method achieves the improvements by benefiting from the encoding method for chemical groups and the local patterns modeled by the representation learning model. In the encoding process, we collected a large of unlabeled data of drugs and targets to train the encoding method. Meanwhile, we found the lack of labeled data limits the improvements of deep models to predict new DTIs. Thus, our future work may focus on the utilization of these unlabeled data, like the unsupervised learning method for DTI learning.
  22 in total

1.  Comprehensive analysis of kinase inhibitor selectivity.

Authors:  Mindy I Davis; Jeremy P Hunt; Sanna Herrgard; Pietro Ciceri; Lisa M Wodicka; Gabriel Pallares; Michael Hocker; Daniel K Treiber; Patrick P Zarrinkar
Journal:  Nat Biotechnol       Date:  2011-10-30       Impact factor: 54.908

2.  Chemical name to structure: OPSIN, an open source solution.

Authors:  Daniel M Lowe; Peter T Corbett; Peter Murray-Rust; Robert C Glen
Journal:  J Chem Inf Model       Date:  2011-03-09       Impact factor: 4.956

3.  Some case studies on application of "r(m)2" metrics for judging quality of quantitative structure-activity relationship predictions: emphasis on scaling of response data.

Authors:  Kunal Roy; Pratim Chakraborty; Indrani Mitra; Probir Kumar Ojha; Supratik Kar; Rudra Narayan Das
Journal:  J Comput Chem       Date:  2013-01-08       Impact factor: 3.376

4.  DeepCDA: deep cross-domain compound-protein affinity prediction through LSTM and convolutional neural networks.

Authors:  Karim Abbasi; Parvin Razzaghi; Antti Poso; Massoud Amanlou; Jahan B Ghasemi; Ali Masoudi-Nejad
Journal:  Bioinformatics       Date:  2020-11-01       Impact factor: 6.937

5.  A network integration approach for drug-target interaction prediction and computational drug repositioning from heterogeneous information.

Authors:  Yunan Luo; Xinbin Zhao; Jingtian Zhou; Jinglin Yang; Yanqing Zhang; Wenhua Kuang; Jian Peng; Ligong Chen; Jianyang Zeng
Journal:  Nat Commun       Date:  2017-09-18       Impact factor: 14.919

6.  SimBoost: a read-across approach for predicting drug-target binding affinities using gradient boosting machines.

Authors:  Tong He; Marten Heidemeyer; Fuqiang Ban; Artem Cherkasov; Martin Ester
Journal:  J Cheminform       Date:  2017-04-18       Impact factor: 5.514

Review 7.  Machine learning approaches and databases for prediction of drug-target interaction: a survey paper.

Authors:  Maryam Bagherian; Elyas Sabeti; Kai Wang; Maureen A Sartor; Zaneta Nikolovska-Coleska; Kayvan Najarian
Journal:  Brief Bioinform       Date:  2021-01-18       Impact factor: 11.622

8.  ChemBoost: A Chemical Language Based Approach for Protein - Ligand Binding Affinity Prediction.

Authors:  Rıza Özçelik; Hakime Öztürk; Arzucan Özgür; Elif Ozkirimli
Journal:  Mol Inform       Date:  2020-12-14       Impact factor: 3.353

9.  On two novel parameters for validation of predictive QSAR models.

Authors:  Partha Pratim Roy; Somnath Paul; Indrani Mitra; Kunal Roy
Journal:  Molecules       Date:  2009-04-29       Impact factor: 4.411

10.  Drug-target interaction prediction via multiple classification strategies.

Authors:  Qing Ye; Xiaolong Zhang; Xiaoli Lin
Journal:  BMC Bioinformatics       Date:  2022-01-20       Impact factor: 3.169

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.