| Literature DB >> 35678684 |
Lei Deng1, Yunyun Zeng1, Hui Liu2, Zixuan Liu3, Xuejun Liu2.
Abstract
Drug-target interactions provide insight into the drug-side effects and drug repositioning. However, wet-lab biochemical experiments are time-consuming and labor-intensive, and are insufficient to meet the pressing demand for drug research and development. With the rapid advancement of deep learning, computational methods are increasingly applied to screen drug-target interactions. Many methods consider this problem as a binary classification task (binding or not), but ignore the quantitative binding affinity. In this paper, we propose a new end-to-end deep learning method called DeepMHADTA, which uses the multi-head self-attention mechanism in a deep residual network to predict drug-target binding affinity. On two benchmark datasets, our method outperformed several current state-of-the-art methods in terms of multiple performance measures, including mean square error (MSE), consistency index (CI), rm2, and PR curve area (AUPR). The results demonstrated that our method achieved better performance in predicting the drug-target binding affinity.Entities:
Keywords: binding affinity; convolutional neural network; multi-head self attention mechanism; residual network; word embedding
Year: 2022 PMID: 35678684 PMCID: PMC9164023 DOI: 10.3390/cimb44050155
Source DB: PubMed Journal: Curr Issues Mol Biol ISSN: 1467-3037 Impact factor: 2.976
Hyperparameter optimization and their tuned values.
| Hyperparameters | Value |
|---|---|
| Batch_size | 32 |
| Embedding_size | 128 |
| Filter length (Protein) | 12 |
| Filter length (Drug) | 4 |
| Number of filters | [32;64;96] |
| num_head | 8 |
| num_block | 2 |
| Learning_rate |
|
| Hidden neurous | [2048;1024;512;256] |
| epcho | 500 |
Comparison of our method with six competitive methods on Davis dataset.
| Method | CI |
| AUPR | MSE |
|---|---|---|---|---|
| KronRLS | 0.871 | 0.407 | 0.661 | 0.379 |
| SimBoost | 0.872 | 0.644 | 0.709 | 0.282 |
| DeepCPI | 0.867 | 0.607 | 0.705 | 0.293 |
| DeepDTA | 0.878 | 0.630 | 0.714 | 0.261 |
| GANsDTA | 0.881 | 0.653 | 0.691 | 0.276 |
| DeepGS | 0.882 | 0.686 | 0.763 | 0.252 |
| DeepMHADTA1 | 0.871 | 0.663 | 0.734 | 0.279 |
| DeepMHADTA2 |
|
|
|
|
Comparison of our method with six competitve methods on KIBA dataset.
| Method | CI |
| AUPR | MSE |
|---|---|---|---|---|
| KronRLS | 0.782 | 0.342 | 0.635 | 0.411 |
| SimBoost | 0.836 | 0.629 | 0.760 | 0.222 |
| DeepCPI | 0.852 | 0.657 | 0.782 | 0.211 |
| DeepDTA | 0.863 | 0.673 | 0.788 | 0.194 |
| GANsDTA | 0.866 | 0.675 | 0.753 | 0.224 |
| DeepGS | 0.860 | 0.684 | 0.801 | 0.193 |
| DeepMHADTA1 | 0.873 | 0.704 | 0.799 | 0.195 |
| DeepMHADTA2 |
|
|
|
|
Figure 1Scatter plots between the predicted and measured binding affinity values.
Performance of DeepMHADTA and the ablation models using structures or sequence alone on Davis dataset.
| Models | CI |
| AUPR | MSE |
|---|---|---|---|---|
| Without Structures | 0.893 | 0.699 | 0.754 | 0.253 |
| Without Sequence | 0.853 | 0.588 | 0.698 | 0.358 |
| DeepMHADTA |
|
|
|
|
Performance of DeepMHADTA with the Ablation model using structures or sequence alone on KIBA dataset.
| Models | CI |
| AUPR | MSE |
|---|---|---|---|---|
| Without Structures | 0.863 | 0.674 | 0.796 | 0.207 |
| Without Sequence | 0.778 | 0.463 | 0.603 | 0.360 |
| DeepMHADTA |
|
|
|
|
Summary of the Davis and KIBA dataset.
| Dataset | Proteins | Compounds | Interactions |
|---|---|---|---|
| Davis | 442 | 68 | 30,056 |
| KIBA | 229 | 2111 | 118,254 |
Figure 2Architecture of our proposed model DeepMHADTA. The model combines sequence and structure information of protein and drug via CNN block for feature extraction, and the embeddings are concatenated as input to fully connected layer for quantitative binding affinity prediction.
Figure 3Data length distribution of Davis and KIBA data sets. (A) Represents the length distribution of SMILES. (B) The number in the middle represents the length distribution of Morgan fingerprints, and (C) represents the length of the distributed protein sequence.
Figure 4Illustrative diagram of the multi-head attention layers.