| Literature DB >> 32063849 |
Yiwei Wang1,2, Lei Huang3,4, Siwen Jiang3, Yifei Wang1, Jun Zou1, Hongguang Fu3, Shengyong Yang1.
Abstract
Capsule networks (CapsNets), a new class of deep neural network architectures proposed recently by Hinton et al., have shown a great performance in many fields, particularly in image recognition and natural language processing. However, CapsNets have not yet been applied to drug discovery-related studies. As the first attempt, we in this investigation adopted CapsNets to develop classification models of hERG blockers/nonblockers; drugs with hERG blockade activity are thought to have a potential risk of cardiotoxicity. Two capsule network architectures were established: convolution-capsule network (Conv-CapsNet) and restricted Boltzmann machine-capsule networks (RBM-CapsNet), in which convolution and a restricted Boltzmann machine (RBM) were used as feature extractors, respectively. Two prediction models of hERG blockers/nonblockers were then developed by Conv-CapsNet and RBM-CapsNet with the Doddareddy's training set composed of 2,389 compounds. The established models showed excellent performance in an independent test set comprising 255 compounds, with prediction accuracies of 91.8 and 92.2% for Conv-CapsNet and RBM-CapsNet models, respectively. Various comparisons were also made between our models and those developed by other machine learning methods including deep belief network (DBN), convolutional neural network (CNN), multilayer perceptron (MLP), support vector machine (SVM), k-nearest neighbors (kNN), logistic regression (LR), and LightGBM, and with different training sets. All the results showed that the models by Conv-CapsNet and RBM-CapsNet are among the best classification models. Overall, the excellent performance of capsule networks achieved in this investigation highlights their potential in drug discovery-related studies.Entities:
Keywords: Capsule network; classification model; convolution-capsule network; deep learning; hERG; restricted Boltzmann machine-capsule networks
Year: 2020 PMID: 32063849 PMCID: PMC6997788 DOI: 10.3389/fphar.2019.01631
Source DB: PubMed Journal: Front Pharmacol ISSN: 1663-9812 Impact factor: 5.810
Figure 1Architecture of convolution-capsule networks (Conv-CapsNet). The input is one-dimensional vector containing 179 components. The convolution layer has 32 filters of size 1×3. The hidden feature layer and PrimaryCaps layer consist of 128 and 64 nodes, respectively. The weight matrix between PrimaryCaps layer and DigitCaps layer is 8×8×2×2, and two dynamic routing iterations were adopted.
Hyperparameter settings of convolution-capsule networks (Conv-CapsNet).
| Hyperparameter | Setting |
|---|---|
| L2 normalization term | 0.001 |
| Activation | Relu |
| Batch size | 148 |
| Iteration epoch | 300 |
| Learning rate of network | 0.001 |
| Optimizer | Adam |
| Filter | 32 |
| Kernel_size | 3 |
| Number of nodes in the hidden feature layer | 128 |
| Number of nodes in the PrimaryCaps layer | 64 |
| Routing time | 2 |
| Dimension of each capsule | 8 |
| Length of PrimaryCaps | 2 |
| Length of DigitCaps | 2 |
Algorithm and training procedure of convolution-capsule networks (Conv-CapsNet).
|
|
|
|
| 1: |
| 15: |
Figure 2Architecture of restricted Boltzmann machine-capsule networks (RBM-CapsNet). The input is one-dimensional vector containing 179 components. The hidden feature layer and PrimaryCaps layer consist of 256 and 128 nodes, respectively. The weight matrix between PrimaryCaps layer and DigitCaps layer is 8×8×2×2, and two dynamic routing iterations were adopted.
Hyperparameter settings of restricted Boltzmann machine-capsule networks (RBM-CapsNet).
| Hyperparameter | Setting |
|---|---|
| Numbers of RBM | 2 |
| Number of nodes in the hidden feature layer | 256 |
| Number of nodes in the PrimaryCaps layer | 128 |
| Iteration of RBM | 100 |
| Iteration of network | 200 |
| Learning rate of RBM | 0.001 |
| Learning rate of network | 0.005 |
| Activation | Relu |
| Batch size | 148 |
| Optimizer | Adam |
| Routing time | 2 |
| Dimension of each capsule | 8 |
| Length of PrimaryCaps | 2 |
| Length of DigitCaps | 2 |
Algorithm and training procedure of restricted Boltzmann machine-capsule networks (RBM-CapsNet).
|
|
|
|
| 1: |
| 7: |
| 19: |
Prediction results of hERG blockers/nonblockers classification models developed by capsule networks with different architectures.
| Capsule network architecture | SE | SP | MCC | SD | Q (%) |
|---|---|---|---|---|---|
| Original CapsNet | 80.4% | 86.7% | 0.673 | 0.0141 | 84.1% |
| FC+FC | 82.6% | 86.7% | 0.694 | 0.0195 | 85.0% |
| Conv+FC | 82.2% | 86.4% | 0.687 | 0.0166 | 84.6% |
| Conv+FC+FC |
|
|
|
|
|
| Conv+Conv+FC+FC | 84.5% | 85.3% | 0.693 | 0.0142 | 84.9% |
| Conv+Conv+Conv+FC+FC | 81.9% | 86.9% | 0.685 | 0.0173 | 84.9% |
| One RBM | 83.1% | 86.5% | 0.694 | 0.0182 | 84.9% |
| Two RBMs |
|
|
|
|
|
| Three RBMs | 84.5% | 85.5% | 0.696 | 0.0160 | 85.0% |
| Four RBMs | 81.2% | 86.0% | 0.673 | 0.0108 | 83.9% |
| Five RBMs | 84.1% | 86.4% | 0.701 | 0.0156 | 85.4% |
*Conv, convolutional operation; FC, fully connected operation; RBM, restricted Boltzmann machine; Conv-CapsNet, convolution-capsule network; RBM-CapsNet, restricted Boltzmann machine-capsule network (The training set used was the Doddareddy's training set, and five-fold cross-validation was used to monitor the training performance. SE (%), sensitivity; SP (%), specificity; MCC, Matthew's correlation coefficient; SD, standard deviation; Q (%), overall accuracy). Conv-CapsNet and Conv-CapsNet showed the best performance.
Prediction accuracies of hERG blockade classification models developed by different methods with the same Doddareddy's training set.
| Model | SE | SP | MCC | Q (%) | AUC |
|---|---|---|---|---|---|
| Doddareddy's test set (255/P:108, N:147) | |||||
| Conv-CapsNet | 94.4% | 89.8% | 0.835 | 91.8% | 0.940 |
| RBM-CapsNet | 91.7% | 92.5% | 0.840 | 92.2% | 0.944 |
| CNN | 87.0% | 85.0% | 0.715 | 85.9% | 0.933 |
| MLP | 82.4% | 86.4% | 0.687 | 84.7% | 0.920 |
| DBN | 72.2% | 80.8% | 0.533 | 80.8% | 0.903 |
| SVM | 90.7% | 84.4% | 0.743 | 87.1% | 0.933 |
| kNN | 69.4% | 96.6% | 0.703 | 85.1% | 0.928 |
| Logistic regression | 88.8% | 83.7% | 0.710 | 85.5% | 0.858 |
| LightGBM | 79.6% | 82.3% | 0.617 | 81.2% | 0.810 |
| Doddareddy's external validation (60/P:18, N:42) | |||||
| Conv-CapsNet | 88.9% | 71.4% | 0.554 | 76.7% | 0.806 |
| RBM-CapsNet | 94.4% | 71.4% | 0.604 | 78.7% | 0.811 |
| CNN | 94.4% | 52.4% | 0.441 | 65.0% | 0.725 |
| MLP | 88.9% | 57.1% | 0.426 | 66.7% | 0.707 |
| DBN | 88.9% | 52.4% | 0.386 | 63.3% | 0.683 |
| SVM | 88.9% | 52.4% | 0.386 | 63.3% | 0.660 |
| kNN | 77.8% | 52.4% | 0.279 | 60.0% | 0.624 |
| Logistic regression | 83.3% | 52.4% | 0.332 | 61.7% | 0.623 |
| LightGBM | 61.1% | 59.5% | 0.190 | 60.0% | 0.609 |
(TP, true positive; TN, true negative; FP, false positive; FN, false negative; SE (%), sensitivity, SE = TP/(TP + FN); SP (%), specificity, SP = TN/(TN + FP); Q (%), overall accuracy, Q = [TP + TN)/(TP + TN + FP + FN)].
Figure 3Receiver operating characteristic (ROC) curves for Doddareddy's test set by (A) convolution-capsule networks (Conv-CapsNet) and (B) restricted Boltzmann machine-capsule networks (RBM-CapsNet), respectively.
Prediction results of various hERG blockade classification models developed with training sets different from Doddareddy's training set.
| Entry | Model | Training set | Test set | SE | SP | Q |
|---|---|---|---|---|---|---|
| 1 | RP ( | Hou's training set 1 (P: 283; N: 109) | Hou's test set 1 (P: 129; N: 66) | 79.8% | 75.8% | 78.5% |
| NB ( | 82.2% | 75.8% | 80.0% | |||
| SVM ( | 90.7% | 65.2% | 82.1% | |||
| Conv-CapsNet | 85.7% | 78.8% | 82.0% | |||
| RBM-CapsNet | 84.1% | 80.3% | 82.0% | |||
| 2 | RP ( | Hou's training set 2 (P: 272; N: 120) | Hou's test set 2 (P: 140; N: 55) | 80.0% | 74.5% | 78.5% |
| NB ( | 81.4% | 80.0% | 81.0% | |||
| SVM ( | 85.0% | 74.5% | 82.1% | |||
| Conv-CapsNet | 82.1% | 81.8% | 82.0% | |||
| RBM-CapsNet | 81.4% | 83.6% | 82.0% | |||
| 3 | Bayesian ( | Hou's training set 3 (P: 314; N: 306) | Hou's test set 3 (P: 63; N: 57) | 86.9% | 83.1% | 85.0% |
| Conv-CapsNet | 87.3% | 86.0% | 86.8% | |||
| RBM-CapsNet | 88.9% | 84.2% | 86.8% | |||
| 4 | SVM ( | Zhang's training set (P: 717; N: 210) | Zhang's test set (P: 188; N: 48) | 95.8% | 34.0% | 83.5% |
| kNN ( | 92.6% | 40.4% | 82.2% | |||
| Conv-CapsNet | 88.8% | 66.7% | 84.5% | |||
| RBM-CapsNet | 90.4% | 64.6% | 85.2% | |||
| 5 | LibSVM ( | Sun's training set (P: 483; N: 2541) | Sun's test set (P: 53; N: 13) | 68.0% | 85.0% | 71.0% |
| RF ( | 72.0% | 85.0% | 74.0% | |||
| Conv-CapsNet | 83.0% | 84.6% | 83.3% | |||
| RBM-CapsNet | 86.8% | 84.6% | 86.3% | |||
| 6 | LibSVM ( | Siramshetty's training set T3 (P: 1406; N: 1708) | Doddareddy's test set (P: 108; N: 147) | 64.0% | 89.0% | 78.0% |
| RF ( | 68.0% | 91.0% | 81.0% | |||
| Conv-CapsNet | 85.2% | 88.4% | 87.1% | |||
| RBM-CapsNet | 83.3% | 91.2% | 87.8% |