| Literature DB >> 28923002 |
Haiou Li1, Jie Hou2, Badri Adhikari3, Qiang Lyu1, Jianlin Cheng4.
Abstract
BACKGROUND: Deep learning is one of the most powerful machine learning methods that has achieved the state-of-the-art performance in many domains. Since deep learning was introduced to the field of bioinformatics in 2012, it has achieved success in a number of areas such as protein residue-residue contact prediction, secondary structure prediction, and fold recognition. In this work, we developed deep learning methods to improve the prediction of torsion (dihedral) angles of proteins.Entities:
Keywords: Deep learning; Deep recurrent neural network; Protein torsion angle prediction; Restricted Boltzmann machine
Mesh:
Substances:
Year: 2017 PMID: 28923002 PMCID: PMC5604354 DOI: 10.1186/s12859-017-1834-2
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1The flowchart of the deep learning methods for protein torsion angle prediction. Five commonly used features and two new features are used as input to build our deep learning method to predict torsion angles
Fig. 2The error distributions of torsion angles predicted from fragments for 20 types of amino acids. The red and green lines represent the distribution of phi and psi angles, respectively. The x-axis represents angles in the range [−180,180], and y-axis is the density
Fig. 3The architectures of four deep learning networks for torsion angle prediction. a Deep neural network (DNN). Features in the input layers are mapped to the hidden layers by sigmoid functions, from hidden layer to hidden layer, and finally propagated to the output layer for prediction. The weights in the network are randomly initialized according to the uniform distribution. The architecture is fine-tuned by back-propagation with SFO algorithm. b Deep recurrent neural network (DRNN). The inputs are connected to the first hidden layer by weight matrix W, and the neighboring positions in the first hidden layer are inter-connected by weight matrix U. The parameters are randomly initialized. The architecture is optimized by back-propagation through time with SFO algorithm. c Deep restricted Boltzmann (belief) network (DRBM). The layers are stacked with each other through Restricted Boltzmann Machine. And the weights are pre-trained by RBM. Predictions are made by forward pass and the network is optimized by back-propagation with SFO algorithm. d Deep Recurrent RBM network. Forward-propagation and backward-propagation follows the same strategy as deep recurrent neural network, while the parameters are pre-trained by RBM
The Mean Absolute Error (MAE) of different feature combinations with the DBRM method
| Number of features | Feature combinationa | phi | psi | avgb |
|---|---|---|---|---|
| 1 |
|
|
|
|
| 8-state secondary structure (8stateSS) | 25.12 | 33.52 | 29.32 | |
| Contacts_number_15_classes (CN15) | 25.58 | 37.26 | 31.42 | |
| Error_distribution_of_fragment_based_angles (fragsion) | 24.24 | 40 | 32.12 | |
| 3-state secondary structure (3SS) | 25.8 | 38.95 | 32.38 | |
| Contacts_number_1_real_value (CN1) | 26.92 | 44.71 | 35.82 | |
| 7 physicochemical properties (7PC) | 27.27 | 52.18 | 39.73 | |
| Solvent_accessibility (SA) | 29.15 | 53.84 | 41.5 | |
| Disorder | 30.8 | 64.69 | 47.75 | |
| 2 |
|
|
|
|
| PSSM_CN15 | 22.41 | 33.14 | 27.78 | |
| PSSM_Fragsion | 22.19 | 34.29 | 28.24 | |
| PSSM_7PC | 22.42 | 35.75 | 29.09 | |
| PSSM_DISORDER | 22.96 | 35.23 | 29.1 | |
| PSSM_SA | 23.47 | 35.53 | 29.5 | |
| 3 |
|
|
|
|
| PSSM_8stateSS_Fragsion | 21.63 | 30.72 | 26.18 | |
| PSSM_8stateSS_CN15 | 21.99 | 30.12 | 26.06 | |
| PSSM_SS8_Disorder | 22.91 | 31.08 | 27 | |
| PSSM_8stateSS_SA | 23.09 | 31.41 | 27.25 | |
| 4 |
|
|
|
|
| PSSM_8stateSS_7PC_SA | 21.88 | 30.89 | 26.39 | |
| PSSM_8stateSS_7PC_Disorder | 22.17 | 30.97 | 26.57 | |
| PSSM_8stateSS_7PC_Fragsion | 22.08 | 31.11 | 26.595 | |
| 5 |
|
|
|
|
| PSSM_8stateSS_7PC_CN15_SA | 21.93 | 30.39 | 26.16 | |
| PSSM_8stateSS_7PC_CN15_Fragsion | 21.81 | 30.83 | 26.32 | |
| 6 |
|
|
|
|
| PSSM_8stateSS_7PC_CN15_Disorder_SA | 22.24 | 30.60 | 26.42 | |
| 7 |
|
|
|
|
aFeatures combination: for example “PSSM_8stateSS” represent the combination of PSSM and 8-state secondary structure as input features. The bold font denotes the best combination selected for a specific number of features in terms of the average MAE of phi and psi angles
bavg.: Average of phi and psi values for each features combination
The prediction performance of using the standard features with two novel features
| Feature set 1 | Feature set 2 | Feature set 3 | Feature set 4 | |
|---|---|---|---|---|
| phi | 22.16 | 21.85 | 22.05 |
|
| psi | 33.21 | 32.81 | 32.76 |
|
Note:
Feature set 1: standard features
Feature set 2: standard features plus contact number
Feature set 3: standard features plus fragsion feature
Feature set 4: standard features plus contact number and fragsion
The prediction performance of phi and psi angles of using different local window sizes with four deep learning methods
| Torsion Angle | Window sizea | Featuresb | DRNN | DReRBM | DNN | DRBM |
|---|---|---|---|---|---|---|
| phi | 1 | 56 | 21.09 | 21.81 | 22.13 | 21.95 |
| 3 |
| 20.52 |
| 21.49 | 21.07 | |
| 5 | 280 | 20.39 | 20.92 | 21.24 | 20.89 | |
| 7 | 392 |
| 21.03 | 21.22 |
| |
| 9 | 504 | 20.40 | 20.95 | 21.28 | 21.62 | |
| 11 | 616 | 20.49 | 20.88 |
| 21.57 | |
| 13 | 728 | 20.56 | 20.98 | 21.27 | 21.79 | |
| 15 | 840 | 20.63 | 21.12 | 21.19 | 21.69 | |
| 17 | 952 | 20.69 | 21.04 | 21.38 | 21.66 | |
| psi | 1 | 56 | 31.68 | 32.93 | 33.55 | 33.02 |
| 3 |
| 29.29 |
| 30.14 | 29.74 | |
| 5 | 280 | 28.96 | 29.94 | 29.25 | 28.92 | |
| 7 | 392 |
| 30.11 | 29.25 |
| |
| 9 | 504 | 28.86 | 29.94 | 29.38 | 29.61 | |
| 11 | 616 | 29.06 | 29.95 |
| 29.75 | |
| 13 | 728 | 29.27 | 30.13 | 29.38 | 30.19 | |
| 15 | 840 | 29.44 | 30.48 | 29.24 | 30.25 | |
| 17 | 952 | 29.72 | 30.36 | 29.54 | 30.33 |
aNumber of window size range from 1 to 17
bNumber of features as input for the deep learning model. For each residue, we used 7 kinds of features, represented by 56 numbers
The bold fond denotes the best result for each method
The prediction performance of phi and psi angles of using different memory lengths for DRNN and DReRBM
| Torsion Angle | Memory Length | DRNN | DReRBM | ||
|---|---|---|---|---|---|
| phi | 5 |
|
|
|
|
|
| 20.53 | 0.600 | 20.95 | 0.581 | |
| 15 | 20.74 | 0.589 | 20.81 | 0.582 | |
|
| 22.20 | 0.539 | 20.90 | 0.580 | |
| 25 | 22.16 | 0.541 | 20.82 | 0.586 | |
| psi | 5 |
|
|
|
|
|
| 29.35 | 0.700 | 30.02 | 0.694 | |
| 15 | 29.74 | 0.696 | 29.94 | 0.694 | |
| 20 | 32.82 | 0.670 | 30.02 | 0.696 | |
| 25 | 32.78 | 0.671 | 29.89 | 0.698 | |
The effect of the number of hidden layers on DRBM and DNN
| Hidden layers | DRBM | DNN | ||
|---|---|---|---|---|
| Phi (MAE) | Psi (MAE) | Phi (MAE) | Psi (MAE) | |
| 2 layers | 21.76 | 30.57 | 21.33 | 31.02 |
| 3 layers | 21.13 | 29.43 | 21.21 | 30.05 |
| 4 layers | 21.77 | 31.02 | 21.65 | 30.88 |
| 5 layers | 21.23 | 29.47 | 21.25 | 29.97 |
The prediction performance of torsion angles on six different methods
| DRNN | DReRBM | DNN | DRBM | SPIDER2 | ANGLOR | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| phi | Psi | phi | psi | phi | psi | phi | Psi | phi | psi | phi | psi |
| 20.78 | 29.85 |
|
| 21.14 | 29.35 | 20.75 | 29.07 | 20.88 | 31.64 | 24.72 | 44.42 |
The prediction performance of torsion angles on CASP12 free modeling targets
| DNN | DRBM | DRNN | DReRBM | SPIDER2 | |
|---|---|---|---|---|---|
| Phi (MAE) | 22.68 | 22.72 | 22.38 | 22.47 | 22.61 |
| Psi (MAE) | 38.37 | 37.56 | 37.54 | 35.99 | 37.67 |
The statistical significance (p-value) of the performance difference between our methods and SPIDER2
| Method | DRBM | DRNN | DReRBM | SPIDER2 | ||||
|---|---|---|---|---|---|---|---|---|
| phi | psi | phi | psi | phi | psi | phi | psi | |
| DNN | 8.5E-02 | 1.2E-05 | 5.1E-03 | 5.8E-05 | 9.3E-03 | 2.6E-08 | 8.2E-02 | 2.0E-03 |
| DRBM | 8.4E-03 | 3.8E-01 | 3.3E-04 | 1.4E-07 | 1.4E-03 | 6.3E-04 | ||
| DRNN | 3.5E-02 | 5.5E-05 | 6.4E-03 | 8.9E-02 | ||||
| DReRBM | 1.5E-03 | 1.5E-06 | ||||||