Literature DB >> 23626696

CNNcon: improved protein contact maps prediction using cascaded neural networks.

Wang Ding1, Jiang Xie, Dongbo Dai, Huiran Zhang, Hao Xie, Wu Zhang.   

Abstract

BACKGROUNDS: Despite continuing progress in X-ray crystallography and high-field NMR spectroscopy for determination of three-dimensional protein structures, the number of unsolved and newly discovered sequences grows much faster than that of determined structures. Protein modeling methods can possibly bridge this huge sequence-structure gap with the development of computational science. A grand challenging problem is to predict three-dimensional protein structure from its primary structure (residues sequence) alone. However, predicting residue contact maps is a crucial and promising intermediate step towards final three-dimensional structure prediction. Better predictions of local and non-local contacts between residues can transform protein sequence alignment to structure alignment, which can finally improve template based three-dimensional protein structure predictors greatly.
METHODS: CNNcon, an improved multiple neural networks based contact map predictor using six sub-networks and one final cascade-network, was developed in this paper. Both the sub-networks and the final cascade-network were trained and tested with their corresponding data sets. While for testing, the target protein was first coded and then input to its corresponding sub-networks for prediction. After that, the intermediate results were input to the cascade-network to finish the final prediction.
RESULTS: The CNNcon can accurately predict 58.86% in average of contacts at a distance cutoff of 8 Å for proteins with lengths ranging from 51 to 450. The comparison results show that the present method performs better than the compared state-of-the-art predictors. Particularly, the prediction accuracy keeps steady with the increase of protein sequence length. It indicates that the CNNcon overcomes the thin density problem, with which other current predictors have trouble. This advantage makes the method valuable to the prediction of long length proteins. As a result, the effective prediction of long length proteins could be possible by the CNNcon.

Entities:  

Mesh:

Substances:

Year:  2013        PMID: 23626696      PMCID: PMC3634008          DOI: 10.1371/journal.pone.0061533

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

It is well known that discovering the three-dimensional (3D) structure of a protein can provide important clues to understand of the mechanism of protein functions. Unfortunately, determination of 3D protein structure through experimental methods, such as X-ray crystallography or NMR spectroscopy, are time consuming and not working effectively with all kinds of proteins, especially membrane proteins [1]. Additionally, there are more than 24 million protein sequences in UniPortKB [2] currently, among which only about 84,508 proteins have had their structures solved experimentally [3]. Furthermore, almost 10,000 entries are newly added into Protein Data Bank (PDB) yearly [3]. That means more than 2,400 years are needed to solve the currently existed protein structures through experimental methods, under the situation of current experimental technology and no more newly discovered proteins. In fact, the number of newly discovered sequences grows much faster than the number of structures solved with experimental methods. The computation method is obviously the only way to bridge the huge protein sequence-structure gap. Although many 3D protein structure predictors (3D-JIGSAW [4], I-TASSER [5], LOMETS [6], MODELLER [7], MODWEB [8], ROBETTA [9], SWISS-MODEL [10] and so on) with different accuracies have been developed in recent years, few predictors can produce desirable resolution structures for applications in medicine, such as drug design. The latest CASP experiment [11] shows that the progress has slowed and even reaches the bottleneck in direct prediction from one-dimensional (sequence) to three-dimensional (structure). With such difficulties, residue contact maps (CM or residue-residue contact) prediction, a matrix representation of protein residue-residue contacts, is the most promising one among recently developed prediction ideas. CM of a protein is a simplified version of the protein structure and provides a new avenue for predicting 3D protein structure [12]. As these two-dimensional representations capture all the important features of a protein fold, the whole complex and difficult 3D structure prediction task can be divided into two steps. That is solving the one-dimensional to two-dimensional prediction firstly and then the final two-dimensional to three-dimensional prediction. This idea of divide and conquer makes the problem much easier and also help reconstruct final 3D structure from predicted contact maps. Protein CM has some advantages below. First, CM conveys strong information about the 3D protein structure. Second, the binary CM nature can be regarded as a classical problem of a two-state classification which has been thoroughly studied. Third, it has been shown that the empirical reconstruction algorithms are quite insensitive to high levels of random noise in CM, so that it is not necessary to predict all contacts correctly for reconstructing the protein 3D structure [12]–[15]. So far, several contact maps prediction methods, such as NNcon [16], PROFcon [17], SVMcon [18], RECON [19], CMWeb [20] and CMAPpro [21], have been developed successfully. An improved multiple neural networks based contact map predictor, CNNcon, was proposed in this paper. It’s composed of six input sub-networks and one output network, which forms a two-level cascaded network architecture. All the networks used are standard back-propagation neural networks. For network inputs, different sources of information were mixed and most of them had been used separately in some way before.

Results and Conclusion

Assessment of the Prediction Efficiency

To score the efficiency of the CNNcon method, two widely used and accepted statistical indices are introduced. Here, we only sketch these scores that are described in detail in [22]–[25]. The first and most frequently used one is accuracy, also referred to as ‘Specificity’, defined as follows:where and are the number of correctly assigned contacts and that of total predicted contacts respectively. They also correspond to the sum of true positives (TP) and the sum of both TPs and false positives (FT) respectively. Routinely the accuracy is evaluated for each test protein and then averaged over the protein set. We also evaluate the performance on the coverage of correct predicted contacts, also referred to as ‘Sensitivity’, defined as:where is the same in equation (1) and is the number of observed contacts, which corresponds to the sum of TPs and false negatives (FN).

Results

Table 1 gives the prediction results of sub-networks and the final cascade-network, respectively. Two conclusions follow from these results. First, the prediction accuracy of each sub-network alone is comparable to other neural network based methods [16]–[18], whose performance is showed in Table 2. It indicates that our idea of assigning different prediction tasks to specific sub-networks corresponding to the protein length is practicable. Second, the remarkable improvement of accuracy from final cascade-network with little coverage loss proves that the CNNcon method is extremely effective and valuable.
Table 1

Performance of sub-networks and final cascade-network.

Separationa Seq-Lenb THRc Chainsd Acce Erracc f Cove Errcov f
Sub-network 1651–700.13046.439.2441.914.89
Sub-network 2771–900.64044.359.8936.437.64
Sub-network 31091–1300.719943.998.0536.334.59
Sub-network 413131–1900.724641.388.3933.955.32
Sub-network 517191–2900.820128.317.7236.572.64
Sub-network 621291–4500.98731.819.9034.472.53
Average 34.01 8.87 35.44 4.60
CNNcon51–450803 57.86 8.07 34.28 4.52

Sequence separation: if value is s, then only contacts between pairs minimally s residues apart are considered, that is .

Length Range of protein sequence of corresponding sub-network training and testing data sets.

Minimal prediction value to determine residues contact or not.

Size of test data set for each sub-network.

Acc: prediction accuracy(%), defined in equation (1) and Cov: coverage(%), defined in equation (2).

Standard error.

Table 2

Comparison results with other current methods.

PredictorAcce Cove Targetsg Method
CNNconh 57.8634.28803Neural network based; Using optimized thresholds.
NNconi 54.5035.00116Neural network based; Top L/5 predicted.
PROFconj 32.4019.60633Neural network based; Top L/2 predicted.
SVMconk 37.0021.0048Support vector machine based;Top L/5 predicted.

As in Table 1.

Size of test data sets.

This work.

Results are summarized from previous works [16]–[18], respectively.

Sequence separation: if value is s, then only contacts between pairs minimally s residues apart are considered, that is . Length Range of protein sequence of corresponding sub-network training and testing data sets. Minimal prediction value to determine residues contact or not. Size of test data set for each sub-network. Acc: prediction accuracy(%), defined in equation (1) and Cov: coverage(%), defined in equation (2). Standard error. As in Table 1. Size of test data sets. This work. Results are summarized from previous works [16]–[18], respectively. In general, it is neither straightforward nor completely fair to compare the performance of different contact map predictors. First, different predictors are usually suitable for different length range proteins. Second, there also not existed a benchmark data set big enough and accepted widely. Therefore, the comparisons with other current contact map predictors in Table 2 are used for reference. The results show that the CNNcon method achieves the best accuracy and the coverage is the second best, which is almost as good as the best one. Moreover, the largest test data set is used in order to make the present results reliable. To further verify the performance of the CNNcon method, we applied all the compared methods on the same test data set, 64 CASP10 targets. This test data set contains all the targets with length from 51 to 450 and valid PDB codes. Since different methods predict different number of contacts, in order to correctly compare them, n predicted contacts with the highest probabilities are selected. To increase the comparison preciseness, instead of being assigned one value, n was assigned to T/2, 2T/3 and T, respectively, where T was the total true contacts of the whole test data set. Then the final compared statistical indices take the average values. The details of the compared results are given in Table 3. Both accuracy and coverage of the present method are better than others.
Table 3

Comparison results on 64 CASP10 targets.

PredictorAcce Erracc f Cove Errcov f
CNNcon55.4817.1336.894.79
NNcon46.3911.7931.709.49
PROFcon39.907.0225.559.87
SVMcon38.159.0225.6210.93

As in Table 1.

As in Table 1. The prediction accuracies upon all proteins in the six test sets by corresponding sub-networks are shown in Figure 1. Clearly, accuracies decrease sharply while protein sequence length increases owing to the density of contacts decreasing greatly as the inverse of the protein length [12], [26]. This also troubles most other current contact predictors. However, the prediction accuracies from the present method almost keep the same with the increase of protein sequence length in Figure 2. That means the CNNcon method overcomes the thin density problem [12], [26], which suggests that it might be a valuable candidate for long length protein prediction.
Figure 1

Prediction results upon all test proteins by corresponding sub-networks.

The X axis is length range of tested proteins. The Y axis is prediction accuracy (%). Each point represents the predicted accuracy of a protein by its belonged sub-network. The average accuracy is as high as 34.01%. However, the accuracies decrease while the length of proteins increases.

Figure 2

Prediction results upon all test proteins by the final cascade-network.

The X axis and Y axis are the same in Figure 1. Each point represents the predicted accuracy of a protein by the final cascade-network. The average accuracy is as high as 57.86%. Moreover, the accuracies keep steady while the length of proteins increases.

Prediction results upon all test proteins by corresponding sub-networks.

The X axis is length range of tested proteins. The Y axis is prediction accuracy (%). Each point represents the predicted accuracy of a protein by its belonged sub-network. The average accuracy is as high as 34.01%. However, the accuracies decrease while the length of proteins increases.

Prediction results upon all test proteins by the final cascade-network.

The X axis and Y axis are the same in Figure 1. Each point represents the predicted accuracy of a protein by the final cascade-network. The average accuracy is as high as 57.86%. Moreover, the accuracies keep steady while the length of proteins increases.

Conclusion

An improved neural network based approach for protein contact map prediction, called CNNcon, was developed in this paper. The method performs better on prediction accuracy than other compared state-of-the-art methods. Further, the CNNcon method has better consistency and stability on prediction accuracy as protein length increases. Although training the six sub-networks and one cascade network costs computationally more than single-network predictors, it is one-time work. While in testing, the CNNcon method can divide the contact map prediction task naturally and run in parallel, based on the specially designed architecture. This advantage makes the method almost as fast as other single network based methods. It is expected that the CNNcon will be used to enhance parallel performance with longer protein length. As the neural network can be improved by adding more input information and training with a larger training data set, next work will be focus on combining more input information (e.g. correlated mutation information) and adding more protein chains to training data set. Parallel version of the CNNcon algorithm will also be implemented and worked on supercomputers in the future.

Discussion

Optimized Thresholds were Crucial for Performance

Table 1 (Column ‘THR’) gives the optimized thresholds for all sub-networks. They are minimal prediction values to determine residues contact or not for corresponding sub-networks. Different thresholds resulting in both different accuracies and different coverages are found. And different sub-networks have their own optimized thresholds. This was probably related to the different contact densities of different protein length ranges, according to which the sub-networks were introduced. Further, it is discovered that the coverage score dropped sharply while the threshold was once greater than a specific value. These specific values were used as our final optimized thresholds for the corresponding sub-networks.

Combining and Balancing Multiple Predictions Improves Accuracies

As expected, the prediction accuracies of sub-networks are at the same level of most single neural network based methods. However, the final prediction accuracy is improved greatly by our cascade-network because of the following two advantages of our model. First, instead of being processed by a single network, each test protein was input to its corresponding sub-network, left-next sub-network and right-next sub-network for prediction in parallel. This increases the opportunity of contacted amino acids to be found. Second, three optimized balancing weights were introduced to balance the predicted results of sub-networks during final cascade-network prediction.

Materials and Methods

Contact Map Definition

The contact map of a protein with N amino acids is an binary symmetric matrix . The components are defined as follows: We define two amino acids as being in contact if the distance between their atoms ( for glycines which having a hydrogen substituent as its side-chain) is less than 8 Å, a standard threshold widely used [12], [22], [27]–[29].

Neural Network Architecture

The finding that number of contacts in a protein is proportional to the protein length N, while the number of possible contacts increases with [12], implies the contact densities in the map decrease as the inverse of the protein length. In other words, long proteins have lower contact densities than short ones [26]. This makes the contact maps of long proteins more difficult to predict and the prediction accuracy is affected by the protein length greatly. Six specific sub-networks for different protein length range respectively and one cascade-network are introduced in order to solve this problem. They are all classical feed-forward 3-layer neural networks trained with the same standard back-propagation algorithm [30]. Architectures of all the six sub-networks are the same and composed of 1747 input nodes, 5 hidden nodes and 1 output node. The cascade-network contains 9 input nodes, 6 hidden nodes and 1 output node. The numbers of middle nodes are actually decided by repeated trials in experiment depending on the balance of computation time and prediction accuracy. Same values are assigned to the number of middle nodes of all sub-networks. In fact, it might be more suitable to assign the parameter of each sub-network with its different and specific values, since each sub-network is designed for proteins with different lengths. In next improved version of CNNcon (v2.0, also parallel and super computer version), this work will be considered and these optimal values of this parameter will be picked out through experiments performed on each sub-network. Among these input nodes, six are coded by prediction results from sub-networks and the remaining three are coded by balanced weights. The whole architecture of the CNNcon method is shown in Figure 3.
Figure 3

Architectures of sub-neural network (left) and cascade-neural network (right).

Since architectures of all the six sub-networks are the same, only one of them is shown here (left).

Architectures of sub-neural network (left) and cascade-neural network (right).

Since architectures of all the six sub-networks are the same, only one of them is shown here (left). Each sub-network was trained and tested with its corresponding data set. The data sets and length range divisions are mentioned in section of data sets below. While for testing, the target protein was coded first and input to its corresponding sub-networks for prediction. We defined the sub-network id as 1 to 6 as increase of its length coverage and the particular sub-network with length range covering the target protein length defined as i. Thus for each target protein, its corresponding sub-networks were , i and , that is just local communication needed while in parallel. After prediction by its corresponding sub-networks, the intermediate results along with three optimized balance factors were input to the cascade-network to finish the final prediction.

Input Codings

The basic input coding method used here is the same as previously introduced in [31]. Each residue pair is characterized by an vector containing 210 elements , representing all the possible ordered couples of residues. The input coding vectors of each residue couple and its symmetric ones are the same. In the method, multiple sequence information instead of single sequence was used, since evolutionary information had been proved to improve prediction performance greatly [17]. Multiple sequence alignment information of each protein sequence was gained from its corresponding HSSP file [32]. Considering the prediction performance and our computing resource, we chose as most as 100 multiple sequence alignment sequences (including the target one) with the identity of each aligned sequence less than 80%. For each sequence in the alignment, a pair of residues in position i and j were counted. The final input coding, representing the frequency of each pair in the alignment, was normalized to the number of the aligned sequences [31]. Conservation weights and secondary structures [33] information from HSSP file were also coded with one and three elements respectively. Thus the length of the input coding vector becomes 218 . To obtain local information of each residue, similar to [17], we used two content windows of size 2 centered around i and j (window of i: , window of j: ) respectively. That means that, for each residue pair , we incorporated information from all residues in those two windows of five consecutive residues. Thus, the length of the input coding vector was increased to 1090 . Further, we introduced a segment window with size of 2 to code information from the segment connecting i and j. For each residue pair , we incorporated information from all residues in the window centered around k , which was the middle position of i and j. Thus, the segment window spanned the interval ) and the length of our input coding vector again was added to 1744 . Finally, we used sequence separation, sequence length and segment separation length to represent the global information from the entire protein. The size of our input coding vector was lastly set to 1747 .

Data Sets

Data set used here for training and testing was extracted from the March 2012 25% pdb_select list [34]–[37] with 5,300 chains and 788,447 residues. For the goal of algorithm design, we removed all protein chains of non-X-ray determined structures, all chains with resolution greater than 1.5 Å, all backbone broken chains (contain missing backbone atoms in the PDB files), all chains containing non-standard residues in its corresponding PDB files and all chains with obsolete PDB ID (e.g. 3G62 is obsolete and replaced by 4F1U). We reduced the data set further by excluding all protein chains longer than 450 residues. Without loss of generality, all chains shorter than 51 residues were removed as well. After above processing, our final data set contains 1,103 chains (1,082 proteins) and 192,640 residues. As prediction performances greatly depend on protein length distribution, here we give the protein length distribution of data set to make assessment more reasonable. 7.25% of the proteins have a length from 51 to 70 residues (sub-network 1), 8.16% comprise from 71 to 90 residues (sub-network 2); 22.57% from 91 to 130 residues (sub-network 3); 26.84% from 131 to 190 residues (sub-network 4); 22.76% from 191 to 290 residues (sub-network 5); 12.42% from 291 to 450 residues (sub-network 6). These distributions are also the partitions of length range coverage of sub-networks. That’s also why six sub-networks are needed in the CNNcon method. The data set was split into six subsets according to the above length range distributions. Each sub-network was trained with 50 samples randomly selected from its corresponding data set and tested by the remaining. We used all the six test subsets (803 protein chains in total) to test the final cascaded network.

Balanced Training

To address the extreme disproportion distribution of true (contacts) and false (non-contacts) samples during the training phase, we used balanced training to reduce back-propagation learning cycles [38]. A balancing probability factor was also introduced to further reduce the false samples and the whole training data set size in a random way.
  34 in total

1.  A neural network based predictor of residue contacts in proteins.

Authors:  P Fariselli; R Casadio
Journal:  Protein Eng       Date:  1999-01

2.  Ab initio modeling of small, medium, and large loops in proteins.

Authors:  S Galaktionov; G V Nikiforovich; G R Marshall
Journal:  Biopolymers       Date:  2001       Impact factor: 2.505

3.  Prediction of contact maps with neural networks and correlated mutations.

Authors:  P Fariselli; O Olmea; A Valencia; R Casadio
Journal:  Protein Eng       Date:  2001-11

4.  Progress in predicting inter-residue contacts of proteins with neural networks and correlated mutations.

Authors:  P Fariselli; O Olmea; A Valencia; R Casadio
Journal:  Proteins       Date:  2001

5.  Tools for comparative protein structure modeling and analysis.

Authors:  Narayanan Eswar; Bino John; Nebojsa Mirkovic; Andras Fiser; Valentin A Ilyin; Ursula Pieper; Ashley C Stuart; Marc A Marti-Renom; M S Madhusudhan; Bozidar Yerkovich; Andrej Sali
Journal:  Nucleic Acids Res       Date:  2003-07-01       Impact factor: 16.971

6.  The SWISS-MODEL workspace: a web-based environment for protein structure homology modelling.

Authors:  Konstantin Arnold; Lorenza Bordoli; Jürgen Kopp; Torsten Schwede
Journal:  Bioinformatics       Date:  2005-11-13       Impact factor: 6.937

7.  Selection of representative protein data sets.

Authors:  U Hobohm; M Scharf; R Schneider; C Sander
Journal:  Protein Sci       Date:  1992-03       Impact factor: 6.725

Review 8.  Knowledge-based protein modeling.

Authors:  M S Johnson; N Srinivasan; R Sowdhamini; T L Blundell
Journal:  Crit Rev Biochem Mol Biol       Date:  1994       Impact factor: 8.250

9.  Predicting residue contacts using pragmatic correlated mutations method: reducing the false positives.

Authors:  Petras J Kundrotas; Emil G Alexov
Journal:  BMC Bioinformatics       Date:  2006-11-16       Impact factor: 3.169

10.  LOMETS: a local meta-threading-server for protein structure prediction.

Authors:  Sitao Wu; Yang Zhang
Journal:  Nucleic Acids Res       Date:  2007-05-03       Impact factor: 16.971

View more
  7 in total

1.  Multi-Dimensional Scaling and MODELLER-Based Evolutionary Algorithms for Protein Model Refinement.

Authors:  Yan Chen; Yi Shang; Dong Xu
Journal:  Proc Congr Evol Comput       Date:  2014-07

2.  Ordering Protein Contact Matrices.

Authors:  Chuan Xu; Guillaume Bouvier; Benjamin Bardiaux; Michael Nilges; Thérèse Malliavin; Abdel Lisser
Journal:  Comput Struct Biotechnol J       Date:  2018-03-16       Impact factor: 7.271

3.  Forecasting residue-residue contact prediction accuracy.

Authors:  P P Wozniak; B M Konopka; J Xu; G Vriend; M Kotulska
Journal:  Bioinformatics       Date:  2017-11-01       Impact factor: 6.937

4.  Sequence-based Gaussian network model for protein dynamics.

Authors:  Hua Zhang; Lukasz Kurgan
Journal:  Bioinformatics       Date:  2013-12-12       Impact factor: 6.937

5.  Characteristics of protein residue-residue contacts and their application in contact prediction.

Authors:  Pawel P Wozniak; Malgorzata Kotulska
Journal:  J Mol Model       Date:  2014-11-06       Impact factor: 1.810

6.  Parallel clustering algorithm for large-scale biological data sets.

Authors:  Minchao Wang; Wu Zhang; Wang Ding; Dongbo Dai; Huiran Zhang; Hao Xie; Luonan Chen; Yike Guo; Jiang Xie
Journal:  PLoS One       Date:  2014-04-04       Impact factor: 3.240

7.  The evolution of logic circuits for the purpose of protein contact map prediction.

Authors:  Samuel D Chapman; Christoph Adami; Claus O Wilke; Dukka B Kc
Journal:  PeerJ       Date:  2017-04-18       Impact factor: 2.984

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.