Literature DB >> 35540205

Protein secondary structure prediction with context convolutional neural network.

Shiyang Long1, Pu Tian2.   

Abstract

Protein secondary structure (SS) prediction is important for studying protein structure and function. Both traditional machine learning methods and deep learning neural networks have been utilized and great progress has been achieved in approaching the theoretical limit. Convolutional and recurrent neural networks are two major types of deep learning architectures with comparable prediction accuracy but different training procedures to achieve optimal performance. We are interested in seeking a novel architectural style with competitive performance and in understanding the performance of different architectures with similar training procedures. We constructed a context convolutional neural network (Contextnet) and compared its performance with popular models (e.g. convolutional neural network, recurrent neural network, conditional neural fields…) under similar training procedures on a Jpred dataset. The Contextnet was proven to be highly competitive. Additionally, we retrained the network with the Cullpdb dataset and compared with Jpred, ReportX, Spider3 server and MUFold-SS method, the Contextnet was found to be more Q3 accurate on a CASP13 dataset. Training procedures were found to have significant impact on the accuracy of the Contextnet. This journal is © The Royal Society of Chemistry.

Entities:  

Year:  2019        PMID: 35540205      PMCID: PMC9075825          DOI: 10.1039/c9ra05218f

Source DB:  PubMed          Journal:  RSC Adv        ISSN: 2046-2069            Impact factor:   4.036


Introduction

Protein secondary structure is the local three dimensional (3D) organization of its peptide segments. In 1951 Pauling and Corey first proposed helical and sheet conformations for protein polypeptide backbones based on hydrogen bonding patterns,[1] and three secondary structure states were defined accordingly. There were two regular secondary structure states: helix (H) and sheet (E), and one irregular secondary structure type: coil (C). In 1983 Sander[2] developed a secondary structure assignment method DSSP (Dictionary of Secondary Structure of Proteins), which classified secondary structure into eight states (H = α-helix, E = extended strand, B = residues in isolated β-bridge, G = 310-helix, I = π-helix, T = hydrogen bonded turn, S = bend and C = coil, remaining). These eight states were often reduced to three states termed helix, sheet and coil respectively. The most widely used convention was that G, H and I were reduced to helix (H); B and E were reduced to sheet (E), and all other states were reduced to coil (C).[3-5] Understanding protein function requires knowledge of their structures. Although many protein structures have been deposited in Protein Data Bank,[6] (http://www.wwpdb.org) far more sequences were known. Additionally, with present second generation and coming more efficient (and accurate) sequencing technologies, the gap between known sequences and known structures was expected to grow with accelerated speed. Considering the high cost of protein structure determination by experiments and rapid increase of available computational power, predicting protein structures using their sequence information computationally was a potentially practical solution. As the first step to predict 3D protein structures, protein secondary structure prediction has been studied for over sixty years.[4] Secondary structure prediction methods can roughly be divided into template-based methods[7-10] which using known protein structures as templates and template-free ones.[3,5,11,12] Template-based methods usually have better performance, but do not work well with sequences that lack homologous templates. Template-free methods utilize sequence information alone. Q3 accuracy (i.e. with secondary structures labeled as H, E and C) based on template-free methods has been increased slowly from <70% to 82–84%,[3,4,13,14] gradually approaching the theoretical limit (88–90%).[13] Three major factors that decide prediction accuracy were input features, predicting methods (algorithms) and training dataset. Input features, as the source of the information, have critical impact on accuracy. Utilizing multiple sequence alignment of homologous sequences rather than a single sequence has long been recognized as a way to improve prediction accuracy. The PSSM (position-specific scoring matrices) calculated by PSI-BLAST (Position Specific Iterative-Basic Local Alignment Search Tool)[15] has been widely utilized, which contributed significantly to the improvement of the Q3 accuracy to over 80% on benchmark datasets.[4] Compared with results from predicting using the sequence information only, Q3 accuracy has been increased by about 10%.[16] Some other input features, such as physio-chemical properties of amino acids and protein profiles generated using HHBlits, have been used in some secondary structure prediction models recently.[5,11] Dataset used for training secondary structure prediction model has grown to several thousand sequences.[3,5,11] The sequence identity within each of these dataset was usually smaller than 25% or 30%. Many different algorithms have been used to predict protein secondary structure in previous investigations, such as hidden markov models[17] and support vector machines.[7,18] In recent studies, neural networks were widely used. DeepCNF (Deep Convolutional Neural Fields)[3] was an integration of CRF (Conditional Random Fields) and shallow neural networks, DeepCNF improved accuracy over several methods including SPINE-X, PSIPRED and Jpred by 1–4% on some datasets. LSTM (Long Short-Term Memory) bidirectional recurrent neural network[11] was developed to predict protein secondary structure states, backbone angles, contact numbers and solvent accessibility. MUFold-SS[5] used inception–inside–inception network to predict secondary structure states. 2C-BRNNs (two-dimensional convolutional bidirectional recurrent neural network)[12] was used to improve the accuracy of 8-state secondary structure prediction. Present focus of protein secondary structure prediction studies was mainly on accuracy of secondary structure state, a very coarse classification. Potential value of these studies would only be fully embodied when combined with 3D structural prediction and design. It might well be that information representations learned in non-output layers can provide additional assistance for later stage structural studies. Therefore, searching for novel architecture flavors is meaningful by providing potentially unique and useful representations even if no significant Q3 accuracy improvement was achieved. We plan to investigate the value of non-output layer representations of various typical neural network architectures for secondary structure prediction in future work. In this work, we constructed a Contextnet (context convolutional neural network) and obtained higher accuracy than a few typical LSTM and CNN based architectures on Jpred and some other frequently used test datasets.

Methodology

Dataset and hardware

Five datasets were utilized in this study. Jpred dataset[19] and CB513 (ref. 20) dataset were downloaded from Jpred server (http://www.compbio.dundee.ac.uk/jpred/about.shtml). Jpred dataset contained a 1348-sequence training set and a 149-sequence test set, these sequences were selected representatives from SCOP superfamilies rather than constructed with a simple sequence identity cutoff. All the test and training protein sequences belong to different superfamilies. CB513 contains 513 non-redundant sequences, all sequences in which had been compared pairwise, and were non redundant to a 5SD cut-off.[20] CASP12 and CASP13 were downloaded from Protein Structure Prediction Center (http://predictioncenter.org/), and the target structures were used. Cullpdb[21] dataset was downloaded from dunbrack lab (http://dunbrack.fccc.edu/PISCES.php). Cullpdb dataset was generated on 2018.11.26 (the percentage identity cutoff was 25%, the resolution cutoff was 2.0 angstroms and the R-factor cutoff was 0.25). We downloaded from PDB (protein data bank, http://www.wwpdb.org) the 9311 chains in the Cullpdb list. All the sequences in Cullpdb, CB513, CASP12 and CASP13 datasets are culled with CD-HIT server, sequence that had more than 25% identity to any sequences in the datasets was removed. So the sequence identity in all the datasets are less than 25%. We also removed sequences that failed in our described PSSM construction procedure (see below). Finally, we have 8601 sequences in Cullpdb dataset, 261 sequences in CB513 dataset, 35 sequences in CASP12 dataset and 15 sequences in CASP13 dataset. The Cullpdb dataset was arbitrarily divided into a 8401-sequence Cullpdb training dataset and a 200-sequence test dataset. All networks were trained with GPU (GTX 1080Ti).

The context convolutional neural network

Both local interactions due to neighboring residues in primary sequence and various non-local interactions due to tertiary interactions and electrostatic interactions participate in deciding secondary structure state of each residue. Explicit tertiary structure input is not available for sequences that do not have corresponding 3D structure available. PSSM in principle convey relevant non-local interaction information, but not in explicit and easy-to-decode form. Both convolution and LSTM architecture have the ability to capture non-local interactions through either maxpooling or recurrent operations. However, maxpooling, while expands the receptive field of convolutional networks, reduces image size and results in loss of information. Recurrent operation makes training process rather computationally intensive when compared with convolutional networks. The context module was constructed to increase the accuracy of state-of-the-art semantic segmentation systems.[22] The module used dilated convolutions to systematically aggregate multiscale contextual information without losing resolution. Dilated convolution can effectively increase the receptive field in the kernel without increasing the model parameters or the amount of calculation. With the dilated convolution, non-local interactions may be captured by fewer convolution layers. We therefore, hoping to better capture non-local interactions without engendering expensive training computation, constructed a context convolutional neural network (Contextnet, see Fig. 1). The first 8 layer convolution results were concatenated together and the features of different receptive field were mixed. The operation concatenates tensors along one dimension, here we put the channels of different layers together and the number of output channels is the sum of the first 8-layer channels. In 5–8 layers the dilated convolution were used, strides of the dilated convolutions in layers 5–8 were 2, 4, 8 and 16 respectively. The kernel size used were 3 × 1, the activation function was relu in hidden layers, and the activation function in the output layer was softmax. The loss function was cross entropy and the network was built with Tensorflow. Detailed code can be found on github.
Fig. 1

Context convolutional neural network architecture. The gray square represented input features, blue squares represented convolution operation, green squares represented dilated convolution operations and yellow squares represented concatenation operation. Numbers in squares were the channel number of the corresponding convolution operation.

Input features and preprocessing

In Jpred dataset, PSSM were used as the only input features, with 20 elements to each residue. For Jpred dataset, both PSSM and labels were downloaded from Jpred server. For Cullpdb dataset, we ran PSI-BLAST with E-value threshold 0.001 and iteration number 3 to search UniRef90 (ref. 23) to generate PSSM. The UniRef90 database was downloaded from Jpred server. For Cullpdb dataset we also used physio-chemical properties[24] and HHBlits profiles[25] as input features. For a residue we had 57 features, 20 from PSSM, 30 from HHBlits profiles and 7 from physio-chemical properties. The HHBlits profiles were generated with uniprot20_2013_03 database. We transformed the features to center it by subtracting the mean value of the training dataset, then scaled it by dividing by their standard deviation. The mean and standard deviation (obtained from the training set) were applied to the test set. Labels were obtained by first calculating 8-state DSSP labels and then reduced to 3-state, H (G, I, H), E (E, B) and C (others). The input features PSSM of each sequence were converted to shape (sequence length, 1, 20).

Comparison with six other networks

In the first part of our performance comparison study, Adam (adaptive moment estimation) optimizer was used to train all the networks. For each network we manually chose an optimal learning rate among various tested values. The relu activation function was selected for all hidden layers and softmax was selected for the output layer, and with cross entropy selected as the loss function for all networks.

Training strategies for the Contextnet

In performance comparison with servers, we trained the Contextnet with SGD (stochastic gradient descent) optimizer. To improve generalization capability of the Contextnet, a number of tricks were utilized. First we added white noise to input features by multiplying a random number sampled from a Gaussian distribution with mean 1.0 and standard deviation 0.5 to each feature element; second a random learning rate sampled from a uniform distribution in the range (0.02, 0.12) was utilized in each epoch of SGD optimization; third a L2 regularization factor of 0.01 was utilized; fourth, for each sequence half of the labels were randomly selected and masked (masked labels did not participate in backpropagation); finally, two dropout layers were added (after the first and the third concatenation layers) in the network. The Q3 accuracy on corresponding test dataset was calculated every epoch during the training. The best test results from 20 epochs and corresponding model parameters were chosen as our final results. We did not use sequences that contain less than 20 residues when training with the Cullpdb dataset. The batch size was one sequence. We did not limit the size of convolutional network, the output feature numbers were the same as the length of the input sequence and each feature corresponds to a multi-class operation. So only one pass of convolution operation was performed on a sequence and results were obtained for all residues. The input feature shape was again (sequence length, 1, 20).

Results

Performance of the Contextnet and six other typical neural networks with the same simple training procedure

To evaluate the performance of the Contextnet and to compare with other published networks, we first used Jpred dataset. To make the comparison relatively fair, all networks were given PSSM as the input features. The labels were 3-state secondary structure and were generated with DSSP calculations followed by a 8-state to 3-state reduction H (H), E (E, B), C (others). We trained these models with 10 Cross-validation on Jpred training set and calculated the Q3 accuracy on test set. 10 to 20-epoch optimization was carried out for each network until apparent overfit was observed. Network parameter set that gave the highest validation accuracy in epoches were chosen as the final result. We constructed seven different networks with Tensorflow. Some of networks were constructed according to previous studies (e.g. DeepCNF, bidirectional recurrent neural networks and inception–inside–inception networks). No dropout and other special training methods (e.g. label masking, random noise addition, random learning rate etc.) were used in the training of all networks. It was important to note that the networks and training methods were not completely consistent with previous studies. Details were listed below: (1) Simple convolutional neural network: A simple convolutional neural network with twelve hidden layers and one output layer was constructed. The kernel size was 3 × 1 and the number of channel was 256 in hidden layers. The SAME padding was utilized for each layer. (2) Bidirectional LSTM neural networks: The network was constructed with a bidirectional LSTM layer, two hidden layers and an output layer. The number of units in LSTM was 256. 1024 units were in the first and 512 units were in the second hidden layer. (3) Convolutional neural network with conditional neural fields layer:[3] The difference between Deep Convolutional Neural Fields (DeepCNF) and this implementation was that we used the relu activation function instead of sigmoid for hidden layers and we trained the network with Adam method instead of L-BFGS. Additionally, Wang trained the network layer by layer, but we trained the whole network directly. L2 regularization was not used here. (4) Inception–inside–inception (Deep3I):[5] In our implementation no dropout layer was used and the input features were PSSM alone. In the original work physio-chemical properties of amino acids and HHBlits profiles were used as input features besides PSSM. (5) Double bidirectional LSTM neural networks:[11] Similar to (4), in contrast to the original work, the network was constructed without dropout layers. Physio-chemical properties of amino acids and HHBlits profiles were not used as input features. (6) Convolutional neural network with bidirectional LSTM layer: A network with five convolution layers, followed by one bidirectional LSTM layer and one output layer was constructed. The kernel size was 3 × 1 and the channel number was 256 in convolution layers. The number of units in LSTM was 256. (7) Context convolutional neural network: We constructed a context convolutional neural network (Contextnet). Concatenation and dilated convolution operations were utilized in the network. The detailed structure was described in the Methods section (see also Fig. 1). We trained each network ten times. The average Q3 accuracy on test set and the standard deviation were shown in Table 1. The Contextnet obtained the highest accuracy at 83.96%.

Q3 accuracy of seven different secondary structure prediction networks on Jpred dataset

Accuracy (%)Standard deviation (%)Learning rate
Simple CNN82.680.350.0001
BiLSTM83.030.200.001
CNN + CRF82.080.320.0001
Deep3I83.040.210.0001
Double BiLSTM83.300.280.001
CNN + BiLSTM83.350.130.0001
Contextnet83.660.240.0001
These results by no means suggested that Contextnet was a superior secondary structure predictor to other tested networks as selection of training procedures and optimization details may change relative ranking of various networks. It was neither meaningful nor realistic to test all possible combinations of training methods. Nevertheless, the results strongly suggested that Contextnet was a competitive network.

Training tricks and the ensemble method applied on the Contextnet

In order to improve the generalization capability of the network, we tested some training tricks on the Contextnet as described in the Methods section. With these tricks applied simultaneously, the Q3 accuracy of Contextnet was increased to 84.14% on the Jpred test set, and the variance also dropped as indicated by reduction of standard deviation from 0.24 to 0.13 (see Table 2).

Improvement of Contextnet by training tricks and the ensemble method

Q3 accuracy (%)Standard deviation (%)
Contextnet83.660.24
Trained with tricks84.140.13
Ensemble of Contextnet84.74
The performance can be further improved by the ensemble method. We use models of 10 Cross-validation to vote the final results. The Q3 accuracy increase to 84.74% as shown in Table 2.

Comparison with Jpred, ReportX, Spider3 server and MUFold-SS on CB513, CASP12 and CASP13 datasets

Jpred,[19] ReportX[26] and Spider3 (ref. 11) and MUFold-ss[5] are widely known servers for secondary structure prediction. To compare performance of our network with these servers, we retrained the Contextnet on the Cullpdb dataset, then tested the resulting model on CB513, CASP12 and CASP13 datasets. The input features used here were physio-chemical properties, HHBlits profiles and PSSM. We trained the model with 10 Cross-validation on Cullpdb training dataset and tested the model performance on Cullpdb test dataset. Here we calculate Q3 accuracy, Q8 accuracy and SOV (Segment of OVerlap) score.[27] SOV score measures how well the observed and the predicted SS segments match. In particular, it assigns a lower score to the prediction deviating from observed SS segment length distribution even if it has high Q3 accuracy. A wrong prediction in the middle region of a SS segment results in a lower SOV score than a wrong prediction at terminal regions. The SOV score used here is calculated with 3-state segments. Mean accuracy of 10 parallel models and ensemble accuracy of these 10 models are list in Table 3. We further tested the model on CB513, CASP12 and CASP13 datasets to compare with Jpred, ReportX, Spider3 and MUFold-SS. The larger size (when compared with Jpred training set) of training datasets implied a potentially better evaluation of the performance for tested networks.

Q3 accuracy, Q8 accuracy and SOV of Cullpdb test set

Q3 accuracy (%)Q8 accuracy (%)SOV
Contextnet84.4173.1377.35
Ensemble of Contextnet85.2974.6880.41
The results of Jpred, ReportX and Spider3 are generated with online services and results of MUFold-SS are generated with a downloaded local package. Q3 accuracy of Jpred, ReportX, Spider3, MUFold-SS and Q8 accuracy of ReportX, MUFold-SS are used here (Jpred and Spider3 servers can only provide 3-state prediction). Jpred, ReportX, Spider3 and MUFold-SS used different ways to reduce 8 secondary structure states to 3 states. Jpred reduced H to H, E and B to E, others to C. ReportX, Spider3 and MUFold-SS reduced H, I and G to H, E and B to E, others to C. Here we used the second reduced method. In Tables 4 and 5 we used culled CB513, CASP12 and CASP13 dataset, the sequences are culled with our training dataset. In Tables 6 and 7 we use complete CB513, CASP12 and CASP13 dataset. The Q3 and Q8 accuracy results were listed in Tables 4 and 6, with the ensemble of Contextnet provides higher accuracy. The SOV score results were listed in Tables 5 and 7.

Q3(Q8) accuracy of Jpred server, ReportX server, DeepCNF server, MUFold-SS and the Contextnet. Culled CB513, CASP12 and CASP13 dataset used here

CB513 (%)CASP12 (%)CASP13 (%)
Jpred server80.1178.5180.01
ReportX server82.34(70.30)80.84(69.16)81.13(67.71)
Spider3 server84.5682.2383.22
MUFold-SS85.12(72.41)80.98(68.87)83.19(72.30)
Contextnet83.98(71.15)81.67(69.87)83.81(71.01)
Ensemble of Contextnet85.04(72.76)82.69(71.20)84.93(72.95)

SOV score of Jpred server, ReportX server, DeepCNF server, MUFold-SS and the Contextnet. Culled CB513, CASP12 and CASP13 dataset used here

CB513CASP12CASP13
Jpred server75.4073.5673.61
ReportX server77.3875.1974.70
Spider3 server81.1975.7377.32
MUFold-SS80.6073.8484.27
Contextnet76.0971.8674.64
Ensemble of Contextnet78.9075.2078.53

Q3(Q8) accuracy of Jpred server, ReportX server, DeepCNF server, MUFold-SS and the Contextnet. Complete CB513, CASP12 and CASP13 dataset used here

CB513 (%)CASP12 (%)CASP13 (%)
Jpred server80.4778.0780.17
ReportX server82.83(70.89)80.72(68.84)80.79(67.38)
Spider3 server84.7181.9982.97
MUFold-SS85.81(73.41)80.94(68.88)83.34(71.89)
Contextnet84.32(71.96)81.43(69.56)83.62(70.73)
Ensemble of Contextnet85.28(73.77)82.56(70.78)84.81(72.32)

SOV score of Jpred server, ReportX server, DeepCNF server, MUFold-SS and the Contextnet. Complete CB513, CASP12 and CASP13 dataset used here

CB513CASP12CASP13
Jpred server76.5672.5473.65
ReportX server78.3674.8474.12
Spider3 server81.3575.5875.73
MUFold-SS82.1274.0083.73
Contextnet77.2772.1474.67
Ensemble of Contextnet80.2775.9178.18

Discussion

Performance of given networks may be improved by using combination of different hyperparameters, training methods (e.g. layer normalization, random noise, various regularization, dropouts etc.) and optimization methods.[28] We demonstrated that the same was true for the Contextnet when used to predict protein secondary structures. Significant impact of performance by application of various training tricks clearly illustrated that the free energy profile of neural network parameters have multiple local minima. Different training procedures, hyperparameters, initializations and optimization methods may take the training trajectory to various local minima with different generalization capability. Consistently, it was observed that all tested neural networks (with Jpred dataset) eventually overfit. We made some effort to improve the Contextnet through various training tricks as described in the Methods section. Certainly, there might be more potential space for improvement that we simply do not have time to explore. The fact that we observed better performance for the Contextnet on nearly all tested datasets by no means suggested that the Contextnet was superior to other typical architectures, which may well be improved further in terms of Q3 accuracy if sufficient effort was made to explore combinations of training, optimization and hyperparameters. All published deep networks have the capability to extract non-local information and were sufficiently complex to overfit. Nonetheless, our contribution was that a new flavor of architecture that have competitive Q3 accuracy for protein secondary structure prediction was constructed and tested. As stated in the Introduction, diversity of network architecture might be of importance since intermediate representations learned might provide additional assistance to downstream 3D structural prediction and design. We were interested in exploring specifics of intermediate representations from various flavors of neural networks with competitive secondary structure prediction performance. Recurrent networks were significantly more expensive in training than convolutional networks while their performance as measured by Q3 accuracy were comparable. Nonetheless, we believed that recurrent networks were of great value as they can provide potentially unique useful information from intermediate representations.

Conclusions

We constructed a context convolutional neural network to predict protein secondary structure state. In this network we used dilated convolutions to capture non-local interactions and used concatenation operation to mix multiscale contextual information. This network achieved competitive performance when compared with other published networks in our tests on seven datasets. In consistency with many neural network studies,[28] we demonstrated the importance of training procedures in determining the generalization capability of the Contextnet. We believed that diverse architectures with competitive protein secondary structure prediction capability were potentially of great importance by providing different intermediate representations which might well be useful in downstream 3D structural studies. We plan to investigate specifics of learned intermediate representations from major network architectures used for protein secondary structure prediction.

Conflicts of interest

There are no conflicts to declare.
  24 in total

1.  The Protein Data Bank.

Authors:  H M Berman; J Westbrook; Z Feng; G Gilliland; T N Bhat; H Weissig; I N Shindyalov; P E Bourne
Journal:  Nucleic Acids Res       Date:  2000-01-01       Impact factor: 16.971

2.  Application of multiple sequence alignment profiles to improve protein secondary structure prediction.

Authors:  J A Cuff; G J Barton
Journal:  Proteins       Date:  2000-08-15

Review 3.  Review: protein secondary structure prediction continues to rise.

Authors:  B Rost
Journal:  J Struct Biol       Date:  2001 May-Jun       Impact factor: 2.867

4.  Prediction of protein secondary structure by mining structural fragment database.

Authors:  Haitao Cheng; Taner Z Sen; Andrzej Kloczkowski; Dimitris Margaritis; Robert L Jernigan
Journal:  Polymer (Guildf)       Date:  2005-05-26       Impact factor: 4.430

5.  Protein secondary structure prediction improved by recurrent neural networks integrated with two-dimensional convolutional neural networks.

Authors:  Yanbu Guo; Bingyi Wang; Weihua Li; Bei Yang
Journal:  J Bioinform Comput Biol       Date:  2018-10       Impact factor: 1.122

6.  Prediction of protein secondary structure by combining nearest-neighbor algorithms and multiple sequence alignments.

Authors:  A A Salamov; V V Solovyev
Journal:  J Mol Biol       Date:  1995-03-17       Impact factor: 5.469

7.  MUFOLD-SS: New deep inception-inside-inception networks for protein secondary structure prediction.

Authors:  Chao Fang; Yi Shang; Dong Xu
Journal:  Proteins       Date:  2018-03-12

8.  Capturing non-local interactions by long short-term memory bidirectional recurrent neural networks for improving prediction of protein secondary structure, backbone angles, contact numbers and solvent accessibility.

Authors:  Rhys Heffernan; Yuedong Yang; Kuldip Paliwal; Yaoqi Zhou
Journal:  Bioinformatics       Date:  2017-09-15       Impact factor: 6.937

9.  The Universal Protein Resource (UniProt).

Authors:  Amos Bairoch; Rolf Apweiler; Cathy H Wu; Winona C Barker; Brigitte Boeckmann; Serenella Ferro; Elisabeth Gasteiger; Hongzhan Huang; Rodrigo Lopez; Michele Magrane; Maria J Martin; Darren A Natale; Claire O'Donovan; Nicole Redaschi; Lai-Su L Yeh
Journal:  Nucleic Acids Res       Date:  2005-01-01       Impact factor: 16.971

10.  RaptorX-Property: a web server for protein structure property prediction.

Authors:  Sheng Wang; Wei Li; Shiwang Liu; Jinbo Xu
Journal:  Nucleic Acids Res       Date:  2016-04-25       Impact factor: 16.971

View more
  1 in total

1.  Protein Science Meets Artificial Intelligence: A Systematic Review and a Biochemical Meta-Analysis of an Inter-Field.

Authors:  Jalil Villalobos-Alva; Luis Ochoa-Toledo; Mario Javier Villalobos-Alva; Atocha Aliseda; Fernando Pérez-Escamirosa; Nelly F Altamirano-Bustamante; Francine Ochoa-Fernández; Ricardo Zamora-Solís; Sebastián Villalobos-Alva; Cristina Revilla-Monsalve; Nicolás Kemper-Valverde; Myriam M Altamirano-Bustamante
Journal:  Front Bioeng Biotechnol       Date:  2022-07-07
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.