Literature DB >> 28492483

PCVMZM: Using the Probabilistic Classification Vector Machines Model Combined with a Zernike Moments Descriptor to Predict Protein-Protein Interactions from Protein Sequences.

Yanbin Wang1, Zhuhong You2, Xiao Li3, Xing Chen4, Tonghai Jiang5, Jingting Zhang6.   

Abstract

Protein-protein interactions (PPIs) are essential for most living organisms' process. Thus, detecting PPIs is extremely important to understand the molecular mechanisms of biological systems. Although many PPIs data have been generated by high-throughput technologies for a variety of organisms, the whole interatom is still far from complete. In addition, the high-throughput technologies for detecting PPIs has some unavoidable defects, including time consumption, high cost, and high error rate. In recent years, with the development of machine learning, computational methods have been broadly used to predict PPIs, and can achieve good prediction rate. In this paper, we present here PCVMZM, a computational method based on a Probabilistic Classification Vector Machines (PCVM) model and Zernike moments (ZM) descriptor for predicting the PPIs from protein amino acids sequences. Specifically, a Zernike moments (ZM) descriptor is used to extract protein evolutionary information from Position-Specific Scoring Matrix (PSSM) generated by Position-Specific Iterated Basic Local Alignment Search Tool (PSI-BLAST). Then, PCVM classifier is used to infer the interactions among protein. When performed on PPIs datasets of Yeast and H. Pylori, the proposed method can achieve the average prediction accuracy of 94.48% and 91.25%, respectively. In order to further evaluate the performance of the proposed method, the state-of-the-art support vector machines (SVM) classifier is used and compares with the PCVM model. Experimental results on the Yeast dataset show that the performance of PCVM classifier is better than that of SVM classifier. The experimental results indicate that our proposed method is robust, powerful and feasible, which can be used as a helpful tool for proteomics research.

Entities:  

Keywords:  position-specific scoring matrix; probabilistic classification vector machines; proteins

Mesh:

Substances:

Year:  2017        PMID: 28492483      PMCID: PMC5454941          DOI: 10.3390/ijms18051029

Source DB:  PubMed          Journal:  Int J Mol Sci        ISSN: 1422-0067            Impact factor:   5.923


1. Introduction

Recognition of protein–protein interactions (PPIs) is essential for elucidating the function of proteins and further understanding the various biological processes in cells. In the last decade, a variety of biological methods have been used for large-scale PPIs detection, such as tandem affinity purification [1], yeast two-hybrid systems [2,3], and protein chip [4]. For the limit of the experimental technique, these methods have some disadvantages, including high cost and time-intensive, as well as high rates of both false-positive and false-negative. Hence, computational methods for the detection of protein interactions have become hot research topics of proteomics research. So far, a number of computational methods have been presented for the detection of PPIs based on different data types, such as protein domains, protein structure information, genomic information and phylogenetic profiles [5,6,7,8,9,10,11,12,13]. However, these approaches cannot be achieved unless prior information of the protein is available. Hence, the mentioned methods are not widespread. Compared to the rapid growth of a large number of protein sequences, other data that can be used to predict the PPIs are scarce. Therefore, computational methods using only protein amino acid sequence information for PPIs prediction is especially interesting [14]. Bock and Gough used a support vector machine (SVM) with protein sequence descriptors to predict PPIs [15]. Martin et al. proposed an approach to predict PPIs by using signature product, which is a descriptor that extends from signature descriptors [16]. Najafabadi et al. attempted to solve this problem with Bayesian network [17]. Shen et al. adopted a SVM model to predict PPI network by combining Skernel function of protein pairs with a conjoint triad feature [18]. Yu-An Huang et al. developed a method by combining discrete cosine transform and using weighted sparse representation-based classifier to predict PPIs, and it has achieved very exciting prediction accuracy when applying this method to detecting yeast PPIs [19]. Yan-Zhi Guo et al. also obtained promising prediction results by adopting support vector machine and auto covariance [20]. Loris Nanni et al. developed several matrix-based protein representation methods, including [21,22,23,24,25]. Other feature extraction approaches based on protein sequence have been proposed in [26,27,28,29,30,31,32,33,34]. In this study, a novel computational approach for predicting PPIs from amino acid sequences based on a probabilistic classification vector machines model (PCVM) and a Zernike moments descriptor (PCVMZM) was proposed. The major improvement is the development of a more accurate protein sequence representation. Specifically, we employed the Zernike moments feature representation on a Position-Specific Scoring Matrix (PSSM) to extract the evolutionary information from protein sequence, and then a probabilistic classification vector machines classifier is used to infer the PPIs. In more detail, a PSSM representation is used to represent each protein. Afterward, for the sake of obtaining more representative information, we apply a Zernike moments descriptor to extract features in each protein PSSM and use Zernike moments of 12-order information and generate a 42-dimensional feature vector. Finally, we adopt the machine learning method called PCVM to accomplish classification. The proposed method was applied to Yeast and H. Pylori PPIs datasets. The experiments have shown that a PCVM prediction model with a Zernike moments descriptor yields fantastic performance. By further contrast experiment, we found that our proposed method was superior to the state-of-the-art SVM, which clearly shows that the proposed approach is trustworthy in predicting PPIs [35,36,37,38,39].

2. Results and Discussion

2.1. Evaluation Measure

The proposed method is evaluated against the following criteria: The Accuracy (Acc), Sensitivity (Sen), Precision (Pre), and Matthew`s correlation coefficient (MCC). All the computational formula is defined as follows: where represents the number of true positive, that true samples are predicted correctly, represents the number of true negative that true noninteracting pairs are predicted correctly. represents the number of false positive that non-interacting pairs are predicted to be interaction. represents the number of false negative that interacting pairs are predicted to be non-interacting. In addition, the receiver operating characteristic (ROC) curve [40] is applied to evaluate the performance of our method. The area under an ROC curve (AUC) [41] also is computed.

2.2. Assessment of Prediction

In order to make our method more reliable, five-fold cross-validation was adopted to divide a whole dataset into five parts. Hence, we obtained five models through separate experiments for each data set. The prediction result of PCVM prediction models with a Zernike moments description of protein sequence on Yeast and H. Pylori datasets are shown in Table 1 and Table 2. From Table 1, we can see that our proposed method achieved a good performance on the Yeast dataset. Its average accuracy, sensitivity, precision, and MCC are 94.48%, 95.13%, 93.92% and 89.58%, respectively. When using our proposed method on the H. Pylori dataset, as shown in Table 2, we also achieved some satisfactory results of average accuracy, sensitivity, precision, and MCC of 91.25%, 92.05%, 90.60% and 84.04%, respectively.
Table 1

Fivefold cross validation results using the proposed method on Yeast dataset.

Testing SetAcc (%)Sen (%)Pre (%)MCC (%)
196.3897.2195.5793.02
294.0595.2392.7788.81
393.0796.7390.2787.06
494.4694.2094.7189.53
594.4292.2696.2689.46
Average94.48 ± 1.295.13 ± 2.093.92 ± 2.489.58 ± 2.2
Table 2

Fivefold cross validation results using the proposed method on H. Pylori dataset.

Testing SetAcc (%)Sen (%)Pre (%)MCC (%)
189.5492.1186.8281.24
292.1192.6891.4185.46
391.0891.1691.1683.75
491.4292.2590.3484.31
592.1292.0493.2385.42
Average91.25 ± 1.192.05 ± 0.690.06 ± 2.484.04 ± 1.7
From the experimental results, it can be seen that our proposed approach is robust, accurate and practical for predicting PPIs. The outstanding performance for detecting PPIs can be put down to the feature extraction and the classification model of our proposed method. It is effective that Zernike moments are used for feature extraction, and the PCVM model is accurate and robust in dealing with classification problems.

2.3. Comparison with the Support Vector Machine (SVM)-Based Method

In order to further evaluate the prediction performance of the proposed entire model, the SVM model is adopted based on the Yeast dataset to predict PPIs using the same Zernike moments to extract feature, and then, we compared the classification result between PCVM and SVM. We employed the SVM through the library for Support Vector Machines (LIBSVM) tool [42]. SVM have two parameters, c and g, respectively. A grid search method is used to optimize parameters c and g. In our experiment, a radial basis function is used as the kernel function and the initial value c and g was set to 0.4 and 0.5. Table 3 gives the prediction results of five-fold cross-validation over two different classification methods on the Yeast dataset. From Table 3, we can see that the classification method of SVM achieved 89.31% average accuracy, 87.54% average sensitivity, 90.81% average precision, 80.91% average MCC. While the classification results of the PCVM method achieved 94.48% average accuracy, 95.13% average sensitivity, 93.92% average precision, 89.58% average MCC. Experimental results show that PCVM classification method is significantly better than the SVM classification method. Comparison of ROC curves performed between RVM and SVM on the Yeast dataset from Figure 1 and Figure 2, we have experimental data obtained that the PCVM classifier is more accurate and robust than the SVM classifier for detecting PPIs.
Table 3

Five-fold cross-validation results by using two models on the Yeast dataset.

ModelTesting SetAcc (%)Sen (%)Pre (%)MCC (%)
Probabilistic Classification Vector Machines (PCVM)196.3897.2195.5793.02
294.0595.2392.7788.81
393.0796.7390.2787.06
494.4694.2094.7189.53
594.4292.2696.2689.46
Average94.48 ± 1.295.13 ± 2.093.92 ± 2.489.58 ± 2.2
Support Vector Machin (SVM)189.2387.7590.2780.76
290.4888.7391.4982.74
387.6287.3788.0778.30
489.6388.0590.9781.40
589.6085.7993.2381.32
Average89.31 ± 1.787.54 ± 1.190.81 ± 1.980.91 ± 1.62
Figure 1

Receiver operating characteristic (ROC) curves performed of a probabilistic classification vector machines model (PCVM) on the Yeast dataset.

Figure 2

ROC curves performed of support vector machine (SVM) on the Yeast dataset.

The main improvement is attributed to three points: (1) the main advantage of PCVM is that the truncated Gaussian priors are adopted to generate robust and sparse results—in other words, the number of weight vectors is less than SVM. Hence, the complexity of the model is reduced, besides, the model is more general; (2) The parameter optimization procedure of the PCVM based on EM algorithm and probabilistic inference not only can improve the performance, but also save the effort to do cross-validation; (3) The PCVM model is simpler and easier to be understood, because the number of basic functions does not grow linearly with the number of training points. In general, the PCVM is a sparse model that makes up the shortcoming of SVM without deskilling the generalization performance and provides probabilistic outputs. Here it is, our proposed approach can produce satisfactory results.

2.4. Comparison with Other Methods

In recent years, many classification methods have been developed to predict PPIs. To further validate the performance of our proposed method, we compared the predictive performance of our method with other existing several well-known methods. The achieved results of five-fold cross-validation of different methods on the Yeast dataset and H. pylori dataset are shown in Table 4 and Table 5. From Table 4, the prediction accuracy of other previous methods on the Yeast dataset varies from 75.08% to 93.92%, while the proposed method achieved higher value of 94.48%. Similarly, the sensitivity and MCC of our method are also higher than those of other methods. We can find similar results on the H. pylori dataset in Table 5. Our proposed method achieves 91.25% accuracy, which is higher than the other five methods with the highest prediction accuracy of 87.50%. The same is true for precision, sensitivity and MCC. All prediction results in Table 4 and Table 5 indicate that the PCVM classifier is stable and robust and can improve the prediction performance compared with the state-of-the-art methods. The improvement of prediction performance of our method may derive from the novel feature extraction method which extracts the highly discriminative information, and the use of PCVM classifier which ensures accurate and stable prediction.
Table 4

Practical predicting results of different methods on the Yeast dataset.

ModelTesting SetAcc (%)Sen (%)Pre (%)MCC (%)
Guo [20]Auto Covariance (ACC)89.33 ± 2.6789.93 ± 3.6888.87 ± 6.16N/A
auto covariance (AC)87.36 ± 1.3887.30 ± 4.6887.82 ± 4.33N/A
Yang [23]Cod175.08 ± 1.1375.81 ± 1.2074.75 ± 1.23N/A
Cod280.04 ± 1.0676.77 ± 0.6982.17 ± 1.35N/A
Cod380.41 ± 0.4778.14 ± 0.9081.66 ± 0.99N/A
Cod486.15 ± 1.1781.03 ± 1.7490.24 ± 1.34N/A
You [24]Principal Component Analysis-Ensemble Extreme Learning Machines (PCA-EELM)87.00 ± 0.2986.15 ± 0.4387.59 ± 0.3277.36 ± 0.44
Wong [30]Rotation Forest (RF) + Property Response-Local Phase Quantization (PR-LPQ)93.92 ± 0.3691.10 ± 0.3196.45 ± 0.4588.56 ± 0.63
Proposed MethodPCVM94.48 ± 1.2095.13 ± 2.0093.92 ± 2.4089.58 ± 2.20
Table 5

Practical predicting results of different methods on the H. Pylori dataset.

ModelAcc (%)Sen (%)Pre (%)MCC (%)
Nanni [23]83.0086.0085.10N/A
Nanni [32]84.0086.0084.00N/A
Nanni and Lumini [25]86.6086.7085.00N/A
Z-H You [29]87.5088.9586.1578.13
L Nanni [24]84.0084.0084.00N/A
Proposed Method91.2592.0590.0684.04

3. Materials and Methodology

3.1. Dataset

Up to now, many databases of PPIs data have been generated, such as Database of Interaction Proteins (DIP) [43], Molecular Interaction Database (MINT) [44], and Biomolecular Interaction Network Database (BIND) [45]. To evaluate our approach, we used two publicly available datasets: Yeast and H. Pylori, which were extracted from Database of Interaction Proteins (DIP). In order to ensure the reliability of the tests, we extract 5594 positive protein pairs to constitute the positive dataset and 5594 negative protein pairs to constitute the negative protein dataset from the Yeast dataset. Analogously, we extract 1458 positive protein pairs to constitute the positive dataset and 1458 negative protein pairs to constitute the negative protein dataset from the H. Pylori dataset. Therefore, the Yeast dataset consists of 11,188 protein pairs and the H. Pylori dataset consists of 2916 protein pairs.

3.2. Position-Specific Scoring Matrix

A Position-Specific Scoring Matrix (PSSM) was usually adopted to find distantly related proteins, protein disulfide, protein quaternary structural attributes and protein folding patterns [46,47,48,49]. In this paper, we also adopt PSSM to predict PPIs. Here, each protein was transformed into a PSSM matrix by employing the Position-Specific Iterated Basic Local Alignment Search Tool (PSI-BLAST) [50,51]. A PSSM is represented as where = , (i = 1, 2, …, 20). A PSSM contains × 20 elements, where denotes the length of an amino acid sequence and 20 columns are owing to 20 amino acids. The of the PSSM element is indicated as a score of th amino acid in the th position of the given protein sequence and it can be expressed as = where is the appearing frequency value of the amino acid at position i of the probe, and represents the value of Dayhoff’s mutation matrix [52] between the and the amino acids. Consequently, the higher the score, the better the conserved position [53,54,55]. In our study, the experiment datasets were built by using PSI-BLAST to transform each protein into a PSSM for detecting PPIs. To obtain more extensive homologous sequences, the e-value parameter of PSI-BLAST was set to 0.001 and chose three iterations. As a result, the PSSM of a protein sequence can be represented as a M × 20 matrix, where M is the number of residues and each column represents an amino acid [56,57,58,59].

3.3. Zernike Moments

Zernike moments have an exciting performance in the field of image recognition for extract image feature, because it is robust against rotation and it can represent information from different angles. In this paper, we first introduced Zernike moments to extract significant information from protein sequences. In this section, Zernike moments and their principal properties are described, and we illustrate how to achieve the rotation invariance. Finally, we describe the process of feature selection.

3.3.1. Invariance of Normalized Zernike Moment

The principle of Zernike moments [60,61,62,63] is Zernike polynomials [64,65,66], that is a set of complete orthogonal polynomials within the unit circle. In two-dimensional space, these polynomials can be expressed as {} and expression is as follows: where n is a nonnegative integer and m is an integer subject to constraints n−|m| even, |m| ≤ n. Here, {} is a radial polynomial in the form of Note that The set of polynomials are orthogonal, i.e., With The two-dimensional Zernike moments for continuous function f () are the projection of f () onto these orthogonal basis function and denoted by Correspondingly, for a digital function, the two-dimensional Zernike moments are represented by To compute the Zernike moments of a PSSM matrix [67,68,69,70], the center of the matrix is taken as the origin and coordinates are mapped into a unit circle, i.e., . Those values of matrix falling outside the unit disk are not used in the computation. Note that

3.2.2. Introduction of a Zernike Moments Descriptor

When we define as the rotated function, the equivalence between original and rotated function is The Zernike moments of the rotated function become Equation (13) indicates that Zernike moments only need phase shift on rotation. Therefore, the magnitude of the Zernike moment, ||, can be adopted as rotation-invariant feature. Therefore, after moving the origin of PSSM matrix into the centroid, we can compute the Zernike moments and the magnitudes of the moments are rotation-invariant [71,72].

3.2.3. Feature Selection

According to the foregoing, we have known that the magnitudes of Zernike moments can be used as rotation-invariant features. One problem that must be considered is how big should N be? The lower-order moments extract gross information and high details information are captured by higher-order moments. In our experiments, N is set to 12. We can obtain 42 features from each protein sequence. The feature vector be represented as: where || represent the Zernike moments magnitude. Here, we do not consider the case of m = 0, because they do not include useful information regarding the PPIs and Zernike moments with m < 0 have not been considered, because they are inferred through Hence, the dimension of the feature vector is 42 [73]. The obtained Zernike moments is shown in Table 6.
Table 6

List of Zernike Moments (ZMs) sorted by n and m in sequence for the case where (n, m) = (12, 12).

NMomentsNo.NMomentsNo.
1A1117A71,A73,A75,A774
2A2218A82,A84,A86,A884
3A31,A3329A91,A93,A95,A97,A995
4A42,A44210A10,2,A10,4,A10,6,A10,8,A10,105
5A51,A53,A55311A11,1,A11,3,A11,5,A11,7,A11,9,A11,116
6A62,A64,A66312A12,2,A12,4,A12,6,A12,8,A12,10,A12,126

3.4. Related Machine Learning Models

In the field of machine learning, the Support Vector Machines (SVM) [74] are acknowledged as an excellent supervision model in pattern recognition, classification, and regression analysis. However, there are certain apparent disadvantages when using this method: (1) the count of support vectors grows linearly with the scale of the training set; (2) Outputs of the SVMs are not probabilistic; (3) The parameters of kernel function need to be optimized by cross-validation, the procedure wastes a lot of computing resources. Compared with SVM, the Relevance Vector Machines (RVM) [75] based on Bayesian technique can avoid these problems. The RVM method takes advantage of the Bayesian automatic relevance determination (ARD) [76] framework and gives a zero-mean Gaussian prior over every weight to produce a sparse solution. However, for a classification problem, the zero-mean Gaussian prior are given over weights for negative and positive classes, which leads to a problem that some training points belonging to negative classes may be given positive weights and vice-versa. Under this circumstance, it may give rise to produce some unreliable vectors for the decision of RVMs. For the sake of addressing this problem and proposing an appropriate probabilistic model for predicting PPIs, we first adopt the Probabilistic Classification Vector Machine (PCVM) classifier which gives different priors over weights for training points that belong to different classes, i.e., the non-negative, left-truncated Gaussian is used for the positive class and the non-positive, right-truncated Gaussian is used for the negative class. PCVM provides many advantages: (1) PCVM produces the probabilistic outputs for each test point; (2) It is effective that PCVM used expectation maximization (EM) algorithm to optimizing kernel parameters; (3) PCVM introduced a sparser model leading to faster performance in the test stage.

3.5. PCVM Algorithm

PCVM is a classification model that supervised learning. Hence, we need a set of input-target training pairs , where = {−1, +1} to train a learning model f (x; w), which is defined by parameters W. The model is a linear combination of N basis functions and is represented as where the {} is basis function, (wherein represent the parameter vector of the basis function), the W = is the parameter of the PCVM model, the b is the bias. In this paper, we adopt the radial basis function (RBF) [77] as the basis and adopt the probit link function to obtain the binary outputs. Finally, mapping the f (x; w) into , the expression of the PCVM model becomes: A truncated Gaussian distribution as a prior is employed over each weight as follow A zero-mean Gaussian distribution as a prior is employed over the bias b: The is a truncated Gaussian function, is the precision of the corresponding parameter , represents the precision of the normal distribution of b. When , the truncated prior is a non-negative, left-truncated Gaussian, and when , the prior is a non-positive, right-truncated Gaussian. This can be represented as The gamma distribution is adopted as the hyper prior of α and β. Using the EM algorithm, assign the parameters of a PCVM model, such as parameters b, W and θ. The EM algorithm is an iterative algorithm, which is used to estimate the maximum likelihood or maximum posterior probability involving latent variables. For more details about the PCVM theory, please refer to [78,79].

3.6. Initial Parameter Selection and Training

The PCVM algorithm has only one parameter, θ, which can be optimized automatically in the training process. However, the EM algorithm is susceptible to initial point and trap in local maxima. Choosing the best initialization point is an effective method to avoid the local maxima. We train a PCVM model with eight initialization points over the five training folds of each data. Hence, we obtain a 5 × 8 matrix of parameters, where the rows represent the folds and the columns represent the initializations. For each row, we select the results of the lowest test error. Hence, we find only five points, and then, we select the medium over those parameters. We have experimental obtained the optimal initial value which is seted as 3.6 on the Yeast dataset and 1.18 on the H. pylori dataset.

4. Conclusions

Considering time, efficiency and economy, the use of computational methods based on protein amino acid sequences to predict PPIs has attracted the attention of researchers. The computational method is playing an important role in proteomics research, because it saves manpower and material resources and is more accurate and efficient. In this paper, we introduce an accurate computational method based on protein sequence. It is established by using a PCVM classifier combined with a Zernike moments descriptor on the PSSM. The experiments showed that the performance of our proposed method achieves a high classification accuracy and is superior to the SVM. The main improvements of the developed approach come from adopting a Zernike moments descriptor as feature extraction approach that can capture multi-angle useful and representative information. More than this, the use of a PCVM classifier ensures more reliable and accurate recognition, because the use of the truncated Gaussian priors can lead to obtaining robust and sparse results—the number of support vectors is less than SVM, and the probabilistic outputs produced by PCVM can assess the uncertainty of prediction on the skewed dataset. In addition, the parameter optimization procedure of the PCVM not only can improve the performance, but also save effort to do cross-validation. Due to the outstanding performance of the Zernike moments descriptor and PCVM, our method can improve the PPIs accuracy rate. All in all, our proposed method is highly efficient and stable and can be a useful tool for predicting PPIs.
  51 in total

1.  Protein secondary structure prediction based on position-specific scoring matrices.

Authors:  D T Jones
Journal:  J Mol Biol       Date:  1999-09-17       Impact factor: 5.469

2.  Predicting protein interaction sites from residue spatial sequence profile and evolution rate.

Authors:  Bing Wang; Peng Chen; De-Shuang Huang; Jing-jing Li; Tat-Ming Lok; Michael R Lyu
Journal:  FEBS Lett       Date:  2005-12-19       Impact factor: 4.124

3.  Predicting protein-protein interactions from protein domains using a set cover approach.

Authors:  Chengbang Huang; Faruck Morcos; Simon P Kanaan; Stefan Wuchty; Danny Z Chen; Jesús A Izaguirre
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2007 Jan-Mar       Impact factor: 3.710

4.  PSI-BLAST tutorial.

Authors:  Medha Bhagwat; L Aravind
Journal:  Methods Mol Biol       Date:  2007

5.  Representation of videokeratoscopic height data with Zernike polynomials.

Authors:  J Schwiegerling; J E Greivenkamp; J M Miller
Journal:  J Opt Soc Am A Opt Image Sci Vis       Date:  1995-10       Impact factor: 2.129

6.  Assessing and predicting protein interactions by combining manifold embedding with multiple information integration.

Authors:  Ying-Ke Lei; Zhu-Hong You; Zhen Ji; Lin Zhu; De-Shuang Huang
Journal:  BMC Bioinformatics       Date:  2012-05-08       Impact factor: 3.169

7.  A semi-supervised learning approach to predict synthetic genetic interactions by combining functional and topological properties of functional gene network.

Authors:  Zhu-Hong You; Zheng Yin; Kyungsook Han; De-Shuang Huang; Xiaobo Zhou
Journal:  BMC Bioinformatics       Date:  2010-06-24       Impact factor: 3.169

8.  Large-scale protein-protein interactions detection by integrating big biosensing data with computational model.

Authors:  Zhu-Hong You; Shuai Li; Xin Gao; Xin Luo; Zhen Ji
Journal:  Biomed Res Int       Date:  2014-08-18       Impact factor: 3.411

9.  Prediction of Protein Structural Class Based on Gapped-Dipeptides and a Recursive Feature Selection Approach.

Authors:  Taigang Liu; Yufang Qin; Yongjie Wang; Chunhua Wang
Journal:  Int J Mol Sci       Date:  2015-12-24       Impact factor: 5.923

10.  An ensemble approach for large-scale identification of protein- protein interactions using the alignments of multiple sequences.

Authors:  Lei Wang; Zhu-Hong You; Xing Chen; Jian-Qiang Li; Xin Yan; Wei Zhang; Yu-An Huang
Journal:  Oncotarget       Date:  2017-01-17
View more
  25 in total

1.  Machine-learning techniques for the prediction of protein-protein interactions.

Authors:  Debasree Sarkar; Sudipto Saha
Journal:  J Biosci       Date:  2019-09       Impact factor: 1.826

2.  Multi-view heterogeneous molecular network representation learning for protein-protein interaction prediction.

Authors:  Xiao-Rui Su; Lun Hu; Zhu-Hong You; Peng-Wei Hu; Bo-Wei Zhao
Journal:  BMC Bioinformatics       Date:  2022-06-16       Impact factor: 3.307

3.  Protein-protein interaction and non-interaction predictions using gene sequence natural vector.

Authors:  Nan Zhao; Maji Zhuo; Kun Tian; Xinqi Gong
Journal:  Commun Biol       Date:  2022-07-02

4.  Special Protein Molecules Computational Identification.

Authors:  Quan Zou; Wenying He
Journal:  Int J Mol Sci       Date:  2018-02-10       Impact factor: 5.923

5.  A Deep Learning Framework for Robust and Accurate Prediction of ncRNA-Protein Interactions Using Evolutionary Information.

Authors:  Hai-Cheng Yi; Zhu-Hong You; De-Shuang Huang; Xiao Li; Tong-Hai Jiang; Li-Ping Li
Journal:  Mol Ther Nucleic Acids       Date:  2018-03-09       Impact factor: 8.886

6.  PCLPred: A Bioinformatics Method for Predicting Protein-Protein Interactions by Combining Relevance Vector Machine Model with Low-Rank Matrix Approximation.

Authors:  Li-Ping Li; Yan-Bin Wang; Zhu-Hong You; Yang Li; Ji-Yong An
Journal:  Int J Mol Sci       Date:  2018-03-29       Impact factor: 5.923

7.  Protein-Protein Interactions Prediction Based on Graph Energy and Protein Sequence Information.

Authors:  Da Xu; Hanxiao Xu; Yusen Zhang; Wei Chen; Rui Gao
Journal:  Molecules       Date:  2020-04-16       Impact factor: 4.411

8.  Computational predictions for protein sequences of COVID-19 virus via machine learning algorithms.

Authors:  Heba M Afify; Muhammad S Zanaty
Journal:  Med Biol Eng Comput       Date:  2021-07-22       Impact factor: 2.602

9.  Protein-Protein Interactions Prediction Using a Novel Local Conjoint Triad Descriptor of Amino Acid Sequences.

Authors:  Jun Wang; Long Zhang; Lianyin Jia; Yazhou Ren; Guoxian Yu
Journal:  Int J Mol Sci       Date:  2017-11-08       Impact factor: 5.923

10.  Identifying RNA N6-Methyladenosine Sites in Escherichia coli Genome.

Authors:  Jidong Zhang; Pengmian Feng; Hao Lin; Wei Chen
Journal:  Front Microbiol       Date:  2018-05-14       Impact factor: 5.640

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.