Literature DB >> 30795499

Prediction of Self-Interacting Proteins from Protein Sequence Information Based on Random Projection Model and Fast Fourier Transform.

Zhan-Heng Chen1,2, Zhu-Hong You3,4, Li-Ping Li5, Yan-Bin Wang6, Leon Wong7,8, Hai-Cheng Yi9,10.   

Abstract

It is significant for biological cells to predict self-interacting proteins (SIPs) in the field of bioinformatics. SIPs mean that two or more identical proteins can interact with each other by one gene expression. This plays a major role in the evolution of protein‒protein interactions (PPIs) and cellular functions. Owing to the limitation of the experimental identification of self-interacting proteins, it is more and more significant to develop a useful biological tool for the prediction of SIPs from protein sequence information. Therefore, we propose a novel prediction model called RP-FFT that merges the Random Projection (RP) model and Fast Fourier Transform (FFT) for detecting SIPs. First, each protein sequence was transformed into a Position Specific Scoring Matrix (PSSM) using the Position Specific Iterated BLAST (PSI-BLAST). Second, the features of protein sequences were extracted by the FFT method on PSSM. Lastly, we evaluated the performance of RP-FFT and compared the RP classifier with the state-of-the-art support vector machine (SVM) classifier and other existing methods on the human and yeast datasets; after the five-fold cross-validation, the RP-FFT model can obtain high average accuracies of 96.28% and 91.87% on the human and yeast datasets, respectively. The experimental results demonstrated that our RP-FFT prediction model is reasonable and robust.

Entities:  

Keywords:  fast Fourier transform; position-specific scoring matrix; random projection; self-interacting proteins

Mesh:

Substances:

Year:  2019        PMID: 30795499      PMCID: PMC6412412          DOI: 10.3390/ijms20040930

Source DB:  PubMed          Journal:  Int J Mol Sci        ISSN: 1422-0067            Impact factor:   5.923


1. Introduction

Protein is an important component of all cells. It is an organic macromolecule and the basic material of life. It also is the main undertaker of activity. Without protein, there is no life. Most proteins often work together with a partner or other proteins. They can interact with two or more copies by themselves, which is termed self-interacting proteins (SIPs). However, for most researchers, whether proteins can interact with each other is a difficult thing to determine. SIPs play a key role in the development of protein interaction networks (PINs) [1,2]. The functions of many proteins, which could control the transport of ions and small molecules that pass through cell membranes, depends on their homo-oligomers [3]. Ispolatov et al. discovered that the average quantity of SIPs is more than twice that of other proteins in the PINs [4]. It is crucial for elucidating the functions of SIPs to comprehend whether a protein can self-interact; this also gives us an insight into the adjustment of protein function and can help us achieve a better comprehension of disease mechanisms [5]. Over the past few years, many studies have shown that homo-oligomerization plays an important role in many biological processes, such as signal transduction, gene expression regulation, immune response, and enzyme activation [6,7,8,9]. Therefore, SIPs will be useful for improving steadiness and preventing against cellular stress and the denaturation of proteins via reducing the surface area [10]. So far, there are many ways to study bioinformatics [11,12,13,14,15,16] and genomics [17,18,19,20,21,22], and a number of previous methods for predicting PPIs have been put forward. For example, Pitre et al. [23] raised a new Protein‒Protein Interaction Prediction Engine (PIPE), which could predict PPIs for any target pair of the yeast S. cerevisiae proteins from their primary structure and without the need for any additional information or predictions about the proteins. Xia et al. [24] put forward a sequence-based multi-classifier system that applied auto-correlation descriptor to encode a protein interaction pair and selected rotation forest as classifier to deduce PPIs. However, these methods are good for PPI detection [25] but have certain limitations in that they must take the correlation between protein pairs into account for Protein Self-interaction detection—for example, co-expression, co-localization, and co-evolution. Nevertheless, this information is useless for SIPs. Moreover, the datasets for PPI detection are balanced and those of SIPs are unbalanced. Besides, prediction of PPIs datasets has no PPIs between the same partners. For these reasons, the above computational models are not suitable for SIPs detection. Accordingly, it is becoming more and more crucial to exploit an effective calculation method to predict SIPs. In our study, a random projection (RP) method for SIPs prediction from protein sequence information with Fast Fourier Transform (FFT) was proposed. Furthermore, the main idea of our proposed method includes four aspects: (1) the protein sequence information could be described as a Position-Specific Scoring Matrix (PSSM); (2) using the fast Fourier transform (FFT) method to extract eigenvectors from protein sequences on a PSSM; (3) using the Principal Component Analysis (PCA) approach to convert the high-dimensional data into useful information after FFT and the noise is removed, so the pattern in the data is found; (4) the RP algorithm is employed to build a training set where the classifier will be trained. Take it in detail as follows: first, the PSSM from each protein sequence is likely to result in a eigenvector whose dimension is 400 by applying the FFT method for extracting important information; then, reduce the dimension of the FFT vector to 300 for improving the performance of prediction by employing the PCA dimensionality reduction method; eventually, perform classification on yeast and human datasets by applying the RP classifier. The results demonstrate that this method outperforms the SVM-based approach and six other existing technologies. This indicates that the proposed model is suitable and performs well for predicting SIPs.

2. Results and Discussion

2.1. Performance Evaluation

In this study, to estimate the stability and availability of our prediction model, we used five measurements that were commonly used in binary classification tasks, including accuracy (Acc.), sensitivity (Sen.), specificity (Spe.), Matthews correlation coefficient (MCC) [26,27,28,29,30,31,32], and Balanced Accuracy (B_Acc.) [33], respectively. They could be defined as follows: where TP represents the count of true positives, that is to say the number of real interacting pairs predicted correctly. FP is the quantity of false positives, defined as the volume of real non-interacting pairs mis-predicted. TN stands for the count of true negatives, which is the quantity of real non-interacting pairs correctly predicted. FN means the quantity of false negatives; in other words, it represents the true sample error predicted to be false samples. On the basis of these parameters, a Receiver Operating Curve (ROC) was plotted to assess the performance of the random projection approach. Then we can compute the area under the curve (AUC) to estimate the quality of the classifier.

2.2. Performance of the Proposed Method

In order to evaluate the performance of the presented model and avoid the overfitting problem, we applied the RP-FFT model to the human dataset. In statistical prediction, three cross-validation (CV) methods, such as an independent dataset test, a sub-sampling (or k-fold CV) test, and a leave-one-out CV (LOOCV) test, are frequently used to calculate the expected success rate of a developed predictor [34,35,36,37,38]. Among the three methods, however, the LOOCV test is deemed the least arbitrary and most objective, as demonstrated by Equations (28)–(32) of [39], and hence it has been widely recognized and increasingly adopted by investigators to examine the quality of various predictors [38,40,41]. However, it seems time- and resource-consuming. Thus, we used 5-fold CV to examine the proposed models. In 5-fold CV, the benchmarking dataset was randomly partitioned into 10 subsets. One subset is used as a test set and the remaining nine subsets are used as the training sets. This procedure is repeated five times, where each subset is used once as a test set. The performance of the five corresponding results is averaged to give the performance of the classifier. To assess the feasibility and stability of our prediction method, we also estimated the prediction performance of RP-FFT model on the yeast dataset. To ensure the fairness of the experiment, we optimized a number of parameters for the RP-FFT prediction model. In this paper, we set up the same parameters for human and yeast datasets. Thus, we classify the training and test sets for B1 = 10 independent projections, each one carefully chosen from a block of size B2 = 30, and then chose the K-Nearest Neighbor (KNN) as the base classifier and the leave-one-out test error estimate, where k = seq (1, 40, by = 3). Our model can not only deal with balanced data, but can also solve the imbalanced data problem to some extent. At first, we employed the undersampling technique, as mentioned in [18], to solve the imbalanced dataset problem. The human dataset included 1441 SIPs as positives and 1441 non-SIPs as negatives. Using the same strategy, the yeast dataset contained 710 positive samples and 710 negative samples. The experimental results can be seen in Table 1 and Table 2.
Table 1

The results of the RP-FFT method with 5-fold cross-validation on the human dataset.

Testing SetAcc. (%)Sen. (%)Spe. (%)MCC (%)
194.4488.28100.0089.36
292.5385.37100.0086.07
392.1985.48100.0085.51
493.7586.76100.0088.08
594.8189.73100.0090.12
Average93.54 ± 1.1587.12 ± 1.87100.00 ± 0.0087.83 ± 2.01
Table 2

The results of the RP-FFT method with 5-fold cross-validation on the yeast dataset.

Testing SetAcc. (%)Sen. (%)Spe. (%)MCC (%)
180.9997.1265.5265.71
283.4592.1475.0068.03
382.0497.8966.2067.57
484.8695.1474.2971.13
583.4592.4174.1067.83
Average82.96 ± 1.4894.94 ± 2.6371.02 ± 4.7368.05 ± 1.95
In addition, the initial imbalanced data collected from DIP, BioGRID, IntAct, InnateDB, and MatrixDB also used to compare our proposed method with previous work. If we use the undersampling technique to reconstruct the dataset, the size of the initial imbalanced data will be substantially reduced. As shown in Table 2 and Table 3, we performed our proposed model on the initial imbalanced data in the experiment.
Table 3

The results of the RP-FFT method with 5-fold cross-validation on the human dataset.

Testing SetAcc. (%)Sen. (%)Spe. (%)MCC (%)B_Acc. (%)
196.2379.5197.7475.7288.63
296.2080.3497.6575.8989.00
396.5882.4997.8978.6190.19
496.4079.7897.7975.4088.79
596.0085.2897.0176.6891.15
Average96.28 ± 0.2281.48 ± 2.4397.62 ± 0.3576.46 ± 1.2989.55 ± 1.08
The experimental results of the RP-FFT prediction model on the human and yeast datasets are listed in Table 3 and Table 4. Table 3 lists the data obtained that the model put forward obtained for average Accuracy (Acc.), Sensitivity (Sen.), Specificity (Spe.), Matthews correlation coefficient (MCC), and Balance accuracy (B_Acc.): 96.28%, 81.48%, 97.62%, 76.46%, and 89.55% for the human dataset and the standard deviations of them 0.22%, 2.43%, 0.35% 1.29%, and 1.08%, respectively. In the same way, we also got good results in Table 4 for average Acc., Sen., Spe., MCC and B_Acc.: 91.87%, 48.81%, 97.42%, 54.62% and 73.12%, and the standard deviations of them are 0.82%, 4.50%, 0.45%, 4.25%, and 2.30% for the yeast dataset, respectively.
Table 4

The results of the RP-FFT method with 5-fold cross-validation on the yeast dataset.

Testing SetAcc. (%)Sen. (%)Spe. (%)MCC (%)B_Acc. (%)
191.3250.0096.7353.0973.37
291.7247.3397.8155.3572.57
392.2049.6397.3954.8073.51
491.0042.3697.3649.0669.86
593.0954.7497.8360.8276.29
Average91.87 ± 0.8248.81 ± 4.5097.42 ± 0.4554.62 ± 4.2573.12 ± 2.30
From the above data, it is obvious that the proposed method could achieve good outcomes for SIPs predictions due to the suitable feature extraction and classifier. It can be summarized that the main improvement of our characteristic extraction technique contains the following factors: (1) The PSSM gives the score for finding a special matching amino acid in a target protein sequence. It is a good tool that can not only represent the protein sequence information but also saves enough prior information. Therefore, a PSSM contains all the major information of one protein sequence for detecting SIPs. (2) We extracted the features from the protein sequence by using the Fast Fourier Transform (FFT) method, which can further increase the performance of the RP-FFT model. (3) In case of ensuring the integrity information of FFT feature vector, we used Principal Component Analysis (PCA) to decrease the dimension of data and influence of noise, and thus the pattern in the data is found. Experimental results revealed that the eigenvector extracted from applying FFT on PSSM is quite suitable for SIP detection.

2.3. Comparison with Other Feature Extraction Methods

In this section, in order to illustrate the use of the FFT feature extraction method, we compared the FFT method with SVD (Singular Value Decomposition), DCT (Discrete Cosine Transform), and COV (Covariance) [42,43] on the Random Projection classifier. The results of RP classifier based on different feature extraction methods with 5-fold cross-validation on the yeast dataset are shown in Table 5. On the whole, it can be seen that the FFT feature extraction method works better than other methods for the yeast dataset.
Table 5

The results of RP classifier based on different feature extraction methods on the yeast dataset.

Feature Extraction MethodsAcc. (%)Sen. (%)Spe. (%)MCC (%)B_Acc. (%)
SVD88.73 ± 0.7510.25 ± 2.9398.86 ± 0.4319.76 ± 2.9654.55 ± 1.31
DCT90.35 ± 0.8420.38 ± 2.6299.36 ± 0.3237.57 ± 1.7459.87 ± 1.18
COV91.93 ± 0.8142.43 ± 4.8298.31 ± 0.2553.10 ± 4.9170.37 ± 2.49
FFT91.87 ± 0.8248.81 ± 4.5097.42 ± 0.4554.62 ± 4.2573.12 ± 2.30

2.4. Comparison with the SVM-Based Method

Though the RP-FFT model achieved better performance for predicting SIPs, we still need to further assess its use with our presented method. The veracity and stability of prediction of the RP classifier were compared with the state-of-the-art SVM method via the same characteristic extraction approach based on the yeast and human datasets, respectively. We applied the LIBSVM packet tool [44] to run the classification. Before the experiment, there are several parameters of SVM classifier should be optimized. In this paper, we chose a radial basis function (RBF) as the kernel function, and then used grid search to optimize the parameters of RBF, whose parameters were set to c = 0.03 and g = 1200. As shown in Table 6 and Table 7, we employed 5-fold cross-validation to train and compare the models of RP and SVM on the yeast and human datasets, respectively. The average Acc., the average Sen., the average Spe., the average MCC and B_Acc. of SVM classifier are 93.68%, 23.80%, 100.00%, 47.13%, and 61.90% on the human dataset in Table 6, respectively. Nevertheless, the RP classifier obtained 96.28% average Acc., 81.48% average Sen., 97.62% average Spe., 76.46% average MCC, and 89.55% average B_Acc. On the human dataset. Similarity, the average Accuracy, the average Sen., the average Spe., the average MCC and B_Acc. of SVM classifier are 90.63%, 17.79%, 100.00%, 39.95%, and 58.90% on the yeast dataset in Table 7. Nevertheless, the RP classifier received 91.87% average Acc., 48.81% average Sen., 97.42% average Spe., 54.62% average MCC and 73.12% average B_Acc. On the human dataset. In a word, it is obvious that the overall prediction result of RP classifier is much better than that of the SVM method.
Table 6

Performance comparison of RP and SVM on the human dataset.

ModelTesting SetAcc. (%)Sen. (%)Spe. (%)MCC (%)B_Acc. (%)
RP + FFT196.2379.5197.7475.7288.63
296.2080.3497.6575.8989.00
396.5882.4997.8978.6190.19
496.4079.7897.7975.4088.79
596.0085.2897.0176.6891.15
Average96.28 ± 0.2281.48 ± 2.4397.62 ± 0.3576.46 ± 1.2989.55 ± 1.08
SVM + FFT193.5522.22100.0045.5761.11
293.6423.79100.0047.1761.90
393.2120.54100.0043.7360.27
494.1924.34100.0047.8662.17
593.8228.09100.0051.3064.05
Average93.68 ± 0.3623.80 ± 2.82100.00 ± 0.0047.13 ± 2.8261.90 ± 1.41
Table 7

Performance comparison of RP and SVM on the yeast dataset.

ModelTesting SetAcc. (%)Sen. (%)Spe. (%)MCC (%)B_Acc. (%)
RP+FFT191.3250.0096.7353.0973.37
291.7247.3397.8155.3572.57
392.2049.6397.3954.8073.51
491.0042.3697.3649.0669.86
593.0954.7497.8360.8276.29
Average91.87 ± 0.8248.81 ± 4.5097.42 ± 0.4554.62 ± 4.2573.12 ± 2.30
SVM+FFT190.1114.58100.0036.2257.29
290.8424.00100.0046.6262.00
390.7614.81100.0036.6457.41
490.5118.06100.0040.3859.03
590.9217.52100.0039.8758.76
Average90.63 ± 0.3317.79 ± 3.80100.00 ± 0.0039.95 ± 4.1758.90 ± 1.90
Meanwhile, the ROC curves between RP and SVM on the human and yeast datasets are displayed in Figure 1 and Figure 2. From Figure 1, it is clear that the average area under the curve (AUC) of SVM classifier is 0.6190 and that of the RP classifier is 0.8955. From Figure 2, we can see that the average AUC of SVM classifier is 0.5890 and that of the RP classifier is 0.7312. It is obvious that the average AUC of RP method is also larger than the AUC of the SVM method. So Random Projection is an accurate and robust method for SIP detection.
Figure 1

Comparison of ROC curves between RP and SVM on human (5-fold cross validation). (a) is the ROC curve of SVM method on human dataset by 5-fold cross validation. (b) is the ROC curve of RP classifier on human dataset by 5-fold cross validation.

Figure 2

Comparison of ROC curves between RP and SVM on yeast (5-fold cross validation). (a) is the ROC curve of SVM method on yeast dataset by 5-fold cross validation. (b) is the ROC curve of RP classifier on yeast dataset by 5-fold cross validation.

2.5. Comparison with Other Existing Methods

In our study, we compared the presented model, termed RP-FFT, with other existing models on the yeast and human datasets to further prove that it can achieve good results. These comparison results of RP-FFT models and other models on the yeast and human datasets are shown in Table 8 and Table 9. From Table 8, it is obvious that the RP-FFT model obtained a higher average accuracy than other existing models on yeast dataset. It is also clear that the other six methods got lower specificity and sensitivity than our proposed model for the same dataset. Accordingly, as is apparent from Table 9, the overall outcomes of our prediction model are also significantly better than the other six models on the human dataset. To sum up, the experimental results of the proposed model called RP-FFT prove its accuracy for predicting SIPs compared with the six approaches. This explains why our prediction model is superior to the other six methods, because it employs a good method of feature extraction and a suitable classifier. It can be further illustrated that our RP-FFT model is suitable for predicting SIPs.
Table 8

Comparison of RP-FFT with the other existing models on the yeast dataset.

ModelAcc. (%)Spe. (%)Sen. (%)MCC (%)B_Acc. (%)
SLIPPER [2]71.9072.1869.7228.4270.95
DXECPPI [45]87.4694.9329.4428.2562.19
PPIevo [46]66.2887.4660.1418.0173.80
LocFuse [47]66.6668.1055.4915.7761.80
CRS [48]72.6974.3759.5823.6866.98
SPAR [48]76.9680.0253.2424.8466.63
Proposed method91.8797.4248.8154.6273.12
Table 9

Comparison of RP-FFT with the other existing models on the human dataset.

ModelAcc. (%)Spe. (%)Sen. (%)MCC (%)B_Acc. (%)
SLIPPER [2]91.1095.0647.2641.9771.16
DXECPPI [45]30.9025.8387.088.2556.46
PPIevo [46]78.0425.8287.8320.8256.83
LocFuse [47]80.6680.5050.8320.2665.67
CRS [48]91.5496.7234.1736.3365.45
SPAR [48]92.0997.4033.3338.3665.37
Proposed method96.2897.6281.4876.4689.55

3. Materials and Methodology

3.1. Datasets

The datasets derived from the UniProt database [49] include 20,199 curated human protein sequences. The PPIs data could be collected from a variety of sources, including DIP [50], BioGRID [51], IntAct [52], InnateDB [53], and MatrixDB [54]. In this experiment, we mainly built the PPIs dataset, which obtains two identical interacting proteins and whose style of interaction was described as ‘direct interaction’ in correlative databases. On this foundation, 2994 human SIPs could be obtained. We built the datasets to estimate the performance of our prediction method, which has three steps [48]: (1) protein sequences with a length less than 50 or more than 5000 residues from the human proteome were removed; (2) to build the human positive dataset, we picked out the SIPs data with high quality, which should meet one of the following requirements: (a) the self-interactions were discovered by at least one small-scale experiment or two types of large-scale experiments; (b) we annotated the protein as a homo-oligomer (comprising homodimer and homotrimer) in UniProt; (c) it has been reported by at least two publications for self-interactions; (3) for the human negative dataset, we eliminated SIPs from all the human proteome (containing proteins labeled as ‘direct interaction’ and much wider ‘physical association’) and the prediction of SIPs in the UniProt database. Eventually, the human dataset contained 1441 SIPs as a positive dataset and 15,938 non-SIPs as a negative dataset [48]. In addition, the yeast dataset was also built to further illustrate the cross-species performance of the RP-FFT model, which included 710 SIPs samples and 5511 non-SIPs samples [48] via the same strategy mentioned above.

3.2. Position-Specific Scoring Matrix

We discovered distantly correlative proteins by applying the Position-Specific Scoring Matrix (PSSM) [55,56,57], which is a helpful tool. Therefore, a PSSM can be transformed from each protein sequence information by applying the Position-Specific Iterated BLAST (PSI-BLAST) [58]. Then, each protein sequence could be transformed into an N × 20 PSSM matrix as follows: where N indicates the size of a protein sequence, and each protein gene was constructed by 20 types of amino acids. For the query protein sequence, a PSSM could arrange the value M that represents the β-th amino acid at the position of α. Thus, M could be described as: where p(α, k) means the occurrence frequency score of the k-th amino acid in the position of α with the probe, and q(β, k) represents the value of Dayhoff’s mutation matrix between the β-th and k-th amino acids. Accordingly, a high value is a strongly conservative position; otherwise, it means a weakly conservative position. In conclusion, PSSM could be a helpful tool for predicting self-interacting proteins. Each PSSM from the protein sequence was generated by employing PSI-BLAST for SIPs detection. For the sake of getting a high degree and a wide range of homologous information, we chose three iterations and assigned the e-value of PSI-BLAST to be 0.001 in this process. Consequently, the PSSM of each protein sequence could be expressed as a matrix consisting of M × 20 elements, where row M of the matrix means the quantity of residues of each protein, and column 20 of the PSSM indicates the 20 different kinds of amino acids.

3.3. Fast Fourier Transform

Fast Fourier Transform (FFT) [59] was first applied in digital signal processing in a number of diverse areas. Afterwards it was used for image processing for a given curve C whose shape was a closed scope. At a certain time t, there is a data sequence F(t), 0 ≤ t < T. Since F(t) is a periodic function, F(t) = F(t + nT). In this study, we used it to extract the eigen values. Hence, we expand F(t) into a Fourier series as much as possible; it can be described as follows: where ω is the Fourier coefficients of F(t). The discrete Fourier transform is given by where , , . F(t) is commonly named the shape signature, which represents the shape boundary of any one-dimensional function. Fourier transform could only capture the architectural characteristics of a shape, which is important to stem FFT from a perceptually meaningful shape signature. FFT stemmed from the centroid distance function is superior to FFT stemmed from other shape signatures. From the centroid (x, y) of the shape, the centroid distance function r(t) could be defined by the distance of the boundary points: where , and N is the quantity of boundary points. It is a matter of great significance to extract informative characteristics based on machine learning approaches. In our study, for the sake of each protein sequence being composed of amounts of amino acids, the eigenvector cannot be directly obtained from a PSSM by PSI-BLAST, which will lead to diverse length of eigenvectors. For solving the question, we multiply the transpose of PSSM by PSSM to obtain a 20 × 20 matrix, and the feature extraction method of fast Fourier transform is employed to generate characteristic vectors from the PSSM profile. In the end, each protein sequence could be calculated to a 400-dimensional vector after FFT. In this study, eventually, each protein sequence from the yeast and human datasets was transformed into a 400-dimensional vector by employing the fast Fourier transform method. In our study, for the sake of obtaining the main important data and advancing the prediction accuracy, we used the Principal Component Analysis (PCA) approach to reduce the size of the yeast and human databases from 400 to 300. Furthermore, reducing the dimensionality of the datasets could remove the complexity of the classifier and improve the generalization error.

3.4. Support Vector Machine

Support vector machine (SVM) was first proposed by Cortes and Vapnik et al. [60] in 1995. SVM inherently do binary classification. SVM is a statistical learning theory method, which is mainly used in the field of pattern recognition. The purpose of SVM is to find the hyperplane that maximizes the distance margin between the two classes. Hence, we can transform it into a convex quadratic programming problem. This idea can be expressed formally as follows: where (x, y) is a training set of instance-label pairs, i = 1, ..., l. x ϵ R are mapped into a higher dimensional space by the function Ø. y ϵ {1, −1}. Furthermore, the kernel function can be described as K(x, x) ≡ Ø(x)Ø(x). It has four basic kernels that can be found in [61]: Linear: K(x, x) = x. Polynomial: K(x, x) = (γx + r), γ > 0. Radial basis function (RBF): K(x, x) = exp(−γ||x − x||2), γ > 0. Sigmoid: K(x, x) = tan h(γx + r). Here, γ, r, and d are kernel parameters. In our experiment, we chose RBF as the kernel function.

3.5. Random Projection Classifier

In mathematics and statistics, random projection (RP) is a technique for dimensionality reduction of some points that exist in Euclidean space. The meaning of the RP method is that projecting N points in N dimensional space can almost always onto a space of dimension ClogN with control on the ratio of distances and the error [62]. This method has been successfully applied for the reestablishment of frequency-sparse signals [63,64], facial recognition [65,66,67], protein mapping [68], and textual and visual information retrieval [69]. Next, we formally describe the random projection technique in detail. First, let be the primitive high-dimensional space dataset, where n is the quantity of high dimension and N is the count of the dataset. The goal of descending dimension is embedding the eigenvectors into a lower dimensional space R from a high-dimensional R where q << n. The output of data is represented as follows: where q approaches the inherent dimensionality of Γ. Thus, the vectors of Γ were regarded as embedding vectors. If we want to reduce the dimension of Γ via the random projection method, a random vector set γ = must first be constructed, where r ∈ R. The random basis can be obtained by two common choices, as follows [62]: The vectors are normally distributed over the q dimensional unit sphere. The components of the vectors are selected Bernoulli +1/−1 distribution and the vectors are standardized so that ||r||2 = 1 for i = 1, …, n. The columns of q × n matrix R consist of the vectors in γ. The embedding result à of A can be got by In our proposed method, random projection is employed to build a training set where the classifier would be trained. We enrich the component of the integration method by using random projection. Next, the dimension of the objective space was set to one part around the space where the training members reside. We built a size of n × N matrix G whose columns are made up the column eigenvectors in Γ. The training set Γ is given in Equation (7). Then, we construct k random matrices whose magnitude is q × n, q and n are mentioned in the above paragraph, and k is the quantity of integration classifiers. Here, the columns of matrices are normalized so the l2 norm is 1. Then, using our method, we constructed training sets by projecting G onto which is the k random matrix. It can be represented as follows: The training sets are imported into an inducer and the export results are a set of classifiers . How do we classify a new dataset I through classifier ℓ? First, we embed I into the dimensionality reduction space R. Then, it can be owned via embedding u in the random matrix R as follows: where Ĩ is the inlaying of u, the classification of Ĩ can be garnered from the classification of I by ℓ. In this ensemble method, the random projection classifier apply a data-driven voting threshold that is employed on the classification results of the whole classifier for the Ĩ to decide produce the ultimate classification result of Ĩ. In this experiment, the random projections were segmented into non-overlapping parts, where B1 = 10 and each one was carefully chosen from a certain part of size B2 = 30 that achieved the smallest estimate of the test error. We chose the k-Nearest Neighbor (KNN) as the base classifier and the leave-one-out test error estimate, where k = seq (1, 40, by = 3). The prior probability of interaction pairs in the training sample dataset was taken as the voting parameter. Our classifier integrates the results of taking advantage of the base classifier on the chosen projection, with the data-driven voting threshold confirming the ultimate mission.

4. Conclusions

In our study, we developed a new prediction model based on protein sequence information to detect SIPs. This model was created by combining Position-Specific Scoring Matrix with Fast Fourier Transform and Random Projection classifier, which was termed RP-FFT. The main point of the experiment is that the datasets used by the classifier are unbalanced. The main improvements of the presented model are: (1) making use of a reasonable feature extraction method that could capture the main information of the data to improve the performance efficiency. (2) The RP classifier is strongly suitable for SIPs prediction. To summarize, the experimental results achieved by the presented method on the yeast and human datasets indicated that our prediction performance is obviously better than that of the SVM-based method and six other existing models. In the future, there will be more and more characteristic extraction techniques and machine learning or deep learning methods attempted for detecting SIPs.
  10 in total

1.  A Novel Ensemble Learning-Based Computational Method to Predict Protein-Protein Interactions from Protein Primary Sequences.

Authors:  Jie Pan; Shiwei Wang; Changqing Yu; Liping Li; Zhuhong You; Yanmei Sun
Journal:  Biology (Basel)       Date:  2022-05-19

2.  MIPDH: A Novel Computational Model for Predicting microRNA-mRNA Interactions by DeepWalk on a Heterogeneous Network.

Authors:  Leon Wong; Zhu-Hong You; Zhen-Hao Guo; Hai-Cheng Yi; Zhan-Heng Chen; Mei-Yuan Cao
Journal:  ACS Omega       Date:  2020-07-09

3.  ACP-DL: A Deep Learning Long Short-Term Memory Model to Predict Anticancer Peptides Using High-Efficiency Feature Representation.

Authors:  Hai-Cheng Yi; Zhu-Hong You; Xi Zhou; Li Cheng; Xiao Li; Tong-Hai Jiang; Zhan-Heng Chen
Journal:  Mol Ther Nucleic Acids       Date:  2019-05-10       Impact factor: 8.886

4.  MLMDA: a machine learning approach to predict and validate MicroRNA-disease associations by integrating of heterogenous information sources.

Authors:  Kai Zheng; Zhu-Hong You; Lei Wang; Yong Zhou; Li-Ping Li; Zheng-Wei Li
Journal:  J Transl Med       Date:  2019-08-08       Impact factor: 5.531

5.  Global Vectors Representation of Protein Sequences and Its Application for Predicting Self-Interacting Proteins with Multi-Grained Cascade Forest Model.

Authors:  Zhan-Heng Chen; Zhu-Hong You; Wen-Bo Zhang; Yan-Bin Wang; Li Cheng; Daniyal Alghazzawi
Journal:  Genes (Basel)       Date:  2019-11-12       Impact factor: 4.096

6.  An effective drug-disease associations prediction model based on graphic representation learning over multi-biomolecular network.

Authors:  Hanjing Jiang; Yabing Huang
Journal:  BMC Bioinformatics       Date:  2022-01-04       Impact factor: 3.169

7.  SIPGCN: A Novel Deep Learning Model for Predicting Self-Interacting Proteins from Sequence Information Using Graph Convolutional Networks.

Authors:  Ying Wang; Lin-Lin Wang; Leon Wong; Yang Li; Lei Wang; Zhu-Hong You
Journal:  Biomedicines       Date:  2022-06-29

8.  Mathematical Modeling and Computational Prediction of High-Risk Types of Human Papillomaviruses.

Authors:  Junchao Zhang; Kechao Wang
Journal:  Comput Math Methods Med       Date:  2022-07-21       Impact factor: 2.809

9.  FFP: joint Fast Fourier transform and fractal dimension in amino acid property-aware phylogenetic analysis.

Authors:  Wei Li; Lina Yang; Yu Qiu; Yujian Yuan; Xichun Li; Zuqiang Meng
Journal:  BMC Bioinformatics       Date:  2022-08-19       Impact factor: 3.307

10.  Identification of self-interacting proteins by integrating random projection classifier and finite impulse response filter.

Authors:  Zhan-Heng Chen; Zhu-Hong You; Li-Ping Li; Yan-Bin Wang; Yu Qiu; Peng-Wei Hu
Journal:  BMC Genomics       Date:  2019-12-27       Impact factor: 3.969

  10 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.