Literature DB >> 28961273

Improved detection of DNA-binding proteins via compression technology on PSSM information.

Yubo Wang1,2, Yijie Ding1,2, Fei Guo1,2, Leyi Wei1,2, Jijun Tang1,3,2.   

Abstract

Since the importance of DNA-binding proteins in multiple biomolecular functions has been recognized, an increasing number of researchers are attempting to identify DNA-binding proteins. In recent years, the machine learning methods have become more and more compelling in the case of protein sequence data soaring, because of their favorable speed and accuracy. In this paper, we extract three features from the protein sequence, namely NMBAC (Normalized Moreau-Broto Autocorrelation), PSSM-DWT (Position-specific scoring matrix-Discrete Wavelet Transform), and PSSM-DCT (Position-specific scoring matrix-Discrete Cosine Transform). We also employ feature selection algorithm on these feature vectors. Then, these features are fed into the training SVM (support vector machine) model as classifier to predict DNA-binding proteins. Our method applys three datasets, namely PDB1075, PDB594 and PDB186, to evaluate the performance of our approach. The PDB1075 and PDB594 datasets are employed for Jackknife test and the PDB186 dataset is used for the independent test. Our method achieves the best accuracy in the Jacknife test, from 79.20% to 86.23% and 80.5% to 86.20% on PDB1075 and PDB594 datasets, respectively. In the independent test, the accuracy of our method comes to 76.3%. The performance of independent test also shows that our method has a certain ability to be effectively used for DNA-binding protein prediction. The data and source code are at https://doi.org/10.6084/m9.figshare.5104084.

Entities:  

Mesh:

Substances:

Year:  2017        PMID: 28961273      PMCID: PMC5621689          DOI: 10.1371/journal.pone.0185587

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

DNA-binding proteins play an important role in a variety of biomolecule functions, such as transcription, the detection of DNA damage and replication. The importance of DNA-binding proteins is facilitating the development of various methods for identifying them. Experimental methods that have been applied to identify DNA-binding proteins include filter binding assays, genetic analysis, chromatin immune precipitation on microarrays and X-ray crystallography [1, 2]. Nevertheless, these experimental methods have some disadvantages, such as expensive and time-consuming. Especially with the development of next-generation high-throughput DNA sequencing techniques [3], protein sequence data are growing rapidly. At present, it is unrealistic to use experimental methods to identify all DNA-binding proteins. Therefore, a lot of computational methods based on machine learning (ML) algorithm or statistical model [4-6] are used to reduce the cost of resources. In order to further facilitate the calculation process, there are some web servers have been developed to generate feature vectors of DNA, RNA or protein sequences, such as a web-server called Pse-in-One [7]. In recent years, computational methods based on machine learning (ML) algorithms have become more and more popular because of their promising performance. According to various feature information, the ML-based approachs are mainly composed of structure information-based [8-18] and sequence information-based method [1, 2, 19–35]. The structural features of proteins are closely related to the functions, and therefore predictors based on the structural information can achieve better performance of DNA-binding protein identification. Nimrod et al. [18] trained a random forest classifier using the average surface electrostatic potentials, dipole moments and cluster-based amino acid conservation patterns of the protein. Ahmad et al. [10] developed a neural network classifier based on the net charge, electric dipole moment and quadrupole moment tensors of the protein. Bhardwaj et al. [13] made use of SVM classifier and three features, including surface and overall composition, overall charge and positive potential surface patches. Some structure-based methods also have the participation of sequence information. For example, Szilágyi and Skolnick [17] extracted feature vectors from the following three perspectives: the relative proportions of certain amino acids, the asymmetry of the spatial distribution of certain other amino acids and the dipole moment of the molecule. However, a large number of proteins can’t be known with the structural information, so structure-based predictors can only be applied to a small portion of the whole protein database. In contrast, sequence information is easier to extract and more convenient to use. We can extract multiple sequence-based features, such as physicochemical properties [20, 36], dipeptide composition [24, 30] and the amino acid composition [21]. Cai and Lin [20] trained SVM classifiers using protein’s amino acid composition, limited range correlation of hydrophobicity and solvent accessible surface area of the protein. Yu et al. [21] developed the binary classifications for rRNA-, RNA-, DNA-binding proteins by feeding these features (being extracted from protein sequence amino acid compositions and physicochemical properties) into the SVM classifier. Liu et al. [36] extracted feature vectors from three sequence features, including overall amino acid composition, pseudo amino acid composition and physicochemical distance transformation. Some researchers also incorporated evolutionary information generated by PSI-BLAST [37] into sequence-based methods to improve prediction performance. For instance, Kumar et al. [38] were the first to use evolutionary information to identify DNA-binding proteins and developed a SVM classifier called DNAbinder. Some similar methods, for example the method of Ho et al. [39], were also proposed to identify DNA-binding proteins. Their results showed that evolutionary information can significantly improve the performance, so evolutionary information is useful in the identification of DNA binding proteins. Liu et al. [25] proposed a predictor called iDNAPro-PseAAC, which incorporates evolutionary information and the pseudo amino acid composition (PseAAC). The method of Waris et al. [28] used features extracted from dipeptide composition, split amino acid composition and position specific scoring matrix (PSSM) to train multiple classifiers and found the classifier that achieved the best predicte performance. As described above, the feature extraction algorithms determine whether protein sequences can be expressed completely by feature vectors. In order to obtain a satisfactory performance, we should select feature extraction algorithms carefully. In this paper, we innovatively combine the 1040-dimension feature vector named PSSM-DWT, the 100-dimension feature vector named PSSM-DCT and the 200-dimension feature vector named NMBAC to predict DNA-binding proteins. Discrete Wavelet Transform (DWT) and Discrete Cosine Transform (DCT) can be used to obtain the effective information by compressing the PSSM matrix. Also, we extract the 200-dimension feature vector according to six physicochemical properties. Then, these features are fed into the training SVM model for predicting DNA-binding proteins. We evaluate our method by three datasets, namely PDB1075, PDB594 and PDB186. The first two datasets are used in Jackknife test and the last dataset, PDB186, is used for independent testing. The results demonstrate the effectiveness of our method in identifying DNA-binding proteins.

Materials and methods

In order to illustrate the overall process, the framework of our method is presented in Fig 1. In the training phase, we extract two features (PSSM-DWT and PSSM-DCT) from the PSSM matrix, and extract NMBAC feature from six physicochemical properties. The prediction model is obtained by feeding these features into the SVM classifier. In the prediction phase, we use the same feature representation algorithm to describe the predictive protein sequence, then use the training SVM model for DNA-binding protein prediction.
Fig 1

The framework of our method.

Datasets

In the present study, we apply three benchmark datasets to evaluate our approach, namely PDB1075, PDB594 and PDB186. These DNA-binding proteins are selected from Protein Data Bank (http://www.rcsb.org/pdb/home/home.do). The protein sequences, which are less than 50 amino acids or contain character “X”, must be removed. We should ensure that no sequence has more than 25% similarity with any other sequences. Concretely, the PDB1075 dataset, constructed by Liu et al. [40], has 525 DNA-binding proteins and 550 DNA-non-binding proteins. The PDB594 dataset, edited by Lou et al. [2], is made up of 297 DNA-binding proteins and 297 DNA-non-binding proteins. These two datasets are applied for Jackknife test. The PDB186 dataset for independent test is also derived from the paper of Lou et al. [2], and contains 93 DNA-binding proteins and 93 DNA-non-binding proteins.

Evolutionary features

Position specific scoring matrix

Position Specific Scoring Matrix (PSSM) generated by PSI-BLAST [37] (BLAST+ [41] options: -num_iterations 3 -db nr -inclusion_ethresh 0.001) stores the evolutionary information of a protein sequence. Suppose the length of a protein sequence is L (L amino acids), the size of the PSSM for this protein is L × 20 (L rows and 20 columns). The form of this matrix is as follows: The formula for each element PSSM(i, j) is as follows: where ω(i, k) is the frequency of k-th amino acid type at the position i, D(k, j) is the rate of mutation from the k-th amino acid to the j-th amino acid in a protein sequence from Dayhoff’s mutation matrix (substitution matrix). The larger values of substitution matrix indicate more strongly conserved positions; otherwise, the reverse.

Discrete Cosine Transform

We use the Discrete Cosine Transform (DCT) [42], which is widely used in data compression to compress PSSM and retain a portion of the compressed PSSM as feature vectors. The DCT is a linear separable transformation and can change the distribution of information density from evenly to unevenly. After compression, we should retain the low frequency part of PSSM, because the low frequency section contains more information than the high frequency section. In this work, 2 dimensions DCT (2D-DCT) is used to compress PSSM. Given an input matrix Mat = PSSM ∈ ℜ, the corresponding conversion formula is as follows: where 0 ≤ i < M, 0 ≤ j < N. According to the above formula for compression, the part that contains most of the information (low frequency section) is distributed in the upper left corner of the compressed PSSM. In the end, we retain the first 100 coefficients as PSSM-DCT feature.

Discrete Wavelet Transform

The Wavelet Transform (WT) is defined as the projection of a signal f(t) onto the wavelet function: where a is a scale variable and b is a translation variable. is the analyzing wavelet function. T(a, b) is the transform coefficients which are found for both specific locations on the signal and specific wavelet periods. Discrete Wavelet Transform (DWT) can decompose the amino acid sequences into coefficients at different dilations and then remove the noise component from the profiles. Nanni et al. [43, 44] proposed an efficient algorithm to perform DWT by assuming that the discrete signal f(t) is x[n], where N is the length of discrete signal. where g is low pass filter and h is high pass filter. y[n] is the approximate coefficient (low-frequency components) of the signal. y[n] is the detailed coefficient (high-frequency components). This decomposition is repeated to further increase the frequency resolution and the approximation coefficients decomposed with high and low pass filters and then down-sampled. With the increase of decomposition level j, more detailed characteristics of the signal can be observed. Inspired by Nanni’s work [43, 44], we use 4-level DWT and calculate the maximum, minimum, mean and standard deviation values of different scales (4 levels of both low and high-frequency coefficients). Because of the high-frequency components are more noisy and hence only the low-frequency components are more important, we also extract the first five discrete cosine coefficients from the approximation coefficients. The schematic diagram of a 4-level DWT is shown in Fig 2.
Fig 2

Schematic diagram of a 4-level DWT.

The Mat = PSSM ∈ ℜ has 20 columns. So, the PSSM consists of 20 types of discrete signals (L lengths). At last, we use above 4 levels DWT to analysis these discrete signals of PSSM (each column) and extract the PSSM-DWT feature from PSSM of protein.

Sequence features

Normalized Moreau-Broto Autocorrelation

We use the Normalized Moreau-Broto Autocorrelation(NMBAC) to extract sequence features from six physicochemical properties for improving the predicte performance. The NMBAC is proposed by Feng at al. [45] for the prediction of membrane protein types. Each physicochemical propertie of 20 amino acid have corresponding values and a protein sequence can be replaced by a vector of physicochemical property values. In our work, the six physicochemical properties are hydrophobicity (H), volumes of side chains of amino acids (VSC), polarity (P1), polarizability (P2), solvent-accessible surface area (SASA) and net charge index of side chains (NCISC) of amino acid, respectively. The physicochemical propertie values of 20 amino acids are shown in Table 1. Before we use these values to represent amino acids, they must be normalized to zero mean and unit standard deviation (SD) as follows: where P is the value of descriptor j for amino acid type i, P is the mean over 20 amino acids of descriptor value j, and S is the corresponding SD.
Table 1

Original values of six physicochemical properties of 20 amino acid types.

Amino acidHVSCP1P2SASANCISC
A0.6227.58.10.0461.1810.007187
C0.2944.65.50.1281.461-0.03661
D-0.940130.1051.587-0.02382
E-0.746212.30.1511.8620.006802
F1.19115.55.20.292.2280.037552
G0.480900.8810.179052
H-0.47910.40.232.025-0.01069
I1.3893.55.20.1861.810.021631
K-1.510011.30.2192.2580.017708
L1.0693.54.90.1861.9310.051672
M0.6494.15.70.2212.0340.002683
N-0.7858.711.60.1341.6550.005392
P0.1241.980.1311.4680.239531
Q-0.8580.710.50.181.9320.049211
R-2.5310510.50.2912.560.043587
S-0.1829.39.20.0621.2980.004627
T-0.0551.38.60.1081.5250.003352
V1.0871.55.90.141.6450.057004
W0.81145.55.40.4092.6630.037977
Y0.26117.36.20.2982.3680.023599
For each physicochemical property, a protein can be represented by a vector composed of normalized physicochemical property values. NMBAC [45] is obtained by inputting these vectors into the following formula: where j represents one descriptor of six descriptor, i is the position in protein sequence X, n is the length of the protein sequence and lag is the sequential distance between one residue and another, a certain number of residues away (lag = 1, 2, …, lg, lg is a parameter determined by an optimization procedure to be described). According to Guo’s work [46], we define the optimal value of lag from 1 to 30. For each protein sequence, we can obtain 30 × 6 = 180 dimensional feature vector. We also add the frequency of 20 amino acids appearing on this sequence to the feature vector. Finally, we can get the 30 × 6 + 20 = 200 dimensional feature vector for a protein sequence.

Classification and feature selection

After feature extraction procedure, all samples in benchmark datasets are converted into numerical feature vectors with the same dimension. The feature space of each protein sequence is composed of PSSM-DWT, PSSM-DCT and NMBAC features. By removing noisy and redundant features from the original feature space (PSSM-DWT + PSSM-DCT + NMBAC), feature selection alleviates the overfitting and improves the performance. In order to reduce feature abundance and computation complexity, we use the Support Vector Machine Recursive Feature Elimination and Correlation Bias Reduction (SVM-RFE+CBR) [47] to select an optimal feature subset. SVM-RFE+CBR is proposed by incorporating the CBR strategy into the feature elimination procedure: (1) less prone to overfitting; (2) able to make full use of the training data; (3) much faster, especially on a lot of candidate features. As a result, it has been successfully applied in many problems, especially in gene selection [48-50]. We can obtain the output of SVM-RFE+CBR with a ranked feature list. Feature selection is achieved by choosing a group of top-ranked features. The ranking criterion of SVM-RFE+CBR is closely related to the SVM model.

Support Vector Machine

Support Vector Machine (SVM) developed by Vapnik [51] is a classification and regression paradigm. In the process of using SVM, samples labeled positive or negative are projected into a high dimensional feature space using a kernel, and the hyper plane in the feature space is optimized to maximize the margin of positive and negative samples. There are some biological problems for example prediction of protein-protein interactions [46, 52–56], homology detection [57], and analysis of gene expression data [58] that can used SVM to solve. Given a training dataset of instance-label pairs {x, γ}, i = 1, 2, …, N with input data x ∈ R and output labels γ ∈ {+1, −1}, the classification decision function implemented by SVM is represented in the following equation: where the coefficient α is obtained by solving the following convex Quadratic Programming (QP) problem: where x is called pupport vector only if the corresponding α > 0, C is a regularization parameter that controls the tradeoff between margin and misclassification error. Under most circumstances, K(x, x) = exp(−γ‖x − x‖2), called the Radial Basis Functions (RBF) kernel, has better boundary response, and most high-dimensional data are approximated by Gaussian-like distributions. We implemented a SVM model using LIBSVM [59] with the radial basis functiona (http://www.csie.ntu.edu.tw/∼cjlin/libsvm/).

Results and discussion

We preform our method on three datasets for predicting DNA-binding protein. In the Jackknife test, we apply our method on the PDB1075 and PDB594 datasets to analyze the effectiveness of feature extraction and feature selection, the performance of our method is also compared with other methods. In the independent test, our prediction model is tested on the independent dataset PDB186 and compared with the results of other methods.

Measurements

We use the Jackknife test to analyze the quality of predictor constructed by our method. Because of the effectiveness of Jackknife test, it is widely used to test the function of predictor(eg., [26, 60]). In the Jackknife test, we use every sample of the benchmark dataset as test dataset one by one, and the rest of the samples are used to train predictor. In addition, We employ four mearsures which are also used in other methods to evaluate the performance of our method, including Accuracy (ACC), Sensitivity (SN), Specificity (SP), and Mathew’s correlation coefficient (MCC). Their formulas are listed blelow: where TP is the number of true positive, TN is the number of true negative, FP is the number of false positive, and FN is the number of false negative.

Parameter optimization

To select the optimal parameters of feature NMBAC and PSSM-DCT, we test the predictive performance for different parameters (NMBAC with different value of lg and PSSM-DCT with different first m coefficients) via a five-fold cross validation. To get the optimal lg, we evaluate values of lg from 5 to 45 (with a step of 5). The results of the prediction on PDB1075 dataset are shown in Fig 3. The ACC of prediction is increasing, when the value of lg is between 5 and 30. After that, the value of ACC is falling. Different value of m may lead to different performance, we test different values of m from 20 to 260 (with a step of 20). The curve of ACC is shown in Fig 4. The value of ACC is rising, when m increases from 20 to 100. But it slightly declines, when m is between 100 and 260. Obviously, PSSM-DCT with m less than 100 (lg less than 30 for NMBAC) would lose some effective features and larger values may introduce noise. Thus, we select lg as 30 (NMBAC) and m as 100 (PSSM-DCT) in our experiments.
Fig 3

The accuracy of different lg values on PDB1075 (Five-fold cross validation).

Fig 4

The accuracy of different m values on PDB1075 (Five-fold cross validation).

Benchmark dataset—PDB1075

Performance of different feartures

We extract three features from the benchmark dataset (PDB1075), namely PSSM-DWT, PSSM-DCT and NMBAC. We need to find a combination of features to achieve the best performance, and analyze the most important feature to get the good prediction. The performance of different feartures by Jackknife test is shown in Table 2. The combination of NMBAC, PSSM-DCT and PSSM-DWT achieves the highest ACC (0.7926), MCC (0.5853), SN (0.8000) and second highest SP (0.7855). In order to obtain the importance of each feature, we compare the AUROC of seven feature combinations obtained by Jackknife cross-validation on PDB1075 dataset, shown in Fig 5. We can see that the highest contribution to the predicton performance is PSSM-DWT, followed by NMBAC, yet the PSSM-DCT is the lowest one. These information show that each feature is useful in prediction of DNA-binding proteins and the combination of three features can achieve the best performance, but the PSSM-DCT feature is not as effective as the other two features.
Table 2

The performance of different features on PDB1075 dataset (Jackknife test evaluation).

FeatureACC(%)MCCSN(%)SP(%)
NMBAC74.050.483677.9070.36
PSSM-DCT70.600.411766.8674.18
PSSM-DWT75.070.501073.3376.73
NMBAC+PSSM-DCT78.050.560677.1478.91
NMBAC+PSSM-DWT78.700.574079.2478.18
PSSM-DWT+PSSM-DCT73.770.475273.5274.00
NMBAC+PSSM-DWT+PSSM-DCT79.260.585380.0078.55
Fig 5

The AUROC comparison of seven feature combinations through Jackknife cross-validation on PDB1075 dataset.

Performance after feature selection

In order to improve the performance on PDB1075 dataset, we remove the noisy and redundant features from the original feature space by SVM-RFE+CBR [47]. Consider the combination of three features, first, we obtain a ranked feature list as shown in Fig 6. NMBAC, PSSM-DCT and PSSM-DWT are divided into three intervals: [1, 200], [201, 300] and [301, 1340]. Then the accuracy of different dimension features by Jackknife test can be seen in Fig 7 and we can find that the best accuracy can be achieved when we select the first 216-dimension features, these features can be obtained from the ranked feature list: 77 features in the interval [1, 200], 22 features in the interval [201, 300] and 117 features in the interval [301, 1340]. The feature selection is applied to each feature combination, and we can obtain the feature dimension with the best performance of each feature combination. As in the previous section, we also analyze the importance of each feature according to the AUROC comparison of seven feature combinations after feature selection, shown in Fig 8. We get the same result that NMBAC and PSSM-DWT are more effective than PSSM-DCT in the prediction of DNA-binding protein.
Fig 6

The feature score through SVM-RFE+CBR on the dataset of PDB1075.

The x-axis represents the feature index.

Fig 7

The accuracy of different dimension features on PDB1075 dataset (Jackknife test evaluation).

Fig 8

The AUROC comparison of seven feature combinations through Jackknife cross-validation on PDB1075 dataset.

* means this feature combination has employed feature selection.

The feature score through SVM-RFE+CBR on the dataset of PDB1075.

The x-axis represents the feature index.

The AUROC comparison of seven feature combinations through Jackknife cross-validation on PDB1075 dataset.

* means this feature combination has employed feature selection. For the results of the previous section, We obtain new results after feature selection which be shown in Table 3. We also find that the combination of three features achieves the best performance and has obviously exceeded the performance without feature selection. It reaches the highest value on all metrics: ACC(0.8623), MCC(0.7250), SN(0.8743) and SP(0.8509). These results strongly demonstrate that feature selection can significantly improve the predict performance.
Table 3

The performance of different features after feature selection on PDB1075 dataset (Jackknife test evaluation).

FeatureACC(%)MCCSN(%)SP(%)
NMBAC*76.090.521876.1976.00
PSSM-DCT*74.600.492876.1973.09
PSSM-DWT*81.020.621382.8679.27
NMBAC+PSSM-DCT*81.400.627680.3882.36
NMBAC+PSSM-DWT*84.930.698785.5284.36
PSSM-DWT+PSSM-DCT*78.330.566478.2978.36
NMBAC+PSSM-DWT+PSSM-DCT*86.230.725087.4385.09

* means this feature combination has employed feature selection.

* means this feature combination has employed feature selection.

Comparision with existing methods

The performance of our method on PDB1075 dataset is compared with other existing methods, including iDNA-Prot|dis [40], iDNA-Prot [61], DNA-Prot [19], PseDNA-Pro [36], DNAbinder [38], iDNAPro-PseAAC [25], Kmer1+ACC [23] and Local-DPP [62]. The performance of different methods by Jackknife test is displayed in Table 4. We can find that four mearsures evaluated by our method are significantly higher than the evaluated mearsures of other methods. The ACC, MCC, SN and SP values of our method are improved by 7.03%, 0.13, 2.63% and 4.73%, respectively, compared with other methods.
Table 4

The performance of our method and other existing methods on PDB1075 dataset (Jackknife test evaluation).

MethodsACC(%)MCCSN(%)SP(%)
IDNA-Prot|dis77.300.5479.4075.27
PseDNA-Pro76.550.5379.6173.63
IDNA-Prot75.400.5083.8164.73
DNA-Prot72.550.4482.6759.76
DNAbinder(dimension = 400)73.580.4766.4780.36
DNAbinder(dimension = 21)73.950.4868.5779.09
iDNAPro = PseAAC76.560.5375.6277.45
Kmer1+ACC75.230.5076.7673.76
Local-DPP(n = 3,λ = 1)79.100.5984.8073.60
Local-DPP(n = 2,λ = 2)79.200.5984.0074.50
Our method86.230.7387.4385.09

Benchmark dataset—PDB594

We compare the performance of our method with several classifiers applied to Lou’s method [2] on the benchmark dataset (PDB594), shown in Table 5. Our method achieve the highest ACC of 86.2%, MCC of 0.724, SN of 87.2% and SP of 85.2%. The ACC, MCC, SN and SP values are improved by 5.7%, 0.114, 1.7% and 7.1%, respectively. This represents the effectiveness of our method for identifying DNA-binding proteins.
Table 5

The performance of our method and other existing methods on PDB594 dataset (Jackknife test evaluation).

MethodsACC(%)MCCSN(%)SP(%)
GNB-based-wrapper80.50.61082.878.1
DT-based-wrapper69.20.38468.470.0
LongR-based-wrapper75.40.51180.570.4
KNN-based-wrapper74.60.49272.177.1
SVM-Poly-based-wrapper77.10.55085.568.7
SVM-RBF-based-wrapper80.10.60584.875.4
Our method86.20.72487.285.2

Independent dataset—PDB186

For the purpose of analyzing the robustness, our method is compared to other methods on the independent dateset (PDB186) (PDB1075 serves as training dataset and PDB186 is applied as test dataset), shown in Table 6. Our method achieves 76.34% of ACC, 0.5566 of MCC, 92.5% of SN and 60.22% of SP. Our approach still performs better than most of existing methods with a certain creditability.
Table 6

The performance of our method and other existing methods on PDB186 dataset.

MethodsACC(%)MCCSN(%)SP(%)
IDNA-Prot|dis72.00.44579.564.5
IDNA-Prot67.20.34467.766.7
DNA-Prot61.80.24069.953.8
DNAbinder60.80.21657.064.5
DNABIND67.70.35566.768.8
DNA-Threader59.70.27923.795.7
DBPPred76.90.53879.674.2
iDNAPro = PseAAC-EL71.50.44282.860.2
Kmer1+ACC71.00.43182.859.1
Local-DPP(n = 3,λ = 1)79.00.62592.565.6
Local-DPP(n = 2,λ = 2)77.40.56890.364.5
Our method76.30.55792.560.2

PDB1075 serves as training dataset and PDB186 is applied as test dataset.

PDB1075 serves as training dataset and PDB186 is applied as test dataset.

Computational time

The computational time of feature extraction and jackknife test evaluation on PDB1075 is shown in Table 7. From the table, we can find that the computational time of jackknife test evaluation which has used feature selection algorithm is significantly shorter than the jackknife test evaluation without feature selection. This prove that the feature selection algorithm can effectively reduces redundant features.
Table 7

The computational time of feature extraction and jackknife test evaluation on PDB1075.

FeatureFE(sec)JT(sec)JT-FS(sec)
NMBAC3.092317.6486.7
DCT187.981357.5352.0
DWT299.7516166.0757.8
NMBAC+DCT+DWT490.8217520.01642.0

The values of column “FE” indicate the computational time of feature extraction on PDB1075. The values of column “JT” indicate the computational time of jackknife test evaluation which has not used feature selection algorithm on PDB1075. The values of column “JT-FS” indicate the computational time of jackknife test evaluation which has used feature selection algorithm on PDB1075.

The values of column “FE” indicate the computational time of feature extraction on PDB1075. The values of column “JT” indicate the computational time of jackknife test evaluation which has not used feature selection algorithm on PDB1075. The values of column “JT-FS” indicate the computational time of jackknife test evaluation which has used feature selection algorithm on PDB1075.

Conclusion

In this paper, we propose a novel feature extraction algorithm to construct a machine learning method of DNA-binding protein prediction. We employ the feature extraction algorithm to extract three feature vectors, namely NMBAC, PSSM-DWT and PSSM-DCT. It is meaningful that we apply the DWT and DCT methods, which are rarely used in bioinformatics to obtain PSSM-DWT and PSSM-DCT. Through these approaches, the effective information is extracted from the PSSM matrix and stored in the feature vectors. In Jackknife test, our method can achieve excellent prediction performances, and our prediction performance has obviously exceeded other existing methods after feature selection. On the independent dataset, our approach still performs better than most of existing methods. Furthermore, we can find that the PSSM-DWT feature makes the greatest contribution to the prediction performance. The performance of our method proves the rationality of feature extraction algorithm and the effectiveness of our method in predicting DNA-binding protein.
  50 in total

1.  Prediction of membrane protein types based on the hydrophobic index of amino acids.

Authors:  Z P Feng; C T Zhang
Journal:  J Protein Chem       Date:  2000-05

2.  Mismatch string kernels for discriminative protein classification.

Authors:  Christina S Leslie; Eleazar Eskin; Adiel Cohen; Jason Weston; William Stafford Noble
Journal:  Bioinformatics       Date:  2004-01-22       Impact factor: 6.937

3.  Wavelet images and Chou's pseudo amino acid composition for protein classification.

Authors:  Loris Nanni; Sheryl Brahnam; Alessandra Lumini
Journal:  Amino Acids       Date:  2011-10-13       Impact factor: 3.520

4.  Structure-based prediction of DNA-binding proteins by structural alignment and a volume-fraction corrected DFIRE-based energy function.

Authors:  Huiying Zhao; Yuedong Yang; Yaoqi Zhou
Journal:  Bioinformatics       Date:  2010-06-04       Impact factor: 6.937

5.  An ensemble of reduced alphabets with protein encoding based on grouped weight for predicting DNA-binding proteins.

Authors:  Loris Nanni; Alessandra Lumini
Journal:  Amino Acids       Date:  2008-02-21       Impact factor: 3.520

6.  Combing ontologies and dipeptide composition for predicting DNA-binding proteins.

Authors:  Loris Nanni; Alessandra Lumini
Journal:  Amino Acids       Date:  2008-01-04       Impact factor: 3.520

7.  PseDNA-Pro: DNA-Binding Protein Identification by Combining Chou's PseAAC and Physicochemical Distance Transformation.

Authors:  Bin Liu; Jinghao Xu; Shixi Fan; Ruifeng Xu; Jiyun Zhou; Xiaolong Wang
Journal:  Mol Inform       Date:  2014-09-26       Impact factor: 3.353

8.  Prediction of DNA-binding protein based on statistical and geometric features and support vector machines.

Authors:  Weiqiang Zhou; Hong Yan
Journal:  Proteome Sci       Date:  2011-10-14       Impact factor: 2.480

9.  DNA-binding protein prediction using plant specific support vector machines: validation and application of a new genome annotation tool.

Authors:  Graham B Motion; Andrew J M Howden; Edgar Huitema; Susan Jones
Journal:  Nucleic Acids Res       Date:  2015-08-24       Impact factor: 16.971

10.  DBD-Hunter: a knowledge-based method for the prediction of DNA-protein interactions.

Authors:  Mu Gao; Jeffrey Skolnick
Journal:  Nucleic Acids Res       Date:  2008-05-31       Impact factor: 16.971

View more
  19 in total

1.  Application of DNA-Binding Protein Prediction Based on Graph Convolutional Network and Contact Map.

Authors:  Weizhong Lu; Nan Zhou; Yijie Ding; Hongjie Wu; Yu Zhang; Qiming Fu; Haiou Li
Journal:  Biomed Res Int       Date:  2022-01-17       Impact factor: 3.411

2.  Research on DNA-Binding Protein Identification Method Based on LSTM-CNN Feature Fusion.

Authors:  Weizhong Lu; Xiaoyi Chen; Yu Zhang; Hongjie Wu; Yijie Ding; Jiawei Shen; Shixuan Guan; Haiou Li
Journal:  Comput Math Methods Med       Date:  2022-06-02       Impact factor: 2.809

3.  Comparative Analysis on Alignment-Based and Pretrained Feature Representations for the Identification of DNA-Binding Proteins.

Authors:  Die Chen; Hua Zhang; Zeqi Chen; Bo Xie; Ye Wang
Journal:  Comput Math Methods Med       Date:  2022-06-28       Impact factor: 2.809

4.  A sequence-based multiple kernel model for identifying DNA-binding proteins.

Authors:  Yuqing Qian; Limin Jiang; Yijie Ding; Jijun Tang; Fei Guo
Journal:  BMC Bioinformatics       Date:  2021-05-31       Impact factor: 3.169

5.  AC2: An Efficient Protein Sequence Compression Tool Using Artificial Neural Networks and Cache-Hash Models.

Authors:  Milton Silva; Diogo Pratas; Armando J Pinho
Journal:  Entropy (Basel)       Date:  2021-04-26       Impact factor: 2.524

6.  iAPSL-IF: Identification of Apoptosis Protein Subcellular Location Using Integrative Features Captured from Amino Acid Sequences.

Authors:  Yadong Tang; Lu Xie; Lanming Chen
Journal:  Int J Mol Sci       Date:  2018-04-13       Impact factor: 5.923

7.  Identification of DNA-protein Binding Sites through Multi-Scale Local Average Blocks on Sequence Information.

Authors:  Cong Shen; Yijie Ding; Jijun Tang; Jian Song; Fei Guo
Journal:  Molecules       Date:  2017-11-28       Impact factor: 4.411

8.  A Novel Computational Method for Detecting DNA Methylation Sites with DNA Sequence Information and Physicochemical Properties.

Authors:  Gaofeng Pan; Limin Jiang; Jijun Tang; Fei Guo
Journal:  Int J Mol Sci       Date:  2018-02-08       Impact factor: 5.923

9.  A Model Stacking Framework for Identifying DNA Binding Proteins by Orchestrating Multi-View Features and Classifiers.

Authors:  Xiu-Juan Liu; Xiu-Jun Gong; Hua Yu; Jia-Hui Xu
Journal:  Genes (Basel)       Date:  2018-08-01       Impact factor: 4.096

10.  Predicting ATP-Binding Cassette Transporters Using the Random Forest Method.

Authors:  Ruiyan Hou; Lida Wang; Yi-Jun Wu
Journal:  Front Genet       Date:  2020-03-25       Impact factor: 4.599

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.