Literature DB >> 25229688

A new supervised over-sampling algorithm with application to protein-nucleotide binding residue prediction.

Jun Hu1, Xue He1, Dong-Jun Yu2, Xi-Bei Yang3, Jing-Yu Yang1, Hong-Bin Shen4.   

Abstract

Protein-nucleotide interactions are ubiquitous in a wide variety of biological processes. Accurately identifying interaction residues solely from protein sequences is useful for both protein function annotation and drug design, especially in the post-genomic era, as large volumes of protein data have not been functionally annotated. Protein-nucleotide binding residue prediction is a typical imbalanced learning problem, where binding residues are extremely fewer in number than non-binding residues. Alleviating the severity of class imbalance has been demonstrated to be a promising means of improving the prediction performance of a machine-learning-based predictor for class imbalance problems. However, little attention has been paid to the negative impact of class imbalance on protein-nucleotide binding residue prediction. In this study, we propose a new supervised over-sampling algorithm that synthesizes additional minority class samples to address class imbalance. The experimental results from protein-nucleotide interaction datasets demonstrate that the proposed supervised over-sampling algorithm can relieve the severity of class imbalance and help to improve prediction performance. Based on the proposed over-sampling algorithm, a predictor, called TargetSOS, is implemented for protein-nucleotide binding residue prediction. Cross-validation tests and independent validation tests demonstrate the effectiveness of TargetSOS. The web-server and datasets used in this study are freely available at http://www.csbio.sjtu.edu.cn/bioinf/TargetSOS/.

Entities:  

Mesh:

Substances:

Year:  2014        PMID: 25229688      PMCID: PMC4168127          DOI: 10.1371/journal.pone.0107676

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Protein-ligand interactions are ubiquitous in virtually all biological processes [1]–[3], and the prediction of protein-ligand interactions using automated computational methods has been an area of intense research in bioinformatics fields [4]–[15]. As important ligand types, nucleotides (e.g., ATP, ADP, AMP, GDP, and GTP) play critical roles in various metabolic processes, such as providing chemical energy, signaling, and replication and transcription of DNA [10]–[15]. The residues in a protein to which nucleotides bind are called protein-nucleotide binding residues. By interacting with the binding residues in a protein, nucleotides can carry out their specific biological functions. Furthermore, protein-nucleotide (e.g., protein-ATP) binding residues are considered valuable targets of therapeutic drugs [12]. Hence, accurate identification of nucleotide-binding residues in protein sequences is of significant importance for protein function analysis and drug design [16], especially in the post-genomic era, as large volumes of protein data have not been functionally annotated. Much effort has been made to identify and characterize nucleotide-binding residues from protein sequences. In the early stages, motif-based methods [17]–[21] dominated this field. For most motif-based methods, conserved motifs in known nucleotide-binding protein sequences or structures are first identified; then, the identified motifs are further utilized to uncover potential binding residues in those un-annotated proteins. Although considerable progress has been achieved in motif-based methods, challenges remain. As Chen et al. [14] reported, motif-based methods often characterize the protein-nucleotide interaction motifs within a relatively narrow range, usually only for a selected interaction mode for a single nucleotide type; in addition, some motif-based methods require tertiary protein structure as the input, which substantially limits their utility, as it is very common in many realistic application scenarios for a given protein target to only have sequence information and no corresponding tertiary structure information [22], [23]. The above-mentioned challenges have motivated researchers in this field to develop machine-learning-based methods for predicting protein-ligand binding residues solely from protein sequences [4]–[6], [13], [14], [22], [24]–[26]. In pioneering work, Chauhan et al. [13] designed a predictor, called ATPint, specifically for predicting protein-ATP binding residues. This group also designed a GTP-specific predictor for protein-GTP binding residue prediction [27], and their earlier studies demonstrated the feasibility of predicting protein-nucleotide binding residues solely from protein sequence information [13], [27]. Later, researchers tended to design predictors that covered a wide range of nucleotide types. For example, Firoz et al. [15] implemented a method of performing binding residue predictions for six nucleotide types, i.e., AMP, GMP, ADP, GDP, ATP and GTP. Recently, Chen et al. [14] presented a predictor, called NsitePred, that could also be used to perform binding residue predictions for multiple nucleotides based on much larger training datasets. All in all, great success has been achieved in this field. Machine-learning-based protein-nucleotide binding residue prediction is, in fact, a typical imbalanced learning problem because the number of negative samples (i.e., non-binding residues) is significantly larger than that of positive samples (i.e., binding residues). Previous studies in the machine-learning field have shown that direct application of traditional machine-learning algorithms tends to result in a bias toward the majority class [28]. Unfortunately, most of the existing machine-learning-based predictors, including ATPint [13], ATPsite [24], and NsitePred [14], have not carefully considered this serious class imbalance phenomenon. Considerable effort has been made to develop effective solutions for imbalanced learning [28]. Roughly speaking, the existing solutions for imbalanced learning can be grouped into three categories: sample rescaling-based methods [29], [30], learning-based methods (e.g., cost-sensitive learning [31], [32], active learning [33], [34], kernel learning [35], [36]), and hybrid methods, which combine both the sampling rescaling and learning methods [37], [38]. Among the above-mentioned solutions, the sample rescaling strategy (e.g., over-sampling [39] and under-sampling [40]) is the basic technique, and it attempts to balance the sizes of different classes by changing the numbers and distributions within them; this strategy has been demonstrated to be effective for imbalanced learning problems [29], [30]. For example, we recently investigated class imbalance in the protein-nucleotide binding prediction problem and found that prediction performance could be improved by balancing the number of samples in different classes via an under-sampling technique [22], [25], [26]. In this study, we seek to overcome the problem of class imbalance via an over-sampling technique. In contrast to the under-sampling technique, which reduces the size of the majority class, an over-sampling technique attempts to balance the sizes of different classes by generating additional samples for the minority class. To date, many over-sampling techniques have emerged, including random over-sampling (ROS), the synthetic minority over-sampling technique (SMOTE) [39], and adaptive synthetic sampling (ADASYN) [41]. Motivated by these existing over-sampling techniques, in this study, we propose a new supervised over-sampling (SOS) algorithm that synthesizes new additional samples for minority classes using a supervised process to guarantee the validity of the synthesized samples. Additionally, a new predictor, called TargetSOS, is developed based on the proposed SOS for performing protein-nucleotide binding residue prediction. The experimental results from two benchmark datasets demonstrate the effectiveness of TargetSOS. TargetSOS and the datasets used in this study are freely available at http://www.csbio.sjtu.edu.cn/bioinf/TargetSOS/.

Materials and Methods

Benchmark Datasets

Two benchmark datasets were chosen to evaluate the efficacy of the proposed SOS algorithm and of the implemented predictor, TargetSOS. The first dataset [13], ATP168, consists of 168 non-redundant, ATP-interacting protein sequences, of which the maximal pairwise sequence identity is less than 40%. In total, ATP168 includes 3104 and 59226 residues for ATP binding and ATP non-binding, respectively. The second dataset [14], NUC5, is a multiple nucleotide-interacting dataset that consists of five training sub-datasets, each for a specific type of nucleotide; more specifically, NUC5 consists of 227, 321, 140, 56, and 105 protein sequences that interact with five types of nucleotides, i.e., ATP, ADP, AMP, GTP, and GDP, respectively, and the maximal pairwise identity of the sequences of each of the five sub-datasets is less than 40%. In addition, for each nucleotide type, Chen et al. [14] constructed a corresponding, independent validation dataset to evaluate the generalization capability of a prediction model. For each independent validation dataset, the maximal pairwise sequence identity is culled to 40%. Furthermore, any sequence in the independent validation dataset shares less than 40% identity to sequences in the corresponding training sub-dataset. Table 1 summarizes the detailed compositions of the two benchmark datasets. All data listed in Table 1 can be found in Supporting Information S1. Further details regarding the construction of the datasets can be found in [13] and [14].
Table 1

Compositions of the two benchmark datasets.

DatasetLigand TypeCross-Validation Dataset (Training Dataset)Independent Validation DatasetTotal No. of Sequences
No. of Sequences(numP, numN)* Ratio No. of Sequences(numP, numN)* Ratio
ATP168 [13] ATP168(3104, 59226)19168
ATP227(3393, 80409)2417(248, 6974)28244
ADP321(4688, 121158)2626(405, 10553)26347
NUC5 [14] AMP140(1756, 44009)2520(263, 6057)23160
GTP56(875, 21401)247(134, 2678)2063
GDP105(1577, 36561)237(94, 2420)26112

* Figures numP, numN in 2-tuple (numP, numN) represent the number of positive (binding residues) and negative (non-binding residues) samples, respectively; △ Ratio = numN/numP.

* Figures numP, numN in 2-tuple (numP, numN) represent the number of positive (binding residues) and negative (non-binding residues) samples, respectively; △ Ratio = numN/numP.

Feature Representation and Classifier

The main purpose of this study is to demonstrate the feasibility of the proposed SOS algorithm and its effectiveness in protein-nucleotide binding residue prediction. To fulfill the aforementioned purpose, only the most commonly used feature representation methods and classifiers in the field of protein-nucleotide binding residue prediction are used. More specifically, the position-specific scoring matrix (PSSM) and predicted protein secondary structure (PSS), both of which have been demonstrated to be especially useful for protein-nucleotide binding residue prediction [13], [14], [25], [26], are taken to extract discriminative feature vectors. Support vector machine (SVM) [42] is used as a classifier for constructing a prediction model.

A. Extract Feature Vector from the Position-Specific Scoring Matrix

Position-specific scoring matrix (PSSM) derived features have been widely used in bioinformatics including intrinsic disorder prediction [43]–[45], protein secondary structure prediction [46], transmembrane helix prediction [47]–[49], protein 3D structure prediction [50], and protein-ligand binding prediction [14], [51]. In this study, we obtain the PSSM of a query protein sequence by performing PSI-BLAST [52] to search the Swiss-Prot database through three iterations and with 0.001 as the E-value cutoff against the query sequence. To facilitate the subsequent computation, we further normalize each score, denoted as , that is contained in the PSSM using the logistic function . Based on the normalized PSSM, the feature vector, denoted LogisticPSSM, for each residue in the protein sequence can be extracted by applying a sliding-window technique, as follows [25], [26]: for a residue at position along the query sequence, its LogisticPSSM feature vector consists of the normalized PSSM scores of the query sequence that correspond to a sequence segment of length that is centered on . It has been demonstrated that W = 17 is a better choice for several protein-ligand binding residue prediction studies [25], [26]. Consequently, the dimensionality of the LogisticPSSM feature vector of a residue is 17×20 = 340-D.

B. Extract Feature Vector from the Predicted Protein Secondary Structure

PSIPRED [53], which has been widely used in bioinformatics [54], [55], can predict the probabilities of each residue in a query protein sequence belonging to three secondary structure classes, i.e., coil, helix, and strand. We obtained the predicted protein secondary structure by performing PSIPRED against the query sequence. The obtained predicted secondary structure is an L×3 probability matrix, where L is the length of the protein sequence. Similar to the LogisticPSSM feature extraction, we can extract a 17×3 = 51-D feature vector, denoted as PSS, for each residue in the protein by applying a sliding window of size 17. The final discriminative feature vector of a residue is formed by serially combining its LogisticPSSM feature with the corresponding PSS feature, and the dimensionality of the obtained feature vector for the residue is 340+51 = 391-D.

C. Support Vector Machine

Support vector machine (SVM), which was proposed by Vapnik [42], has been widely used in a variety of bioinformatics fields, including the protein-nucleotide binding residue prediction [13], [14] considered in this study. In view of this, we will also use SVM as the base-learning model to evaluate the efficacy of the proposed SOS algorithm. Here, we will briefly introduce the basic idea of SVM. Let be the set of samples, where and are the feature vector and the corresponding label of the i-th sample, respectively, and +1 and −1 are the labels of positive class and negative class, respectively. In linearly separable cases, SVM constructs a hyperplane that separates the samples of two classes with a maximum margin. The optimal separating hyperplane (OSH) is constructed by finding another vector, , and a parameter, , that minimizes and satisfies the following conditions:where is a vector normal to the hyperplane, and is the Euclidean norm of . The solution is a unique, globally optimized result with the following expansion: Support vectors are those , whose corresponding . Once the and are found, a query input can be classified as follows: To allow for mislabeled examples, Corinna Cortes and Vladimir N. Vapnik suggested a modified maximum margin idea, i.e., “soft margin” technique [56]. For each training sample, a corresponding slack variable is introduced: , . Accordingly, the relaxed separation constraint is given as: Then, the OSH can be solved by minimizing.where is the regularization parameter. Furthermore, to address non-linearly separable cases, the “kernel substitution” technique is introduced as follows: first, the input vector is mapped into a higher dimensional Hilbert space, H, by a non-linear kernel function, ; then, the OSH in the mapped space, H, is solved using a procedure similar to that for a linear case, and the decision function is given by: To train a SVM on a given data set, the kernel function and the regularity parameter need to be specified in advance. In this study, LIBSVM [57] (http://www.csie.ntu.edu.tw/~cjlin/libsvm/) is taken. The Gaussian kernel , which is one of the most commonly used kernel functions, is chosen as the kernel function. The regularization parameter and the kernel width parameter are optimized based on 10-fold cross-validation using a grid search strategy in the LIBSVM [57] software.

Dealing with Class Imbalance: A New Supervised Over-Sampling Method

As described in the introduction section, protein-nucleotide binding residue prediction is a typical imbalanced learning problem. By revisiting Table 1, we can easily find that a severe class imbalance phenomenon does exist among both training datasets and independent validation datasets: the ratio of the number of non-binding residues to that of binding residues is often larger than 20. In this study, we propose a new SOS algorithm for relieving the severity of class imbalance to facilitate the subsequent statistical machine learning methods. To demonstrate the effectiveness of the proposed SOS, several popular over-sampling methods, including ROS, SMOTE [39], and ADASYN [41], are used to perform comparisons with the proposed SOS.

A. Random Over-sampling

In the ROS technique, the minority set is augmented by replicating randomly selected samples within the set. Although ROS is simple and easy to perform, a potential problem is that the resulting dataset tends to be over-fitted because ROS simply appends replicated samples to the original dataset; thus, multiple instances of certain samples become “tied” [58]. In view of this issue, several improved over-sampling techniques, e.g., SMOTE [39] and ADASYN [41], have been proposed and have shown promising results in various imbalanced applications. In this study, two improved over-sampling techniques, i.e., SMOTE [39] and ADASYN [41], were considered.

B. Synthetic Minority Over-sampling Technique

The SMOTE method [39] augments the minority class set by creating artificial samples based on the feature space similarities between existing minority samples. The SMOTE procedure is briefly described below. For each sample in , let be the set of the K-nearest neighbors of in under the Euclidian distance metric. To synthesize a new sample, an element in , denoted as , is selected and then multiplied by the feature vector difference between and and by a random number between [0, 1]. Finally, this vector is added to :where [0, 1] is a random number. These synthesized samples help break the ties introduced by ROS and augment the original dataset in a manner that, in general, significantly improves subsequent learning [28].

C. Adaptive Synthetic Sampling

SMOTE creates the same number of synthetic samples for each original minority sample without considering the neighboring majority samples, which increases the occurrence of overlapping between classes [28]. In view of this limitation, various adaptive over-sampling methods, e.g., ADASYN [41], have been proposed. ADASYN uses a systematic method to adaptively create different numbers of synthetic samples for different original minority samples according to their distributions. The ADASYN procedure is briefly described below. The number of samples that must be synthesized for the entire minority class is computed first:where is a parameter that determines the balance level after the ADASYN process. Then, for each original sample, , its K-nearest neighbors are found according to the Euclidean distance metric, and the distribution function, , which is defined as:is calculated, where is the number of samples in the K-nearest neighbors of that belong to , and is a normalization constant so that is a distribution function, i.e., . Next, the number of synthetic samples that must be generated for each is computed: Finally, for each , synthetic samples are generated according to Eq. (7), as in SMOTE. The key difference between ADASYN and SMOTE is that the former uses a density distribution, , as a criterion to automatically decide the number of synthetic samples that must be generated for each minority sample by adaptively changing the weights of the different minority samples to compensate for the skewed distributions [28], [41]. The latter generates the same number of synthetic samples for each original minority sample.

D. Proposed Supervised Over-sampling

Let be the training dataset, where is the minority class sample set, and is the majority class sample set. The purpose of the proposed SOS algorithm is to obtain a relatively balanced dataset, denoted as , by synthesizing additional minority class samples under a supervised process. Let be the parameter of the over-sampling coefficient, which is a scalar quantity that measures the ratio of the size of the minority class sample set after over-sampling to that of the original minority class sample set. In other words, controls how many additional minority samples will be generated. More additional minority samples will be synthesized with larger values of . The process of the proposed SOS is described as follows: Step I: Training an initial classifier model, denoted as , on the original training dataset : The trained classifier model will be used to judge whether a synthesized minority class sample is valid. Step II: Synthesizing an additional minority sample: First, two samples, denoted as and , will be randomly selected from the minority class sample set : According to the two randomly selected minority class samples, an additional sample can be synthesized:where is a random value ranging from 0 to 1. Then, the confidence of the synthesized sample, , being a minority class sample is predicted using the trained initial classifier model : The validity of the synthesized sample depends on its confidence. More specifically, the synthesized sample is a valid minority class sample if and only if , i.e., its confidence lies within the prescribed confidence interval . Step II is repeated until the valid minority class samples have been synthesized. Algorithm 1 summarizes the proposed SOS. Note that the three parameters, i.e., , , and , are problem-dependent. In this study, we set , , and . Note that in Step II, it is straightforward and reasonable that a synthesized sample will not be considered valid when its confidence is less than the prescribed lower confidence, . However, a synthesized sample will also be considered invalid if its confidence is larger than the prescribed upper confidence, . The underlying reason for this choice is that we believe that a synthesized sample with confidence that is too high tends to become “tied” with those true minority class samples, thus potentially leading to an over-fitting problem. Algorithm 1. Supervised Over-Sampling (SOS) INPUT: - The training dataset, where is the minority class sample set and is the majority class sample set; - The over-sampling coefficient, which is the size of the minority class after over-sampling, divided by that of the original minority class; - The confidence interval, which is used to determine whether a synthetic sample belongs to the minority class. OUTPUT: - The over-sampled training dataset, where is the minority class sample set after over-sampling. 1. Training a classifier model, denoted as , using the original training set : 2. 3. WHILE 4. Randomly select two samples, denoted as and , from : 5. Synthesize a new sample: where is a random value ranging 0 from 1; 6. Predict the confidence of being a minority class sample: 7. IF 8. 9. END IF 10. END WHILE 11. 12. 13. RETURN

Evaluation Indexes

Let , , , and be the abbreviations for true positive, false positive, true negative, and false negative, respectively. Then, (Sen), (Spe), (Acc), and the Matthews correlation coefficient (MCC) can be defined as follows: However, these four evaluation indexes are threshold-dependent, i.e., the values of these indexes vary with the threshold that is used in the prediction model. Considering that the MCC measures the overall quality of the binary predictions, we reported these threshold-dependent evaluation indexes by choosing the threshold that maximizes the value of the MCC of the predictions (termed MaxMCC Evaluation in this study). It has not escaped our notice that several predictors reported their performances by selecting the threshold that balances the values of Sen and Spe [13], [25], [26] (termed Balanced Evaluation in this study). For the purpose of a fair comparison, we also used Balanced Evaluation when comparing the proposed method with these predictors. In addition, the Area Under the receiver operating characteristic (ROC) Curve (AUC), which is threshold-independent and increases in direct proportion to prediction performance, was used to evaluate the overall prediction qualities of the considered prediction models.

Experimental Results and Analysis

Supervised Over-Sampling Helps to Enhance Prediction Performance

In this section, we empirically demonstrate that the performance of protein-nucleotide binding residue prediction can be further improved by applying the proposed SOS algorithm. Tables 2 and 3 summarize the performance comparisons between with-SOS and without-SOS for ATP168 and ATP227 over five-fold cross-validation under Balanced Evaluation and MaxMCC Evaluation, respectively. Figure 1 (a) and (b) illustrate the ROC curves of with-SOS and without-SOS for ATP168 and ATP227 over five-fold cross-validation. The results listed in Tables 2 and 3 show that the prediction performances are remarkably improved after SOS is applied. An improvement in the AUC of over 2% is observed for both the ATP168 and ATP227 datasets. In addition, the other four indexes, i.e., Sen, Spe, Acc, and MCC, of the with-SOS predictions are consistently higher than that of the without-SOS predictions. Taking MCC as an example, improvements of 5% and 4% are observed for ATP168 and ATP227, respectively, under Balanced Evaluation, whereas improvements of 12% and 8% are achieved for ATP168 and ATP227, respectively, under MaxMCC Evaluation.
Table 2

Performance comparisons of with-SOS and without-SOS predictions for ATP168 and ATP227 over five-fold cross-validation under Balanced Evaluation.

DatasetUpper-Sampling Sen (%) Spe (%) Acc (%) MCC AUC
ATP168with-SOS 80.0 80.1 80.1 0.311 0.878
without-SOS75.277.277.10.2620.843
ATP227with-SOS 81.3 81.7 81.7 0.306 0.893
without-SOS79.079.179.10.2660.871
Table 3

Performance comparisons of with-SOS and without-SOS predictions for ATP168 and ATP227 over five-fold cross-validation under MaxMCC Evaluation.

DatasetUpper-Sampling Sen (%) Spe (%) Acc (%) MCC AUC
ATP168with-SOS 42.3 99.2 96.3 0.536 0.878
without-SOS35.298.595.30.4150.843
ATP227with-SOS 46.3 99.2 97.0 0.553 0.893
without-SOS40.198.996.50.4730.871
Figure 1

ROC curves of with-SOS and without-SOS predictions for ATP168 and ATP227 over five-fold cross-validation.

(a) ROC curves for ATP168; (b) ROC curves for ATP227.

ROC curves of with-SOS and without-SOS predictions for ATP168 and ATP227 over five-fold cross-validation.

(a) ROC curves for ATP168; (b) ROC curves for ATP227.

Comparisons with Other Over-Sampling Methods

In this section, we compare the proposed SOS with several other popular over-sampling methods, including ROS, SMOTE [39], and ADASYN [41]. Table 4 shows comparisons of the performance of SOS, ROS, SMOTE, and ADASYN for ATP168 and ATP227 over five-fold cross-validation under MaxMCC Evaluation. The results for the four other types of nucleotide ligands, i.e., ADP, AMP, GTP, and GDP, can be found in Supporting Information S2.
Table 4

Performance comparisons between SOS and ROS, SMOTE, and ADASYN for ATP168 and ATP227 over five-fold cross-validation under MaxMCC Evaluation.

DatasetOver-Sampling Method Sen (%) Spe (%) Acc (%) MCC AUC
SOS 42.3 99.2 96.3 0.536 0.878
ATP168ADASYN [41] 41.799.096.10.5120.877
SMOTE [39] 41.499.096.10.5110.860
ROS39.298.895.80.4740.846
SOS46.3 99.2 97.0 0.553 0.893
ATP227ADASYN [41] 46.5 98.996.80.537 0.896
SMOTE [39] 44.799.096.80.5260.880
ROS42.999.196.90.5220.876
From Table 4, it is clear that the proposed SOS significantly outperforms ROS for both ATP168 and ATP227. Taking AUC and MCC, which are two overall measurements of prediction quality, as examples, average improvements of approximately 3% and 5% are observed. We also found that the proposed SOS achieves comparable performance to ADASYN and slightly outperforms SMOTE for ATP168 and ATP227. Similar phenomenon could also be found for the four other types of nucleotide ligands (refer to Supporting Information S2). The results listed in Table 4 and Supporting Information S2 show that the proposed SOS performs much better than ROS and can achieve comparable performances to ADASYN and SMOTE, which demonstrates the efficacy of the proposed SOS.

Comparisons with Existing Predictors

In this section, we compare the proposed predictor, called TargetSOS, to the existing popular protein-nucleotide binding residue predictors to demonstrate its efficacy. TargetSOS performs predictions using a SVM model, which is trained with the proposed SOS algorithm in the NUC5 dataset and uses the LogisticPSSM+PSS feature as the model input. The comparisons are performed for both the cross-validation test and the independent validation test. Note that when cross-validation comparisons are performed for ATP168, only the Balanced Evaluation results are reported because the results for most existing predictors that are constructed from ATP168 are reported under Balanced Evaluation. For the same reason, cross-validation comparisons for the NUC5 dataset are reported under MaxMCC Evaluation.

A. Cross-Validation Test

Table 5 lists the performance comparisons of the proposed TargetSOS, TargetATP [26], TargetATPsite [25], and ATPint [13] for ATP168 over five-fold cross-validation under Balanced Evaluation. By observing Table 5, we find that the proposed TargetSOS significantly outperforms ATPint and is the best performer among the four considered predictors that were specifically designed for protein-ATP binding residue prediction. An over 5% improvement is observed for each of the five considered evaluation indexes, i.e., Sen, Spe, Acc, MCC, and AUC. In addition, TargetSOS performs better, although not significantly better, than the two most recently released predictors, i.e., TargetATP [26] and TargetATPsite [25].
Table 5

Performance comparisons between the proposed TargetSOS, TargetATP, and TargetATPsite for ATP168 over five-fold cross-validation under Balanced Evaluation.

Predictor Sen (%) Spe (%) Acc (%) MCC AUC
TargetSOS 80.0 80.1 80.1 0.311 0.878
TargetATP [26] 79.179.879.80.3080.873
TargetATPsite [25] 78.278.478.40.2900.860
ATPint [13] 74.475.875.10.2490.823
Table 6 summarizes the performance comparisons between the proposed TargetSOS and several other popular protein-nucleotide binding residue predictors for the NUC5 dataset over five-fold cross-validation under MaxMCC Evaluation. It is found that the proposed TargetSOS almost always achieves the best performance, with only one exception for ATP concerning MCC and AUC, which are two evaluation indexes that measure the overall prediction quality of a predictor. Taking MCC as an example, TargetSOS achieves improvements of approximately 3%, 8%, 6%, 7%, and 3% for ATP, ADP, AMP, GDP, and GTP, respectively, compared with the second-best performer (i.e., TargetATPsite [25] for ATP and NsitePred [14] for ADP, AMP, GDP, and GTP). The underlying reason for the improvement in MCC is that the TargetSOS can achieve much higher performance with respect to the true positive rate (i.e., Sen) while simultaneously achieving comparable or even slightly better performances for the true negative rate (i.e., Spe). We believe that this improvement may be a result of the SOS technique.
Table 6

Performance comparisons between the proposed TargetSOS and other popular predictors for the NUC5 dataset over five-fold cross-validation under MaxMCC Evaluation.

Ligand TypePredictor Sen (%) Spe (%) Acc (%) MCC AUC
TargetSOS 46.3 99.2 97.0 0.553 0.893
TargetATP [26] 41.299.096.60.501 0.895
ATPTargetATPsite [25] 44.598.996.60.5200.881
NsitePred* 44.498.296.00.4600.861
SVMPred* 36.198.896.20.4330.854
TargetSOS 60.5 99.1 97.7 0.653 0.914
ADPNsitePred* 54.498.897.10.5720.893
SVMPred* 45.8 99.3 97.30.5550.885
TargetSOS 38.1 98.8 96.4 0.440 0.850
AMPNsitePred* 30.498.896.20.3770.829
SVMPred* 20.8 99.6 96.60.3600.820
TargetSOS 66.1 99.5 98.2 0.744 0.923
GDPNsitePred* 64.699.197.60.6750.910
SVMPred* 62.398.997.70.6550.905
TargetSOS 47.3 99.5 97.4 0.598 0.850
GTPNsitePred* 47.399.196.80.5620.844
SVMPred* 37.3 99.7 97.00.5510.836

* Data excerpted from [14].

* Data excerpted from [14].

B. Independent Validation Test

It has been routine procedure to evaluate the generalization capability of a predictor using an independent validation test because evaluating a newly developed predictor by only comparing it to existing predictors and by using the same datasets may potentially lead to optimistically biased results, in the sense that the new predictor’s characteristics over-fit the used datasets [59]. Considering this potential bias, we also performed independent validation tests for the proposed TargetSOS and compared their performances with those of several other popular sequence-based protein-nucleotide binding residue predictors, as shown in Table 7.
Table 7

Performance comparisons between the proposed TargetSOS and other popular predictors for the independent validation dataset of NUC5.

Ligand TypePredictor Sen (%) Spe (%) Acc (%) MCC AUC
TargetSOS 53.6 99.2 97.6 0.603 0.912
TargetATP [26] 48.998.996.90.542 0.912
ATPTargetATPsite [25] 45.899.197.20.5300.882
NsitePred* 46.098.596.70.4760.875
SVMPred* 36.799.196.90.4510.868
TargetSOS 60.0 98.597.0 0.585 0.912
ADPNsitePred* 47.498.796.80.5120.893
SVMPred* 38.8 99.3 97.1 0.5000.886
TargetSOS 45.6 98.996.7 0.522 0.880
AMPNsitePred* 42.398.7 96.9 0.5010.876
SVMPred* 33.5 99.4 96.70.4780.870
TargetSOS49.1 99.1 97.20.5620.866
GDPNsitePred* 58.5 98.597.0 0.576 0.867
SVMPred* 51.198.8 97.1 0.5530.855
TargetSOS 61.9 98.8 97.1 0.655 0.900
GTPNsitePred* 60.498.896.90.640 0.909
SVMPred* 48.5 99.3 96.90.6020.887

*Data excerpted fdrom [14].

*Data excerpted fdrom [14]. From Table 7, we find that the AUCs for ATP, ADP, AMP, GDP, and GTP when using TargetSOS in the corresponding independent validation datasets are 0.912, 0.912, 0.880, 0.866, and 0.900, respectively. By revisiting Table 6, it is found that the AUCs of TargetSOS for ATP, ADP, AMP, GDP, and GTP on the training datasets are 0.893, 0.914, 0.850, 0.923, and 0.850, respectively. In other words, TargetSOS achieves similar overall prediction performances (measured by AUCs) on the training dataset and the corresponding independent validation dataset for all five nucleotide ligands, indicating that the generalization capability of the TargetSOS that is derived from the knowledge buried in the training datasets has not been under- or over-estimated. In addition, we find that the proposed TargetSOS achieves comparable overall performance (AUC) to the state-of-the-art sequence-based predictors considered in this study. On the other hand, TargetSOS almost always achieves the best performances for MCC, with only one exception for GDP, and an average improvement of approximately 3% is observed compared with the second-best performer (i.e., TargetATP [26] for ATP and NsitePred [14] for ADP, AMP, GDP, and GTP).

Conclusion

In this study, a new SOS algorithm that balances the samples of different classes by synthesizing additional samples for minority class with a supervised process is proposed to address imbalanced learning problems. We apply the proposed SOS algorithm to protein-nucleotide binding residue prediction, and a web-server, called TargetSOS, is implemented. Cross-validation tests and independent validation tests on two benchmark datasets demonstrate that the proposed SOS algorithm helps to improve the performance of protein-nucleotide binding residue prediction. The findings of this study enrich the understanding of class imbalance learning and are sufficiently flexible to be applied to other bioinformatics problems in which class imbalance exists, such as protein functional residue prediction and disulfide bond prediction. Datasets used in this study. (DOC) Click here for additional data file. Performance comparisons between different over-sampling techniques on the ADP, AMP, GTP, GDP sub-datasets in NUC5. (DOC) Click here for additional data file.
  38 in total

1.  Protein secondary structure prediction based on position-specific scoring matrices.

Authors:  D T Jones
Journal:  J Mol Biol       Date:  1999-09-17       Impact factor: 5.469

2.  On the molecular discrimination between adenine and guanine by proteins.

Authors:  I Nobeli; R A Laskowski; W S Valdar; J M Thornton
Journal:  Nucleic Acids Res       Date:  2001-11-01       Impact factor: 16.971

3.  Prediction and analysis of nucleotide-binding residues using sequence and sequence-derived structural descriptors.

Authors:  Ke Chen; Marcin J Mizianty; Lukasz Kurgan
Journal:  Bioinformatics       Date:  2011-11-29       Impact factor: 6.937

4.  Understanding and predicting druggability. A high-throughput method for detection of drug binding sites.

Authors:  Peter Schmidtke; Xavier Barril
Journal:  J Med Chem       Date:  2010-08-12       Impact factor: 7.446

5.  Scoring function based approach for locating binding sites and understanding recognition mechanism of protein-DNA complexes.

Authors:  M Michael Gromiha; Kazuhiko Fukui
Journal:  J Chem Inf Model       Date:  2011-02-28       Impact factor: 4.956

6.  SOMRuler: a novel interpretable transmembrane helices predictor.

Authors:  Dongjun Yu; Hongbin Shen; Jingyu Yang
Journal:  IEEE Trans Nanobioscience       Date:  2011-07-07       Impact factor: 2.935

7.  Prediction of GTP interacting residues, dipeptides and tripeptides in a protein from its evolutionary information.

Authors:  Jagat S Chauhan; Nitish K Mishra; Gajendra P S Raghava
Journal:  BMC Bioinformatics       Date:  2010-06-03       Impact factor: 3.169

Review 8.  The ATP-binding site of type II topoisomerases as a target for antibacterial drugs.

Authors:  Anthony Maxwell; David M Lawson
Journal:  Curr Top Med Chem       Date:  2003       Impact factor: 3.295

9.  Alignment of helical membrane protein sequences using AlignMe.

Authors:  Marcus Stamm; René Staritzbichler; Kamil Khafizov; Lucy R Forrest
Journal:  PLoS One       Date:  2013-03-04       Impact factor: 3.240

10.  Intrinsic disorder in the BK channel and its interactome.

Authors:  Zhenling Peng; Yoshihisa Sakai; Lukasz Kurgan; Bernd Sokolowski; Vladimir Uversky
Journal:  PLoS One       Date:  2014-04-11       Impact factor: 3.240

View more
  8 in total

1.  Prediction of Protein-Protein Interaction Sites with Machine-Learning-Based Data-Cleaning and Post-Filtering Procedures.

Authors:  Guang-Hui Liu; Hong-Bin Shen; Dong-Jun Yu
Journal:  J Membr Biol       Date:  2015-11-12       Impact factor: 1.843

2.  ATPbind: Accurate Protein-ATP Binding Site Prediction by Combining Sequence-Profiling and Structure-Based Comparisons.

Authors:  Jun Hu; Yang Li; Yang Zhang; Dong-Jun Yu
Journal:  J Chem Inf Model       Date:  2018-02-08       Impact factor: 4.956

3.  Seabed Modelling by Means of Airborne Laser Bathymetry Data and Imbalanced Learning for Offshore Mapping.

Authors:  Tomasz Kogut; Arkadiusz Tomczak; Adam Słowik; Tomasz Oberski
Journal:  Sensors (Basel)       Date:  2022-04-19       Impact factor: 3.847

4.  Incorporating efficient radial basis function networks and significant amino acid pairs for predicting GTP binding sites in transport proteins.

Authors:  Nguyen-Quoc-Khanh Le; Yu-Yen Ou
Journal:  BMC Bioinformatics       Date:  2016-12-22       Impact factor: 3.169

5.  Imbalance learning for the prediction of N6-Methylation sites in mRNAs.

Authors:  Zhixun Zhao; Hui Peng; Chaowang Lan; Yi Zheng; Liang Fang; Jinyan Li
Journal:  BMC Genomics       Date:  2018-08-01       Impact factor: 3.969

6.  SXGBsite: Prediction of Protein-Ligand Binding Sites Using Sequence Information and Extreme Gradient Boosting.

Authors:  Ziqi Zhao; Yonghong Xu; Yong Zhao
Journal:  Genes (Basel)       Date:  2019-11-22       Impact factor: 4.096

7.  Machine learning classification of texture features of MRI breast tumor and peri-tumor of combined pre- and early treatment predicts pathologic complete response.

Authors:  Lal Hussain; Pauline Huang; Tony Nguyen; Kashif J Lone; Amjad Ali; Muhammad Salman Khan; Haifang Li; Doug Young Suh; Tim Q Duong
Journal:  Biomed Eng Online       Date:  2021-06-28       Impact factor: 2.819

8.  Web tools to perform long non-coding RNAs analysis in oncology research.

Authors:  Shixing Gu; Guangjie Zhang; Qin Si; Jiawen Dai; Zhen Song; Yingshuang Wang
Journal:  Database (Oxford)       Date:  2021-07-23       Impact factor: 3.451

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.