Literature DB >> 34220299

An Improved Computational Prediction Model for Lysine Succinylation Sites Mapping on Homo sapiens by Fusing Three Sequence Encoding Schemes with the Random Forest Classifier.

Samme Amena Tasmia1, Fee Faysal Ahmed1, Parvez Mosharaf1, Mehedi Hasan1, Nurul Haque Mollah1.   

Abstract

BACKGROUND: Lysine succinylation is one of the reversible protein post-translational modifications (PTMs), which regulate the structure and function of proteins. It plays a significant role in various cellular physiologies including some diseases of human as well as many other organisms. The accurate identification of succinylation site is essential to understand the various biological functions and drug development.
METHODS: In this study, we developed an improved method to predict lysine succinylation sites mapping on Homo sapiens by the fusion of three encoding schemes such as binary, the composition of k-spaced amino acid pairs (CKSAAP) and amino acid composition (AAC) with the random forest (RF) classifier. The prediction performance of the proposed random forest (RF) based on the fusion model in a comparison of other candidates was investigated by using 20-fold cross-validation (CV) and two independent test datasets were collected from two different sources.
RESULTS: The CV results showed that the proposed predictor achieves the highest scores of sensitivity (SN) as 0.800, specificity (SP) as 0.902, accuracy (ACC) as 0.919, Mathew correlation coefficient (MCC) as 0.766 and partial AUC (pAUC) as 0.163 at a false-positive rate (FPR) = 0.10 and area under the ROC curve (AUC) as 0.958. It achieved the highest performance scores of SN as 0.811, SP as 0.902, ACC as 0.891, MCC as 0.629 and pAUC as 0.139 and AUC as 0.921 for the independent test protein set-1 and SN as 0.772, SP as 0.901, ACC as 0.836, MCC as 0.677 and pAUC as 0.141 at FPR = 0.10 and AUC as 0.923 for the independent test protein set-2. It also outperformed all the other existing prediction models.
CONCLUSION: The prediction performances as discussed in this article recommend that the proposed method might be a useful and encouraging computational resource for lysine succinylation site prediction in the case of human population.
© 2021 Bentham Science Publishers.

Entities:  

Keywords:  Protein sequences; encoding schemes; feature selection; fusion model; lysine succinylation site; prediction; random forest

Year:  2021        PMID: 34220299      PMCID: PMC8188582          DOI: 10.2174/1389202922666210219114211

Source DB:  PubMed          Journal:  Curr Genomics        ISSN: 1389-2029            Impact factor:   2.236


Introduction

Succinylation is one type of reversible protein post-translational modifications (PTMs) in which a succinyl group (-CO-CH2-CH2-CO2H) is transferred to a protein molecule at its lysine (K) residue. This modification is transpired in many protein molecules to change their structure and functions in both eukaryotic and prokaryotic organisms [1-3] and altered lysine's charge from +1 to −1 (at physiological pH) and produced comparatively large structural moiety (100 Da), bigger than acetylation (42 Da) or methylation (14 Da) [4]. It is a central regulatory PTM in many biological processes and plays a significant role in various cellular physiologies of human as well as many other organisms [2, 4, 5]. Recently, several empirical methods have been built to detect succinylated protein including high-performance liquid chromatography assays, chromatography-mass spectrometry and spectrophotometric assays [6, 7]. In the previous years, different massive proteomic technologies were widely developed to detect succinylated protein in various organisms, including Escherichia coli, Mycobacterium tuberculosis, Toxoplasma gondii, Saccharomyces cerevisiae, Homo sapiens and Mus musculus, and currently in plants [1, 4, 8-14]. However, the experiential technologies are often time-consuming, cost-effective and difficult to detect exact modifications of protein. There were few computational predictors in the literature for predicting succinylated protein using a web server [14-18]. Hasan et al. developed two predictors termed SuccinSite2.0 [14] and SuccinSite [15] with the combination of RF classifier scores based on the amino acid frequency and properties. Huang et al developed a computational predictor, named CNN-SuccSite, which has been developed based on deep learning architectures with different encoding schemes [16]. Recently, Ning et al. developed HybridSucc using Group-based Prediction System (GPS) via diverse encoding systems including k-space amino acid pair composition (CKSAAP), amino acid index (AAindex) physicochemical properties and pseudo amino acid composition (PseAAC) [17]. More recently, Hasan et al. also suggested a predictor termed GPSuc, by combining five sequence encoding schemes i.e. profile-based composition of k-spaced amino acid pairs (pCKSAAP), Amino acid composition (AAC), AAindex, binary amino acid codes (BE), and position-specific scoring matrix (PSSM). They used the selected feature vectors and random forest (RF) classifier to build the predictor [18]. Their sensitivity was reported to be less than 70%. So, further improvement was expected that can supplement the existing Lysine succinylation site prediction techniques. In this study, an attempt was made to develop an improved prediction model by fusing few encoding schemes with the machine learning approaches to predict lysine succinylation sites mapping on Homo sapiens. We introduced the necessary materials and methods for the development of the proposed computational procedure in section 2. The summary results and their discussions were given in sections 3 and 4, respectively, and section 4 presents the conclusion of this study.

Materials and Methods

Data Sources and Descriptions

In this study, we considered two protein datasets. Both datasets were collected from the Uni-ProtKB/Swiss-Prot and NCBI protein sequence databases. The dataset-1 contained experimentally validated 1700 lysine succinylated protein sequences mapping on Homo sapiens. The dataset-2 contained experimentally validated 704 lysine succinylated protein sequences mapping on human pathogen Histoplasma capsulatum. The succinylated sites for both datasets are also known as positive windows and the remaining lysine residues in the protein sequence for both datasets were considered as the non-succinylated sites (negative windows) as suggested [15, 18, 19]. Dataset-1 was used to develop the prediction model and dataset-2 was used to investigate the performance of the prediction model as described in detail in the next subsection 2.2.

Data Preparation and Overview on the Development of the Proposed Computational Predictor

The prediction performance not only depends on the estimation of the model parameters but also depends on the selection of tuning parameters like CD-HIT (Cluster Database at High Identity with Tolerance) threshold [20-23], window size of protein sequence, ratio of positive and negative windows, encoding scheme, features and classifier. The appropriate value of CD-HIT threshold (CHT), window size (WS) and ratio of positive and negative windows depends on the dataset [15-18, 24-26]. As for example, Hasan and Kurata (2018) [18] used CHT=30%, WS=41 and ratio=1:2; Manavalan et al. (2018) [25] used CHT=30%, WS=41 and ratio=1:2; Chen et al. (2019) [26] used CHT=80%, WS=21 and ratio=1:2; and Mosharaf et al. (2020) [24] used CHT=40%, WS=27 and ratio=1:1 to develop their prediction models. Thus, different authors used different CHTs, WSs and ratios of positive and negative windows to develop their prediction models. However, none of them discussed how they selected these tuning parameters. In this case, we tried to discuss this issue slightly. At first, we considered three CHTs 30%, 40% and 50% to select the more appropriate one of them for removing identical sequences from the protein dataset-1 to solve the overprediction problem. The protein dataset-1 was reduced to three datasets of sizes 550, 665 and 824 based on 3 values of CHT, respectively. We observed that all predictors showed better performance with the protein dataset of size 550 corresponding to CHT=30%. Thereafter, we considered the dataset-1 of size p11=550 corresponding to CHT=30% to develop the predictor. The dataset-1 was partitioned into a training dataset and independent test dataset-1, where the test dataset-1 was constructed by taking p12=50 proteins from the protein dataset-1 that were also used in the previously published prediction model as the independent test dataset [14, 16-18] for a fair comparison with our proposed model. Rest of the 500 proteins were used to construct the training dataset. There were 1195 positive and 16842 negative window sites within all 500 proteins of the training dataset. On the other hand, there were 54 positive and 2004 negative window sites within all 50 proteins of the independent test dataset-1. We also constructed independent test dataset-2 by taking p21=202 proteins randomly from the protein dataset-2 for a more fair investigation on the performance of the prediction model. There were 463 positive and 6742 negative window sites within all 202 proteins of the independent test dataset-2. The training dataset of size p11=500 was used to train different candidate predictors as described below. The performance of prediction models depends on the window size as mentioned previously [15-18, 24-26]. Therefore, the training dataset was partitioned into positive and negative window samples corresponding to each of the window sizes 19, 25, 31, 37 and 41, respectively to select one of them as a more appropriate window size. Each site was defined as a peptide segment of 2w+1 length with lysine (K) in the center, where 2w+1 is the window size. We observed that the predictor was optimized corresponding to the window size 2w+1=25. Thus, we considered positive and negative window samples corresponding to the window size 25 to develop a better predictor. Obviously, positive window samples (n1=1195) and negative window samples (n2=16842) in the training dataset were unbalanced. To construct a comparatively balanced dataset, the training datasets were constructed at three ratios 1:1, 1:2, and 1:3 of positive and negative window samples, respectively by randomly taking the negative windows out of n2=16842 for each ratio case. The training dataset of ratio 1:1 was constructed by taking all 1195 positive windows with the randomly selected 1195 negative windows out of 16842. The training dataset of ratio 1:2 was constructed by taking all 1195 positive windows with the randomly selected 1195x2=2390 negative windows out of 16842. Similarly, the training dataset of ratio 1:3 was constructed by taking all 1195 positive windows with the randomly selected 1195x3=3585 negative windows out of 16842. We developed the prediction model with each training dataset and investigated their performance by using 20-fold cross-validation (CV) and two independent test datasets as introduced previously. We observed that the predictor was optimized corresponding to the 1:2 ratio of positive and negative window samples. To convert the protein window sequence data to numeric data, we considered three popular encoding schemes i.e. CKSAAP, binary and AAC (see section 2.4). To remove the unimportant encoded features from the dataset for reducing the computational load of the predictor, we considered a non-parametric feature selection approach known as Kruskal-Wallis (KW) [22] to select a better one of them. We trained 3 different classifiers ADA [23], SVM [27] and RF [28] (see section 2.6) separately based on the selected features of each encoding system. Then we improved the prediction model by fusing three encoding schemes (see section 2.7). We observed that the proposed RF-based fusion model is optimized corresponding to the 1:2 ratio of positive and negative window samples. Thus, we built an improved computational prediction model as displayed in Fig. ( (see the results section for further more discussion).
Fig. (1)

Overview on the development of the proposed predictor.

Two Sample Logo (TSL) Analyses

The TSL analysis of protein sequences is used to visualize significant differences between the amino acid samples of positive and negative window groups. It determines the statistically significant residues (amino acids) around the protein PTM site for finding the differences between the two window groups. Statistical evidence is computed for each amino acid at every position between two window groups under the null hypothesis (Ho) that the residue samples follow the same distribution in both positive and negative window groups. Let A and B be two groups of protein sequences based on the positive and negative windows. Let |A| and |B| be the number of sequences in these groups. Let N denote the length of each window in both groups. Let A be the ith sequence in group A and let A be the jth position in A. where r is the symbol of a residue. The vector is formed conversely. Then we calculated the p-value of H0 that both vectors and follow the same distribution. To test H0, one of the two types of statistical tests (i) two-sample t-test and (ii) binomial test, is used usually. It should be noted here that the t-test is less accurate but significantly faster, while the binomial test is more accurate, but significantly slower [29]. The TSL exhibits two kinds of graphical image: (i) significant amino acid symbols are plotted using the size of the symbol that is proportional to the difference between the two amino acid samples; (ii) significant amino acids are plotted based on the same size for each amino acid symbol. Amino acids are divided into two groups: (i) enriched in the positive window samples, and (ii) depleted in the positive window samples.

Data Encoding Scheme

To construct a robust prediction model, a numeric feature vector is essential to train the classifiers. There are several encoding approaches in the compositions to covert the sequence data into numeric data. Here, we considered three encoding approaches as discussed below:

CKSAAP Encoding

In various PTMs site prediction, there is a powerful feature encoding scheme named the composition of k-spaced amino acid pairs (CKSAAP) [19]. It has been predominantly used in the numerous bioinformatics work [30-34]. In this paper, a sequence fragment of 25 amino acids is detected from the Succinylation or non- Succinylation site. For every single k (k denotes the gap between two amino acids), it may create (21 21) = 441 (21 means 21 kinds of amino acids with the gap (O)) types of amino acid pairs (i.e. AA, AC, AD, . . ., OO), if window size of a fragment is 2r + 1. For each sequence, there is 21 (kmax + 1) 21 = 2646 distinct amino acid combination developed for the highest k taking k max =5. Then, the feature vectors are calculated using the following equation: (,,, …, )441, (1) where, Ntotal denotes the total composition residues length. ,,..., are fragments’ frequency of the amino acid pair. More details are available somewhere [14, 34]. A schematic diagram of CKSAAP encoding was depicted in Fig. (.
Fig. (2)

A schematic diagram of CKSAAP encoding. (A higher resolution/colour version of this figure is available in the electronic copy of the article).

Binary Encoding

In the binary encoding scheme, the number of residues may be less than 25 for the sites placed in k-terminal. In this binary encoding approach, 21 amino acids (including gap (O)) are reshaped to numeric vectors. Thus, 21 different amino acids (such as, ACDEFGHIKLMNPQRSTVWYO) are organized in this encoding scheme. Each amino acid is displayed in the query proteins by a 21-dimensional binary vector, e.g. A: 100000000000000000000, C: 010000000000 000000 000, …, O: 000000000000000000001, etc. In each window of succinylation site, the central location is always K, which is unnecessary to be taken into account. The whole dimension of this encoding scheme is (21X(25-1)) = 504 when we select a window of size 25. Details are described in previous studies [15, 16].

AAC Encoding

Amino acid composition (AAC) is a typical attribute used to examine substrate site motifs. AAC determines the probability of amino acids occurring in the flanking region of PTM sites [35, 36]. It can create protein arrangements data by recreating amino acid event frequencies. In this research, AAC was determined dependent on amino acid event frequencies in the arrangement sections encompassing the succinylation and non-succinylation sites (the site itself is not tallied). For every grouping section, 20 frequencies were determined for 20 types of amino acids. Given a divided arrangement x with a 25-mer string length, nx(m) is the quantity of a particular amino acid, m, occurring in the section, where m indicates the 20 amino acids. Thus, the probability Px(m) of particular amino acid m is [37] Px(m) = ; k=1,…, 20 Then, the creation of the 20 amino acids can be changed to a 20-dimensional numeric vector Vx for the divided sequence x: Vx = [Px(1), Px(2), …, Px(20)]

Feature Selection from the Encoded Data

A high dimensional feature vector is created by encoding each of the succinylated and non-succinylated fragments as discussed in section 2.4. However, some features are not so important to develop the prediction model. These unimportant features are removed from the dataset to reduce the computation burden from the prediction model. An encoded feature is said to be unimportant if its mean difference between succinylated and non-succinylated groups is statistically insignificant. In this study, we considered the top 1500 and 400 most important features based on CKSAAP and binary encoding schemes respectively to develop the prediction model. To remove the unimportant features from the prediction model, we considered the non-parametric Kruskal-Wallis (KS) test procedure [22], since the distribution of feature components is unknown.

Learning Classifiers

To build a better predictor for protein Succinylation site prediction, we considered 3 popular classifiers namely, Random Forest (RF), AdaBoost (ADA), & Support Vector Machine (SVM), for a comparison based on the encoded protein sequences. Let us consider a dataset consisting of n training data (,), (,)... (,), where is an input vector in space X⊆ and is the response variable that takes value +1 (succinylated site) and -1 (non-succinylated site). The objective is to classify a new window sample into one of two classes (+1, -1). For convenience of the readers, let us introduce together those classifiers as follows:

Random Forest (RF)

The random forest (RF) classifier is a statistical learning algorithm. It is widelyhttps://www.thesaurus.com/browse/ magnificently used in protein bioinformatics [19, 24, 28, 30-32, 36-43]. Generally, it has two stages: one is random forest creation, and the other is to make a prediction from the random forest classifier created in the first stage. For the convenience of presentation, let (X, Y) ={(,), (,)... (,)}. Then B (b = 1, …, B) times selects a random sample (, with replacement from the given dataset (X,Y) and trains a regression tree on (, ) to fit trees to these samples. After training, predictions for new samples ’ can be written as, (2) In this study, we have applied the RF classifier through the ‘randomForest’ R package [44].

AdaBoost

AdaBoost is an adaptive boosting machine learning meta-algorithm [23]. It is adaptive in the sense that subsequent weak learners are tweaked in favor of those instances misclassified by previous classifiers. Its technical description is as follows: Feature vector: (x xx) Training dataset: { (xi,yi); i=1,2,…,n } Suppose there are T weak classifiers defined by satisfying, Then the AdaBoost classifier is defined by: , , Then the classification rule is defined as: In this paper, R package ‘ada’ was used to execute the AdaBoost algorithm [45].

Support Vector Machine (SVM)

The main objective of SVM is to find a hyperplane in an m-dimensional space that clearly classifies the data points [17, 19, 27, 46, 47]. Let us consider the data points consist of n training data (,), (,)... (,), where is an input vector in space X⊆ and is the output variable that takes value 1 (succinylated site) and -1 (non-succinylated site). The SVM approach constructs a hyperplane in high dimensional space, which can be used in both classification and regression. The hyperplane can be written as follows: (4) Where b is scalar and W is an m-dimensional normalized vector and perpendicular to the separating hyperplane. If the data are linearly separable, the two classes are as follows: and. If the data is not linearly separable, then SVM uses kernel functions to transform the original data space with high dimension space that can easily separate the classes as succinylated site and non-succinylated site. In such a case, the hyperplane can be written as: (5) where, is Lagrange multiplier, is the class label that belongs to (-1, 1), and is the Kernel function between and . In this study, we have adopted kernel as a radial basis function (RBF). In this paper, R package ‘e1071’ was used to execute the SVM algorithm [48].

Fusion Model (Proposed)

Several authors used fusion technique to improve the performance of their prediction models [33, 40, 41]. In this article, we also attempted to improve the performance of our prediction model by fusing 3 encoding schemes Binary, CKSAAP and AAC with RF classifier as follows: RF(CKSAAP, Binary, AAC) = w1×RF(CKSAAP) + w2×RF(Binary) + w3×RF(AAC),……. .. (6) where RF(Binary), RF(CKSAAP) and RF(AAC) represent the RF classification scores estimated with binary, CKSAAP and AAC encoding schemes, respectively. The values of w1, w2 and w3 were selected based on the ratio of individual prediction performance of RF(Binary), RF(CKSAAP) and RF(AAC) satisfying w1 + w2 + w3= 1. Similarly, we improved the prediction performance of ADA and SVM classifiers by fusing binary, CKSAAP and AAC encoding schemes to compare with the proposed RF-based prediction model.

Performance Evaluation Measures

In order to compare the performance of different candidate prediction models, we considered some popular widely used measurements including true positive rate (TPR), which is also known as sensitivity (SN), true negative rate (TNR), which is also known as specificity (SP), false-positive rate (FPR), false-negative rate (FNR), accuracy (ACC), misclassification rate (MCR) and Mathew correlation coefficient (MCC), receiving operating characteristics (ROC) curve, area under the ROC curve (AUC) and partial AUC (pAUC). These measurements are defined as follows: (7) (8) (9) (10) (11) (12) (13) where, n(TN): number of true negative, n(TP): number of true positive, n(FN): number of false negative, n(FP): number of false positive. The ROC curve is created by plotting TPR against FPR. Obviously TPR+FNR=1, TNR+FPR=1, (FPR, FNR)→(0,0) implies MCR→0 and (TPR, TNR, ACC, MCC, AUC)→(1,1,1,1,1), conversely (FPR, FNR)→(1,1) implies MCR→1 and (TPR, TNR, ACC, MCC, AUC)→(0,0,0,0,0,-1). Therefore, the larger scores of TPR, TNR, ACC, MCC, AUC and conversely the smaller scores of FPR, FNR and MCR indicate better prediction performance. The more detailed descriptions about these evaluation measures were given in Supplementary file (S1).

K-fold Cross-Validation (CV)

In order to perform K-fold CV, the training dataset ‘D’ was randomly partitioned into k =20 mutually exclusives subsets (D1, D2,…,DK) of almost equal sizes. Among K=20 subsets, the (K– 1)=19 subsets were used as training data to train the prediction model, and the remaining one subset was used as the test subset to evaluate the performance of the prediction models. This procedure was then repeated K=20 times by replacing one subset from the training dataset with the test subset such that each subset is used only one time as the test dataset. Then, the average value of K =20 scores was used as a single score for each of the performance measures (TPR/SN, TNR/SP, FPR, FNR, MCR, ACC, MCC & AUC).

Results

As mentioned previously in section 2.2, the prediction performance is not only influenced by the estimation of the model parameters but also depends on the selection of tuning parameters like CD-HIT threshold (CHT), window size (WS) and ratio of positive and negative windows. To select comparatively better combination of CHT, WS and ratio; we considered three CHTs (30%, 40%, 50%), five WSs (19, 25, 31, 37, 41) and three ratios (1:1, 1:2, 1:3). We developed seven-candidate prediction models based on three encoding schemes (Binary, CKSSAP and AAC) with each of three classifiers (ADA, SVM and RF) for each of the 45 combinations of CHT, WS and ratio. By the trial and performance checking strategy, we observed that the combination of CHT=30%, WS=25 and ratio=1:2 with each prediction model optimizes their performance scores (TPR/SN, TNR/SP, FPR, FNR, MCR, ACC, MCC & AUC) out of different 45 (=3×5×3) combinations. That is, the prediction performance scores (SN, SP, ACC, MCC & AUC) were maximized and conversely, the alternative performance scores (FPR, FNR & MCR) were minimized at CHT=30%, WS=25 and ratio=1:2. Thereafter, we only discussed the performance scores of 7×3=21 different prediction models based on three classifiers (ADA, SVM and RF) corresponding to the combination of tuning parameters at CHT=30%, WS=25 and ratio=1:2 to select the better prediction model. To examine the adequacy of the training dataset for the window size 25, two sample logo (TSL) analyses were performed in section 3.1. To select the better prediction model in a comparison of the others, first, we assessed the training performance scores in section 3.2. Then we investigated the prediction performance by using 20-fold CV and discussed the results in section 3.3. The performance scores based on two independent test datasets were discussed in section 3.4. Then we compared the performance of the proposed prediction model with four existing models [14, 16-18] in section 4.

The TSL Analysis

We performed two-sample logo (TSL) analyses to investigate the adequacy of the dataset. By TSL software, the neighboring succinylation and non-Succinylation sites are shown for the training dataset in Fig. ( [49]. The positive or negative samples represented residues at each location above and under the X-axis plotted respectively, in proportion to the percentage of over represented (if positive) or underrepresented samples (if negative) shown by the height of the letter denoting the resultant residue. Y-axis indicates the cumulative percentage of these positive / negative residues. TSL logos represent the amino acid occurrences between positive and negative samples of succinylation protein. Only residues that were significantly enriched or depleted (t-test, P<0.05) flanking the centered succinylation sites are shown in Fig. (. Significant differences were observed between positive and negative samples. We found that at specific points, some amino acids are over / under represented, which shows that the positional amino acid encoding is essential to identify the succinylation sites.
Fig. (3)

The occurrences of amino acid propensities of surrounding positive windows (succinylation site) and negative windows (non-succinylation sites) of size 25 are presented by the Two-Sample Logos software (Vacie et al., 2006) [29]. (A higher resolution/colour version of this figure is available in the electronic copy of the article).

The binary encoding is encoded position-wise. Thus, the efficiency of binary encoding (Table ) is sufficient to exactly identify the succinylation sites. In the following calculation and operation, we selected 25-mer (-12, +12) window size and Fig. ( shows the position-specific difference of amino acid compositions between succinylation sites and non-succinylation sites. We also provided the two sample logos for other window sizes (19, 31, 37 & 41) in the Supplementary file (Figs. -) for a visual representation of patterns of amino acid conserved-ness around the lysine succinylation sites.

Assessment of Prediction Performance with the Training Dataset

To investigate the prediction performance of the proposed model in comparison of the other candidate models, all prediction models were trained by the training dataset corresponding to the ratio 1:2 of 1195 positive and 2390 negative windows as discussed in section 2.2. Then we computed prediction performance scores SN, SP, ACC, MCC and AUC as well as the alternative prediction performance scores FPR, FNR and MCR to investigate the training performance of the proposed fusion prediction model RF(Binary, CKSAAP, AAC) in a comparison of the other 26 candidate models ADA(Binary), ADA(CKSAAP), ADA(AAC), ADA(CKSAAP, Binary), ADA(CKSAAP, AAC), ADA(Binary, AAC), ADA(CKSAAP, Binary, AAC), SVM (Binary), SVM(CKSAAP), SVM(AAC), SVM (CKSAAP, Binary), SVM(CKSAAP, AAC), SVM(Binary, AAC), SVM(CKSAAP, Binary, AAC), RF(Binary), RF(CKSAAP), RF(AAC), RF(CKSAAP, Binary), RF(CKSAAP, AAC) and RF(Binary, AAC). Table shows the summary of different training performance scores (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC) with all prediction models. At first, we observed that binary encoding produced slightly better results with all three classifiers (ADA, SVM and RF) than the CKSAAP and AAC encodings. Then we observed that the fusion model RF(CKSAAP, Binary, AAC) produced the highest prediction performance scores of TPR (0.869), TNR (0.902), ACC (0.937), MCC (0.781), pAUC (0.198) at FPR=0.10 and AUC (0.959), and the smallest alternative prediction performance score of FNR (0.131) and MCR (0.100) at FPR=0.10 against any of the other prediction models. Similar performance trend was also observed at FPR=0.20 (See Table in the Supplementary file). Thus, the proposed RF-based fusion model shows better prediction performance than the other candidate prediction models with the training dataset.

Assessment of Prediction Performance based on 20-fold Cross-validation (CV)

To investigate the test performance of the proposed RF-based fusion prediction model in a comparison of the other candidate prediction models by 20-fold CV, the training dataset corresponding to 1:2 ratio of 1195 positive and 2390 negative window samples was partitioned into 20 mutually exclusive subgroups (G1, G2, G3,…, G20) of almost equal sizes such that each subgroup contains 1:2 ratio of positive and negative samples. Obviously, each subgroup consisted of around 59 (5% of 1195) positive window samples and around 118 (95% of 2390) negative window samples. The 20-fold CV was completed within 20 steps. At step 1, nineteen (19) groups (G2, G3, G4,…, G20), those contained around 95% window samples satisfying 1:2 ratio of positive and negative window samples, were used to train all prediction models ADA(Binary), ADA(CKSAAP), ADA(AAC), ADA(CKSAAP, Binary), ADA(CKSAAP, AAC), ADA(Binary, AAC), ADA(CKSAAP, Binary, AAC), SVM(Binary), SVM(CKSAAP), SVM(AAC), SVM(CKSAAP, Binary), SVM(CKSAAP, AAC), SVM(Binary, AAC), SVM(CKSAAP, Binary, AAC), RF(Binary), RF(CKSAAP), RF(AAC), RF(CKSAAP, Binary), RF(CKSAAP, AAC), RF(Binary, AAC), and proposed RF(CKSAAP, Binary, AAC). The other group G1 was used to evaluate the trained models by using different performance measures (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC). At step 2, nineteen (19) groups (G1, G3, G4,…, G20), those also contained around 95% window samples were used to train all the prediction models as before. The other group G2 was used to evaluate the trained models by using different performance measures (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC) as before. In this step, G1-G2 subgroups interchanged their positions between the training and test sets only. Similarly, at other eighteen (18) steps/loops, subgroup pairs G2-G3, G3-G4, G4-G5, …, G19-G20 interchanged their positions between the training and test sets, respectively. Then we calculated the average performance scores of SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC. Table shows the summary of average performance scores of different candidate predictors based on 20-fold CV with a 1:2 ratio of positive and negative window samples in the training dataset with a cutoff point at FPR=0.10. We also calculated the standard error (SE) for each performance measure to observe the stability of the estimates. The values within the first bracket in Table indicate the SE. A prediction model with the smallest SE of performance scores is known as a more stable predictor. We observed that binary encoding produces slightly better results with all three classifiers (ADA, SVM and RF) than the AAC and CKSAAP encoding as before. Then we observed that the fusion model RF(CKSAAP, Binary, AAC) produces the highest average prediction performance scores of TPR (0.800), TNR (0.902), ACC (0.919), MCC (0.766), AUC (0.958) and pAUC (0.163) at FPR=0.10, and the smallest average of alternative prediction performance scores of MCR (0.141) in a comparison of the other prediction models. Similar performance trend was also observed at FPR=0.20 (See Table in the supplementary file). The ROC curves given in the Supplementary Figs. also supported these results. The SE values for each performance are measured corresponding to the RF(CKSAAP, Binary, AAC) prediction model as much smaller compared to any of the other twenty (20) prediction models. So RF (CKSAAP, Binary, AAC) prediction model would be more efficient than the other prediction model. Thus, the RF-based fusion model shows better prediction performance compared to other candidate prediction models by cross-validation.

Assessment of Prediction Performance with the 
Independent Test Datasets

The prediction performance of the proposed model was compared with the other candidate prediction models by using two independent test datasets as discussed below:

Performance with the Independent Test Dataset-1

All candidate prediction models including the proposed RF(CKSAAP, Binary, AAC) were trained by the training dataset corresponding to a 1:2 ratio of 1195 positive and 2390 negative windows as discussed previously in section 3.2. We considered independent test dataset-1 that was introduced in section 2.2 to demonstrate the performance of the proposed RF(CKSAAP, Binary, AAC) prediction model in a comparison of the other candidate twenty (20) prediction models ((ADA(Binary), ADA(CKSAAP), ADA(AAC), ADA(CKSAAP, Binary), ADA(CKSAAP, AAC), ADA(Binary, AAC), ADA(CKSAAP, Binary, AAC), SVM The independent test dataset-1 was consisted based on 50 proteins with 54 positive windows and 2004 negative windows as mentioned in section 2.2. It was used to evaluate the trained models by using different performance measures (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC) as before. Table shows different performance scores (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC) of different candidate prediction models with the independent test dataset-1. It is clearly observed that the proposed RF(CKSAAP, Binary, AAC) prediction model gives the highest correct prediction performance scores of TPR (0.798), TNR (0.902), ACC (0.891), MCC (0.629), AUC (0.921) and pAUC (0.139) at FPR=0.10 and the smallest incorrect prediction performance scores of FNR (0.201) and MCR (0.145) against the other prediction models. Fig. ( also supports the results of Table . A similar performance trend was also observed at FPR=0.20 (Table in the Supplementary file). Thus, the proposed RF(CKSAAP, Binary, AAC) prediction model shows better performance than the other candidate prediction models with the independent test dataset-1 also. A comparative discussion between the proposed prediction model and two existing models has been given in section 4.
Fig. (4)

ROC curve to display the performance of different candidate predictors based on independent test dataset-1. (A higher resolution/colour version of this figure is available in the electronic copy of the article).

Performance with the Independent Test Dataset-2

To investigate the performance of the proposed RF(CKSAAP, Binary, AAC) prediction model in a comparison of the others more fairly, we considered the independent test dataset-2 that was introduced in section 2.2. For the convenience of presentation, again it should be mentioned here that the training dataset and the independent test dataset-1 were generated from the same data source, while the independent test dataset-2 was collected from the other independent data source. This independent test dataset-2 was based on 202 lysine succinylated proteins, which contained 463 positive and 6742 negative window sites. Table shows the summary of different prediction performance scores (SN, SP, FNR, ACC, MCR, MCC, ROC, AUC and pAUC) of different candidate predictors with the independent test-2 dataset. It is clearly observed that proposed RF(CKSAAP, Binary, AAC) prediction model gives the highest correct prediction performance scores of TPR (0.772), TNR (0.901), ACC (0.886), MCC (0.677), AUC (0.923) and pAUC (0.141) at FPR=0.10 and the smallest incorrect prediction performance score of FNR (0.227) and MCR (0.163) against to other twenty prediction models (ADA (Binary), ADA(CKSAAP), ADA(AAC), ADA (CKSAAP, Binary), ADA(CKSAAP, AAC), ADA(Binary, AAC), ADA(CKSAAP, Binary, AAC), SVM(Binary), SVM (CKSAAP), SVM(AAC), SVM(CKSAAP, Binary), SVM (CKSAAP, AAC), SVM (Binary, AAC), SVM (CKSAAP, Binary, AAC), RF (Binary), RF (CKSAAP), RF (AAC), RF (CKSAAP, Binary), RF (CKSAAP, AAC), RF (Binary, AAC)) as like as 20-fold CV results. Fig. ( also supports the results of Table . A similar performance trend was also observed at FPR=0.20 (Table in the Supplementary file). Thus, the proposed RF(CKSAAP, Binary, AAC) prediction model shows better performance than the other candidate prediction models also with the independent test dataset-2.
Fig. (5)

ROC curve to display the performance of different candidate predictors based on independent test dataset-2. (A higher resolution/colour version of this figure is available in the electronic copy of the article).

Discussion

We developed an improved computational prediction model by maximizing the prediction performance scores (SN, SP, ACC, MCC & AUC) and conversely minimizing the alternative prediction performance scores (FPR, FNR & MCR) with respect to different model parameters and some tuning parameters like a cut-off value of CD-HIT at 30%, window size at 25, a ratio of positive and negative windows at 1:2 based on the combined model of three encoding schemes (binary, CKSAAP and AAC) with the random forest (RF) classifier to predict lysine succinylation sites mapping on Homo sapiens. We observed that all candidate prediction models show much better performance with both 1:2 and 1:3 ratios of positive and negative window samples than 1:1 ratio. We also observed that their performance with 1:2 and 1:3 ratios of positive and negative window samples was not so different significantly. Therefore, we considered the ratio 1:2 of positive and negative window samples to develop the prediction model by reducing the computational cost. The performance of the proposed predictor in a comparison of the other candidate predictors was investigated by using the prediction performance scores with the training dataset and independent test performance scores based on 20-fold cross-validation and two independent test datasets that were collected from two different sources. It should be mentioned here that the training dataset and the independent test dataset-1 were collected from the same data source, while the independent test dataset-2 was collected from the other data source. We observed that binary encoding produces slightly better performance scores with all three classifiers (ADA, SVM and RF) than the CKSAAP and AAC encoding in all cases (Tables -, -, Figs. and ). So, we imposed more weight on the binary encoding than the CKSAAP and AAC encoding during the construction of the fusion model. Then we observed from Table and as discussed in the previous section 3.2 that the proposed RF(CKSAAP, Binary, AAC) model shows much better prediction performance compared to other candidate prediction models (ADA(Binary), ADA(CKSAAP), ADA(AAC), ADA(CKSAAP, Binary), ADA(CKSAAP, AAC), ADA(Binary, AAC), ADA(CKSAAP, Binary, AAC), SVM(Binary), SVM(CKSAAP), SVM(AAC), SVM(CKSAAP, Binary), SVM(CKSAAP, AAC), SVM(Binary, AAC), SVM(CKSAAP, Binary, AAC), RF(Binary), RF(CKSAAP), RF(AAC), RF(CKSAAP, Binary), RF(CKSAAP, AAC), RF(Binary, AAC)). Similarly, Tables 2, and as discussed in the previous section 3.3 show that the proposed RF-based fusion prediction model outperforms the other candidate prediction models in the case of 20-fold CV. Tables , , - and Figs. ( and ) showed that the proposed RF(CKSAAP, Binary, AAC) fusion prediction model outperforms the other candidate prediction models also with both independent test datasets-1&2. Furthermore, we considered the test dataset-1 to investigate the performance of the proposed RF(CKSAAP, Binary, AAC) model with four existing species-wise prediction models known as SuccinSite2.0 [14], CNN-SuccSite [16]), HybridSucc [17] and GPSuc [18] for the prediction of lysine succinylation sites using the same independent test dataset as mentioned previously in section 2.2. We considered five important performance measures (SN, SP, ACC, MCC and AUC) to compare the proposed method with these four existing methods. Table shows 5 performance scores with the same independent dataset-1 for the mentioned four existing methods and the proposed method. We observed that the proposed method greatly improves the performance over the existing four methods with respect to all 5 measures. Especially, MCC values increased and the AUC value was 4-8% higher than the existing prediction models. Thus, our proposed prediction model outperformed the existing prediction models. Noted that, until now, only four predictors are publicly available. All of the other existing predictors are not publicly available. Therefore, to make a fair comparison, we compared these four existing prediction models. In addition, to examine our prediction model robustly, we considered the independent test dataset-2 which was already introduced in section 2.2. It consisted of 202 human pathogen Histoplasma capsulatum succinylated proteins which contained 463 modification sites. We found from Tables and , and Fig. ( that our prediction model can be applied for Histoplasma capsulatum succinylation site prediction.

Conclusion

We developed an improved predictor based on the sequence information to predict lysine succinylation sites mapping on Homo sapiens by fusing three encoding schemes (binary, CKSAAP and AAC) with the random forest machine learning framework. We performed a comparative study on the prediction of succinylation sites by using the empirically developed succinylated protein sequences of Homo sapiens. The investigational results by 20-fold CV and two independent test-sets show that the proposed method can identify succinylated sites more accurately than the other candidate prediction models. Moreover, the benchmarking experiments demonstrated that our proposed predictor gives a competitive performance compared to the existing four predictors. The proposed method may be a useful and encouraging computational resource for lysine succinylation site prediction in the case of Human PTMs. To implement the proposed method, the computational codes and necessary instructions can be downloaded from http://www.ru.ac.bd/biorgru/software/succinsitefusing-zip/. To further improve the prediction performances, we may use newly proposed encoding schemes and integrated approaches [26, 50-54].
Table 1

Summary performance at FPR=0.10 for different candidate predictors based on the training dataset corresponding to 1:2 ratio of positive and negative windows.

Predictors Classifier(Encodings) TPR (SN) TNR (SP) FNR ACC MCC MCR AUC pAUC
ADA(Binary)0.7520.9010.2480.8670.6520.1870.8810.131
ADA(CKSAAP)0.7470.9000.2530.8530.6370.2210.8650.110
ADA(AAC)0.7500.9020.2500.8620.6430.1980.8760.115
ADA(CKSAAP, Binary)0.7630.9010.2370.8710.6870.1780.9020.142
ADA(CKSAAP, AAC)0.7570.9010.2430.8680.6560.1890.8870.133
ADA(Binary, AAC)0.7610.9000.2390.8690.6580.1860.8990.139
ADA(CKSAAP,Binary,AAC)0.7690.9010.2310.8730.6900.1710.9080.144
SVM(Binary)0.6430.9020.3570.7560.5430.2660.8220.069
SVM(CKSAAP)0.6320.9030.3680.7340.5320.2860.8120.073
SVM(AAC)0.6380.9010.3620.7370.5410.2680.8200.079
SVM(CKSAAP, Binary)0.6650.9010.3350.7770.5730.2500.8410.069
SVM(CKSAAP, AAC)0.6680.9020.3320.7790.5750.2430.8480.071
SVM(Binary, AAC)0.6750.9010.3250.7810.5780.2410.8500.072
SVM(CKSAAP,Binary,AAC)0.6780.9020.3220.7880.5810.2340.8520.076
RF(Binary)0.8330.9020.1670.9120.8580.1210.9400.195
RF(CKSAAP)0.8010.9030.1990.9030.7030.1340.9040.181
RF(AAC)0.7610.9020.2390.9050.6270.1590.9130.129
RF(CKSAAP, Binary)0.8540.9000.1460.9300.7760.1180.9400.156
RF(CKSAAP, AAC)0.8570.9010.1430.9320.7780.1130.9420.157
RF(Binary, AAC)0.8590.9010.1410.9340.7790.1100.9470.161
RF(CKSAAP,Binary,AAC)0.8690.9020.1310.9370.7810.1000.9650.198
Table 2

Summary of average performance at FPR=0.10 for different candidate predictors based on 20-fold CV with 1:2 ratio of positive and negative window samples in the training dataset.

Predictors Classifier(Encodings) TPR TNR FNR ACC MCC MCR AUC pAUC
ADA(Binary)0.3780.9020.6210.5930.2050.4070.7340.048
ADA(CKSAAP)0.3640.9030.6350.5890.1990.4100.7260.050
ADA(AAC)0.3440.9010.6550.6070.2510.3920.7020.030
ADA(CKSAAP, Binary)0.5570.8990.4430.6830.3760.3180.7830.069
ADA(CKSAAP, AAC)0.4780.9010.5210.6700.3680.3290.7740.059
ADA(Binary, AAC)0.5590.9020.4410.6850.3770.3860.7880.078
ADA(CKSAAP,Binary,AAC)0.6120.9010.4870.7210.4560.3020.8260.110
SVM(Binary)0.4560.9000.5150.6500.3170.3500.7450.058
SVM(CKSAAP)0.3430.9010.6570.5780.1780.4210.7190.044
SVM(AAC)0.2810.9010.7160.5740.1820.4250.7080.034
SVM(CKSAAP, Binary)0.5570.9030.4420.6850.3840.3140.7790.068
SVM(CKSAAP, AAC)0.5430.9020.4560.6670.3760.3560.7660.054
SVM(Binary, AAC)0.5670.9010.4320.6840.3820.3240.7710.069
SVM(CKSAAP,Binary,AAC)0.5980.9020.4010.7000.4220.3120.8140.100
RF(Binary)0.7250.9020.2750.8710.6880.1750.8950.132
RF(CKSAAP)0.6910.9010.3080.7860.5840.1830.8690.123
RF(AAC)0.6810.8990.3190.8590.6720.1920.8770.124
RF(CKSAAP, Binary)0.7890.9020.2110.9170.7610.1530.9480.153
RF(CKSAAP, AAC)0.7110.8990.2490.8690.6820.1720.8910.134
RF(Binary, AAC)0.7250.9020.2750.8710.6880.1750.8950.132
RF(CKSAAP,Binary,AAC)0.8000.9020.2000.9190.7660.1410.9580.163

(Binary), SVM(CKSAAP), SVM(AAC), SVM(CKSAAP, Binary), SVM(CKSAAP, AAC), SVM(Binary, AAC), SVM(CKSAAP, Binary, AAC), RF(Binary), RF(CKSAAP), RF(AAC), RF(CKSAAP, Binary), RF(CKSAAP, AAC), RF(Binary, AAC)) along with two existing predictors [14, 18].

Table 3

Summary performance at FPR=0.10 for different candidate predictors based on independent test dataset-1.

Predictors Classifier(Encodings) TPR (SN) TNR (SP) FNR ACC MCC MCR AUC pAUC
ADA(Binary)0.6450.9010.3450.7720.5630.2270.8410.101
ADA(CKSAAP)0.6370.9020.3630.7690.5580.2310.8120.100
ADA(AAC)0.6980.9020.3010.7610.5260.2380.8200.102
ADA(CKSAAP,Binary)0.7030.9020.2970.8110.6660.2080.8660.122
ADA(CKSSAP,AAC)0.7010.9010.2980.8120.6670.2090.8560.121
ADA(Binary,AAC)0.7320.9020.2670.7910.5870.2080.8810.123
ADA(CKSSAP,Binary,AAC)0.7390.9010.2600.7990.6020.2000.8890.138
SVM(Binary)0.3930.9010.6070.6460.3390.3530.7520.069
SVM(CKSAAP)0.3880.9010.6110.6440.3350.3550.7420.073
SVM(AAC)0.3890.9010.6100.6450.3360.3530.7470.075
SVM(CKSAAP,Binary)0.4250.9010.5750.6690.3820.3100.7620.069
SVM(CKSSAP,AAC)0.3890.9010.6100.6450.3360.3530.7480.075
SVM(Binary,AAC)0.4790.9020.5200.6770.3870.3220.7930.098
SVM(CKSSAP,Binary,AAC)0.4820.9010.5180.6790.3970.3110.7960.103
RF(Binary)0.7420.9020.2670.8510.6030.1890.8980.125
RF(CKSAAP)0.7010.9010.2990.8100.5930.2200.8640.121
RF(AAC)0.7320.9020.2670.7710.5440.2280.8680.124
RF(CKSAAP,Binary)0.7610.9020.2390.8600.6270.1590.9130.129
RF(CKSSAP,AAC)0.7420.9020.2570.7810.5540.2080.8780.128
RF(Binary,AAC)0.7610.9020.2390.8600.6270.1590.9130.129
RF(CKSSAP,Binary,AAC)0.7980.9020.2010.8910.6290.1450.9210.139
Table 4

Summary performance at FPR=0.10 for different candidate predictors based on independent test dataset-2.

Predictors Classifier(Encoding) TPR (SN) TNR (SP) FNR ACC MCC MCR AUC pAUC
ADA(Binary)0.6270.9010.3730.7630.5470.2360.8620.110
ADA(CKSAAP)0.6260.9020.3740.7630.5470.2360.7930.093
ADA(AAC)0.7020.9010.2970.7560.5150.2430.8160.100
ADA(CKSAAP,Binary)0.6430.9030.3570.8120.6020.1890.8740.113
ADA(CKSSAP,AAC)0.6310.9010.3680.7830.5920.2170.8460.109
ADA(Binary,AAC)0.6980.9020.3010.8150.6540.1820.8930.129
ADA(CKSSAP,Binary,AAC)0.7000.9020.2990.8160.6560.1810.8950.130
SVM(Binary)0.6040.9020.3960.7040.5580.3020.7750.077
SVM(CKSAAP)0.4530.9030.6520.6130.3240.543O.6890.05
SVM(AAC)0.5810.9010.4180.6720.3510.3270.7400.069
SVM(CKSAAP,Binary)0.6040.9010.3960.7040.5580.3020.7760.077
SVM(CKSSAP,AAC)0.5810.9010.4180.6720.3510.3270.7400.069
SVM(Binary,AAC)0.6670.9020.3320.7270.4580.2720.8070.089
SVM(CKSSAP,Binary,AAC)0.6780.9010.3210.7310.4660.2680.8100.099
RF(Binary)0.7240.9020.2760.8450.6800.2040.9020.125
RF(CKSAAP)0.6930.9020.3070.8040.6660.2250.8470.114
RF(AAC)0.6150.9010.3840.8010.6510.1980.8770.121
RF(CKSAAP,Binary)0.7450.9010.2550.8810.6010.1480.9100.138
RF(CKSSAP,AAC)0.7020.9020.2980.7970.6080.2020.8780.122
RF(Binary,AAC)0.7490.9020.2500.8840.6690.1690.9200.140
RF(CKSSAP,Binary,AAC)0.7720.9010.2270.8860.6770.1630.9230.141
Table 5

Performance comparison of the proposed predictor with other existing Predictors.

Prediction methods Independent Test Dataset TPR (SN) TNR (SP) ACC MCC AUC
Succinsite2.0Dataset-10.6320.8720.8660.2410.845
GPSucDataset-10.6930.8770.8720.2790.885
HybridSuccDataset-10.8220.8550.8590.5620.891
CNN-SuccSiteDataset-10.7160.8440.8420.4430.839
Proposed PredictorDataset-1Dataset-20.7980.7720.9020.9010.8910.8860.6290.6770.9210.923
  44 in total

1.  Two Sample Logo: a graphical representation of the differences between two sets of sequence alignments.

Authors:  Vladimir Vacic; Lilia M Iakoucheva; Predrag Radivojac
Journal:  Bioinformatics       Date:  2006-04-21       Impact factor: 6.937

2.  i4mC-ROSE, a bioinformatics tool for the identification of DNA N4-methylcytosine sites in the Rosaceae genome.

Authors:  Md Mehedi Hasan; Balachandran Manavalan; Mst Shamima Khatun; Hiroyuki Kurata
Journal:  Int J Biol Macromol       Date:  2019-12-02       Impact factor: 6.953

Review 3.  Machine intelligence in peptide therapeutics: A next-generation tool for rapid disease screening.

Authors:  Shaherin Basith; Balachandran Manavalan; Tae Hwan Shin; Gwang Lee
Journal:  Med Res Rev       Date:  2020-01-10       Impact factor: 12.944

4.  Succinylome analysis reveals the involvement of lysine succinylation in metabolism in pathogenic Mycobacterium tuberculosis.

Authors:  Mingkun Yang; Yan Wang; Ying Chen; Zhongyi Cheng; Jing Gu; Jiaoyu Deng; Lijun Bi; Chuangbin Chen; Ran Mo; Xude Wang; Feng Ge
Journal:  Mol Cell Proteomics       Date:  2015-01-20       Impact factor: 5.911

5.  ProIn-Fuse: improved and robust prediction of proinflammatory peptides by fusing of multiple feature representations.

Authors:  Mst Shamima Khatun; Md Mehedi Hasan; Watshara Shoombuatong; Hiroyuki Kurata
Journal:  J Comput Aided Mol Des       Date:  2020-09-23       Impact factor: 3.686

6.  Computational Identification of Protein Pupylation Sites by Using Profile-Based Composition of k-Spaced Amino Acid Pairs.

Authors:  Md Mehedi Hasan; Yuan Zhou; Xiaotian Lu; Jinyan Li; Jiangning Song; Ziding Zhang
Journal:  PLoS One       Date:  2015-06-16       Impact factor: 3.240

7.  Meta-4mCpred: A Sequence-Based Meta-Predictor for Accurate DNA 4mC Site Prediction Using Effective Feature Representation.

Authors:  Balachandran Manavalan; Shaherin Basith; Tae Hwan Shin; Leyi Wei; Gwang Lee
Journal:  Mol Ther Nucleic Acids       Date:  2019-04-30

8.  Proteome-Wide Identification of Lysine Succinylation in the Proteins of Tomato (Solanum lycopersicum).

Authors:  Weibo Jin; Fangli Wu
Journal:  PLoS One       Date:  2016-02-01       Impact factor: 3.240

9.  A systematic identification of species-specific protein succinylation sites using joint element features information.

Authors:  Md Mehedi Hasan; Mst Shamima Khatun; Md Nurul Haque Mollah; Cao Yong; Dianjing Guo
Journal:  Int J Nanomedicine       Date:  2017-08-28
View more
  2 in total

1.  Prediction of serine phosphorylation sites mapping on Schizosaccharomyces Pombe by fusing three encoding schemes with the random forest classifier.

Authors:  Samme Amena Tasmia; Md Kaderi Kibria; Khanis Farhana Tuly; Md Ariful Islam; Mst Shamima Khatun; Md Mehedi Hasan; Md Nurul Haque Mollah
Journal:  Sci Rep       Date:  2022-02-16       Impact factor: 4.379

2.  Deep_KsuccSite: A novel deep learning method for the identification of lysine succinylation sites.

Authors:  Xin Liu; Lin-Lin Xu; Ya-Ping Lu; Ting Yang; Xin-Yu Gu; Liang Wang; Yong Liu
Journal:  Front Genet       Date:  2022-09-29       Impact factor: 4.772

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.