Literature DB >> 29255285

iDTI-ESBoost: Identification of Drug Target Interaction Using Evolutionary and Structural Features with Boosting.

Farshid Rayhan1, Sajid Ahmed1, Swakkhar Shatabda2, Dewan Md Farid1, Zaynab Mousavian3, Abdollah Dehzangi4, M Sohel Rahman5.   

Abstract

Prediction of new drug-target interactions is critically important as it can lead the researchers to find new uses for old drugs and to disclose their therapeutic profiles or side effects. However, experimental prediction of drug-target interactions is expensive and time-consuming. As a result, computational methods for predictioning new drug-target interactions have gained a tremendous interest in recent times. Here we present iDTI-ESBoost, a prediction model for identification of drug-target interactions using evolutionary and structural features. Our proposed method uses a novel data balancing and boosting technique to predict drug-target interaction. On four benchmark datasets taken from a gold standard data, iDTI-ESBoost outperforms the state-of-the-art methods in terms of area under receiver operating characteristic (auROC) curve. iDTI-ESBoost also outperforms the latest and the best-performing method found in the literature in terms of area under precision recall (auPR) curve. This is significant as auPR curves are argued as suitable metric for comparison for imbalanced datasets similar to the one studied here. Our reported results show the effectiveness of the classifier, balancing methods and the novel features incorporated in iDTI-ESBoost. iDTI-ESBoost is a novel prediction method that has for the first time exploited the structural features along with the evolutionary features to predict drug-protein interactions. We believe the excellent performance of iDTI-ESBoost both in terms of auROC and auPR would motivate the researchers and practitioners to use it to predict drug-target interactions. To facilitate that, iDTI-ESBoost is implemented and made publicly available at: http://farshidrayhan.pythonanywhere.com/iDTI-ESBoost/ .

Entities:  

Mesh:

Substances:

Year:  2017        PMID: 29255285      PMCID: PMC5735173          DOI: 10.1038/s41598-017-18025-2

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Targeted drug design is one of the key techniques in therapeutic drug discovery[1]. Prediction of new drug target interactions can help researchers to find new uses for old drugs and to discover their therapeutic profiles or side effects[2-4]. Since experimental prediction of drug-target interaction is expensive and time-consuming, computational methods have been gaining increasing popularity in recent years[5,6]. During the past two decades, a wide range of computational approaches such as ligand-based methods[7,8], target or receptor based methods[9,10], gene ontology based methods[11], literature text mining methods[12,13] have been proposed to address the drug-target interaction. The performance and effectiveness of the ligand-based methods degrade due to the decrease in the the number of known ligands of a target protein. Receptor based methods often use docking simulation[14] and heavily rely on the available three dimensional native structure of the protein targets. It is important to note that finding three-dimensional structures of the proteins is by it self is a costly and time-consuming task which is done using experimental methods such as NMR and X-ray Crystallography. Moreover, three dimensional structures are very difficult to predict for ion channel proteins and G-protein coupled receptors (GPCRs). In addition, the tremendous growth in the Biomedical literature has increased the redundancy problem of the compound names or the gene names and has been the main obstacle for literature based systematic text mining methods. Recently, chemo-genomic methods[15] have been attempted to identify drug-target interactions. This type of methods are mainly based on machine learning[16,17], graph theory[18,19] and network methods[20,21]. In the literature of the supervised learning setting, several classification algorithms have been found to be applied for this task. Examples include support vector machine[22,23], deep learning[24], fuzzy logic[25], and nearest neighbor[26]. Yamanishi et al.[16] first proposed a mechanism to formalize the inference of the drug–target identification as a supervised learning problem. In that pioneering work, they also proposed a gold standard dataset that had been later used extensively in the literature[22,24,27]. In a subsequent work, the same authors[27] explored the association among pharmacological space and chemical space with the network topology of drug-target interactions and applied distance-based learning. Wang et al. proposed RLS-KF[28] that uses regularized least squares method integrated with nonlinear kernel fusion. Drug-based similarity inference (DBSI) was proposed in[21] utilizing two dimensional chemical structural similarity. Another method, KBMF2K, was proposed in[29] that used chemical and genomic kernels and bayesian matrix factorization. Later on other noteworthy methods such as NetCBP[30], DASPfind[31], SELF-BLM[23] have been proposed to solve thi sproblem. Recently[22] used bigram based features extracted from Position Specific Scoring Matrix (PSSM) as molecular fingerprint to tackle this problem. Since the three dimensional native structure of most of the protein targets are not available, most of the supervised learning methods in the literature do not exploit the structure based features. Huang et al.[32] used extremely randomized trees model and represented the proteins as pseudo substitution matrix generated from its amino acid sequence information and the drugs as moelcular fingerprint. In another recent work, Wang et al.[33] explored PSSM based features and drug fingerprints and used rotation forest based predictor. Among other recent works are similarity based method used by Yuan et al.[34], self organizing theory used by Duran et al.[35,36] and ensemble method used by Ezzat et al.[37,38]. Recently, a comprehensive literature review on the computational methods in drug-target interaction prediction was conducted by Chen et al.[39]. In this paper, we present iDTI-ESBoost, a method for identification of Drug Target Interaction Using Evolutionary and Structural Information with Boosting. We exploit the structural features along with the evolutionary features to predict drug-protein interactions. Our work was inspired due to the modern successful secondary structural prediction tools like SPIDER2[40,41] and its use to generate features in supervised learning and classification[42,43]. Our proposed method uses a novel set of features extracted using structural information along with the evolutionary features and molecular fingerprints of drugs. To handle the large amount of imbalance in the data, we propose a novel balancing method and use it along with a boosting algorithm. As a result, iDTI-ESBoost has shown to be superior due to its prediction results on a widely used gold standard data set compared to the other existing methods found in the literature. Our method is publicly available to use at: http://farshidrayhan.pythonanywhere.com/iDTI-ESBoost/. The rest of the paper is organised as follows which is suggested in[44]: description of dataset, formulation of statistical samples, selection and development of a powerful classification algorithm, demonstration of the performance of the predictor using cross-validation, implementation of web server followed by a conclusion.

Results and Discussion

In this section, we present the results of our experiments. All the methods were implemented in Python language using Python3.4 version and Scikit-learn library[45] of Python was used for the implementation of the machine learning algorithms. All experiments were conducted on a Computing Machine hosted by CITS, United International University. Also, each of the experiments was carried out 5 times and the average of the results is reported. We perform several types of experiments. In particular, we conduct four different sets of experiments as follows. First we investigate effectiveness and applicability of the different feature groups as mentioned in Table 1. Note that, in Table 1, four different feature groups, namely, A, B, C and D, were formed. Secondly, we investigate the effectiveness of the classifiers used in our research. Subsequently, we investigate the effectiveness of the balancing methods applied on our highly imbalanced datasets. Finally, we compare iDTI-ESBoost against the state-of-the-art methods found in the literature.
Table 1

Summary of evolutionary and structural features used for protein targets and fingerprint features for drugs.

Feature GroupNumber of FeaturesFeature TypeGroup
Molecular finger print881drug
PSSM bigram400targetA
Secondary Structure Composition3targetB
Accessible Surface Area Composition1
Torsional Angles Composition8
Torsional Angles Auto-Covariance80targetC
Structural Probabilities Auto-Covariance30
Torsional Angles bigram64targetD
Structural Probabilities bigram9
Total1476

The “Group” column shows different feature groups used in our experiments and will be discussed in a later section.

Summary of evolutionary and structural features used for protein targets and fingerprint features for drugs. The “Group” column shows different feature groups used in our experiments and will be discussed in a later section.

Effectiveness of Feature Groups

We created four different feature groups to see the effects of the different sets of features on the classifier performance. The feature groups have already been reported in Table 1. Group A contains 1281 features and was previously used in[22]. We further added other groups, namely, B, C and D, incrementally in that order with the base feature group i.e., Group A and achieve features of size 1293, 1403 and 1476, respectively. We have performed two sets of experiments to test the effectiveness of the feature groups. In both of these experiments we changed the feature groups and used different classifiers and applied different balancing methods on the data to analyze the effect. Results of these experiments are reported respectively in Tables 2 and 3.
Table 2

Performance of AdaBoost, Random Forest and Support Vector Machine classifiers on the gold standard datasets in terms of area under Receiver Operating Characteristic (ROC) curve (auROC) and area under precision recall curve (auPR) using different feature group combinations and random under sampling.

DatasetFeature CombinationClassifierauPRauROC
enzymesAAdaBoost0.540.9530
Random Forest0.430.9457
SVM 0.64 0.9647
A, BAdaBoost 0.51 0.9431
Random Forest0.490.9445
SVM0.48 0.9502
A, B, CAdaBoost 0.66 0.9638
Random Forest0.480.9334
SVM0.410.9360
A, B, C, DAdaBoost 0.65 0.9689
Random Forest0.500.9493
SVM0.630.9628
ion channelsAAdaBoost 0.36 0.9271
Random Forest0.330.9232
SVM0.25 0.9467
A, BAdaBoost 0.33 0.9191
Random Forest0.300.8898
SVM0.23 0.9213
A, B, CAdaBoost 0.34 0.9202
Random Forest0.310.8734
SVM0.23 0.9213
A, B, C, DAdaBoost 0.43 0.9369
Random Forest0.400.9234
SVM0.140.6723
GPCRsAAdaBoost 0.29 0.8856
Random Forest0.230.8743
SVM0.180.7832
A, BAdaBoost 0.29 0.8834
Random Forest0.220.8698
SVM0.150.7802
A, B, CAdaBoost 0.35 0.9116
Random Forest0.310.9034
SVM0.150.7945
A, B, C, DAdaBoost 0.31 0.9128
Random Forest0.300.9168
SVM0.210.7896
nuclear receptorsAAdaBoost 0.41 0.8145
Random Forest0.230.7519
SVM0.190.7898
A, BAdaBoost 0.43 0.7969
Random Forest0.290.7723
SVM0.200.6789
A, B, CAdaBoost 0.36 0.7590
Random Forest0.210.7234
SVM0.210.6971
A, B, C, DAdaBoost 0.33 0.7946
Random Forest0.290.7145
SVM0.200.7287
Table 3

Performance of Adaboost classifier on different datasets in terms of area under Receiver Operating Characteristic (ROC) curve (auROC) and area under precision recall curve (auPR) using different feature group combinations and balancing methods.

DatasetFeature CombinationBalancing MethodauPRauROC
enzymesArandom0.540.9530
clustered0.580.9493
A, Brandom0.510.9431
clustered0.590.9353
A, B, Crandom0.660.9638
clustered0.630.9577
A, B, C, Drandom0.65 0.9689
clustered 0.68 0.9598
ion channelsArandom0.360.9271
clustered0.380.8982
A, Brandom0.330.9191
clustered0.410.8902
A, B, Crandom0.340.9202
clustered0.450.9021
A, B, C, Drandom0.43 0.9369
clustered 0.48 0.9051
GPCRsArandom0.290.8856
clustered0.480.9189
A, Brandom0.290.8834
clustered0.490.8968
A, B, Crandom0.350.9116
clustered 0.50 0.8890
A, B, C, Drandom0.310.9128
clustered0.48 0.9322
nuclear receptorsArandom0.410.8145
clustered0.790.9270
A, Brandom0.430.7969
clustered0.320.8715
A, B, Crandom0.360.7590
clustered0.570.8935
A, B, C, Drandom0.330.7946
clustered 0.79 0.9285
Performance of AdaBoost, Random Forest and Support Vector Machine classifiers on the gold standard datasets in terms of area under Receiver Operating Characteristic (ROC) curve (auROC) and area under precision recall curve (auPR) using different feature group combinations and random under sampling. Performance of Adaboost classifier on different datasets in terms of area under Receiver Operating Characteristic (ROC) curve (auROC) and area under precision recall curve (auPR) using different feature group combinations and balancing methods. Table 2 reports the performance of three different classifiers on the four datasets during our experiments. In this step, we have produced the results for different combination of feature groups by adding them in a forward selection scheme by sorting them based on their individual performance for Nuclear Receptors. For the Nuclear Receptors, the best results achieved using Feature group A, and followed by B, C, and D, consequently. We used the individual performance for Nuclear Receptors as it produced the most distinguished performance for different input feature group. Therefore, we first evaluate the performance using feature group A, then added B, C, and finally D, sequentially. We have produced the results for each feature group as well as different combination of these feature gorups based on sequential forward selection. These results are available and provided as a supplementary material (Supplementary File 1). Note that, though this experiment was intended for classifier selection, we clearly see that the best results in terms of auPR and auROC were found only when the structural features were added. For enzymes dataset, the best result in terms of auPR was 0.66 found with the combination A, B, C which is using structural composition and structural auto-covariance groups with PSSM-bigram and molecular fingerprint based features. This result is slightly better than the case when we use all the features A, B, C, D and get auPR of 0.66. In terms of auROC, the results are somewhat comparable to each other; however, the best result is achieved when all the four feature groups are used together. Thus enzyme dataset shows the effectiveness of structural information based features. Using iDTI-ESBoost for ion channels and GPCRs datasets we attain similar performance in terms of auPR. However it achieved highest auPR value when only the composition features (Group B) were added with the base features for Nuclear receptors dataset. The increase in the value of auROC clearly reveals the effectiveness of the structural features (Groups B, C, D) when added to the base feature (Group A). As it is shown in Table 2, due to differences in the explored problems in this study, the results for different combination of feature groups, and different classifiers are not fully consistent. However, The aim of this experiment is to compare and analyse the performance of our method with respect to different combination of feature groups and different problems and find the best combination for that given specific problem. By conducting this comprehensive comparison, we are able to find the best classifier among those that we investigated here with respect to the combination of feature groups that we studied for a given problem. The next set of experiments were run to show the performance of different balancing or under sampling methods in the training data using various feature groups. These results are shown in Table 3. These experiments were run using the AdaBoost classifier. The results in Table 3 clearly shows that for all the datasets, the best results in terms of auPR and auROC were found when structural features have been added. In case of the GPCRs, the auPR was found to be the highest at 0.5 when three feature groups, namely, Groups A, B, and C have been combined. Apart from this, in all other datasets, the combination of all four groups has shown superior performance both in terms of auPR and auROC. Our hypothesis that the added structural features play a significantly important role in the prediction of drug-target interaction is thus justified according to these experiments.

Effectiveness of the AdaBoost Classifier

To test and select the suitable classifier for our problem, we test three different classifiers namely, AdaBoost ensemble classifier[46] with decision tree as its weak learner, Random Forest[47] and Support Vector Machines (SVM)[48]. For these experiments, we used random under sampling as the balancing method. As features, four different combinations were used as has been mentioned already. The results in terms of auPR and auROC are presented in Table 2. Here for each of the datasets and feature groups combinations bold faced values in the table represents the highest values achieved for that combination. It is evident that except for one case in the enzymes dataset, AdaBoost classifier has shown superior performance in terms of auPR across all feature groups combinations. It is also worth-noting that for all datasets, the highest auPR value was achieved by AdaBoost. The precision-recall curves for these experiments across all feature groups combinations are illustrated in Fig. 1.
Figure 1

Precision-Recall curves of different classifier algorithms using random under sampling and all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr).

Precision-Recall curves of different classifier algorithms using random under sampling and all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr). In case of the ROC curves, achieved results are also in support of the selection of AdaBoost as a classifier. AdaBoost provides the highest auROC values for all the four datasets and it gives better auROC values for 11 out of 16 dataset-feature groups combinations. In other cases, SVM has achieved the highest auROC values, but only marginally so. The ROC curves for different classifiers across all feature groups combinations are illustrated in Fig. 2.
Figure 2

Receiver operating characteristic (ROC) curves of different classifier algorithms using random under sampling and all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr).

Receiver operating characteristic (ROC) curves of different classifier algorithms using random under sampling and all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr). Considering the values of auPR and auROC curves on different datasets as shown in Table 2 and illustrated through the curves in Figs 1 and 2, we select AdaBoost as the classifier for iDTI-ESBoost. Note that, because of the huge imbalance in the datasets, with positive samples being much lower than the negative ones, the auPR curve is more important compared to the auROC curve and AdaBoost clearly outperforms the other two classifiers in terms of auPR values.

Effectiveness of the Balancing Methods

The next set of experiments were run to test the effectiveness of the two different sampling methods on the datasets. The parameters used with AdaBoost classifier for random and cluster based under sampling are reported in Table 4.
Table 4

Parameters of AdaBoost Algorithm used with decision tree as weak classifier along with different balancing methods on four datasets.

Balancing methodDatasetMax depthMin sample splitMin samples LeafCriterion
randomenzymes100161Gini impurity
ion channels841Gini impurity
GPCRs631Gini impurity
nuclear receptors572Gini impurity
clusteredenzymes11011Gini impurity
ion channels921Gini impurity
GPCRs631Gini impurity
nuclear receptors15021Gini impurity
Parameters of AdaBoost Algorithm used with decision tree as weak classifier along with different balancing methods on four datasets. For each of the datasets, we used four feature group combinations and used random and cluster based under sampling and report auPR and auROC values from cross-validation experiments in Table 3. We also show the ROC curves and auPR curves for all four datasets using all the features in Figs 3 and 4.
Figure 3

Receiver operating characteristic (ROC) curves of AdaBoost classifier showing differences between random under sampling and cluster based sampling using all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr).

Figure 4

Precision vs Recall curves of AdaBoost classifier showing differences between random under sampling and cluster based sampling using all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr).

Receiver operating characteristic (ROC) curves of AdaBoost classifier showing differences between random under sampling and cluster based sampling using all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr). Precision vs Recall curves of AdaBoost classifier showing differences between random under sampling and cluster based sampling using all the feature combinations on four datasets: (a) enzymes (b) ion channels (ic) (c) GPCRs (d) nuclear receptors (nr). From the results reported in Table 3, it is worth-mentioning that in terms of auPR for all four datasets, cluster based sampling significantly outperforms random under sampling method. However, in terms of auROC curve, random sampling is slightly better than cluster based sampling in enzymes and ion channel datasets but the situation is in favor of cluster based sampling in GPCRs and nuclear receptors where it outperforms the random sampling method. We have also analyzed the effect of the imbalance ratio of the four datasets with the different balancing methods used in this paper. Note that the performance of the random sampling and clustered sampling are similar in terms of auPR in the dataset enzymes with higher imbalance ratio. The performance drustically falls for random sampling for the nuclear receptor dataset which have the lowest imbalance ratio. In case of the other two datasets ion channels and GPCRs, though their imbalance ratio is similar the difference in the auPR resutls differ significantly for these two balancing methods. Thus we can not conclude any correlation of imbalance ratio with that of the performance of the balacning methods. Rather the number of instances seems to affect the performance of the balancing methods. Enzymes and ion channels datasets with larger number of samples seems to favor random sampling and relatively smaller datasets GPCRs and nuclear receptors produces best resutls when using clustered sampling.

Comparison with Other Methods

Since the pioneering work of Yamanishi et al.[16], many supervised learning methods have been applied to predict drug-target interactions on these standard benchmark gold standard datasets. However, a few of these methods[24,28] do not use cross validation techniques and others[3,23] do not use the same standard datasets. Our method uses molecular fingerprints and evolutionary and structural features for this supervised classification problem. Similar methods, albeit without utilizing the structural features and balancing techniques are reported in[22,49]. Most of the papers in the literature have used auROC curve as the main evaluation metric. We have compared the performance of our method on these four datasets with that of DBSI[21], KBMF2K[29], NetCBP[30], Yamanishi et al.[16], Yamanishi et al.[27], Wang et al.[18] and Mousavian et al.[22] using auROC. The auROC values for all these methods along with iDTI-ESBoost are reported in Table 5.
Table 5

Performance of iDTI-ESBoost on the four benchmark gold datasets in terms on area under receiver operating characteristic curve (auROC) with comparison to other state-of-the-art methods.

DatasetDBSI[21] KBMF2K[29] NetCBP[30] Yamanishi et al.[16] Yamanishi et al.[27] Wang et al.[18] Mousavian et al.[22] iDTI-ESBoost
enzymes0.80750.83200.82510.9040.89200.88600.9480 0.9689
ion channels0.80290.79900.80340.85100.81200.89300.8890 0.9369
GPCRs0.80220.85700.82350.89900.82700.87300.8720 0.9322
nuclear receptors0.75780.82400.83940.84300.83500.82400.8690 0.9285
Performance of iDTI-ESBoost on the four benchmark gold datasets in terms on area under receiver operating characteristic curve (auROC) with comparison to other state-of-the-art methods. From the values shown in bold in Table 5, we see that for all the datasets iDTI-ESBoost is able to significantly outperform all other previous state-of-the-art methods in terms of auROC. All the auROC values are greater than 90% which indicates the effectiveness of the classifier, balancing methods and the novel features proposed in this paper. Moreover, in[22] the authors argued in favor of auPR curve as a measure of evaluating the performance of classifiers for skewed datasets, especially in drug-target interaction where negative samples outnumber the positive samples. This argument does have merit as, logically, a mis-classification of positive samples or false negative should be more penalized in the score. To compare the performance in terms of auPR among iDTI-ESBoost with that in[22], we reported the auPR values of the two predictors in Table 6. The results clearly shows that our method iDTI-ESBoost outperforms the predictor in[22] in terms of auPR as well.
Table 6

Comparison of the performance of iDTI-ESBoost on the four benchmark gold datasets in terms on area under the precision-recall curve (auPR) with the state-of-the-art method in Mousavian et al.[22].

PredictorEnzymesIon channelsGPCRsNuclear receptors
Mousavian et al.[22] 0.5460.3900.2820.411
iDTI-ESBoost 0.680 0.480 0.500 0.790
Comparison of the performance of iDTI-ESBoost on the four benchmark gold datasets in terms on area under the precision-recall curve (auPR) with the state-of-the-art method in Mousavian et al.[22]. In Table 7, we report specificity, sensitivity, precision, MCC and F1-Score for four datasets using different feature group combinations as achieved by iDTI-ESBoost in experiments. Specificity and sensitivity are very high as reported in this table.
Table 7

Specificity, Sensitivity, Precision, MCC and F1 score for four datasets as achieved by iDTI-ESBoost using different feature groups.

DatasetFeature GroupSpecificitySensitivityPrecisionMCCF1 score
enzymesA0.830.90.050.19620.10
A, B0.820.890.050.18120.09
A, B, C0.830.870.050.17620.09
A, B, C, D0.850.850.150.18890.10
Ion channelsA0.780.810.130.26150.22
A, B0.780.840.140.2560.24
A, B, C0.80.860.120.29800.20
A, B, C, D0.780.840.130.29130.20
GPCRsA0.780.840.120.2540.20
A, B0.80.850.110.27600.20
A, B, C0.790.890.110.27970.19
A, B, C, D0.80.840.110.26470.19
Nuclear receptorsA0.850.910.160.21410.27
A, B0.770.880.110.21540.19
A, B, C0.810.880.120.17980.20
A, B, C, D0.920.870.140.22530.24
Specificity, Sensitivity, Precision, MCC and F1 score for four datasets as achieved by iDTI-ESBoost using different feature groups.

Predicting New Interactions

In addition to these, we have analyzed the results produced by the classification algorithm. From the false negatives predicted by iDTI-ESBoost, we noticed that there are a number of false negatives for which the prediction probability is very high for it to be considered as a negative sample. Similar approaches were adopted in[16,27]. In this paper, we suggest that the false negative interactions which are labeled as positive by our method with a very high prediction probability could be potential candidates for finding new positive interactions. A list of such interactions for four group of targets are given in Table 8. Ten interactions are reported for each of the datasets with highest prediction probability.
Table 8

New prediction made by iDTI-ESBoost for four gold standard datasets used in this paper.

DatasetProtein IdDrug IdDrug NameScore
Enzymeshsa:10825D00041Threonine (USP)0.7207
hsa:4759D00041Threonine (USP)0.7163
hsa:129807D00041Threonine (USP)0.7163
hsa:4953D00041Threonine (USP)0.7095
hsa:1845D00041Threonine (USP)0.7078
hsa:9610D00041Threonine (USP)0.7073
hsa:6652D00041Threonine (USP)0.7034
hsa:1734D00136Haloperidol (JP17/USP/INN)0.6995
hsa:1178D03643Dalvastatin (USAN/INN)0.6985
hsa:8435D03643Dalvastatin (USAN/INN)0.6962
Ion channelshsa:285242D00294Diazoxide (JAN/USP/INN)0.9407
hsa:779D00294Diazoxide (JAN/USP/INN)0.9366
hsa:2561D00294Diazoxide (JAN/USP/INN)0.9357
hsa:785D00294Diazoxide (JAN/USP/INN)0.9353
hsa:11254D00294Diazoxide (JAN/USP/INN)0.935
hsa:3775D00225Alprazolam (JP17/USP/INN)0.9339
hsa:6263D00294Diazoxide (JAN/USP/INN)0.932
hsa:6324D02261Quinine hydrochloride hydrate (JP17)0.9305
hsa:6324D02262Quinine sulfate (USP)0.9305
hsa:6332D02262Quinine sulfate (USP)0.8464
GPCRshsas:9052D04625Isoetharine (USP)0.9311
hsa:9052D00632Dobutamine hydrochloride (JP17/USP)0.9311
hsa:9052D03880Dobutamine lactobionate (USAN)0.9311
hsa:9052D03881Dobutamine tartrate (USP)0.9311
hsa:1909D03621Cyclizine (INN)0.931
hsa:57105D01712Theophylline sodium acetate (JAN)0.9307
hsa:155D02671Mesoridazine (USAN/INN)0.9306
hsa:148D02614Denopamine (JAN/INN)0.9303
hsa:155D00480Promethazine hydrochloride (JP17/USP)0.9302
hsa:1909D00480Promethazine hydrochloride (JP17/USP)0.9302
Nuclear receptorshsa:2099D01132Tazarotene (JAN/USAN/INN)0.9792
hsa:2101D00956Nandrolone phenpropionate (USP)0.9755
hsa:2101D00443Spironolactone (JP17/USP/INN)0.9758
hsa:2099D00316Etretinate (JAN/USAN/INN)0.9602
hsa:9971D00316Etretinate (JAN/USAN/INN)0.9593
hsa:2101D00327Fluoxymesterone (JP17/USP/INN)0.9591
hsa:2101D00088Hydrocortisone (JP17/USP/INN)0.9571
hsa:2101D00075Testosterone (JAN/USP)0.9558
hsa:2099D00565Fenofibrate (JAN/INN)0.9557
hsa:2101D00462Oxandrolone (JAN/USP/INN)0.9557
New prediction made by iDTI-ESBoost for four gold standard datasets used in this paper.

Web Server Implementation

We have also implemented our method as shown in Fig. 5 as a separate web server. The web server is freely available to use at: http://farshidrayhan.pythonanywhere.com/iDTI-ESBoost/. The mechanism of the web-server is very simple. We also provide the pre-learned models for each of the datasets. The interface of the web server is easy to use. It requires an user first to select the target group and provide the PSSM and SPD files for the target protein. These files can be easily generated by PSI-BLAST and SPIDER2 software using their online available tools.
Figure 5

Schematic diagram of the training module of iDTI-ESBoost showing the steps of the training phase.

Schematic diagram of the training module of iDTI-ESBoost showing the steps of the training phase. To specify drug, one can select from a drop down list. The drugs are pre-fetched in our system from KEGG website. After selecting the drug and specifying target files, one can click the prediction button to find the prediction for that drug-target pair. The web-server also have a simple page with easy to-use instructions. We have made all the code and programs necessary for implementation of this paper freely available at: https://github.com/farshidrayhanuiu/iDTI-ESBoost to facilitate the use of the other users.

Materials and Methods

In this section, we provide the detail information of the benchmark datasets, feature extraction and balancing methods, classifiers and evaluation metrics used in this study. Figure 5 depicts the training module of our proposed method, iDTI-ESBoost. The training dataset of iDTI-ESBoost contains both interacting (positive) and non-interacting drug-target pairs. For each instance of drug-target pair, a drug is searched in the DrugBank database[50] to fetch the drug chemical structure in SMILES format. Similarly, a target protein sequence is first fetched from KEGG database[51] and then fed to SPIDER2[40,41] and PSI-BLAST[52] in order to receive, respectively, structural information as an SPD file and position specific scoring matrix (PSSM) based profile containing evolutionary information. A feature extraction module then uses these files to generate three types of features: drug molecular fingerprints, PSSM bigram and structural features based on the output of the secondary structure prediction software namely SPIDER2. Features generated in this phase is then fed to an AdaBoost classifier that learns the model for prediction purposes. The prediction module is very similar to that of the training module shown in Fig. 5. For prediction, a query drug-target pair is fed to the system in a similar way to extract three types of features and then the trained and stored model is used to predict whether the given drug-target pair is interacting or non-interacting.

Drug-target Interaction Datasets

In this paper, we have used the gold standard datasets introduced by Yamanishi et al. in[16]. These datasets are publicly available at: http://web.kuicr.kyoto-u.ac.jp/supp/yoshi/drugtarget/. Yamanishi et al. used DrugBank[53], KEGG BRITE[54], BRENDA[55] and SuperTarget[56] to extract information about drug-target interactions. They used the known drugs to four types of protein targets, namely, enzymes, ion channels, g-protein coupled receptors (GPCRs) and nuclear receptors. The number of target proteins in these groups are 664, 204, 95 and 26, respectively that interact with 445, 210, 223 and 54 drugs through 2926, 1476, 635 and 90 known interactions, respectively. A brief summary of these datasets are given in Table 9. These benchmark datasets have been used in many studies in the literature[21,22,24,27] and are referred to as the ‘gold’ standard.
Table 9

Description of the gold standard datasets[16].

DatasetDrugsProteinsPositive InteractionsImbalance Ratio
Enzyme445664292699.98
Ion Chanel210204147628.02
GPCR2239563532.36
Nuclear Receptor54269014.6
Description of the gold standard datasets[16].

Graph Construction from the Dataset

Based on the interactions of four types of proteins with known drugs, we build positive and negative samples for each dataset using a method similar to the one used in[22] as follows. The drug-target interaction network for each dataset is a bipartite graph, G = (V, E), where the set of vertices is such that D is the set of drugs and T is the set of targets, and the set of edges is E. Here, any edge e = (d, t)   E denotes an interaction only between a drug, d    D with a protein target, t    T. Now, for a particular graph from a dataset, all the known interactions in the graph represented by its edges are considered to be positive samples and the non-existent edges are taken as negative samples. Note that, here, non-existent edges refer to the possible valid edges only that are not there; i.e., they do not include edges among the vertices of the same partite set. Formally, a dataset is an union of positive and negative sets as follows: Here, , and . For example, in the nuclear receptor, there are 54 drugs and 26 proteins with possible 54 × 26 = 1404 interactions. Since 90 interactions are known, these are treated to be positive and the rest 1314 as negative. The same procedure was followed for each of the datasets. As expected, the constructed datasets using this technique are imbalanced as the number of negative samples far outnumbers that of positive samples. This issue is attended to later by applying some balancing techniques. The majority class is the class denoting negative interaction. We define imbalance ratio as the number of instances in the majority class to the number of instances in the minority class. The imbalance ratio of each of the datasets used in this paper is reported in Table 9. Note that the enzymes dataset is with highest imbalance ratio near 100 and nuclear receptor dataset has got the lowest imbalance ratio.

Feature Extraction

A dataset constructed in this way has drug-target pairs as instances. In the feature extraction phase, a drug identifier is looked up in the KEGG databased[54] and the corresponding SMILES format is downloaded from the DrugBank database[50]. The features based on drugs are generated using this SMILES data. Similarly, a protein target of each pair is first searched within the KEGG database[54] to fetch the protein sequence. This protein sequence is then fed to two different software: Position Specific Iterated BLAST (PSI-BLAST)[52] to fetch evolutionary profile based Position Specific Scoring Matrix (PSSM) and a secondary structure prediction tool called SPIDER2[40,41] to generate SPD files that contains the structural information. Three groups of features are extracted using these three files. The details are described in the rest of this section.

SMILES Based Features

Several descriptors are used to represent the features or properties of drug compounds[57]. To this end, one of the most popular features is molecular fingerprints which is widely used for similarity searching[58], clustering[59], and classification[22]. Each drug compound is represented by 881 chemical substructures defined in PubChem database[60]. The presence (absence) of a particular substructure is encoded as 1 (0). Thus the length of this molecular fingerprint based feature is 881. We used the rcdk package of R[61] to extract these molecular fingerprints based features.

PSSM Based Features

We used the PSSM matrix returned by the PSI-BLAST software to generate evolutionary features from the protein target sequences. Each PSSM file contains a PSSM matrix that is constructed after multiple sequence alignment using the non redundant (NR) database. The PSSM file contains a matrix M of dimension L × 20, where L is the length of the protein and each of the entries in this matrix, m , represents the probability of substitution of the j-th amino acid in the i-th location of the given protein sequence. We first convert this matrix M to a normalized matrix using a normalization technique similar to that proposed in[62,63]. The dimension of this matrix is same as the original matrix M. After that we generate PSSM-bigram features using the following equation:Bigram features for PSSM were first proposed in[64] and subsequently used successfully in drug-target interaction prediction in[22]. Total number of features generated using this method is 400.

Structure Based Features

The traditional drug discovery is a lock-key problem, where the lock is the target. The structure of the target thus plays a very important role in traditional drug discovery and is at the center of the docking based software. We make a hypothesis that even if the full structure is not present for the targets, estimated structural properties still can play an important role in drug-target interaction prediction. Structural features are generated using the structural information generated and stored in SPD files by SPIDER2 software. The information generated by SPIDER2 are: accessible surface area (ASA), secondary structural (SS) motifs, torsional angles (TA) and structural probabilities (SP)[65-67]. Following features are generated using these information: Secondary Structure Composition: This feature is the normalized count or frequency of the structural motifs present at the amino-acid residue positions. There are three types of motifs: α-helix (H), β-sheet (E) and random coil (C). SPIDER2 returns a vector SS of dimension L × 1 containing this information. Thus we can define this feature as following:Here, L is the length of the protein andHere, SS is the structural motif at position j of the protein sequence and f is one of the 3 different motif symbols. Accessible Surface Area Composition: The accessible surface area composition is the normalized sum of accessible surface area defined by:Here ASA is the vector of accessible surface area of dimension L × 1 containing the values of accessible surface area for all the amino acid residues. Torsional Angles Composition: Four different types of torsional angles: φ, ψ, τ and θ are returned by SPIDER2 for each residue. First, we convert each of them into radians from degree angles and then take sign and cosine of the angles at each residue position. Thus we get a matrix of dimension L × 8. We denote this matrix by T. Torsional angles composition is defined as: Torsional Angles Bigram: The Bigram for the torsional angles is similar to that of the PSSM matrix and is defined as: Structural Probabilities Bigram: Structural probabilities for each position of the amino-acid residue are given in the SPD2 file as a matrix of dimension L × 3, which we denote by P. The Bigram of the structural probabilities is similar to that of PSSM matrix and is defined as: Torsional Angles Auto-Covariance: This feature is also derived from the torsional angles and is defined as:This feature group depends on parameter DF which is the distance factor. In this study, we used DF = 10 and this value was selected for the parameter DF as it was shown as the most effective window size to extract features based on torsion angles and similar properties[40,65,67,68]. Structural Probablities Auto-Covariance: This feature is also derived from the structural probabilities and is defined as: A brief summary of the three group of features derived from each drug-target pair is given in Table 1. Note that there are two types of features. Drug related features and target related features in four groups A, B, C and D.

Balancing Methods

As it was specified earlier, our employed datasets are all imbalanced. Several sampling techniques in the literature have been employed to balance these data such as: random under sampling[22], synthetic over sampling[69], balanced random sampling (BRS)[68,70], neighborhood cleaning rule[71], and cluster based under sampling[72,73]. In this paper, we explore random under sampling (RUS) method as done previously for drug-target interaction prediction in[22]. We also propose a novel modified cluster based under sampling method based on[73] as follows. In this method, the dataset is first divided into two subsets as major class and minor class. In the major class k-means clustering is applied to divide the major class samples in k clusters while the minor class samples are kept unchanged. After that from the k clusters of major class samples, subsamples are chosen randomly to represent the entire major class. We denote this method as cluster based under sampling (CUS) throughout this paper. The random under sampling will be denoted as random under sampling (RUS). The pseudo-code for the CUS algorithm is given in Algorithm 1. Our CUS algorithm depends on two parameters, namely, k and h. In our experiments, we have varied k for values from and found the the best performing value to be 23. However, more sophisticated clustering algorithms can be applied on this data. The role of the parameter h is to control the random under sampling of the clustered majority class samples. The details of the experimental results for selecting this hyper parameter h is given as supplementary material (Supplementary File 2). Cluster Based Under Sampling (dataset, k, h).

Description of the classifier

We have selected the adaptive boosting algorithm (AdaBoost)[46] as our classification algorithm. Adaptive boosting is a meta or ensemble classifier that uses several weak learning algorithms or weak classifiers and improves over their performance. We choose decision tree classifiers as the weak classifiers. AdaBoost is a meta-classifier of the following form: AdaBoost iteratively adds up a weak classifier h (x) at each iteration of the algorithm weighted by α where α is the weight achieved from the error function ε for the weak classifier h (x) at iteration t. Each of these weak classifiers is chosen in a way so as to minimize the error on the training sample weighted by the distribution D : The algorithm of AdaBoost[46] is sketched in Algorithm 2 following the notations of [74]. AdaBoost (dataset = (X, Y)).

Performance Evaluation

A large variety of performance metrics are used in the literature to compare the performance of supervised learning methods[75]. The gold datasets that are used in the literature of drug-target interaction prediction are largely imbalanced and the number of negative samples largely outnumbers that of the positive samples. Therefore, the typical measures like accuracy does not make much sense. Moreover, the output of the classifier generating probabilistic outputs depends on the thresholds or the values predicted by it for each of the predicting classes. In such cases, thresholds or values play an important role on the sensitivity and specificity of the classifiers. Two measures that are independent of the values or thresholds set for decision making are area under curve for Receiver Operating Characteristic (auROC) and area under precision recall curve (auPR). These two measures are widely used in the literature of drug-target interaction prediction[22,24,30,76] and thus have become standard metrics for comparison. Lets assume, P is the total number of positive samples in a dataset and N is the total number of negative samples in a dataset. Let TP denote the number of true positives, TN denote the number of true negatives, FN denote the number of false negatives and FP denote the number of false positives predicted by a classifier. True positives (negatives) are correctly classified positive (negative) samples by the classifier. Conversely, false positives (negatives) are negative (positive) samples incorrectly predicted as positives (negatives) by the classifier. Following these notions, we can define sensitivity or true positive rate as follows: Therefore, sensitivity is the ratio of correctly predicted positive samples to the total number of positive samples. Precision is defined as the positive predictive rate (PPV) as follows: Therefore, precision shows the percentage of positive predictions by the classifiers that are accurate. Another important measure is specificity (SPC) or true negative rate defined as follows: Fall-out or false positive rate (FPR) is the ration of the number of wrongly classified negative instances to the total number of negative instances defined as follows:F1 Score is a statistical score based on precision and sensitivity and defined as follows: All theses performance measures have values within the range [], 0 being the worst and 1 being the best. Another score that is often used in comparison is called defined as follows:Value of this coefficient ranges from −1 to +1, where +1 means a perfect predictor and −1 means a total disagreement. Receiver operating characteristic (ROC) curve plots true positive rate or sensitivity against false positive rate or (1-specificity) at various threshold values. The performance of a predictor is calculated by the area under the ROC curve (auROC). A perfect classifier have a auROC value of 1 and a random classifier have a value of 0.5. However, for imbalanced datasets like ours, area under precision recall curve (auPR) is of more significance[22] as follows: auPR curve plots the precision rate vs the recall rate at different threshold values. This score penalizes the false positives more as compared to auROC and thus more suitable for skewed datasets. The value of auPR ranges from 0 to 1 and the higher the value is the better. It is very important to test the methods to check and balance the bias-variance trade-off[77]. Various methods of sampling are used to measure the performance of supervised learning algorithms[78]. Among them mostly used are k–fold cross validation and jack knife tests. Because of the high imbalance, dimensionality and cardinality of the datasets, in most of the methods in the literature, 5-fold cross validation have been preferred and used as the sampling method[22,24,30,76]. We also use the 5-fold cross validation to test our method for the sake of fair comparison with the other state-of-the-art methods. In the 5-fold cross validations, first the dataset is randomly split into five equal parts retaining the ratio of imbalance in each split same to the original dataset. Each time one part of the dataset is used as test and the other four are used as training data. First the balancing techniques are applied to the training data (clustered or random) and then the classifier is used to train the data into a model. The stored model is used subsequently to predict the labels for the test data. Thus all the drug-target pairs in the datasets are used in testing the classifier performance using cross-validation. The measures reported are the average of all 5-fold results.

Data and Material availability

All the data and materials used in this paper are available at: http://farshidrayhan.pythonanywhere.com/iDTI-ESBoost/.

Conclusion

In this paper, we have presented iDTI-ESBoost, a novel method to predict and identify drug-target interactions. iDTI-ESBoost is unique in its exploitation of structural features along with the evolutionary features to predict drug-protein interactions. It also uses a novel balancing technique and a boosting technique. We have conducted extensive experiments to test and analyze the performance of iDTI-ESBoost. On four benchmark datasets known as the gold standard data in the literature, iDTI-ESBoost outperforms the state-of-the-art methods in terms of area under Receiver Operating Characteristic (auROC) curve. Notably, the gold standard datasets used in the literature as benchmarks to analyze the performance of the methods for drug-target interactions prediction and identification are highly imbalanced with negative samples far outnumbering the positive samples. In the literature it has been argued that area under Precision Recall (auPR) curve is the most appropriate metric for comparison for such imbalanced datasets. To this end, iDTI-ESBoost also outperforms the latest and the best-performing method in the literature to-date in terms of area under precision recall (auPR) curve. We believe that the excellent performance of iDTI-ESBoost both in terms of auROC and auPR would motivate the researchers and practitioners to use it to predict drug-target interactions. To facilitate that, iDTI-ESBoost is publicly available for use at: http://farshidrayhan.pythonanywhere.com/iDTI-ESBoost/. In addition of target proteins, there are some types of RNA molecules so called non-coding RNAs and ncRNAs which are not translated into proteins. These RNA molecules als can make a new class of drug targets. Recently, a new database called NRDTD has been developed to collect the experimentally validated associations between drugs and ncRNAs[79]. In the future, our aim is to use NRDTD database as a gold standard dataset for predicting new associations between drugs and ncRNAs which have not been experimentally verified. By replacing features of target proteins with a set of informative features for ncRNAs, which have been published in the literature, the model presented in this study can also be used for predicting drug-ncRNA interaction prediction. Supplementary Information 1 Supplementary Information 2
  63 in total

1.  KEGG: kyoto encyclopedia of genes and genomes.

Authors:  M Kanehisa; S Goto
Journal:  Nucleic Acids Res       Date:  2000-01-01       Impact factor: 16.971

2.  Predicting drug-target interactions from chemical and genomic kernels using Bayesian matrix factorization.

Authors:  Mehmet Gönen
Journal:  Bioinformatics       Date:  2012-06-23       Impact factor: 6.937

3.  PubChem as a source of polypharmacology.

Authors:  Bin Chen; David Wild; Rajarshi Guha
Journal:  J Chem Inf Model       Date:  2009-09       Impact factor: 4.956

4.  Prediction of polypharmacological profiles of drugs by the integration of chemical, side effect, and therapeutic space.

Authors:  Feixiong Cheng; Weihua Li; Zengrui Wu; Xichuan Wang; Chen Zhang; Jie Li; Guixia Liu; Yun Tang
Journal:  J Chem Inf Model       Date:  2013-04-08       Impact factor: 4.956

5.  Predicting backbone Cα angles and dihedrals from protein sequences by stacked sparse auto-encoder deep neural network.

Authors:  James Lyons; Abdollah Dehzangi; Rhys Heffernan; Alok Sharma; Kuldip Paliwal; Abdul Sattar; Yaoqi Zhou; Yuedong Yang
Journal:  J Comput Chem       Date:  2014-09-12       Impact factor: 3.376

6.  SucStruct: Prediction of succinylated lysine residues by using structural properties of amino acids.

Authors:  Yosvany López; Abdollah Dehzangi; Sunil Pranit Lal; Ghazaleh Taherzadeh; Jacob Michaelson; Abdul Sattar; Tatsuhiko Tsunoda; Alok Sharma
Journal:  Anal Biochem       Date:  2017-03-28       Impact factor: 3.365

7.  Sequence-Based Prediction of Protein-Carbohydrate Binding Sites Using Support Vector Machines.

Authors:  Ghazaleh Taherzadeh; Yaoqi Zhou; Alan Wee-Chung Liew; Yuedong Yang
Journal:  J Chem Inf Model       Date:  2016-09-22       Impact factor: 4.956

8.  SELF-BLM: Prediction of drug-target interactions via self-training SVM.

Authors:  Jongsoo Keum; Hojung Nam
Journal:  PLoS One       Date:  2017-02-13       Impact factor: 3.240

9.  Some remarks on protein attribute prediction and pseudo amino acid composition.

Authors:  Kuo-Chen Chou
Journal:  J Theor Biol       Date:  2010-12-17       Impact factor: 2.691

10.  Identification of chemogenomic features from drug-target interaction networks using interpretable classifiers.

Authors:  Yasuo Tabei; Edouard Pauwels; Véronique Stoven; Kazuhiro Takemoto; Yoshihiro Yamanishi
Journal:  Bioinformatics       Date:  2012-09-15       Impact factor: 6.937

View more
  16 in total

1.  PreDTIs: prediction of drug-target interactions based on multiple feature information using gradient boosting framework with data balancing and feature selection techniques.

Authors:  S M Hasan Mahmud; Wenyu Chen; Yongsheng Liu; Md Abdul Awal; Kawsar Ahmed; Md Habibur Rahman; Mohammad Ali Moni
Journal:  Brief Bioinform       Date:  2021-03-12       Impact factor: 11.622

2.  DeepStack-DTIs: Predicting Drug-Target Interactions Using LightGBM Feature Selection and Deep-Stacked Ensemble Classifier.

Authors:  Yan Zhang; Zhiwen Jiang; Cheng Chen; Qinqin Wei; Haiming Gu; Bin Yu
Journal:  Interdiscip Sci       Date:  2021-11-03       Impact factor: 2.233

3.  AttentionSiteDTI: an interpretable graph-based model for drug-target interaction prediction using NLP sentence-level relation classification.

Authors:  Mehdi Yazdani-Jahromi; Niloofar Yousefi; Aida Tayebi; Elayaraja Kolanthai; Craig J Neal; Sudipta Seal; Ozlem Ozmen Garibay
Journal:  Brief Bioinform       Date:  2022-07-18       Impact factor: 13.994

4.  Binding affinity prediction for binary drug-target interactions using semi-supervised transfer learning.

Authors:  Betsabeh Tanoori; Mansoor Zolghadri Jahromi; Eghbal G Mansoori
Journal:  J Comput Aided Mol Des       Date:  2021-06-30       Impact factor: 3.686

Review 5.  Applications of artificial intelligence to drug design and discovery in the big data era: a comprehensive review.

Authors:  Neetu Tripathi; Manoj Kumar Goshisht; Sanat Kumar Sahu; Charu Arora
Journal:  Mol Divers       Date:  2021-06-10       Impact factor: 2.943

6.  DTiGEMS+: drug-target interaction prediction using graph embedding, graph mining, and similarity-based techniques.

Authors:  Maha A Thafar; Rawan S Olayan; Haitham Ashoor; Somayah Albaradei; Vladimir B Bajic; Xin Gao; Takashi Gojobori; Magbubah Essack
Journal:  J Cheminform       Date:  2020-06-29       Impact factor: 5.514

Review 7.  Machine learning approaches and databases for prediction of drug-target interaction: a survey paper.

Authors:  Maryam Bagherian; Elyas Sabeti; Kai Wang; Maureen A Sartor; Zaneta Nikolovska-Coleska; Kayvan Najarian
Journal:  Brief Bioinform       Date:  2021-01-18       Impact factor: 11.622

8.  iBCE-EL: A New Ensemble Learning Framework for Improved Linear B-Cell Epitope Prediction.

Authors:  Balachandran Manavalan; Rajiv Gandhi Govindaraj; Tae Hwan Shin; Myeong Ok Kim; Gwang Lee
Journal:  Front Immunol       Date:  2018-07-27       Impact factor: 7.561

9.  A Machine Learning Method for Drug Combination Prediction.

Authors:  Jiang Li; Xin-Yu Tong; Li-Da Zhu; Hong-Yu Zhang
Journal:  Front Genet       Date:  2020-08-25       Impact factor: 4.599

10.  Human Microbe-Disease Association Prediction Based on Adaptive Boosting.

Authors:  Li-Hong Peng; Jun Yin; Liqian Zhou; Ming-Xi Liu; Yan Zhao
Journal:  Front Microbiol       Date:  2018-10-09       Impact factor: 5.640

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.