Literature DB >> 31800601

Machine learning approach to single nucleotide polymorphism-based asthma prediction.

Joverlyn Gaudillo1,2, Jae Joseph Russell Rodriguez3, Allen Nazareno1, Lei Rigi Baltazar1,2, Julianne Vilela4, Rommel Bulalacao5, Mario Domingo5, Jason Albia1,2.   

Abstract

Machine learning (ML) is poised as a transformational approach uniquely positioned to discover the hidden biological interactions for better prediction and diagnosis of complex diseases. In this work, we integrated ML-based models for feature selection and classification to quantify the risk of individual susceptibility to asthma using single nucleotide polymorphism (SNP). Random forest (RF) and recursive feature elimination (RFE) algorithm were implemented to identify the SNPs with high implication to asthma. K-nearest neighbor (kNN) and support vector machine (SVM) algorithms were trained to classify the identified SNPs whether associated with non-asthmatic or asthmatic samples. Feature selection step showed that RF outperformed RFE and the feature importance score derived from RF was consistently high for a subset of SNPs, indicating the robustness of RF in selecting relevant features associated with asthma. Model comparison showed that the integration of RF-SVM obtained the highest model performance with an accuracy, precision, and sensitivity of 62.5%, 65.3%, and 69%, respectively, when compared to the baseline, RF-kNN, and an external MeanDiff-kNN models. Furthermore, results show that the occurrence of asthma can be predicted with an Area under the Curve (AUC) of 0.62 and 0.64 for RF-SVM and RF-kNN models, respectively. This study demonstrates the integration of ML models to augment traditional methods in predicting genetic predisposition to multifactorial diseases such as asthma.

Entities:  

Year:  2019        PMID: 31800601      PMCID: PMC6892549          DOI: 10.1371/journal.pone.0225574

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Risk prediction of complex diseases such as asthma proves to be challenging due to the combinatorial effects of environmental and genetic factors. The major challenges are: 1) obtaining a sufficient sample size to develop univariate or multivariate predictor and, 2) equipping the predictor with biological variables associated with complex disease. In general, biological modelling studies for disease risk assessment using genetic information have been supplemented by several technologies and study designs. For example, genome-wide association studies (GWAS) use single-nucleotide polymorphism (SNP), a single-base pair change in the DNA sequence, as biomarkers to locate genetic regions associated with a particular trait. Recent GWAS [1-5] identified SNPs linked to candidate genes implicated to asthma. However, GWAS only accounts for a small fraction of individual traits, while heritability related to complex diseases remains unresolved, thus, the emergence of “mystery of missing heritability” phenomenon. Based on [6, 7], the estimated missing heritability of asthma from twin studies ranges from 35% to 70%. One possible explanation for the “mystery of missing heritability” is the compounding biological interactions inherent to a complex disease [8]. Most of the biological modeling studies [9-11] which established SNP association with asthma still adopt the traditional statistical tests on individual SNP subject to a certain threshold, consequently neglecting the SNP-SNP interactions. Machine learning has the potential to unravel complex genetic and environment interactions constituting a disease by considering SNP-SNP interactions in model development. In genomic analysis, machine learning is employed to identify genetic variants and predict individual’s susceptibility to disease based on the identified biomarkers [12]. Despite the limited attempts, recent works have shown that machine learning-based approach augments the analysis on asthma at the genomic level, i.e., from genotype-phenotype association to phenotype prediction. For instance, Xu et al. [12] utilized random forest (RF) algorithm to identify relevant SNPs and to classify asthma case and control samples based on clinical data and SNPs. Due to the well-known curse of dimensionality inherent to the SNP data, i.e., higher feature size compared to sample size, constructing accurate machine learning-based predictive models for complex disease presents a challenge [12]. One way to circumvent the curse of dimensionality problem is to integrate various machine learning algorithms. The general approach is to employ feature selection method on the SNP data before classifying samples using machine learning-based classifiers. For example, Mieth et al. [13] used support vector machine (SVM) to select relevant SNPs associated with breast cancer. SVM along with Naive Bayes and decision trees have also been used to identify breast cancer cases using SNPs selected via information gain [14]. Furthermore, mean difference calculation and k-nearest neighbor (kNN) have also been employed to quantify SNP relevance and to perform classification task on the breast cancer dataset [15], respectively. In this work, we developed a novel machine learning-based approach to quantify the individual’s susceptibility to asthma based on SNP data. We implemented a two-step approach which involves feature selection and classification steps. The RF and RFE were utilized to select SNPs highly associated to asthma, while SVM and kNN were employed to quantify the propensity of an individual to develop asthma. The predictive models were evaluated using the following metrics: accuracy, sensitivity, precision, and area under the curve (AUC). In the next section, we describe the procedure of data collection, preprocessing, SNP selection, and classification. Finally, the results of integrating feature selection and classification methods will be compared with baseline and external models.

Materials and methods

Data collection and preprocessing

The SNP data was obtained from openSNP database [16], a public repository from 23andMe, deCODEme, and FamilyTreeDNA. The raw dataset is composed of an MS Excel file containing phenotypes information of users, and text files containing SNP data encoded in Variant Call Format (VCF). Each text file contains meta-information lines and a header line with the following field names: SNP ID, chromosome number, position, and genotype. Phenotypes irrelevant to the study were eliminated and users with incomplete report on asthma were discarded. The cleaned data is composed of 143 samples which represents the users with complete asthma report, 1,088,991 SNPs as features of each sample, and class column that labels each sample based on their phenotype (non-asthmatic or asthmatic). We adopted similar quality control procedure from [15] which involves missing call filtering and deviation from Hardy-Weinberg equilibrium to select the SNPs. For the missing call filtering, we formulated criteria based on the quality and characteristic of the collected data. The first criterion excludes SNPs whose percentage of present samples is ≤ 98%. For the second criterion, the samples were ranked based on the percentage of present SNP in which 90% of the samples were retained. The Hardy-Weinberg equilibrium principle states that in an infinitely sized population with no selection, migration and mutation, random mating will bring about constant allele and genotype frequencies across generations. Genotypes at a bi-allelic locus can then be predicted to have the following distribution of frequencies: where p2 is the frequency of homozygotes for allele p, 2pq is the frequency of heterozygotes for alleles p and q, and q2 is the frequency of homozygotes for allele q. SNPs that deviated from Hardy-Weinberg equilibrium due to selection, population admixture, cryptic relatedness, genotyping error and genuine genetic association [17] were excluded in the dataset. Chi-square test was implemented using Hardy-Weinberg package in R [18] on allele frequencies of each SNP from the control samples, wherein SNPs with p − value < 0.001 were discarded. After data preprocessing and quality control, 128 samples (57 samples with asthma condition during childhood and adulthood and 71 control samples are unaffected with asthma) were retained. A total of 176,288 SNPs was retained with genotypes comprised of minor and major alleles. Finally, genotypic encoding was implemented wherein homozygous, heterozygous, and variant homozygous SNPs were represented by 1, 2, 3, respectively.

Feature selection

Disease detection driven by high-dimensional data entails the use of feature selection method to identify the features highly associated to the phenotype [19]. In this study, RF algorithm which involves building multiple decision trees based on relevant SNPs was used. RF outperforms other feature selection methods in terms of robustness as a result of its ensemble trait of measuring the predictive importance of each feature to build efficient decision trees [19]. The RF algorithm was trained using the following input matrix, where N is the number of samples, and X and Y are the SNP values and the output class of i sample, respectively. For each data point, an out-of-bag error (OOBE) and the mean over the forest were calculated. After training, the j SNP was permuted among the training data and the difference the OOBE before and after permutation was calculated. To obtain the importance score of the j SNP, the mean of the differences were calculated and normalized.

Model construction

To predict the individual’s susceptibility to asthma, two different integrated machine learning models were constructed. These models are RF-kNN and RF-SVM which both employed RF as the feature selection algorithm to identify the relevant SNPs. Subsequently, the selected SNPs were used as inputs to kNN and SVM classifiers.

k-Nearest Neighbor (kNN)

kNN algorithm is a nonparametric lazy learner due to the absence of a training phase when performing classification tasks. In this work, the kNN algorithm was implemented on the SNP data to construct a predictive model that classifies non-asthmatic and asthmatic samples. A particular set of SNPs associated to a sample is represented as a continuous variable that can be plotted into the feature space. Hence, given a sample with SNP as features and an unknown label, the kNN algorithm gives prediction by performing a majority voting on the output class of the sample’s k nearest neighbors. For further discussion of kNN using SNP data, refer to [15]. To implement kNN, we define X be the p-dimensional input vector of the i sample, and is defined as, where p is the total number of SNPs. The corresponding output vector Y is given by, where N is the total number of samples. The dataset are plotted on a multi-dimensional feature space with their corresponding class label. To determine the class of an unseen sample U, the k nearest neighbor is determined by calculating the Euclidean distance d(U, X) relative to the other labeled samples. The Euclidean distance between U and X is given by, where j represents a certain SNP. We note that our selection of Euclidean distance as a similarity measure is motivated by numerous literatures [14, 15, 20] which extensively used this metric to analyze SNPs. Furthermore, to make our work more comprehensive, we tested several distance metrics and compared their performance with the Euclidean distance (see S2 Table).

Support vector machine (SVM)

SVM is a learning algorithm used in tasks such as non-linear classification [11, 13, 14], function estimation, and density estimation. The learning algorithm constructs a hyperplane that provides maximum separation or margin, between classes. For detailed discussion of SVM implementation on SNP data, refer to [13, 14]. The construction of the classification model can be viewed as an optimization problem. To implement SVM, the training vector representing SNP values and its corresponding label is given by, where N is the total number of samples, and X and Y represent the SNP values and class label of the i sample, respectively. The algorithm minimizes, subject to the following condition, where w is learning weight vector, b is the bias, ε is the error term, and C is the tradeoff parameter between the error and margin. After finding the optimal solution, the function, is used to predict the class of a unseen sample. The sign of the function is given by, wherein positive (+) and (-) values indicate non-asthmatic and asthmatic, respectively.

Hyperparameter selection and performance evaluation

One crucial step in constructing a machine learning predictive model is the selection of hyperparameters to achieve the highest model performance. Shown in Table 1 are the tested hyperparameter and their corresponding test values to train and optimize the machine learning algorithms.
Table 1

List of hyperparameter values.

MethodHyperparameterRange of values
Random Forestoptimum number of trees in the forest, n_estimators{20, 30, 40, 50, 60, 70, 80, 90, 100, 110, 120, 130, 140, 150, 160, 170, 180, 190, 200}
maximum number of features considered for splitting a node to achieve least uncertainty when creating a tree, max_features{400, 500, 600}
SVMkernel{linear, sigmoid, rbf, poly}
regularization parameter, C{1, 10, 100, 1000}
tolerance, ε{0.002, 0.1, 0.001, 0.0001}
kNNnumber of neighbors, k1, 3, 5, 7, 9, 11, 13, 15, 17, 19, 21, 23, 25, 27, 29, 31, 33, 35
Cross-validation techniques were employed to estimate the model performance on the unseen data. The RF classifier was evaluated using stratified k-fold cross validation (k = 10) which ensures that each fold contains roughly the same number of samples per class. RF-SVM and RF-kNN were evaluated using leave-one-out cross-validation (LOOCV), which discards a single sample to be used as a test set before training the model on the remaining samples; hence, k is equal to the number of samples (k = 128). The performance metrics used to evaluate the models were accuracy (ratio of correctly predicted observations to the total observations), precision (ratio of correctly predicted positive observations to the total predicted positive observations), and sensitivity (ratio of correctly predicted positive observations to the actual positive observations). The AUC—receiver operating characteristic (ROC) curve is another measure of performance of a machine learning classifier model. The ROC is a probability curve while the AUC represents the capability of the model to distinguish among classes. The ROC curve is constructed by plotting the true positive rate versus false positive rate. To further evaluate the results, we compare the integrated models with baseline model and an external model [15] implemented for breast cancer prediction. Scripts were developed using Python programming language and simulations were executed in a machine with 2.3 GHz Intel Core i5 processor, 8Gb random access memory (RAM), and 2 cores.

Results and discussion

We optimized the hyperparameters to obtain the best RF architecture. Results show that the highest performance having accuracy, precision, and sensitivity of 68.75%, 83.3%, and 68%, respectively, is obtained when n_estimators = 40 and max_features = 400. This optimum RF architecture is then used to select the SNPs which will serve as input to train the SVM and kNN classifiers. Training phase revealed that the highest accuracy for RF-SVM and RF-kNN were obtained when the optimum number of features are 310 and 400, respectively, see Fig 1.
Fig 1

Accuracy vs. number of SNP for SVM (left) and kNN (right).

To further evaluate the performance of RF in selecting highly associated SNPs, we implemented recursive feature elimination (RFE) in combination with SVM and kNN. As a feature selection method, RFE initially builds a model on the entire dataset and computes feature importance score for each predictor. For the next iteration, the predictor with the lowest feature importance score is omitted from the dataset that will be used to rebuild the model. Result shows (see Table 2) that RF when integrated to SVM and kNN classifiers has a better performance as a feature selection method compared with RFE.
Table 2

Comparison of feature selection methods.

ModelAccuracy (%)
RF-SVM62.17
RF-kNN61.70
RFE-SVM50.90
RFE-kNN49.68
Shown in Fig 2 is the feature importance scores of top 20 SNPs. Table 3 shows the characteristics of the top five SNPs that were selected by the RF based on feature importance score. Based on the feature importance score, the top five SNPs were consistently selected by RF in every fold during the cross-validation procedure. This result is indicative that integration of RF with classifiers is robust in selecting important SNPs. Favoring a robust feature selection algorithm apart from classification performance is highly desirable to further gain insight on the SNP data. Moreover, the robustness of the RF can be attributed to its stability in selecting important features, i.e., RF is insensitive to perturbations in the dataset [19].
Fig 2

Feature importance score of top 20 SNPs.

Table 3

Characteristics of top five SNPs selected by RF based on importance score [21].

SNP IDChromosome LocationGeneFunctional ConsequenceGeneCard Summary
rs75419501:147903855GJA8intron variantGJA8 (Gap Junction Protein Alpha 8) is a Protein Coding gene. Diseases associated with GJA8 include Cataract 1, Multiple Types and Cataract Microcornea Syndrome. Among its related pathways are Development Slit-Robo signaling and Vesicle-mediated transport. Gene Ontology (GO) annotations related to this gene include channel activity and gap junction channel activity
rs75419561:111366426LOC105378904intron variantRNA Gene, and is affiliated with the ncRNA class
rs75420251:40643890RIMS3intron variantRIMS3 (Regulating Synaptic Membrane Exocytosis 3) is a Protein Coding gene. Gene Ontology (GO) annotations related to this gene include ion channel binding
rs75420281:168718584DPTintron variantDPT (Dermatopontin) is a Protein Coding gene. Diseases associated with DPT include Commensal Bacterial Infectious Disease and Anisometropia
rs75420821:118617643NANANA
While the asthma-associated SNPs selected by RF has low feature importance scores, it is interesting to note that prior studies have not identified these SNPs to be associated to asthma, nor included in the 19 asthma-associated SNPs [22] present in the dataset. This observation is attributed to: 1) various sequencing platforms were used which may lead to the exclusion of lung-function associated SNPs or affect the phenotypic variance within the population; 2) as demonstrated in [23-26], RF takes into account SNP-SNP interactions while the traditional approach only performs direct genotype-phenotype association testing. Considering the result of our feature selection scheme, in which the selected SNPs are currently excluded in the list of known SNPs associated to asthma [22], we posit that the integration method can be utilized as a preliminary method to discover new biomarkers associated with complex diseases.

Model comparison

The SVM and kNN were implemented on both entire and selected dataset to establish baseline models. Furthermore, a more stringent test on the performance of SVM and kNN classifiers were implemented on the 19 asthma-associated SNPs. Considering the higher accuracy of RF as a feature selection method when integrated to SVM and kNN as discussed above, we compare the performance of RF-SVM and RF-kNN with the baseline model. Table 4 shows that the integrated models obtained higher classification performance than the baseline models.
Table 4

Comparison of performance between RF-SVM and RF-kNN to the baseline SVM and kNN models.

The selected SNPs for Baseline SVM and kNN refers to the 19 asthma-associated SNPs.

ModelsAccuracy (%)
Entire DatasetSelected SNPs
Baseline SVM52.8351.81
Baseline kNN49.6949.09
RF-SVM56.3062.17
Baseline kNN54.7061.70

Comparison of performance between RF-SVM and RF-kNN to the baseline SVM and kNN models.

The selected SNPs for Baseline SVM and kNN refers to the 19 asthma-associated SNPs. To further evaluate the performance of the integrated models, MeanDiff-kNN derived from [15] was implemented on the dataset. Fig 3 shows the comparison of the three integrated models RF-SVM, RF-kNN, MeanDiff-kNN. The RF-SVM attained the highest performance among the models with an accuracy, precision, and sensitivity of 62.17%, 65.3%, and 69%, respectively. We note that the baseline model [15] previously proposed for breast cancer detection when applied to the asthma dataset achieved 55% accuracy. Similar performance metrics percentage of 61.7% and 55% were attained for RF-kNN and MeanDiff-kNN, respectively. The SVM outperformed other classifiers in predicting asthma susceptibility by considering the interactions among SNPs to construct a more suitable predictive model tailored to the input data [13, 14, 27]. Compared to other machine learning methods, SVM is very powerful in detecting intricate trends and patterns in complex genomic dataset [20]. While SVM has better performance in high-dimensional data with few samples, kNN works well with data points in low-dimensional space. Shown in Table 5 is the optimal hyperparameters for each model.
Fig 3

Performance of the three ML models (RF-SVM, RF-kNN, MeanDiff-kNN).

Table 5

Optimal hyperparameters determined for the models.

MethodHyperparameter values
SVMkernel = rbf
regularization parameter, C = 100
tolerance, ε = 0.001
number of selected SNP = 310
kNNnumber of nearest neighbors, k = 7
number of selected SNP = 400
Fig 4 shows the ROC curves of RF-SVM, RF-kNN, and MeanDiff-kNN. Among the integrated models, RF-kNN attained the highest AUC of 0.64. The AUC of the RF-SVM and RF-kNN were comparable to the results in [12] which reported an AUC of 0.66 using 160 SNPs and clinical data to predict severe asthma exacerbations. Despite the absence of clinical data in the predictor variables, our integrated models obtained good performance in distinguishing case from control samples. Moreover, the integrated models achieved good prediction results, notwithstanding the limited sample size. This further indicate that the combination of machine learning algorithms used in the proposed models are able to predict asthma susceptibility with satisfactory performance despite the limited biological variables used in the model construction.
Fig 4

ROC Curves of RF-SVM (a), RF-kNN (b), and MeanDiff-kNN (c).

Conclusion

We developed machine learning-based predictive models which utilized SNP data to quantify the individual susceptibility to asthma. More precisely, we employed an integrated approach consisting of a feature selection step to identify the SNPs highly associated to asthma, and a classification step to predict asthma susceptibility. Results show that the integrated RF-SVM model achieves the highest accuracy, precision, sensitivity, and AUC compared to RF-kNN and baseline models. Furthermore, this study demonstrated that the integration of various machine learning methods is a well-suited approach to investigate high dimensional SNP data for genotype-phenotype association and phenotype prediction. To a large extent, we have shown how machine learning approach can augment the insights derived from GWAS in analyzing massive and complex biological data obtained from next generation sequencing (NGS) platforms. The integrated approach can be naturally extended on a wide variety of biological data such genetic mutations, copy number variations as well as clinical data which may take into account environmental factors to come up with a more holistic understanding of genetic etiology of complex diseases.

Source code

The Python codes as well as the SNP data used in this study is available in this github repository.

Feature importance table.

(PDF) Click here for additional data file.

Distance metrics performance table.

(PDF) Click here for additional data file.
  21 in total

1.  A pilot study on the application of statistical classification procedures to molecular epidemiological data.

Authors:  Holger Schwender; Manuela Zucknick; Katja Ickstadt; Hermann M Bolt
Journal:  Toxicol Lett       Date:  2004-06-15       Impact factor: 4.372

2.  Graphical tests for Hardy-Weinberg equilibrium based on the ternary plot.

Authors:  Jan Graffelman; Jair Morales Camarena
Journal:  Hum Hered       Date:  2007-09-26       Impact factor: 0.444

Review 3.  Applications of Support Vector Machine (SVM) Learning in Cancer Genomics.

Authors:  Shujun Huang; Nianguang Cai; Pedro Penzuti Pacheco; Shavira Narrandes; Yang Wang; Wayne Xu
Journal:  Cancer Genomics Proteomics       Date:  2018 Jan-Feb       Impact factor: 4.069

4.  A population-based study of bronchial asthma in adult twin pairs.

Authors:  M M Nieminen; J Kaprio; M Koskenvuo
Journal:  Chest       Date:  1991-07       Impact factor: 9.410

5.  Genetics of asthma and hay fever in Australian twins.

Authors:  D L Duffy; N G Martin; D Battistutta; J L Hopper; J D Mathews
Journal:  Am Rev Respir Dis       Date:  1990-12

6.  Association of IL33-IL-1 receptor-like 1 (IL1RL1) pathway polymorphisms with wheezing phenotypes and asthma in childhood.

Authors:  Olga E Savenije; Jestinah M Mahachie John; Raquel Granell; Marjan Kerkhof; F Nicole Dijk; Johan C de Jongste; Henriëtte A Smit; Bert Brunekreef; Dirkje S Postma; Kristel Van Steen; John Henderson; Gerard H Koppelman
Journal:  J Allergy Clin Immunol       Date:  2014-02-22       Impact factor: 10.793

7.  Predictive models for breast cancer susceptibility from multiple single nucleotide polymorphisms.

Authors:  Jennifer Listgarten; Sambasivarao Damaraju; Brett Poulin; Lillian Cook; Jennifer Dufour; Adrian Driga; John Mackey; David Wishart; Russ Greiner; Brent Zanke
Journal:  Clin Cancer Res       Date:  2004-04-15       Impact factor: 12.531

8.  The challenge for genetic epidemiologists: how to analyze large numbers of SNPs in relation to complex diseases.

Authors:  A Geert Heidema; Jolanda M A Boer; Nico Nagelkerke; Edwin C M Mariman; Daphne L van der A; Edith J M Feskens
Journal:  BMC Genet       Date:  2006-04-21       Impact factor: 2.797

9.  Genome-wide association study implicates chromosome 9q21.31 as a susceptibility locus for asthma in mexican children.

Authors:  Dana B Hancock; Isabelle Romieu; Min Shi; Juan-Jose Sienra-Monge; Hao Wu; Grace Y Chiu; Huiling Li; Blanca Estela del Rio-Navarro; Saffron A G Willis-Owen; Saffron A G Willis-Owens; Scott T Weiss; Benjamin A Raby; Hong Gao; Celeste Eng; Rocio Chapela; Esteban G Burchard; Hua Tang; Patrick F Sullivan; Stephanie J London
Journal:  PLoS Genet       Date:  2009-08-28       Impact factor: 5.917

10.  Genome-wide prediction of childhood asthma and related phenotypes in a longitudinal birth cohort.

Authors:  Ben D Spycher; John Henderson; Raquel Granell; David M Evans; George Davey Smith; Nicholas J Timpson; Jonathan A C Sterne
Journal:  J Allergy Clin Immunol       Date:  2012-08       Impact factor: 10.793

View more
  5 in total

1.  Machine learning model for predicting Major Depressive Disorder using RNA-Seq data: optimization of classification approach.

Authors:  Pragya Verma; Madhvi Shakya
Journal:  Cogn Neurodyn       Date:  2021-09-22       Impact factor: 5.082

2.  DeePred-BBB: A Blood Brain Barrier Permeability Prediction Model With Improved Accuracy.

Authors:  Rajnish Kumar; Anju Sharma; Athanasios Alexiou; Anwar L Bilgrami; Mohammad Amjad Kamal; Ghulam Md Ashraf
Journal:  Front Neurosci       Date:  2022-05-03       Impact factor: 5.152

3.  Building a model for predicting metabolic syndrome using artificial intelligence based on an investigation of whole-genome sequencing.

Authors:  Nai-Wei Hsu; Kai-Chen Chou; Chien-Feng Kuo; Yu-Ting Tina Wang; Chung-Lieh Hung; Shin-Yi Tsai
Journal:  J Transl Med       Date:  2022-04-28       Impact factor: 8.440

4.  A machine learning-based SNP-set analysis approach for identifying disease-associated susceptibility loci.

Authors:  Princess P Silva; Joverlyn D Gaudillo; Julianne A Vilela; Ranzivelle Marianne L Roxas-Villanueva; Beatrice J Tiangco; Mario R Domingo; Jason R Albia
Journal:  Sci Rep       Date:  2022-09-22       Impact factor: 4.996

5.  Risk score prediction model based on single nucleotide polymorphism for predicting malaria: a machine learning approach.

Authors:  Kah Yee Tai; Jasbir Dhaliwal; KokSheik Wong
Journal:  BMC Bioinformatics       Date:  2022-08-07       Impact factor: 3.307

  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.