Literature DB >> 35390039

Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis.

Mohammad Ziaul Islam Chowdhury1,2,3, Iffat Naeem1, Hude Quan1, Alexander A Leung1,4, Khokan C Sikdar5, Maeve O'Beirne2, Tanvir C Turin1,2.   

Abstract

OBJECTIVE: We aimed to identify existing hypertension risk prediction models developed using traditional regression-based or machine learning approaches and compare their predictive performance.
METHODS: We systematically searched MEDLINE, EMBASE, Web of Science, Scopus, and the grey literature for studies predicting the risk of hypertension among the general adult population. Summary statistics from the individual studies were the C-statistic, and a random-effects meta-analysis was used to obtain pooled estimates. The predictive performance of pooled estimates was compared between traditional regression-based models and machine learning-based models. The potential sources of heterogeneity were assessed using meta-regression, and study quality was assessed using the PROBAST (Prediction model Risk Of Bias ASsessment Tool) checklist.
RESULTS: Of 14,778 articles, 52 articles were selected for systematic review and 32 for meta-analysis. The overall pooled C-statistics was 0.75 [0.73-0.77] for the traditional regression-based models and 0.76 [0.72-0.79] for the machine learning-based models. High heterogeneity in C-statistic was observed. The age (p = 0.011), and sex (p = 0.044) of the participants and the number of risk factors considered in the model (p = 0.001) were identified as a source of heterogeneity in traditional regression-based models.
CONCLUSION: We attempted to provide a comprehensive evaluation of hypertension risk prediction models. Many models with acceptable-to-good predictive performance were identified. Only a few models were externally validated, and the risk of bias and applicability was a concern in many studies. Overall discrimination was similar between models derived from traditional regression analysis and machine learning methods. More external validation and impact studies to implement the hypertension risk prediction model in clinical practice are required.

Entities:  

Mesh:

Year:  2022        PMID: 35390039      PMCID: PMC8989291          DOI: 10.1371/journal.pone.0266334

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Hypertension is a common medical condition affecting about 1 in 4 people [1] and is a significant risk factor for heart attack, stroke, kidney disease, and mortality [2]. Hypertension has been linked to 13% of deaths globally [3] and is a significant health burden that affects all population segments. Considering the high prevalence and global burden, hypertension prevention, and control strategies need to be a top priority. Hypertension can be prevented by applying strategies that target the general population or individuals and groups at higher risk for hypertension [4]. The need for early identification of at-risk individuals who could benefit from preventive interventions has led to a growing interest in hypertension risk prediction. Predicting the risk of developing hypertension through modeling can help identify important risk factors contributing to hypertension, provide reasonable estimates about future hypertension risk [5], and help identify high-risk individuals targeted for healthy behavioral changes and medical treatment to prevent hypertension [6-8]. Many prediction models have been developed to predict the risk of hypertension in the general population over the years. Models were developed using either a traditional regression-based approach or a modern machine learning approach. Although machine learning approaches are known to produce better predictive performance, their performance often varies, and it is not clear if they perform better than the traditional regression-based models in predicting hypertension. Through a systematic review and subsequent meta-analysis, a pooled synthesis of performance measures of different models produced in multiple studies can be compared and measured [9]. This methodology provides an overview of these models’ predictive ability and allows the models’ performance measures based on the reported data to be explored quantitatively [9]. Two prior studies systematically analyzed hypertension risk prediction models in adults [10, 11]. Both studies performed a narrative synthesis of the evidence to summarize hypertension prediction models’ existing knowledge, and one study also performed a meta-analysis without assessing heterogeneity. None of the prior studies stratified models according to how they were developed. This stratification is important because there are inherent differences in these two types of models’ developmental methods in computation, complexity, interpretability, and accuracy. A formal assessment of study quality was also absent in prior studies. In addition to these two prior reviews, a systematic review was also carried out on prediction models to classify children at an elevated risk of developing hypertension [12]. With this in mind, we aimed to 1) systematically review the literature to identify hypertension risk prediction models that have been applied to the general adult population and the risk factors that were considered in those models; 2) characterize the study populations in which these models were derived and validated, 3) compare the predictive performance of traditionally developed regression-based models and machine learning models, and 4) assess the quality of these prediction models to better inform the selection of models for clinical implementation.

Materials and methods

Data sources and searches

We conducted a systematic review and meta-analysis to identify existing hypertension risk prediction models and associated risk factors and evaluated the models’ predictive performance. We searched MEDLINE, EMBASE, Web of Science, and Scopus (each from inception to December 2020) to identify studies predicting the risk of incident hypertension in the general adult population. Google Scholar and ProQuest (theses and dissertations) were searched for grey literature. Additionally, we explored the reference lists of all relevant articles. The search strategy focused on two key concepts: hypertension and risk prediction. We used proper free-text words and Medical Subject Headings (MeSH) terms to identify relevant studies for each key concept. Certain text words were truncated, or wildcards were used when required. The Boolean operators “AND”, “OR”, and “NOT” were used to combine the words and MeSH terms. A detailed search strategy for MEDLINE is provided in S1 Table.

Eligibility criteria

Although risk prediction models are generally developed using a cohort-based study design with follow-up information, we considered all types of study designs, anticipating that machine learning-based models may use other types of study design. Only original studies were included in this review: this excluded reviews, editorials, commentaries, and letters to the editor. Studies written in languages other than English and French were also excluded. The Population, Prognostic Factors (or models of interest), and Outcome [13] framework was used to outline eligibility criteria.

Population

The study population consisted of people free of hypertension at baseline and those around which hypertension risk prediction models were developed. No restrictions were imposed on the geographic region, time, or gender of the study participants. Nevertheless, only models developed on the adult population were considered, as outcome essential hypertension is expected in adults.

Prognostic factors (or models of interest)

We considered studies where risk prediction models for hypertension in the general adult population were developed. Studies that focused solely on the added predictive value of new risk factors to an existing prediction model, studies presenting a prediction model developed in patients with previous hypertension, or studies that derived risk prediction tools other than score-type tools (e.g., risk charts) were not considered. Further, we did not consider studies that only assessed bivariate association between predictors and hypertension. Instead, we focused on those studies where risk prediction models for hypertension were built incorporating risk factors that demonstrated significant prognostic contribution in predicting incident hypertension. When a model was assessed on more than one external population, information from all reported models was considered. However, when the model was presented both in a derivation and validation cohort, only data from the validation cohort were considered for meta-analysis.

Outcome

Our outcome of interest was hypertension, and we considered all definitions of hypertension to capture the maximum number of studies.

Study selection

Two reviewers (MC and IN) independently identified eligible articles using a two-step process. First, the title and abstracts of non-duplicated records were screened by two reviewers. Studies retained (based on eligibility criteria) during this stage of screening went to a full-text screening. Full-text articles were further screened for eligibility by the same two reviewers independently. Lastly, articles containing extractable data on hypertension prediction models and hypertension risk factors were selected for data extraction. Inter-rater reliability (Kappa coefficient) was estimated to measure agreement between the independent reviewers. Any disagreement between reviewers was resolved through consensus.

Data extraction

Two reviewers (MC and IN) independently extracted data from each study using standardized forms. We classified the identified models into two categories: models developed using a traditional regression-based approach and models developed using machine learning algorithms. Separate data extraction sheets were used for each model type and included study name, the location where the model was developed/location of data used for the model developed and participants’ ethnicity, study design used, sample size, age, and gender of the study participants, risk factors included in the model, number of events and total participants, an outcome considered, the definition used for hypertension, duration of follow-up, modeling method used, measures of discrimination and calibration of the prediction model, and the validation of the prediction model. In a separate form, information about the externally validated hypertension risk prediction models was extracted, including study name/model validated, the total number of validation studies, location of the validation study, follow-up period, number of events, and total participants, the definition of outcome and discrimination and calibration of the model. We also extracted information about risk factors, particularly how many times a specific risk factor was considered in the models. Each reviewer assessed study quality according to the Prediction model Risk Of Bias ASsessment Tool (PROBAST) checklist [14, 15]. The PROBAST is designed to evaluate the risk of bias and concerns regarding diagnostic and prognostic prediction model studies’ applicability. The PROBAST contains 20 questions under four domains: participants, predictors, outcome, and analysis, facilitating judgment of risk of bias and applicability. The overall risk of bias of the prediction models was judged as “low”, “high”, or “unclear,” and overall applicability of the prediction models was considered as “low concern”, “high concern”, and “unclear” according to the PROBAST checklist [14, 15].

Data analysis

We summarized the number of studies identified and those included and excluded (with the reason for exclusion) from the systematic review and subsequent meta-analysis using the PRISMA flow diagram [16]. In data synthesis, we performed a meta-analysis on the performance measure of the traditional regression type’s prediction modeling (e.g., logistic regression model and Cox proportional hazard regression model) and a more complicated modeling strategy (e.g., machine learning tools). Discrimination and calibration are the two most common statistical measures of predictive performance. Discrimination is commonly quantified by the concordance (C) statistic. In this review, we performed a meta-analysis on the C-statistic or AUC (area under the receiver operating characteristic curve) to evaluate the models’ predictive performance and provided a comprehensive summary of the models’ predictive ability. We did not undertake a meta-analysis of the calibration due to the unavailability of relevant data. We logit transformed the C-statistics before pooling as per recommendation [17, 18] and then back-transformed the results to the original scale for interpretation. We used a random-effects meta-analysis with REML estimation and Hartung-Knapp-Sidik-Jonkman (HKSJ) confidence interval (CI) to obtain the pooled weighted average of the logit C-statistic [19]. Forest plots were generated to show the pooled C-statistic together with the 95% CI, 95% approximate prediction interval (indicates an expected performance range of the considered models in a new population) for the summary C-statistic, the author’s name, publication year, and study weights. In studies that only provided a C-statistic but no measure of its variance or confidence intervals, the standard error (SE) and 95% CI of the logit C-statistic (or area under the receiver operating characteristic curve (AUC)) was calculated using the appropriate formula [19]. However, when the C-statistics’ confidence intervals (CIs) were available, standard errors (SE’s) of the logit C-statistics were derived from the CIs [19]. The presence of heterogeneity (primarily due to differences in the study setting, participants, and methodology) was assessed using Cochran’s Q statistic and quantified with the I2 statistic. A p-value of less than 0.05 was considered statistically significant heterogeneity and was categorized as low, moderate, and high when the I2 values were below 25%, between 25% and 75%, and above 75%, respectively [20]. Sources of heterogeneity were further explored using meta-regression and stratified analyses according to modeling type and study characteristics (sex of the participants, age of the participants, number of risk factors considered in the model, sample size considered in the model, and ethnicity of the study participants). We calculated 95% prediction intervals to provide a likely range of performance of a prediction model in a new population and setting. We did not assess publication bias by any statistical tests or funnel plot asymmetry. We used Stata version 16.1 (StataCorp LP, College Station, TX, USA) to perform statistical analysis using the following commands: meta, metan and metareg.

Results

Study identification and selection

We identified 14,730 articles through our electronic database search and an additional 48 articles through our grey literature search. After removing duplicates, titles, and abstracts screening and full-text screening 52 articles were finally selected for the systematic review. Within the chosen final studies, 32 studies provided sufficient information for synthesis through a meta-analysis. The detailed study selection process is summarized in Fig 1. Agreement between reviewers on the initial screening and final articles eligible for inclusion in the systematic review was good (κ = 0.81, and κ = 0.89, respectively). A total of 117 models were identified from the finally selected articles predicting the risk of hypertension in the general adult population, of which 75 were developed using traditional regression-based modeling and 42 using machine learning tools.
Fig 1

PRISMA diagram for systematic review of studies presenting hypertension prediction models developed in the general population.

Study characteristics of traditional regression-based models

Study characteristics of traditional regression-based models are presented in Tables 1 and 2. A total of 573,268 participants were used to develop 75 traditional models in 34 studies. Models mainly were developed either in white Caucasian or Asian populations. There was no model derived from African populations and only one [21] from Latin American populations. Two studies considered only male participants, one study considered only female participants, and the remaining studies considered both to develop the models. The number of risk factors considered to create the models ranged from 1 to 19, with a median of 7 risk factors per model. Age was the most common risk factor considered in 61 models, followed by body mass index (BMI) (32 models), diastolic blood pressure (DBP) (28 models), systolic blood pressure (SBP) (27 models), and sex (21 models). The distribution of the conventional risk factors considered in the different models is presented in Fig 2A. Duration of follow-up time (mean/median/total) considered to develop the models varied between 1.6 years to 30 years. The age of the study participants ranged from 15 to 90 years. SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg, or use of antihypertensive medication was the standard definition used to define hypertension in almost all the studies, except one study where SBP ≥ 130 mm Hg, DBP ≥ 80 mm Hg, or use of any antihypertensive drug was used. Logistic regression was the most used methodology to develop the model (15 studies), followed by Cox proportional-hazards regression (11 studies) and Weibull regression (6 studies). Calibration of the prediction model was not reported by most of the studies (19 studies). Studies those reported calibration measures (15 studies) were mainly using the Hosmer-Lemeshow test. Discrimination was assessed using the C-statistic (or AUC) and reported by almost all studies with values ranging from 0.57 to 0.97. Only one model was externally validated by the same study when they developed the model. Only eight models [22-29] were converted into a risk score after model development.
Table 1

Characteristics of included studies that describe traditional regression-based hypertension prediction models.

StudyLocation Model Developed/ EthnicityStudy DesignAgeGenderEvents (n)/Total Participants (N)Definition of Outcome Predicted/HypertensionDuration of Follow-up
Pearson et al. [41] 1990USA/Mixed, mainly WhitesProspective cohort≤ 25 yearsMale only114/1130Self-reported use of blood pressure-lowering medications30 years
Parikh et al. [22] 2008USA/Mainly WhitesProspective cohort20–69 yearsBoth796/1717SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or use of BP-lowering medicationsMedian 3.8 years
Paynter et al. [42] 2009USA/ Whites and BlacksProspective cohort45–64 yearsFemale onlyDerivation cohort: 1935/9427 Validation cohort: 1068/5395Self-report or SBP ≥ 140 mmHg or DBP ≥ 90 mmHg8 years
Kivimäki et al. [43] 2009England/Mainly WhitesProspective cohort35–68 yearsBoth1258/8207SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or use of BP-lowering medicationsMedian 5.6 years
Kivimäki et al. [44] 2010England/Mainly WhitesProspective cohort36–68 yearsBothDerivation cohort: 614/4135 Validation cohort: 438/2785SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or use of antihypertensive medicationsMedian 5.8 years
Kshirsagar et al. [45] 2010USA/Mixed but mainly WhitesProspective cohort45–64 yearsBoth3795/11,407 (7610 for derivation sample and 3692 for the validation sample)SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or reported use of BP-lowering medicationsUp to 9 years
Bozorgmanesh et al., [25] 2011Iran/AsiansProspective cohort≥ 20 yearsBoth805/4656SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or reported use of BP-lowering medications6 years
Chien et al. [24] 2011Taiwan/ChineseProspective cohort≥ 35 yearsBoth1029/2506SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or reported use of BP-lowering medicationsMedian 6.15 years
Fava et al. [46] 2013Sweden/WhitesProspective cohortMiddle-agedBothNR/10,781SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or reported use of BP-lowering medicationsOver average 23-years
Lim et al. [30] 2013Korea/AsiansProspective cohort40–69 yearsBoth819/4747. Derivation cohort: 483/2840 Validation cohort: 336/1907SBP ≥ 140 mmHg or DBP ≥ 90 mmHg or reported use of BP lowering medications4 years
Choi et al. [47] 2014USA/MexicansProspective cohortNRBothNR/443SBP >140 mm Hg, DBP >90 mm Hg, or use of antihypertensive medicationNR
Lim et al. [48] 2015Korean/AsiansProspective cohort40–69 yearsBothNR/5632SBP ≥140 mm Hg or DBP ≥90 mm Hg or use of antihypertensive medication4-year
Otsuka et al. [23] 2015Japan/AsiansProspective cohort19–63 yearsMale only1633/15,025SBP ≥140 mm Hg or DBP ≥90 mm Hg or use of antihypertensive medicationMedian 4 years
Asgari et al. [49] 2015Iran/AsiansProspective cohort≥ 20 yearsBothISH: 235/4574 IDH: 470/4809ISH: SBP ≥ 140 mmHg and DBP < 90 mmHg IDH: SBP <140 mmHg and DBP ≥ 90 mmHgISH: Median 9.57 years, IDH: Median 9.62 years
Sathish et al. [29] 2016India/AsiansProspective cohort15–64 yearsBoth70/297SBP ≥140 mm Hg or DBP ≥90 mm Hg or use of antihypertensive medicationMean 7.1 years
Lee et al. [50] 2015Korea/AsiansProspective cohort40–69 yearsBothMen: 384/2128 Women: 374/2326SBP ≥140 mm Hg or DBP ≥90 mm Hg or use of antihypertensive medication4 years
Lee et al. [51] 2014Korea/AsiansCross-sectional21–85 yearsBothNR/12,789SBP ≥ 140 mmHg and/or DBP ≥ 90 mmHg or physician-diagnosed hypertensionNR
Kanegae et al. [32] 2018Japan/AsiansProspective cohort18–83 yearsBoth7402/63,495SBP/DBP ≥ 140/90 mm Hg and/or the initiation of antihypertensive medications with self-reported hypertensionMean 3.4 years
Chen et al. [52] 2016China/AsiansProspective cohortAverage age 41.73 years (men), 39.49 years (women)Both2021 (men), 764 (women) 7537 (men), 4960 (women)First occurrence at any follow-up medical check-up of SBP > 140 mm Hg or DBP > 90 mm Hg or of the person taking antihypertensive medicationMedian 4.0 years
Díaz-Gutiérrez et al. [28] 2019Spain/SpanishProspective cohortAge presented according to the number of healthy lifestyle factorsBoth1406/14057SBP ≥ 130 mmHg, DBP≥ 80 mmHg, or use of any antihypertensive drugMedian 10.2 years
Wang et al. [53] 2018China/AsiansLongitudinal18–90 yearsBoth882/5265 (derivation) NR/1597 (validation)Taking antihypertensive drugs or SBP at least 140 mmHg or DBP at least 90 mmHgAverage follow-up of 8.05 ± 5.27 years
Niiranen et al. [54] 2016Finland/WhitesProspective cohort≥ 30 yearsBothNR/2045BP ≥ 140/90 mm Hg and/or antihypertensive medication11 years
Yeh et al. [55] 2001Taiwan/ChineseProspective cohort≥ 20 yearsBoth88/2374SBP ≥140 mm Hg or DBP ≥90 mm HgAverage 3.23 years
Syllos et al. [21] 2020Brazil/South AmericansProspective cohort35–74 yearsBoth1088/8027; Derivation: 4825 Validation: 3202SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg or the use of blood pressure-lowering medications4 years
Wang et al. [27] 2020China/AsiansProspective cohort≥ 18 yearsBoth1658/9034SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg or the use of blood pressure-lowering medicationsMedian 6 years
Xu et al. [56] 2019China/AsiansProspective cohort35–74 yearsBoth1036/4796 (Training)SBP ≥ 140 mm Hg and/or DBP ≥ 90 mm Hg, and/or a diagnosis of hypertension by a physician and currently receiving anti-hypertension treatment6 years
Kadomatsu et al. [26] 2019Japan/AsiansProspective cohortMean age 51.3 yearsBoth324/3936SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg, or use of antihypertensive medicationMedian 5 years
Wang et al. [57] 2015USA/Multi-ethnicTelephone-based health survey≥ 18 yearsBothNR/308,711NRNR
Muntner et al. [58] 2010USA/ Multi-ethnic (Whites, Blacks, Hispanics, and Asians)NR45–84 yearsBoth849/3013The first study visit, subsequent to baseline, at which SBP ≥ 140 mm Hg and/or DBP ≥ 90 mm Hg and/or the initiation of antihypertensive medicationMedian of 1.6 years and 4.8 years
Ture et al. [59] 2005Turkey/ EuropeansRetrospectiveAverage 48.2 years (hypertension) 46.5 (control)Both694 (452 patients with hypertension and 242 controls)Average of 3 or more DBP measurements on at least 3 subsequent visits is ≥ 90 mmHg or when the average of multiple SBP readings on 3 or more subsequent visits is consistently ≥ 140 mmHgNR
Yamakado et al. [60] 2015Japan/AsiansProspective cohort≥ 20 yearsBoth424/2637SBP ≥ 140 mm Hg or DBP ≥ 90 mm Hg or use of antihypertensive medication4 years
Qi et al. [61] 2014China/AsiansCase-controlCase: 64.48 ± 8.53 years; Control: 64.23 ± 10.13 yearsBothPatients: NR/1009 Controls = NR/756SBP ≥ 140 mm Hg or DBP ≥ 90 mm Hg or use of antihypertensive medicationNR
Lu et al. [62] 2015China/AsiansProspective cohort35–74 yearsBoth2559/7724SBP ≥ 140 mm Hg or DBP ≥ 90 mm Hg or use of antihypertensive medicationMean 7.9 years
Zhang et al. [63] 2015China/AsiansProspective cohort18–88 yearsBoth3793/17,471SBP ≥ 140 mm Hg or DBP ≥ 90 mm Hg or use of antihypertensive medication5 years

NR, not reported; SBP, systolic blood pressure; DBP, diastolic blood pressure; BP, blood pressure; ISH, isolated systolic hypertension; IDH, isolated diastolic hypertension

Table 2

The features of hypertension prediction models developed using a traditional regression-based modeling approach.

StudyRisk Factors IncludedModeling MethodDiscriminationCalibrationModel Validation: Internal or External
Pearson et al. [41] 1990Age, SBP at baseline, paternal history of hypertension, and BMICox regressionNRNRNR
Parikh et al. [22] 2008Age, sex, SBP, DBP, BMI, parental hypertension, and cigarette smokingWeibull regressionC-statistic = 0.788 [0.733–0.803]HL Chi-square = 4.35 (p = 0.88)Internal, apparent
Paynter et al. [42] 2009Inclusive Model: Age, ethnicity, BMI, total grain intake, SBP, DBP, apolipoprotein B, lipoprotein (a), and C-reactive protein. Simplified Model with Lipids: Age, BMI, SBP, DBP, ethnicity, and total to HDL- cholesterol ratio Simplified Model: Age, BMI, ethnicity, SBP, and DBPLogistic regressionInclusive Model: C-statistic = 0.705; Simplified Model with Lipids: C-statistic = 0.705; Simplified Model: C-statistic = 0.703Inclusive Model: HL Chi-square = 24.6 (p = 0.002), Simplified Model with Lipids: HL Chi-square = 20.7 (p = 0.008), Simplified Model: HL Chi-square = 12.3 (p = 0.140)Internal, split-sample 2:1
Kivimäki et al. [43] 2009Age, sex, SBP, DBP, BMI, parental hypertension, and cigarette smokingWeibull regressionC-statistic = 0.804HL Chi-square = 14.3 (p = 0.88)Internal, split-sample 6:4
Kivimäki et al. [44] 2010Repeat Measure BP Model: Age, sex, BMI, parental hypertension, repeat measures of BP, and cigarette smoking Average BP Model: Age, sex, BMI, parental hypertension, average BP, and cigarette smokingWeibull regressionRepeat Measure BP Model: C-statistic = 0.799; Average BP Model: C-statistic = 0.794Repeat Measure BP Model: HL Chi-square = 6.5; Average BP Model: NRInternal, split-sample 6:4
Kshirsagar et al. [45] 2010Age, level of SBP or DBP, smoking, family history of hypertension, diabetes mellitus, BMI, female sex, and lack of exerciseLogistic regressionAUC = 0.742 (3years), 0.750 (6 years), 0.791 (9 years), and 0.775 (ever)NRInternal, split-sample 2:1
Bozorgmanesh et al., [25] 2011For Women: age, waist circumference, DBP, SBP, and family history of premature CVDFor Men: age, DBP, SBP, and smokingWeibull regressionC-statistic = 0.731 [0.706–0.755] for women, C-statistic = 0.741 [0.719–0.763] for menHL Chi-square = 7.8 (p = 0.554) for women; HL Chi-square = 8.8 (p = 0.452) for menNR
Chien et al. [24] 2011Clinical Model: Age, gender, BMI, SBP, and DBPBiochemical Model: Age, gender, BMI, SBP, DBP, white blood count, fasting glucose, uric acidWeibull regressionClinical Model: AUC = 0.732 [0.712–0.752] (point based), AUC = 0.737 (coefficient based); Biochemical Model: AUC = 0.735 [0.715–0.755] (point based), AUC = 0.74 (coefficient based)Clinical Model: HL Chi-square = 8.3, p = 0.40 (point based), 10.9, p = 0.21 (coefficient based); Biochemical Model: HL Chi-square = 13.2, p = 0.11 (point based), 6.4, p = 0.60 (coefficient based)Internal, fivefold cross- validation
Fava et al. [46] 2013Age, sex, sex times age, heart rate, obesity, diabetes, hypertriglyceridemia, prehypertension, family history of hypertension, sedentary in spare time, problematic alcohol behavior, married or living as a couple, high-level non-manual work, smokingLogistic regressionAUC = 0.662 [0.651–0.672]NRNR
Lim et al. [30] 2013Age, sex, smoking, SBP, DBP, parental hypertension, BMIWeibull regressionAROC = 0.791 [0.766–0.817]HL Chi-square = 4.17 (p = 0.8415)Internal, split-sample 6:4
Choi et al. [47] 2014Age, gender, smoke, age x gender, Rs10510257 (AA), Rs10510257 (AG), Rs1047115 (GT)GEE for marginal model and logistic random effect model for conditional modelMarginal model: AUC = 0.839 (with SNPs), 0.826 (without SNPs) Conditional model: AUC = 0.973 (with SNPs), 0.973 (without SNPs)NRNR
Lim et al. [48] 2015Traditional variables: age, gender, SBP, current smoking status, family history of hypertension, BMI, and one genetic variable (cGRS or wGRS derived from the 4 SNPs): rs995322, rs17249754, rs1378942, rs12945290Logistic regressionDerivation cohort: C-statistic = 0.810 [0.796–0.824] (model without wGRS, C-statistic = 0.811 [0.797–0.825] (model with wGRS); Validation cohort: Mean C-statistic = 0.811 [0.809–0.816]HL Chi-square = 6.916 (model without wGRS), HL Chi-square = 5.711 (model with wGRS)Internal validation, fivefold cross-validation
Otsuka et al. [23] 2015Age, BMI, SBP and DBP, current smoking status, excessive alcohol intake, parental history of hypertensionCox regressionValidation cohort: C-statistic = 0.861 [0.844–0.877] (model), C-statistic = 0.858 [0.840–0.876] (score)Validation cohort: HL Chi-square = 15.2 (p = 0.085) (model), HL Chi-square = 9.30 (p = 0.41) (score)Internal validation, split sample 4:1
Asgari et al. [49] 2016ISH: Age, SBP, BMI, 2 hours post-challenge plasma glucose IDH: Age, DBP, waist circumference, marital status, gender, HDL-CCox regressionISH: C-statistic = 0.91, IDH: C-statistic = 0.76NRNR
Sathish et al. [29] 2016Age, sex, years of schooling, daily intake of fruits or vegetables, current smoking, alcohol use, BP, prehypertension, central obesity, history of high blood glucoseLogistic regressionAUC = 0.802 [0.748–0.856]Hosmer-Lemeshow p = 0.940NR
Lee et al. [50] 2015BMI, waist circumference, waist-to-hip ratio, waist-to-height ratioCox regressionMen: AROC = 0.58 [0.56–0.60] (BMI), 0.62 [0.60–0.64] (WC, WHR, WHtR) Women: AROC = 0.57 [0.55–0.59] (BMI), 0.66 [0.64–0.68] (WC), 0.68 [0.66–0.70] (WHR, WHtR)NRNR
Lee et al. [51] 2014Women: Height, age, neckC, axillaryC, ribC, waistC, pelvicC, rib_hip, waist_hip, pelvic_hip, rib_pelvic, axillary_rib, chest_rib, axillary_chest, forehead_neck (CFS), height, weight, BMI, age, chestC, forehead_hip, waist_hip, chest_pelvic, waist_pelvic, axillary_waist, forehead_rib, neck_axillary (LR-wrapper)Men: Age, foreheadC, neckC, axillaryC, chestC, ribC, waistC, pelvicC, hipC, rib_hip, waist_hip, rib_pelvic, waist_pelvic, chest_waist, forehead_rib, chest_rib, axillary_chest, forehead_neck (CFS), height, foreheadC, neckC, axillaryC, ribC, pelvicC, forehead_hip, chest_hip, rib_hip, pelvic_hip, forehead_waist, axillary_waist, rib_waist, neck_rib, axillary_rib, chest_rib, forehead_axillary, forehead_neck, WHtR (LR-wrapper)Logistic regressionWomen: AUC = 0.713 (LR-CFS), 0.721 (LR-wrapper) Men: AUC = 0.637 (LR-CFS), 0.652 (LR-wrapper)NRInternal, 10-fold cross- validation
Kanegae et al. [32] 2018Age, sex, BMI, SBP, DBP, low-density lipoprotein cholesterol, uric acid, proteinuria, current smoking, alcohol intake, eating rate, DBP by age, and BMI by ageCox regressionC-statistic = 0.885 [0.865–0.903]Greenwood-Nam-D’Agostino χ2 statistic = 13.6)External validation
Chen et al. [52] 2016Men: Age, BMI, SBP, DBP, gamma-glutamyl transferase, fasting blood glucose, drinking, age x BMI, age x DBPWomen: Age, BMI, SBP, DBP, fasting blood glucose, total cholesterol, neutrophil granulocyte, drinking, smokingCox regressionDerivation: AUC = 0.761 [0.752–0.771] (men), 0.753 [0.741–0.765] (women) Validation: AUC = 0.760 [0.751–0.770] (men), 0.749 [0.737–0.761] (women)NRInternal, 10-fold cross-validation
Díaz-Gutiérrez et al. [28] 2019No smoking, moderate-to-high physical activity, Mediterranean diet adherence, healthy BMI, moderate alcohol intake, and no binge drinkingCox regressionNRNRNR
Wang et al. [53] 2018Age, sex, education, marriage, smoking, drinking, BMI, energy, carbo, fat, proteinMultistate Markov modelNRNRTemporal validation
Niiranen et al. [54] 2016Model 1: GRS Model 2: Model 1 + age + sex Model 3: Model 2 + smoking, diabetes, education, hyper-cholesterolemia, leisure-time exercise, and BMIMultiple linear and logistic regressionC-index = 0.731 (Model 1)C-index = 0.733 (Model 3)NRNR
Yeh et al. [55] 2001Age, DM, and fibrinogen concentration (Men)Age and APTT (activated partial thromboplastin time) (Women)Cox regressionNRNRNR
Syllos et al. [21] 2020Age, sex, educational level, parental history of hypertension, leisure-time physical activity, BMI, neck circumference, smoking, SBP, DBPLogistic regressionAUC = 0.830 [0.810–0.849]H-L Chi-square = 8.22, p = 0.41Internal, split sample 6:4 ratio
Wang et al. [27] 2020Age, parental hypertension, SBP, DBP, BMI, and age by BMILogistic regressionC-index = 0.795 [0.7733–0.810] (Training set), C-index = 0.7914 [0.773–0.809] (Testing set)H–L Chi-square = 7.747, P = 0.459 (Training set)H–L Chi-square = 14.366, P = 0.073 (Testing set)Internal, Bootstrap validation
Xu et al. [56] 2019M1 Model: Age, SBP, DBP, hypertension parental history, WC, interaction item of age with WC, and interaction item of age with DBP W1 Model: Age, SBP, DBP, WC, fruit and vegetable intake, hypertension parental history, interaction item of age with WC, and interaction of age with DBP were included in W1 modelCox regressionTesting Set Men: AUC = 0.771 [0.750–0.791] (M1)Testing Set Women: AUC = 0.765 [0.746–0.783] (W1), 0.764 [0.746–0.783] (W2)Testing Set Men: Modified Nam-D’Agostino test Chi-square = 6.305, p = 0.708 (M1) Testing Set women: Modified Nam-D’Agostino test Chi-square = 6.783, p = 0.147(W1); 7.404, p = 0.115 (W2)Internal, 10-fold cross-validation in training data and external in the testing data
Kadomatsu et al. [26] 2019Age, sex, BMI, current smoking habit, ethanol consumption, presence of DM, parental hypertension history, SBP, DBPLogistic regressionAUC = 0.826 [0.804–0.848] (Entire cohort validation) Median AUC = 0.83 [0.828–0.832] (Cross-validation)H–L Chi-square = 7.06, p = 0.53, (Entire cohort validation); H–L Chi-square = 12.2 (Cross-validation)Internal, split-sample cross-validation 6:4 ratio
Wang et al. [57] 2015Exercise, diabetes, hyperlipemia, age, marriage, education, income, weight, height, sex, smoke, drinkLogistic regressionAccuracy, sensitivity, specificity, and AUC. AUC = 0.74±0.001 (logistic), Accuracy = 71.96% (logistic)NRInternal, split sample 7:3 ratio
Muntner et al. [58] 2010SBP-alone model (7 SBP categories)Age-specific categories of DBP model (20 categories)Repeated-measures Poisson regression modelSBP model: C-statistic = 0.768 [0.751–0.785] (1.6 years follow-up), 0.773 [0.775–0.791] (4.8 years follow-up); Age-specific DBP Model: C-statistic = 0.699 [0.681–0.717] (1.6 years follow-up), 0.691 [0.671–0.711] (4.8 years follow-up)NR NR
Ture et al. [59] 2005Age, sex, family history of hypertension, smoking habits, lipoprotein (a), triglyceride, uric acid, total cholesterol, and BMILogistic regression, Flexible discriminant analysis, multivariate additive regression splines (degree 1), multivariate additive regression splines (degree 2)Sensitivity, specificity, and predictive rate (PR)NRInternal, split sample 3:1 ratio
Yamakado et al. [60] 2015PFAA Index 1: Leucine, alanine, tyrosine, asparagine, tryptophan, and glycine; PFAA Index 2: Isoleucine, alanine, tyrosine, phenylalanine, methionine, and histidineLogistic regressionNRNRInternal, LOOCV and validation in a cohort dataset
Qi et al. [61] 2014rs17030613, rs16849225, rs1173766, rs11066280, rs35444, rs880315, rs16998073, rs11191548, rs17249754Logistic regressionNRNRNR
Lu et al. [62] 2015Model1: GRS+ (age, sex, and BMI); Model2: GRS +Model1 + smoking, drinking, pulse rate, and education; Model3: GRS+ Model2 + SBP and DBPLogistic regression and Cox regressionModel1: C-statistic = 0.650 [0.637–0.663] (without GRS), 0.655 [0.642–0.668] (with GRS) Model 2: C-statistic = 0.683 [0.670–0.695] (without GRS), 0.687 [0.675–0.700] (with GRS) Model 3: C-statistic = 0.774 [0.763–0.785] (without GRS), 0.777 [0.766–0.787] (with GRS)NRNR
Zhang et al. [63] 2015Five latent factors extracted from 11 biomarkers (BMI, SBP, DBP, FBG, TG, HDL-C, Hb, HCT, WBC, LC, NGC): inflammatory factor, blood viscidity factor, insulin resistance factor, blood pressure factor, lipid resistance factor, and ageCox regressionDerivation cohort: AUC = 0.755 [0.746–0.763] (men), AUC = 0.801 [0.792–0.810] (women) Validation cohort: AUC = 0.755 [0.746–0.763] (men), AUC = 0.800 [0.791–0.810] (women)NRInternal, 10-fold cross- validation

NR, not reported; SBP, systolic blood pressure; DBP, diastolic blood pressure; BP, blood pressure; BMI, body mass index; CVD, cardiovascular disease; HDL, high-density lipoprotein; WC, waist circumference; DM, diabetes mellitus; WHR, waist to hip ratio; WHtR, waist to height ratio; ISH, isolated systolic hypertension; IDH, isolated diastolic hypertension; AUC, area under the curve; AROC, area under the receiver operating characteristic curve; LR, logistic regression; GEE, Generalized estimating equations; LOOCV, leave-one-out cross-validation: HL, Hosmer Lemeshow; GRS, genetic risk score; SNP, single-nucleotide polymorphism; CFS, correlation-based feature subset selection; FBG, fasting blood glucose; TG, triglycerides; HDL-C, high-density lipoprotein cholesterol; Hb, hemoglobin; HCT, hematocrit; WBC, white blood cell count; LC, lymphocyte count; NGC, neutrophil granulocyte count

Fig 2

Conventional risk factors considered by traditional regression-based models (A) and by machine learning-based models (B).

Conventional risk factors considered by traditional regression-based models (A) and by machine learning-based models (B). NR, not reported; SBP, systolic blood pressure; DBP, diastolic blood pressure; BP, blood pressure; ISH, isolated systolic hypertension; IDH, isolated diastolic hypertension NR, not reported; SBP, systolic blood pressure; DBP, diastolic blood pressure; BP, blood pressure; BMI, body mass index; CVD, cardiovascular disease; HDL, high-density lipoprotein; WC, waist circumference; DM, diabetes mellitus; WHR, waist to hip ratio; WHtR, waist to height ratio; ISH, isolated systolic hypertension; IDH, isolated diastolic hypertension; AUC, area under the curve; AROC, area under the receiver operating characteristic curve; LR, logistic regression; GEE, Generalized estimating equations; LOOCV, leave-one-out cross-validation: HL, Hosmer Lemeshow; GRS, genetic risk score; SNP, single-nucleotide polymorphism; CFS, correlation-based feature subset selection; FBG, fasting blood glucose; TG, triglycerides; HDL-C, high-density lipoprotein cholesterol; Hb, hemoglobin; HCT, hematocrit; WBC, white blood cell count; LC, lymphocyte count; NGC, neutrophil granulocyte count

Meta-analysis of traditional regression-based models

The overall pooled C-statistics of the traditional regression-based models was 0.75 [0.73–0.77] with high heterogeneity in the discriminative performance of these models (I2 = 99.3, Cochran Q-statistic p < 0.001) (Fig 3). Stratified pooled results by modeling type showed pooled C-statistics were 0.73 [0.69–0.77], 0.77 [0.74–0.81], 0.73 [0.69–0.78], and 0.77 [0.75–0.79] for Cox, logistic, repeated Poisson, and Weibull respectively (Fig 3). The heterogeneity was still observed to be high within the different types of models (Fig 3). The 95% approximate prediction interval for the overall C-statistics was from 0.63 to 0.84.
Fig 3

Forest plot of traditional regression-based models with 95% prediction interval.

To explore possible sources of heterogeneity in the overall pooled C-statistics, we performed a meta-regression. We initially considered the following potential sources of heterogeneity: the definition of hypertension used (the cut-off level used to define hypertension), sex of the participants in included studies (categorized as female-only, male-only, and both male and female), age of the participants (study participants below average age versus above average age), number of risk factors considered in the model (below median versus above median), sample size considered in the model (below median versus above median), and ethnicity of the study participants (Whites versus Asians). However, we excluded the definition of hypertension as a heterogeneity source, as all studies except one used the same definition for hypertension. Meta-regression identified the participants’ sex, that is, being male compared to female (p = 0.044), participants’ age (p = 0.011), and the number of risk factors considered in the model (p = 0.001) as potential sources of high heterogeneity in the C-statistic. Sex of the participants’ when both male and female compared to female-only (p = 0.351), sample size considered in the model (p = 0.395), and ethnicity of the study participants (p = 0.899) were not identified as a statistically significant source of observed heterogeneity in the C-statistic of these models.

Critical appraisal of traditional regression-based models

We assessed study quality using the PROBAST checklist. A detailed assessment of the risk of bias (ROB) and applicability is presented in S2 Table and Fig 4. Overall, ROB was “low” in 19 studies, “high” in 5 studies, and “unclear” in 10 studies. Overall applicability was “low concern” in 12 studies, “high concern” in 21 studies, and “unclear concern” in 1 study. Within the ROB domains, the “low” risk of bias was observed in most of the domains except the “analysis” domain, where a large portion of studies (more than 30%) was “unclear” (Fig 4). Similarly, within the applicability domains, the “participants” domain seems to be a concern, as a large portion of studies (more than 30%) were at “high concern” or “unclear concern” (Fig 4). We also presented the different PROBAST signaling questions’ distribution of responses by the various studies in S1 and S2 Figs.
Fig 4

Graphical summary presenting the percentage of hypertension risk prediction studies rated by level of concern, risk of bias (ROB), and applicability for each domain.

Study characteristics of machine learning-based models

Study characteristics of machine learning-based models are presented in Table 3. A total of 1,211,093 participants were used to develop 42 machine learning-based models in 20 studies. Models were primarily developed either in white Caucasian or Asian populations. The number of risk factors/features considered to create the model ranged from 2 to 169, with a median of 7 risk factors per model. Age was the most common risk factor considered in 25 models, followed by sex/gender (8 models), BMI (7 models), DBP (6 models), smoking (6 models), and parental history of hypertension (6 models). The distribution of the conventional risk factors considered in machine learning models is presented in Fig 2B. Hypertension was predominantly defined using SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg, or antihypertensive medication. Artificial neural network (ANN) was the most common method used to develop the models. Different studies reported different performance measures, and accuracy and AUC/C-statistic were the two most commonly reported measures. Most of the studies did not report calibration measures. In studies that reported discrimination, the AUC (or C-statistic) values range from 0.64 to 0.93.
Table 3

Information about existing hypertension prediction models developed using machine learning algorithms from selected studies.

StudyData LocationSample SizeRisk Factors IncludedOutcome ConsideredDefinition of Outcome PredictedModeling Method UsedPerformance Measure
Falk CT [64] 2003USA300 records each for training and validatingSeven input values: sex; age; total cholesterol; fasting glucose; fasting HDL; fasting triglycerides; body mass index (BMI)High blood pressureSBP > 140 mm Hg or DBP > 90 mm HgTwo neural network programs: NNdriver and SNNSClassification success rate. Training: 91%-98%, (Strategy 1), 70%-87% (Strategy 2); Validation: 59% (Strategy 1), 63% (Strategy 2)
Farran et al. [65] 2013Kuwait10,632 (6759 hypertensive and 3873 non-hypertensive)BMI, age, ethnicity, and diagnosis for diabetesIncident hypertension, type 2 diabetes, and comorbidityNRLogistic regression (LR), k-nearest neighbors, support vector machines, and multifactor dimensionality reduction (MDR)Classification accuracy: 90% (hypertension)
Huang et al. [35] 2010ChinaTraining: 2438, Validation: 616High educational level, predominantly sedentary work, positive family history of HTN, overweight, dysarteriotony, alcohol intake, salty diet, more vegetable and fruit intake, meat consumption, and regular physical exerciseHypertensionAverage SBP or DBP > 139 mmHg or > 89 mmHg, respectivelyLogistic regression model (LRM) and artificial neural network (ANN) model (back-propagated delta rule networks)AUC: 0.900 ± 0.014 (ANN model)AUC: 0.732 ± 0.026 (LRM)
Kwong et al. [66] 2018NR498Age, BMI, exercise level, alcohol consumption level, smoking status, stress level, and salt intake levelSystolic blood pressure (SBP)BP readings > 140 mmHgTwo artificial neural networks (ANN): Back-propagation (BP) neural network and radial basis function (RBF) neural network validate the prediction systemAverage Accuracy, BP ANN: 94.28% (male), 93.74% (female)RBF ANN: 91.06% (male), 90.44% (female)
Polak et al. [67] 2008USA159,989 recordsHigh blood cholesterol, number of cigarettes smoked now, age, weight, height, sexHypertensionNRArtificial neural network (ANN): Around 250 architectures of backpropagation (BP) and fuzzy networksClassification rate and AUROC, different values for different Nets architecture
Priyadarshini et al. [68] 2018USANRSBP, DBP, total cholesterol (TC), high-density lipoprotein (HDL), low-density lipoprotein (LDL), plasma glucose concentration (PGC), and heart rate (HR)Hypertension attackDBP or SBP > 90 mm Hg or > 120 mm Hg, respectively, for at least two measuring instancesDeep neural network modelConfusion/performance matrix formed out of four evaluating parameters: accuracy 88%, precision 92%, recall 82%, and F1 score 76% (average value over 20 iterations)
Sakr et al. [36] 2018USA23,095Age, METS, resting systolic blood pressure, peak diastolic blood pressure, resting diastolic blood pressure, HX coronary artery disease, the reason for the test, history of diabetes, percentage HR achieved, race, history of hyperlipidemia, Aspirin use, hypertension responseHypertensionNRSix machine learning techniques: LogitBoost (LB), Bayesian network classifier (BN), locally weighted naïve Bayes (LWB), artificial neural network (ANN), support vector machine (SVM), and random tree forest (RTF)AUC, F-Score, Sensitivity, Specificity, Precision, and RMSE. AUC (0.93), F-Score (86.70%), Sensitivity (69,96%) and Specificity (91.71%) for RTF model in 10-fold cross-validation AUC (0.88), Sensitivity (74.30%), Precision (73.50%), and F-Score (73.90%) for RTF model in holdout method
Tayefi et al. [69] 2017Iran9078Age, gender, BMI, marital status, level of education, occupation status, depression and anxiety status, physical activity level, smoking status, LDL, triglyceride, total cholesterol, fasting blood glucose, uric acid, and hs-CRP in Model 1Age, gender, white blood cell, red blood cell, hemoglobin, hematocrit, mean corpuscular volume, mean corpuscular hemoglobin, platelets, red cell distribution width and platelet distribution width in Model 2HypertensionSBP of 140 mm Hg, DBP of 90 mm Hg, and/or current use of antihypertensive drugsDecision treeAccuracy, sensitivity, specificity, and area under the ROC curve (AUC): For Model 1, the values are 73%, 63%, 77% and 0.72, respectively, and for Model 2 were 70%, 61%, 74% and 0.68, respectively
Wu et al. [70] 2015USA75 females and 165 malesAge, gender, serum cholesterol, fasting blood sugar and electrocardiographic signal, heart rateSystolic blood pressureSBP and DBP > 140 mm Hg and 90 mm Hg, respectivelyTwo neural network algorithms: back-propagation neural network and radial basis function networkThe absolute difference (error) between the real value and predicted values
Wu et al. [71] 2016NR498Age, BMI, gender, exercise level, alcohol consumption, stress level, salt intake level, smoke status, cholesterol, and blood glucoseSystolic blood pressureSBP > 140 mm HgTwo artificial neural networks: back-propagation neural network and radial basis function neural networkThe average prediction errors (absolute difference between the predicted value and measured value): 51.9% for men and 52.5% for women (backpropagation neural network)51.8% for men and 49.9% for women (radial basis function network)
Ye et al. [37] 2018USA823,627 (training cohort/retrospective cohort), 680,810 (validation cohort/prospective cohort)Total 169 features: 2 demographic features, 14 socioeconomic characteristics, 30 diagnostic diseases, 6 laboratory tests, 98 medication prescriptions, and 19 clinical utilization measuresIncident essential hypertensionICD, 9th Revision, Clinical Modification (ICD-9-CM) diagnosis codes from category 401A supervised machine learning and data mining tool, XGBoostAUC = 0.917 (retrospective cohort), AUC = 0.870 (prospective cohort)
Zhang et al. [72] 2018NRA total of 15,628,501 sets of valid characteristic attributes dataSeven input features: right atrium (AVR), left atrium (AVL), anterior atrium (AVF), photoplethysmography (PPG), oxygen saturation (SPO2), pulse transit time (PTT), heart rate (HR)Blood pressureNRCART (classification and regression tree) modelFour evaluation indexes: accuracy rate, root mean square error (RMSE), deviation rate, and the Theil inequality coefficient (TIC)
Völzke et al. [31] 2013GermanyTraining set: 803 Validation set: 802External validation cohort: 2887Age, mean arterial pressure, rs16998073, serum glucose, and urinary albuminconcentrations, the interaction between age and serum glucose, interaction between rs16998073 and urinary albumin concentrationsIncident hypertensionSBP ≥ 140 mmHg and DBP ≥ 90 mmHgBayesian networkTraining set: AUC = 0.78 [0.74–0.82], Validation set: AUC = 0.79 [0.75–0.83], External validation set: AUC = 0.77 [0.74–0.80];Training set: HL Chi-square = 11.82 (p = 0.16), Validation set: HL Chi-square = 11.65 (p = 0.17), External validation set: H-L Chi-square = 40.6(p < 0.01)
Lee et al. [51] 2014Korea12,789Women: Height, age, neckC, axillaryC, ribC, waistC, pelvicC, rib_hip, waist_hip, pelvic_hip, rib_pelvic, axillary_rib, chest_rib, axillary_chest, forehead_neck (CFS), height, age, foreheadC, neckC, hipC, axillary_hip, axillary_pelvic, chest_pelvic, chest_rib (NB-wrapper)Men: Age, foreheadC, neckC, axillaryC, chestC, RibC, waistC, pelvicC, hipC, rib_hip, waist_hip, rib_pelvic, waist_pelvic, chest_waist, forehead_rib, chest_rib, axillary_chest, forehead_neck (CFS), height, age, foreheadC, neckC, axillaryC, hipC, rib_hip, pelvic_hip, neck_pelvic, waist_pelvic, chest_waist, chest_rib, neck_chest, forehead_neck (NB-wrapper)Hypertension and hypotensionSBP ≥ 140 mmHg and/or DBP ≥ 90 mmHg or physician-diagnosed hypertensionNaive Bayes algorithm (NB)Women: AUC = 0.696 (NB-CFS), 0.713 (NB-wrapper)Men: AUC = 0.64 (NB-CFS), 0.646 (NB-wrapper)
Xu et al. [56] 2019China4796M1 Model: Age, SBP, DBP, hypertension parental history, WC, interaction item of age with WC, and interaction item of age with DBPW1 Model: Age, SBP, DBP, WC, fruit and vegetable intake, hypertension parental history, interaction item of age with WC, and interaction of age with DBPHypertensionSBP ≥ 140 mm Hg and/or DBP ≥ 90 mm Hg and/or a diagnosis of hypertension by a physician and currently receiving anti-hypertension treatmentArtificial neural network (ANN), naive Bayes classifier (NBC), and classification and regression tree (CART)Testing Set Men: AUC = 0.773 [0.752–0.793] (ANN), 0.760 [0.738–0.781] (NBC), 0.722 [0.699–0.743] (CART)Testing Set Women: AUC = 0.756 [0.737–0.775] (ANN), 0.761 [0.742–0.779] (NBC), 0.698 [0.677–0.717] (CART)Testing Set Men: ModifiedNam-D’Agostino test Chi-square = 29.274, p = 0.0006 (ANN); 82.269, p < 0.00001 (NBC); 5.249, p = 0.072 (CART)Testing Set women: ModifiedNam-D’Agostino test Chi-square = 4.744, p = 0.314 (ANN); 189.754, p < 0.00001 (NBC); 19.733, p = 0.00005 (CART)
Wang et al. [57] 2015USA308,711Exercise, diabetes, hyperlipemia, age, marriage, education, income, weight, height, sex, smoke, drinkHypertensionNRMulti-layer perception neural networkAccuracy, sensitivity, specificity, and AUC. Average AUC = 0.77 with h vary from 8 to 11 (neural network); Accuracy = 72% (neural network)
Ture et al. [59] 2005Turkey694Age, sex, family history of hypertension, smoking habits, lipoprotein (a), triglyceride, uric acid, total cholesterol, and BMIEssential hypertensionThe average of 3 or more DBP measurements on at least 3 subsequent visits is ≥ 90 mmHg, or when the average of multiple SBP readings on 3 or more subsequent visits is consistently ≥ 140 mmHgThree decision trees (Chi-squared automatic interaction detector. Classification and regression tree, quick, unbiased, efficient statistical tree); two neural networks (multi-layer perceptron, radial basis function)Sensitivity, specificity, and predictive rate (PR). Values not reported.
Zhao et al. [73] 2008China/AsiansTotal: 4759 (2411 hypertensive and 2,348 age-matched and sex-matched healthy controls)MDR Model: 4-locus model consisted of the SNP KCNMB1-rs11739136, RGS2-rs34717272, PRKG1-rs1881597, and MYLK-rs36025624; CART Model: RGS2, PRKG1, KCNMB1, and MYLKHypertensionCHECKAverage SBP ≥ 150 mm Hg, an average DBP ≥ 95 mm Hg, or current use of antihypertensive medicationMultifactor-dimensionality reduction (MDR) and classification and regression trees (CART)MDR Model: Accuracy = 52.98%, cross-validation consistency = 9.7
Wang et al. [57] 2014China/Asians1009 hypertensive patients and 756 normotensive controlsGenesHypertensionMean SBP ≥ 140 mmHg and/or DBP ≥ 90 mmHg on two occasions and/or the current usage of antihypertensive drug treatmentMultifactor dimensionality reduction (MDR) modelThe best MDR model testing accuracy = 0.6331, cross-validation consistency = 10
Zhao et al. [74] 2014China/Asians1009 hypertensive patients and 756 normotensive controlsThe best MDR model included rs5804 and BMIHypertensionMean SBP of at least 140 mmHg or a mean DBP of at least 90 mmHg or the current intake of antihypertensive drugsMultifactor dimensionality reduction (MDR) modelThe best MDR model: testing accuracy of 0.7309 and a maximum cross-validation consistency of 10 (P < 0.001)

ICD, international classification of diseases

ICD, international classification of diseases

Meta-analysis of machine learning-based models

The overall pooled C-statistics of the machine learning-based models was 0.76 [0.72–0.79] with high heterogeneity in the discriminative performance of these models (I2 = 99.9, Cochran Q-statistic p < 0.001) (Fig 5). Like traditional regression-based models, we did not perform stratified pooled results by modeling type due to diversity in the modeling method. The 95% approximate prediction interval for the overall C-statistics was from 0.63 to 0.84 (Fig 5).
Fig 5

Forest plot of machine regression-based models with 95% prediction interval.

We explored possible sources of heterogeneity in the overall pooled C-statistics through meta-regression as before. However, meta-regression did not identify any of age of the participants (p = 0.358), the number of risk factors considered in the model (p = 0.812), sex of the participants, that is being male compared to female (p = 0.886) and both male and female compared to female-only (p = 0.787), sample size considered in the model (p = 0.577), or ethnicity of the study participants (p = 0.326) as the potential source of high heterogeneity in the C-statistic.

Study characteristics of externally validated models

Only four models [22, 30–32] were found to be externally validated in a different population. Detailed characteristics of the studies that validated these four models are presented in S3 Table. The Framingham hypertension risk model (FHRS) is the only validated model in more than one external population. The FHRS [22] model was validated by eight different studies in diverse populations of 122,348 participants. Study participants had an age range of 18 to 84 years with follow-up time (mean/median/total) from 1.6 years to 25 years. Almost all studies reported performance measures of the FHRS. The Hosmer-Lemeshow test was used to report calibration, while the C-statistic (or AUC) was used to report discrimination. The values of the reported C-statistic ranged from 0.54 to 0.84. Models by Lim et al. [30], Völzke et al. [31], and Kanegae et al. [32] were validated only once in an external population by the same authors. Within these three models, performances were best for the model by Kanegae et al. [32], with a C-statistic of 0.85 [0.76–0.91].

Meta-analysis of externally validated models

The pooled C-statistic of the FHRS [22] model was 0.75 [0.68–0.80] with high heterogeneity in the discriminative performance of this model (I2 = 99.6, Cochran Q-statistic p < 0.001) (S3 Fig). The 95% approximate prediction interval for the C-statistic in the FHRS [22] was from 0.47 to 0.91 (S3 Fig). As the other three models were externally validated only once, pooling their performance measure was irrelevant. We explored possible sources of heterogeneity in the pooled C-statistics through meta-regression, and only the ethnicity (Whites versus Asians) of the study participants (p = 0.044) was identified as a source of high heterogeneity in the C-statistic of the FHRS model [22].

Models developed using genetic risk factors/biomarkers

Genetic risk factors/biomarkers often contribute significantly to developing hypertension, and models were developed considering both conventional risk factors and biomarkers. In addition, there were models where biomarkers were used primarily in model building. Information about models developed using biomarkers (e.g., genetic risk scores) is presented in S4 Table. There were 11 studies where genetic risk factors/biomarkers were used in model building. Biomarkers are often considered very important for increasing the predictive performance of models. However, the pooled predictive performance (C-statistic) of the models that considered biomarkers primarily was 0.76 [0.71–0.80] (S4 Fig) and did not show an overall improvement in the models’ predictive performance. Including genetic factors/biomarkers in the model has some drawbacks. Because information on those biomarkers is frequently unavailable and interpreting the models becomes difficult, the models become less suitable for daily clinical practice.

Discussion

Many hypertension risk prediction models with reasonable predictive performance were identified in this systematic review, but only a few had external validation. Bias and applicability were noted as major concerns in many studies. Overall, there was little difference in the predictive performance of traditional statistical and machine learning models. Our findings are expanded on in the sections that follow. The models were developed mostly in Caucasian or Asian populations. Because certain ethnic groups are more prone to hypertension (e.g., people of African descent [33]), research should include a diverse range of patients to create hypertension risk prediction models. Most of the traditionally developed models considered conventional risk factors for hypertension, which are readily available in clinical practice. Some models also used genetic risk factors, although the inclusion of genetic risk factors into the model did not improve the overall predictive performance of the models. The pooled analysis identified the overall predictive performance of the traditional regression-based models was good but with high heterogeneity. Stratified analysis by modeling methodology (e.g., logistic, Cox) within traditional regression-based models did not show much difference in predictive performance, and heterogeneity was still observed within the modeling methodology. The traditional models we identified in our search were mostly internally validated, often considered not enough for models’ generalizability [34]. The FHRS [22] was the only model that had multiple external validations and good/acceptable pooled predictive performance. However, because the FHRS [22] showed high heterogeneity in its predictive performance, with ethnicity serving as a source of heterogeneity, and the model was built predominantly in a White population, we must proceed with caution when applying it to a completely different population. Models that have only single, or no validation need external validation, preferably by a different group of investigators, to guarantee the model’s generalizability to a different population. Only a few traditional models were converted into risk score after their development. Presenting the risk derived from the model through scoring instead of a complex mathematical formula may facilitate the use of prediction models and subsequently improve the uptake of prediction models in clinical practice. The risk of bias (ROB) was "high" or "unclear" in a large portion of traditional model studies. This is primarily because many studies failed to meet the criteria in the "analysis" domain of ROB. In many studies, the applicability of the models was rated as "high concern" or "unclear concern" due to a failure to properly fulfil the "participants" criteria. Several models were developed in a specific population, making the models less applicable to the general adult population. Since machine learning tools are more recent, advanced, and have a reputation for producing more accurate predictive performance, we assumed that models developed with these tools would outperform traditional regression-based models. However, we did not notice much difference in predictive performance between these two types of models. A few machine learning-based models (e.g., models by Huang et al. [35], Sakr et al. [36], and Ye et al. [37]) showed excellent discriminative performance; however, none of these models has ever been externally validated in an entirely different new population. In fact, none of the machine learning-based models have been externally validated. Consequently, the performance of those models in a new setting/population is quite uncertain. We also noticed high heterogeneity in the predictive performance (C-statistic) of machine learning models. Meta-regression using potential sources of heterogeneity failed to identify the real source of heterogeneity. One possible explanation is a difference in the methodology used to develop the machine learning-based models. Due to the various methods considered in different models, we were unable to investigate this potential source. We did not notice higher expected variability in machine learning-based models’ future predictive performance compared to traditional regression-based models, as the 95% prediction interval for machine learning-based models was similar to traditional regression-based models. We did not find any studies in this review that assessed the impact of adopting hypertension risk prediction models in clinical settings. Ideally, a prediction model, regardless of its development, should have an impact study to assess whether it improves clinical decision-making and patient health outcomes [5, 38]. There were two previous reviews on a similar topic where hypertension risk prediction models were identified through a systematic search and described their characteristics. Our review is different from previous studies and contributes to information on the prediction of hypertension risk and the identification of associated risk factors in the following ways: 1) we synthesized performance of the prediction models through meta-analysis and explored potential sources of heterogeneity; 2) we compared the performance of the prediction models developed using traditional statistical regression-based models and more recent machine learning-based models; 3) we provided a thorough evaluation of the quality of the studies among traditionally developed regression-based models; and 4) we described several additional models that have recently been derived. One of our study’s strengths is the extent of the systematic search, which includes four different databases, grey literature, and extensive use of the reference lists of the identified studies. To the best of our knowledge, this is the first study where a meta-analysis of predictive performance, together with assessment of heterogeneity, comparison of the predictive performance of traditional regression based-models and machine learning-based models, and a detailed critical appraisal of studies in hypertension risk prediction models has been performed. Nevertheless, our study also has limitations. We excluded non-English and non-French publications. While it is widely perceived that the English language is the primary language of science, the choice of scientific results in a particular language can incorporate language bias and may lead to incorrect conclusions [39]. We were only able to use C-statistics to compare the model performance, which could be insensitive to distinguish a model’s ability to correctly stratify patients into clinically relevant risk groups [39, 40]. Calibration was quantified by different measures, and different studies often reported different calibration measures. This led to difficulty in synthesizing calibration measures through meta-analysis. A meta-analysis of calibration measures (e.g., O/E ratio) along with C-statistics could provide a comprehensive summary of the performance of these models [19]. Failing to assess publication bias amongst the studies is another potential limitation of this study. Recent guidelines [19] did not emphasize the need to assess publication bias for prediction model performance, which encouraged us not to do so. Although studies have considered publication bias in a similar scenario before, we believe existing traditional publication bias assessment tools (e.g., funnel plot, Egger’s test, Begg’s test) are more appropriate for studies assessing statistically significant results (e.g., randomized controlled trial (RCT)) than studies assessing predictive performance (e.g., C-statistic) of the prognostic models. Instead, we assessed ROB using the PROBAST checklist. We also could not appraise studies that use machine learning algorithms to predict hypertension. Although most of the PROBAST signaling questions also apply to appraise machine learning algorithms, additional signaling questions are recommended to add due to differences in data analysis methods for machine learning algorithms and regression-based models [14, 15]. Machine learning algorithms use different variable selection strategies, different estimation techniques for variable–outcome estimations, and different ways to adjust for overfitting [14, 15]. When additional questions are added to the PROBAST, these questions need to be appropriately phrased, and specific guidance on assessing these signaling questions also needs to be provided [14, 15]. Considering these additional works, we refrain from appraising studies considered machine learning algorithms. Finally, despite our attempt to capture potential sources of heterogeneity in our study, we asked readers to be cautious while interpreting our findings as there may be a potential bias in our findings due to a limited number of studies included in the analysis and the study’s failure to incorporate additional potential sources of bias in the analysis. In summary, we attempted to provide a comprehensive evaluation of hypertension risk prediction models. We identified many models with acceptable-to-good predictive performance. We did not notice significant differences in the predictive performance of traditional regression-based models and machine learning-based models. Including genetic risk factors/biomarkers also did not show much improvement in the models’ predictive performance. The quality of the studies was reasonable, with areas where further improvement is needed. Only a few of the multiple models developed had been externally validated, which is a concern. Also, there is a lack of impact studies. Models with external validation and impact studies are required to implement a prediction model in a clinical practice guideline. A model with accurate prediction is not beneficial if it is not generalizable to a different population or improves clinical decision-making and patient health outcomes.

PRISMA 2020 checklist.

(DOCX) Click here for additional data file.

The number of PROBAST criteria satisfied by different studies.

(DOC) Click here for additional data file.

Response to different signaling questions by the number of studies.

(DOC) Click here for additional data file.

Forest plot of externally validated models with 95% prediction interval.

(DOC) Click here for additional data file.

Forest plot of models primarily developed using genetic risk factors/biomarkers with a 95% prediction interval.

(DOC) Click here for additional data file.

Keywords used to search in MEDLINE.

(DOC) Click here for additional data file.

Study quality assessment using PROBAST.

(DOC) Click here for additional data file.

Information about external validation studies of existing traditional hypertension prediction models from selected studies.

(DOC) Click here for additional data file.

Information about existing hypertension prediction models developed using biomarkers (genetic risk score) from the selected studies.

(DOC) Click here for additional data file. 8 Nov 2021
PONE-D-21-31564
Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis
PLOS ONE Dear Dr. Turin, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Dec 23 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Antonio Palazón-Bru, PhD Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Thank you for stating the following financial disclosure: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” At this time, please address the following queries: a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution. b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” c) If any authors received a salary from any of your funders, please state which authors and which funders. d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.” Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 3. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability. Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. 4. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide. 5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: No ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors compared the predictive performance of two types of hypertension risk prediction models: those developed using traditional regression-based and those using machine learning approaches. They searched the MEDLINE, EMBASE, Web of Science, Scopus, and the grey literature for studies predicting the risk of hypertension among the general adult population. They used the C-statistic, and a random-effects meta-analysis was used to obtain pooled estimates from the individual studies The potential sources of heterogeneity was assessed using meta-regression, and study quality was assessed using the PROBAST (Prediction model Risk Of Bias ASsessment Tool) checklist. They selected 52 articles for systematic review and 32 for meta-analysis out of the 14,778 citations that they retrieved. They observed modest and similar overall pooled C-statistics of 0.75 [0.73 – 0.77] for the traditional regression-based models and 0.76 [0.72 – 0.79] for the machine learning-based models. There was high heterogeneity in the C-statistic in both methods. The age (p = 0.011), and sex (p = 0.044) of the participants and the number of risk factors considered in the model (p = 0.001) were identified as sources of heterogeneity in traditional regression-based models. The authors concluded that only a few models were externally validated, that the risk of bias and applicability was a concern in many studies that many models with acceptable-to-good predictive performance were identified that overall discrimination was similar between models derived from traditional regression analysis and machine learning methods and that external validation and of the hypertension risk prediction model in clinical practice are required. The authors may wish to consider the following. 1. Selecting a small number of studies may have led to biased conclusions. 2. The variability in the duration of follow-up time (1.6 years to 30 years), the age of the participants (15 to 90 years), SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg, or SBP ≥ 130 mm Hg, DBP ≥ 80 mm Hg, and or use of antihypertensive medication may have led to biased conclusions. 3. In addition, the variability on the geographic region, time, or gender of the study participants may have led to biased conclusions. 4. The authors may wish to expand the limitations section of the Discussion in page 18 to include items 1, 2 and 3 above. 5. Would the authors agree to include the last sentence of the manuscript “we attempted to provide a comprehensive evaluation of hypertension risk prediction models” in the Abstract? Reviewer #2: My review is attached as a document for ease of reading., but I also include it here: Review: Chowdhury et al “Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis” Overview In this paper Chowdhury et al provide a systematic review and meta-analysis comparing prediction models for the development of hypertension in the general population derived using traditional regression-based and machine learning approaches. Meta-analysis was only possible for measures of discrimination. Overall the pooled c-statistics on meta-analysis are similar and of moderate-good performance between traditional regression-based and machine learning derived models. High heterogeneity was found, with sources identified for traditional regression-based models through meta-regression. Only one model has been extensively externally validated (Framingham Hypertension risk model) but it showed significant heterogeneity in meta-analysis. Performance of risk models for hypertension have only been appropriately checked in Asian and Caucasian populations and clinical implementation has not been assessed. Overall impression I would like to congratulate the reviewers on an extremely thorough and methodologically sound systematic review and meta-analysis. My main concerns relate to the structure and writing of the discussion section, and the presentation the table. Major issues • The aims of the study are clearly delineated in the introduction (point 1-4). However I do not feel the structure of the discussion follows these aims or highlights the most salient findings of the analysis. Furthermore in my opinion the discussion section is too long. It would be better presented: o Major findings of the study (3-4 points) o Discussion of previous literature and how this differs o Future areas for research / gaps in knowledge o Limitations o Final conclusion • The presentation of table 1 is extremely difficult to follow. The presentation of so many columns means that some of the entries for each study take up an entire page. It would be better to break this up into at least 2/3 tables e.g. between study population characteristics, model development characteristics/performance, variables used in model; and all these tables do not need to be in the main file (eg Himmelreich et al -> https://academic.oup.com/europace/article/22/5/684/5721485) • Why are traditional regression model study characteristics included in main paper but not machine learning counterparts. It would be better to present them more equally • There are wide prediction intervals suggesting significant heterogeneity. Have you considered a Bayesian approach for meta-analysis? Frequentist methods can produce prediction intervals with poor coverage when there is a mixture of study sizes (https://pubmed.ncbi.nlm.nih.gov/30032705/) Minor issues • I note some of the models for predicting hypertension use systolic blood pressure and diastolic blood pressure. Does this not appear ‘double-dipping’ to include a variable that may well be an outcome? Does this not require some comment? • Page 13 line 330 – please be more specific than ‘basically’ • Page 15 line 388 – I belive it should be ‘models’ • Page 17 line 446-447 does not make sense • Figure 1 – I believe the reasons for exclusion would be better ordered alphabetically or in descending number of records excluded ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: John B. Kostis Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. Submitted filename: Review.docx Click here for additional data file. 8 Mar 2022 Response to journal requirements and reviewers’ comments Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf RESPONSE: Thank you. We have revised our manuscript accordingly. 2. Thank you for stating the following financial disclosure: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” At this time, please address the following queries: a) Please clarify the sources of funding (financial or material support) for your study. List the grants or organizations that supported your study, including funding received from your institution. b) State what role the funders took in the study. If the funders had no role in your study, please state: “The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.” c) If any authors received a salary from any of your funders, please state which authors and which funders. d) If you did not receive any funding for this study, please state: “The authors received no specific funding for this work.” Please include your amended statements within your cover letter; we will change the online submission form on your behalf. RESPONSE: Thank you. None of the authors received any funding for this study. We have now stated, “The authors received no specific funding for this work” in our revised manuscript and in the cover letter. 3. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability. Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. RESPONSE: Thank you. Since our study is a systematic review and we did not use any primary data in our analysis, we have now revised our data availability statement as follows: “All relevant data are within the manuscript and its Supporting information files”. We have included this statement in our revised manuscript and in the cover letter. 4. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide. RESPONSE: Thank you. Yes, we would like to make changes to our Data Availability statement. Since our study is a systematic review and we did not use any primary data in our analysis, we have now revised our data availability statement as follows: “All relevant data are within the manuscript and its Supporting information files”. We have included this statement in our revised manuscript and in the cover letter. 5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. RESPONSE: Thank you. We have now included captions for Supporting Information files at the end of our manuscript. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly ________________________________________ 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ________________________________________ 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ________________________________________ 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: No ________________________________________ 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: COMMENT. The authors compared the predictive performance of two types of hypertension risk prediction models: those developed using traditional regression-based and those using machine learning approaches. They searched the MEDLINE, EMBASE, Web of Science, Scopus, and the grey literature for studies predicting the risk of hypertension among the general adult population. They used the C-statistic, and a random-effects meta-analysis was used to obtain pooled estimates from the individual studies The potential sources of heterogeneity was assessed using meta-regression, and study quality was assessed using the PROBAST (Prediction model Risk Of Bias ASsessment Tool) checklist. They selected 52 articles for systematic review and 32 for meta-analysis out of the 14,778 citations that they retrieved. They observed modest and similar overall pooled C-statistics of 0.75 [0.73 – 0.77] for the traditional regression-based models and 0.76 [0.72 – 0.79] for the machine learning-based models. There was high heterogeneity in the C-statistic in both methods. The age (p = 0.011), and sex (p = 0.044) of the participants and the number of risk factors considered in the model (p = 0.001) were identified as sources of heterogeneity in traditional regression-based models. The authors concluded that only a few models were externally validated, that the risk of bias and applicability was a concern in many studies that many models with acceptable-to-good predictive performance were identified that overall discrimination was similar between models derived from traditional regression analysis and machine learning methods and that external validation and of the hypertension risk prediction model in clinical practice are required. RESPONSE: Thank you so much for your excellent comment. COMMENT. The authors may wish to consider the following. 1. Selecting a small number of studies may have led to biased conclusions. 2. The variability in the duration of follow-up time (1.6 years to 30 years), the age of the participants (15 to 90 years), SBP ≥ 140 mm Hg, DBP ≥ 90 mm Hg, or SBP ≥ 130 mm Hg, DBP ≥ 80 mm Hg, and or use of antihypertensive medication may have led to biased conclusions. 3. In addition, the variability on the geographic region, time, or gender of the study participants may have led to biased conclusions. 4. The authors may wish to expand the limitations section of the Discussion in page 18 to include items 1, 2 and 3 above. RESPONSE: Thank you so much for your excellent comments. We agree with the reviewer that items 1, 2, and 3 could be potential sources of bias. However, we would like to point out here that we considered most of those listed items as potential sources of heterogeneity in C-statistics in our analysis. For example, age, gender (sex), the definition of hypertension used (the cut-off level used to define hypertension as the reviewer indicated), and ethnicity (which reflected the influence of geographic region) were considered as the potential sources of heterogeneity in the C-statistics in our analysis. However, we acknowledge that variations on these items may lead to biased conclusions in study findings, and we have included these as limitations in our revised manuscript. The following lines were added to the revised manuscript: “Finally, despite our attempt to capture potential sources of heterogeneity in our study, we asked readers to be cautious while interpreting our findings as there may be a potential bias in our findings due to a limited number of studies included in the analysis and the study's failure to incorporate additional potential sources of bias in the analysis.” Please see Page 18, lines 461-464 in the revised manuscript. COMMENT. 5. Would the authors agree to include the last sentence of the manuscript “we attempted to provide a comprehensive evaluation of hypertension risk prediction models” in the Abstract? RESPONSE: Thank you. We have included this sentence in the abstract. Please see Page 3, lines 73-74 in the revised manuscript. Reviewer #2: My review is attached as a document for ease of reading., but I also include it here: Review: Chowdhury et al “Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis” COMMENT. Overview In this paper Chowdhury et al provide a systematic review and meta-analysis comparing prediction models for the development of hypertension in the general population derived using traditional regression-based and machine learning approaches. Meta-analysis was only possible for measures of discrimination. Overall the pooled c-statistics on meta-analysis are similar and of moderate-good performance between traditional regression-based and machine learning derived models. High heterogeneity was found, with sources identified for traditional regression-based models through meta-regression. Only one model has been extensively externally validated (Framingham Hypertension risk model) but it showed significant heterogeneity in meta-analysis. Performance of risk models for hypertension have only been appropriately checked in Asian and Caucasian populations and clinical implementation has not been assessed. Overall impression I would like to congratulate the reviewers on an extremely thorough and methodologically sound systematic review and meta-analysis. My main concerns relate to the structure and writing of the discussion section, and the presentation the table. RESPONSE: Thank you so much for your comments and suggestions COMMENT. Major issues • The aims of the study are clearly delineated in the introduction (point 1-4). However I do not feel the structure of the discussion follows these aims or highlights the most salient findings of the analysis. Furthermore in my opinion the discussion section is too long. It would be better presented: o Major findings of the study (3-4 points) o Discussion of previous literature and how this differs o Future areas for research / gaps in knowledge o Limitations o Final conclusion RESPONSE: Thank you so much for taking the time to make such an insightful observation. It is true that the discussion portion is overly lengthy, as stated by the reviewer. However, we would want to point out that our objective was to provide a full explanation of the existing hypertension risk prediction models, which we have done. We discovered 117 models that are extremely huge as a result of our search and addressing the primary conclusions of these models took up a significant amount of space in the discussion section. We hope that offering a full discussion will assist readers in understanding the silent characteristics of the models that have been found. We appreciate your suggestions for the layout of the discussion part, and we acknowledge that we have made every effort to provide the discussion sections in the suggested manner. In addition, we have reduced the length of the discussion part by deleting redundant content whenever possible, as indicated by the reviewer. Please see the revised discussion section. Please see Pages 15-19, lines 374-474 in the revised manuscript. COMMENT. • The presentation of table 1 is extremely difficult to follow. The presentation of so many columns means that some of the entries for each study take up an entire page. It would be better to break this up into at least 2/3 tables e.g. between study population characteristics, model development characteristics/performance, variables used in model; and all these tables do not need to be in the main file (eg Himmelreich et al -> https://academic.oup.com/europace/article/22/5/684/5721485) RESPONSE: Thank you for your comment. We agree with the reviewer. As per the reviewer’s suggestion, we have now split the information in Table 1 into two tables, Table 1 and Table 2. Please see Table 1 and Table 2 in the revised manuscript. Pages 32 – 44. COMMENT. • Why are traditional regression model study characteristics included in main paper but not machine learning counterparts. It would be better to present them more equally. RESPONSE: Thank you for your comment. We agree with the reviewer. As per the reviewer’s suggestion, we have now added the study characteristics of the machine learning models in the main paper. Please see the newly added Table 3 in the revised manuscript. Pages 45- 52. COMMENT. • There are wide prediction intervals suggesting significant heterogeneity. Have you considered a Bayesian approach for meta-analysis? Frequentist methods can produce prediction intervals with poor coverage when there is a mixture of study sizes (https://pubmed.ncbi.nlm.nih.gov/30032705/) RESPONSE: Thank you for making such an astute insight. Unfortunately, we did not take into consideration the Bayesian technique for meta-analysis in our research. In this case, we employed the classic frequentist strategy because we did not expect to see such a significant degree of heterogeneity. We would like to express our gratitude to the reviewer for drawing our attention to this innovative technique. When the study sizes are heterogeneous and the data are sparse, the Bayesian approach to meta-analysis appears to be a promising method of analysis. Considering the Bayesian technique in such a case is something we will look into in the future. COMMENT. Minor issues • I note some of the models for predicting hypertension use systolic blood pressure and diastolic blood pressure. Does this not appear ‘double-dipping’ to include a variable that may well be an outcome? Does this not require some comment? RESPONSE: Thank you for noting this good point. The predictor systolic blood pressure and diastolic blood pressure are highly correlated with the outcome of hypertension. Please note that the models were used to predict incident (new-onset) hypertension. The people that the models were applied to did not have known hypertension at baseline. As would be expected, people with higher baseline blood pressure levels on the initial measurement were more likely to have sustained high blood pressure (or hypertension) long-term. While the predictor is highly correlated with the outcome, it is not synonymous with it. COMMENT. • Page 13 line 330 – please be more specific than ‘basically’ RESPONSE: Thank you. We have changed the word now as suggested. Please see page 13, line 316 in the revised manuscript. COMMENT. • Page 15 line 388 – I belive it should be ‘models’ RESPONSE: Thank you. We have changed the word now as suggested. Please see page 15, line 376 in the revised manuscript. COMMENT. • Page 17 line 446-447 does not make sense RESPONSE: Thank you. We have now removed the lines from the manuscript. Please see page 17, lines 420-423 in the revised manuscript. COMMENT. • Figure 1 – I believe the reasons for exclusion would be better ordered alphabetically or in descending number of records excluded. RESPONSE: Thank you. We have now changed Figure 1. The reasons for exclusion are now presented in descending order on the number of records excluded. Please see the revised figure 1. ________________________________________ 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: John B. Kostis Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. Submitted filename: Response to Reviewers.DOC Click here for additional data file. 21 Mar 2022 Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis PONE-D-21-31564R1 Dear Dr. Turin, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Antonio Palazón-Bru, PhD Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: In my opinion this manuscript is suitable for publication in PLOS ONE. The choice of the topic is timely and appropriate and the methodology used is correct in my opinion. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: John B. Kostis 28 Mar 2022 PONE-D-21-31564R1 Prediction of hypertension using traditional regression and machine learning models: A systematic review and meta-analysis Dear Dr. Turin: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Antonio Palazón-Bru Academic Editor PLOS ONE
  63 in total

1.  Replication of the top 10 most significant polymorphisms from a large blood pressure genome-wide association study of northeastern Han Chinese East Asians.

Authors:  Yue Qi; Hongye Zhao; Yanli Wang; Yuefei Wang; Changzhu Lu; Yu Xiao; Jun Cao; Nan Jia; Bin Wang; Wenquan Niu
Journal:  Hypertens Res       Date:  2013-11-07       Impact factor: 3.872

2.  A guide to systematic review and meta-analysis of prediction model performance.

Authors:  Thomas P A Debray; Johanna A A G Damen; Kym I E Snell; Joie Ensor; Lotty Hooft; Johannes B Reitsma; Richard D Riley; Karel G M Moons
Journal:  BMJ       Date:  2017-01-05

3.  The application of a decision tree to establish the parameters associated with hypertension.

Authors:  Maryam Tayefi; Habibollah Esmaeili; Maryam Saberi Karimian; Alireza Amirabadi Zadeh; Mahmoud Ebrahimi; Mohammad Safarian; Mohsen Nematy; Seyed Mohammad Reza Parizadeh; Gordon A Ferns; Majid Ghayour-Mobarhan
Journal:  Comput Methods Programs Biomed       Date:  2016-10-24       Impact factor: 5.428

4.  Comparison of the Framingham Heart Study hypertension model with blood pressure alone in the prediction of risk of hypertension: the Multi-Ethnic Study of Atherosclerosis.

Authors:  Paul Muntner; Mark Woodward; Devin M Mann; Daichi Shimbo; Erin D Michos; Roger S Blumenthal; April P Carson; Haiying Chen; Donna K Arnett
Journal:  Hypertension       Date:  2010-05-03       Impact factor: 10.190

5.  Prediction model and assessment of probability of incident hypertension: the Rural Chinese Cohort Study.

Authors:  Bingyuan Wang; Yu Liu; Xizhuo Sun; Zhaoxia Yin; Honghui Li; Yongcheng Ren; Yang Zhao; Ruiyuan Zhang; Ming Zhang; Dongsheng Hu
Journal:  J Hum Hypertens       Date:  2020-02-27       Impact factor: 3.012

6.  Incidence and predictors of isolated systolic hypertension and isolated diastolic hypertension in Taiwan.

Authors:  C J Yeh; W H Pan; Y S Jong; Y Y Kuo; C H Lo
Journal:  J Formos Med Assoc       Date:  2001-10       Impact factor: 3.282

7.  Genetic predisposition to higher blood pressure increases risk of incident hypertension and cardiovascular diseases in Chinese.

Authors:  Xiangfeng Lu; Jianfeng Huang; Laiyuan Wang; Shufeng Chen; Xueli Yang; Jianxin Li; Jie Cao; Jichun Chen; Ying Li; Liancheng Zhao; Hongfan Li; Fangcao Liu; Chen Huang; Chong Shen; Jinjin Shen; Ling Yu; Lihua Xu; Jianjun Mu; Xianping Wu; Xu Ji; Dongshuang Guo; Zhengyuan Zhou; Zili Yang; Renping Wang; Jun Yang; Weili Yan; Dongfeng Gu
Journal:  Hypertension       Date:  2015-08-17       Impact factor: 10.190

8.  The Role of Genetic Risk Score in Predicting the Risk of Hypertension in the Korean population: Korean Genome and Epidemiology Study.

Authors:  Nam-Kyoo Lim; Ji-Young Lee; Jong-Young Lee; Hyun-Young Park; Myeong-Chan Cho
Journal:  PLoS One       Date:  2015-06-25       Impact factor: 3.240

9.  Summarising and synthesising regression coefficients through systematic review and meta-analysis for improving hypertension prediction using metamodelling: protocol.

Authors:  Mohammad Ziaul Islam Chowdhury; Iffat Naeem; Hude Quan; Alexander A Leung; Khokan C Sikdar; Maeve O'Beirne; Tanvir C Turin
Journal:  BMJ Open       Date:  2020-04-09       Impact factor: 2.692

10.  Development and validation of prediction models for hypertension risks in rural Chinese populations.

Authors:  Fei Xu; Jicun Zhu; Nan Sun; Lu Wang; Chen Xie; Qixin Tang; Xiangjie Mao; Xianzhi Fu; Anna Brickell; Yibin Hao; Changqing Sun
Journal:  J Glob Health       Date:  2019-12       Impact factor: 4.413

View more
  1 in total

1.  Development and validation of a hypertension risk prediction model and construction of a risk score in a Canadian population.

Authors:  Mohammad Ziaul Islam Chowdhury; Alexander A Leung; Khokan C Sikdar; Maeve O'Beirne; Hude Quan; Tanvir C Turin
Journal:  Sci Rep       Date:  2022-07-27       Impact factor: 4.996

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.