| Literature DB >> 33912652 |
Dineo Mpanya1,2, Turgay Celik3,2, Eric Klug4, Hopewell Ntsinjana5.
Abstract
OBJECTIVE: The partnership between humans and machines can enhance clinical decisions accuracy, leading to improved patient outcomes. Despite this, the application of machine learning techniques in the healthcare sector, particularly in guiding heart failure patient management, remains unpopular. This systematic review aims to identify factors restricting the integration of machine learning derived risk scores into clinical practice when treating adults with acute and chronic heart failure.Entities:
Keywords: Heart failure; Hospitalization; Machine learning; Mortality; Predictive modelling; Risk score; Sub-Saharan Africa
Year: 2021 PMID: 33912652 PMCID: PMC8065274 DOI: 10.1016/j.ijcha.2021.100773
Source DB: PubMed Journal: Int J Cardiol Heart Vasc ISSN: 2352-9067
Fig. 1Flow chart of the systematic literature search.
Characteristics of the included studies.
| Adler, E.D (2019) | 2006–2017 | 5 822 | Inpatient and outpatient | EHR and Trial | 8 | All-cause mortality |
| Ahmad, T (2018) | 2000–2012 | 44 886 | Inpatient and outpatient | Registry | 8 | 1-year all-cause mortality |
| Allam, A (2019) | 2013 | 272 778 | Inpatient | Claims dataset | 50 | 30-day all-cause readmission |
| Angraal, S (2020) | 2006–2013 | 1 767 | Inpatient | Trial | 26 | All-cause mortality and HF hospitalization |
| Ashfaq, A (2019) | 2012–2016 | 7 655 | Inpatient and outpatient | EHR | 30-day all-cause readmission | |
| Awan, SE (2019) | 2003–2008 | 10 757 | Inpatient and outpatient | EHR | 47 | 30-day HF-related readmission and mortality |
| Chen, R (2019) | 2014–2017 | 98 | Inpatient | Prospective Clinical and MRI | 32 | Cardiac death, heart transplantation and HF-related hospitalization |
| Chicco, D (2020) | 2015 | 299 | Inpatient | Medical records | 13 | One year survival |
| Chirinos, J (2020) | 2006–2012 | 379 | Inpatient | Trial | 48 | Risk of all-cause death or heart failure-related hospital admission |
| Desai, R.J (2020) | 2007–2014 | 9 502 | Inpatient and outpatient | Claims data and EHR | 62 | All-cause mortality and HF hospitalization, total costs for hospitalization, outpatient visits, and medication |
| Frizzell, J.D (2017) | 2005–2011 | 56 477 | Inpatient | Registry and claims data | All-cause readmission 30-days after discharge | |
| Gleeson, S (2017) | 2010–2015 | 295 | Inpatient | Echo database & EHR | 291 | All-cause mortality and heart failure admissions |
| Golas, S.B (2018) | 2011–2015 | 11 510 | Inpatient and outpatient | EHR | 3 512 | All-cause 30-day readmission, healthcare utilization cost |
| Hearn, J (2018) | 2001–2017 | 1 156 | EHR and Cardiopulmonary stress test data | All-cause mortality | ||
| Hsich, E (2011) | 1997–2007 | 2 231 | Cardiopulmonary stress test data | 39 | All-cause mortality | |
| Jiang, W (2019) | 2013–2015 | 534 | Inpatient | EHR | 57 | 30-day readmission |
| Kourou, K (2016) | 71 | Pre and post-operative data | 48 | 1-year all-cause mortality | ||
| Krumholz, H (2019) | 2013–2015 | 716 790 | Inpatient | Claims dataset | All-cause death within 30-days of admission | |
| Kwon, J (2019) | 2016–2017 | 2 165 (training dataset) | Inpatient | Registry | 12 and 36-month in-hospital mortality | |
| Liu, W (2020) | 303 233 (heart failure) | Inpatient | Readmission database | Admission 3H myocardial infarction, congestive heart failure and pneumonia 30-day readmission | ||
| Lorenzoni, G (2019) | 2011–2015 | 380 | Inpatient | Research data | Hospitalization among patients with heart failure | |
| Maharaj, S.M (2018) | 2015 | 1 778 | Inpatient | EHR | 56 | 30-day readmission |
| McKinley, D (2019) | 2012–2015 | 132 | Inpatient | EHR | 29 | All-cause readmission within 30-days |
| Miao, F (2017) | 2001–2007 | 8 059 | Public database | 32 | 1-year in-hospital mortality | |
| Nakajima, K (2020) | 2005–2016 | 526 | Multicentre database | 13 | 2-year life-threatening arrhythmic events and heart failure death | |
| Shameer, K (2016) | 1 068 | Inpatient | EHR | 4 205 | 30-day readmission | |
| Shams, I (2015) | 2011–2012 | 1 674 | Inpatient | EHR | 30-day readmission | |
| Stampehl, M (2020) | 2010–2014 | 206 644 | Inpatient | EHR | 30-day and one-year post-discharge all-cause mortality | |
| Taslimitehrani, V (2016) | 1993–2013 | 5 044 | Inpatient | EHR | 43 | 1,2 and 5-year survival after HF diagnosis |
| Turgeman, L (2016) | 2006–2014 | 4 840 | Inpatient | EHR | Readmission |
CVD = cardiovascular disease; EHR = electronic health record; HF = heart failure; MRI = magnetic resonance imaging.
Fig. 2Study population region.
Characteristics of heart failure patients included in the 30 models predicting mortality and hospitalization.
| Adler, E.D (2019) | USA and Europe | 5 822 | 60.3 | ||||
| Ahmad, T (2018) | Europe | 44 886 | 73.2 | 63 | |||
| Allam, A (2019) | USA and Europe | 272 778 | 73 ± 14 | 51 | |||
| Angraal, S (2020) | USA, Canada, Brazil, Argentina, Russia, Georgia | 1 767 | 72 (64–79) | 50 | |||
| Ashfaq, A (2019) | Europe | 7 655 | 78.8 | 57 | |||
| Awan, SE (2019) | Australia | 10 757 | 82 ± 7.6 | 49 | 67 | 55 | |
| Chen, R (2019) | China | 98 | 47 ± 14 | 79 | 23 | ||
| Chicco, D (2020) | Pakistan | 299 | 40–95* | 65 | |||
| Chirinos, J (2020) | USA, Canada, Russia | 379 | 7.4 | 70 (62–77) | 53.5 | 94.5 | 30.6 |
| Desai, R.J (2020) | USA | 9 502 | 5.1 | 78 ± 8 | 45 | 87.1 | 22 |
| Frizzell, J.D (2017) | USA | 56 477 | 10 | 80 (74–86) | 45.5 | 75.7 | 58 |
| Gleeson, S (2017) | New Zealand | 295 | 62 | 74 | 43 | ||
| Golas, S.B (2018) | USA | 11 510 | 7.9 | 75.7 (64–85) | 52.8 | ||
| Hearn, J (2018) | Canada | 1 156 | 54 | 74.6 | |||
| Hsich, E (2011) | USA | 2 231 | 54 ± 11 | 73 | 41 | ||
| Jiang, W (2019) | USA | 534 | 28 | 74.8 | 46 | ||
| Kourou, K (2016) | Belgium | 71 | 48.07 ± 14.82 | 80.3 | |||
| Krumholz, H (2019) | USA | 716 790 | 11.3 | 81.1 ± 8.4 | 45.6 | ||
| Kwon, J (2019) | Asia | 2 165 | 69.8 | 59.7 | |||
| Liu, W (2019) | USA | 303 233 | 72.5 | 50.9 | |||
| Lorenzoni, G (2019) | Italy | 380 | 78 (72–83) | 42.9 | 18.9 | ||
| Maharaj, S.M (2018) | USA | 1 778 | 0.95 | 72.3 ± 12.1 | 97.6 | 14 | |
| McKinley, D (2019) | USA | 132 | 100 | 59.25 | 100 | 91 | |
| Miao, F (2017) | USA | 8 059 | 73.7 | 54 | 25 | 23.2 | |
| Nakajima, K (2020) | Japan | 526 | 66 ± 14 | 72 | 53 | 37 | |
| Shameer, K (2016) | USA | 1 068 | |||||
| Shams, I (2015) | USA | 1 674 | 70.4 | 69.9 | 96 | ||
| Stampehl, M (2020) | USA | 206 644 | 12.6 | 80.5 ± 11.2 | 38.3 | 96.5 | 0.4 |
| Taslimitehrani, V (2016) | USA | 5 044 | 78 ± 10 | 52 | 81 | 70.2 | |
| Turgeman, L (2016) | USA | 4 840 | 69.3 ± 11.02 | 96.5 | 84.9 |
Age showed as mean ± standard deviation, median (25th-75th percentile interquartile range) or minimum and maximum value.* IHD: ischaemic heart disease; USA: United States of America.
Fig. 3Number of studies using machine learning algorithms.
Performance metrics of algorithms predicting mortality and hospitalization in heart failure.
| Adler, E.D (2019) | Boosted decision trees | 0.88 (0.85–0.90) | ||||
| Ahmad, T (2018) | Random forest | 0.83 | ||||
| Allam, A (2019) | Recurrent neural network | 0.64 (0.640–0.645) | ||||
| Logistic regression l2-norm regularization (LASSO) | 0.643 (0.640–0.646) | |||||
| Angraal, S (2020) | Logistic regression | 0.66 (0.62–0.69) | 0.73 (0.66–0.80) | |||
| Logistic regression with LASSO regularization | 0.65 (0.61–0.70) | 0.73 (0.67–0.79) | ||||
| Gradient descent boosting | 0.68 (0.66–0.71) | 0.73 (0.69–0.77) | ||||
| Support vector machines (linear kernel) | 0.66 (0.60–0.72) | 0.72 (0.63–0.81) | ||||
| Random forest | 0.72 (0.69–0.75) | 0.76 (0.71–0.81) | ||||
| Ashfaq, A (2019) | Long Short-Term Memory (LSTM) neural network | 0.77 | 0.51 | |||
| Awan, SE (2019) | Multi-layer perceptron (MLP) | 48.4 | 0.62 | |||
| Chen, R (2019) | Naïve Bayes | 0.827 | 0.855 | |||
| Naïve Bayes + IG | 0.857 | |||||
| Random forest | 0.817 | |||||
| Random forest + IG | 0.827 | |||||
| Decision trees (bagged) | 0.827 | |||||
| Decision trees (bagged) + IG | 0.816 | |||||
| Decision trees (boosted) | 0.735 | |||||
| Decision trees (boosted) + IG | 0.806 | |||||
| Chicco, D (2020) | Random forest | 0.740 | 0.800 | 0.547 | ||
| Decision tree | 0.737 | 0.681 | 0.554 | |||
| Gradient boosting | 0.738 | 0.754 | 0.527 | |||
| Linear regression | 0.730 | 0.643 | 0.475 | |||
| One rule | 0.729 | 0.637 | 0.465 | |||
| Artificial neural network | 0.680 | 0.559 | 0.483 | |||
| Naïve Bayes | 0.696 | 0.589 | 0.364 | |||
| SVM (radial) | 0.690 | 0.749 | 0.182 | |||
| SVM (linear) | 0.684 | 0.754 | 0.115 | |||
| K-nearest neighbors | 0.624 | 0.493 | 0.148 | |||
| Chirinos, J (2020) | Tree-based pipeline optimizer | 0.717 (0.643–0.791) | ||||
| Desai, R.J (2020) | Logistic regression (traditional) | 0.749 (0.729–0.768) | 0.738 (0.711–0.766) | |||
| LASSO | 0.750 (0.731–0.769) | 0.764 (0.738–0.789) | ||||
| CART | 0.700 (0.680–0.721) | 0.738 (0.710–0.765) | ||||
| Random forest | 0.757 (0.739–0.776) | 0.764 (0.738–0.790) | ||||
| GBM | 0.767 (0.749–0.786) | 0.778 (0.753–0.802) | ||||
| Frizzell, J.D (2017) | Random forest | 0.607 | ||||
| GBM | 0.614 | |||||
| TAN | 0.618 | |||||
| LASSO | 0.618 | |||||
| Logistic regression | 0.624 | |||||
| Gleeson, S (2017) | Decision trees | 0.7505 | ||||
| Golas, S.B (2018) | Logistic regression | 0.626 | 0.664 | 0.435 | ||
| Gradient boosting | 0.612 | 0.650 | 0.425 | |||
| Maxout networks | 0.645 | 0.695 | 0.454 | |||
| Deep unified networks | 0.646 | 0.705 | 0.464 | |||
| Hearn, J (2018) | Staged LASSO | 0.827 (0.785–0.867) | ||||
| Staged neural network | 0.835 (0.795–0.880) | |||||
| LASSO (breath-by-breath) | 0.816 (0.767–0.866) | |||||
| Neural network (breath-by-breath) | 0.842 (0.794–0.882) | |||||
| Hsich, E (2011) | Random survival forest | 0.705 | ||||
| Cox proportional hazard | 0.698 | |||||
| Jiang, W (2019) | Logistic and beta regression (ML) | 0.73 | ||||
| Kourou, K (2016) | Naïve Bayes | 85 | 0.86 | |||
| Bayesian network | 85.9 | 0.596 | ||||
| Adaptive boosting | 78 | 0.74 | ||||
| Support vector machines | 90 | 0.74 | ||||
| Neural networks | 87 | 0.845 | ||||
| Random forest | 75 | 0.65 | ||||
| Krumholz, H (2019) | Logistic regression (ML) | 0.776 | ||||
| Kwon, J (2019) | Deep learning | 0.813 (0.810–0.816) | ||||
| Random forest | 0.696 (0.692–0.700) | |||||
| Logistic regression | 0.699 (0.695–0.702) | |||||
| Support vector machine | 0.636 (0.632–0.640) | |||||
| Bayesian network | 0.725 (0.721–0.728) | |||||
| Liu, W (2019) | Logistic regression | 0.580 (0.578–0.583) | ||||
| Gradient boosting | 0.602 (0.599–0.605) | |||||
| Artificial neural networks | 0.604 (0.602–0.606) | |||||
| Lorenzoni, G (2019) | GLMN | 77.8 | 0.812 | 0.86 | ||
| Logistic regression | 54.7 | 0.589 | 0.646 | |||
| CART | 44.3 | 0.635 | 0.586 | |||
| Random forest | 54.9 | 0.726 | 0.691 | |||
| Adaptive Boosting | 57.3 | 0.671 | 0.644 | |||
| Logitboost | 66.7 | 0.625 | 0.654 | |||
| Support vector machines | 57.3 | 0.699 | 0.695 | |||
| Artificial neural networks | 61.6 | 0.682 | 0.677 | |||
| Maharaj, S.M (2018) | Boosted tree | 0.719 | ||||
| Spike and slab regression | 0.621 | |||||
| McKinley, D (2019) | K-nearest neighbor | 0.773 | 0.768 | |||
| K-nearest neighbor (randomized) | 0.477 | 0.469 | ||||
| Support vector machines | 0.545 | 0.496 | ||||
| Random forest | 0.682 | 0.616 | ||||
| Gradient boosting machine | 0.614 | 0.589 | ||||
| LASSO | 0.614 | 0.576 | ||||
| Miao, F (2017) | Random survival forest | 0.804 | ||||
| Random survival forest (improved) | 0.821 | |||||
| Nakajima, K (2020) | Logistic regression | 0.898 | ||||
| Random forest | 0.917 | |||||
| GBT | 0.907 | |||||
| Support vector machine | 0.910 | |||||
| Naïve Bayes | 0.875 | |||||
| k-nearest neighbors | 0.854 | |||||
| Shameer, K (2016) | Naïve Bayes | 0.832 | 0.78 | |||
| Shams, I (2015) | Phase type Random forest | 91.95 | 0.836 | 0.892 | ||
| Random forest | 88.43 | 0.802 | 0.865 | |||
| Support vector machine | 86.16 | 0.775 | 0.857 | |||
| Logistic regression | 83.40 | 0.721 | 0.833 | |||
| Artificial neural network | 82.39 | 0.704 | 0.823 | |||
| Stampehl, M (2020) | CART | |||||
| Logistic regression | ||||||
| Logistic regression (stepwise) | 0.74 | |||||
| Taslimitehrani, V (2016) | CPXR(Log) | 0.914 | ||||
| Support vector machine | 0.75 | |||||
| Logistic regression | 0.89 | |||||
| Turgeman, L (2016) | Naïve Bayes | 48.9 | 0.676 | |||
| Logistic regression | 28.1 | 0.699 | ||||
| Neural network | 8.9 | 0.639 | ||||
| Support vector machine | 23.0 | 0.643 | ||||
| C5 (ensemble model) | 43.5 | 0.693 | ||||
| CART (boosted) | 22.6 | 0.556 | ||||
| CART (bagged) | 9.0 | 0.579 | ||||
| CHAID Decision trees (boosted) | 30.3 | 0.691 | ||||
| CHAID Decision trees (bagged) | 10.5 | 0.707 | ||||
| Quest decision tree (boosted) | 20.3 | 0.487 | ||||
| Quest decision tree (bagged) | 7.2 | 0.579 | ||||
| Naïve network + Logistic regression | 38.2 | 0.653 | ||||
| Naïve network + Neural network | 26.3 | 0.635 | ||||
| Naïve network + SVM | 35.8 | 0.649 | ||||
| Logistic regression + Neural network | 16.8 | 0.59 | ||||
| Logistic regression + SVM | 26.2 | 0.607 | ||||
| Neural network + SVM | 16.5 | 0.577 | ||||
AUC: area under the receiver operating characteristic curve; CART: classification and regression tree; CPXR: contrast pattern aided logistic regression; GBM: gradient-boosted model; HR: hazard ratio; IG: information gain; LASSO: least absolute shrinkage and selection operator; ML: machine learning; SVM: support vector machine; TAN: tree augmented Bayesian network. The AUC is displayed under both the mortality and hospitalization column if the authors did not specify the outcome predicted.