| Literature DB >> 33952192 |
Yinan Huang1, Ashna Talwar1, Satabdi Chatterjee1, Rajender R Aparasu2.
Abstract
BACKGROUND: Advances in machine learning (ML) provide great opportunities in the prediction of hospital readmission. This review synthesizes the literature on ML methods and their performance for predicting hospital readmission in the US.Entities:
Keywords: Hospital readmission; Machine learning; Prediction; Scoping review
Mesh:
Year: 2021 PMID: 33952192 PMCID: PMC8101040 DOI: 10.1186/s12874-021-01284-z
Source DB: PubMed Journal: BMC Med Res Methodol ISSN: 1471-2288 Impact factor: 4.615
Fig. 1Flow diagram for study selection
ML algorithms used in the studies and corresponding featuring studies. (N = 43 studies)
| Type of ML Algorithms | Number of Studiesf (Percent) | Featuring Studies |
|---|---|---|
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| [ | ||
| NN (with multiple hidden layers, e.g. deep learning)b | 10 | [ |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| NN (with a single or unclear number of hidden layers, or unclear) | 5 | [ |
| | [ | |
| | [ | |
| | [ | |
| | [ |
Abbreviations: ML machine learning, Lasso least absolute shrinkage and selection operator, NN neural networks, CNN convolutional neural network, RNN recurrent neural network, DL deep learning, KNN The k-nearest neighbors. ait includes adaboost, gradient boosting, gradient descent boosting, boosting, XGBoost; bit includes CNN, RNN, DNN, deep stacking networks, and ensemble of DL methods; cDT ensembled with SVM, RF combined with SVM, tree-augmented naïve Bayesian network; done hidden layers; edid not specify number of layers
fSince most studies have applied more than 1 machine learning algorithms, therefore the sum of the number of studies by machine learning method is greater than 43
Fig. 2Boxplot and Beeswarm plot of AUC by ML category. Abbreviations: ML: machine learning; NNs: neural networks; RF: random forest, DT: decision tree; SVM: support vector machine
Descriptive statistics of AUC by ML category
| ML category | Number of Studiesa | Mean (STD) | Median | Min | Max | IQR |
|---|---|---|---|---|---|---|
| 15 | 0.71 (0.07) | 0.71 | 0.61 | 0.81 | 0.64–0.78 | |
| 17 | 0.70 (0.06) | 0.7 | 0.59 | 0.81 | 0.66–0.75 | |
| 16 | 0.68 (0.09) | 0.64 | 0.53 | 0.9 | 0.63–0.72 | |
| 9 | 0.70 (0.10) | 0.67 | 0.59 | 0.88 | 0.63–0.77 | |
| 16 | 0.69 (0.08) | 0.65 | 0.58 | 0.84 | 0.64–0.75 | |
| 10 | 0.70 (0.11) | 0.68 | 0.5 | 0.86 | 0.65–0.78 | |
| 10 | 0.68 (0.04) | 0.68 | 0.62 | 0.77 | 0.66–0.71 |
Abbreviations: ML machine learning, NNs neural networks, RF random forest, DT decision tree, SVM support vector machine, STD standard deviation, IQR the interquartile range. athe total number of studies is larger than total number of included studies, because some studies used more than 1 ML algorithms. aIt includes adaboost, gradient boosting, gradient descent boosting, boosting, XGBoost; bIt includes Lasso (L1 regularization), ridge regression (L2 regularization), and elastic-net algorithms; cIt includes: DT ensembled with SVM, RF combined with SVM, tree-augmented naïve Bayesian network
Overview of methods for model validation across studies (N = 43)
| Type of validation | Number of studies (Percent) | Featuring studies |
|---|---|---|
| [ | ||
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| | [ | |
| [ |