| Literature DB >> 36082120 |
Anxing Zhao1,2, Mohamed Elgendi1, Carlo Menon1.
Abstract
An acute hypotensive episode (AHE) can lead to severe consequences and complications that threaten patients' lives within a short period of time. How to accurately and non-invasively predict AHE in advance has become a hot clinical topic that has attracted a lot of attention in the medical and engineering communities. In the last 20 years, with rapid advancements in machine learning methodology, this topic has been viewed from a different perspective. This review paper examines studies published from 2008 to 2021 that evaluated the performance of various machine learning algorithms developed to predict AHE. A total of 437 articles were found in four databases that were searched, and 35 full-text articles were included in this review. Fourteen machine learning algorithms were assessed in these 35 articles; the Support Vector Machine algorithm was studied in 12 articles, followed by Logistic Regression (six articles) and Artificial Neural Network (six articles). The accuracy of the algorithms ranged from 70 to 96%. The size of the study sample varied from small (12 subjects) to very large (3,825 subjects). Recommendations for future work are also discussed in this review.Entities:
Keywords: anesthesia; digital health; emergency and critical care; hypertension; hypotension; intensive care unit; low blood pressure; obstetric and gynecologic
Year: 2022 PMID: 36082120 PMCID: PMC9445248 DOI: 10.3389/fcvm.2022.937637
Source DB: PubMed Journal: Front Cardiovasc Med ISSN: 2297-055X
Figure 1Flow chart of the methodology used to screen the articles. Thirty-five articles published between 2008 and 2021 were included in the review.
Overview of studies included in the systematic review.
|
|
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|---|---|
| Zhang et al. ( | 1,055 | ABP | N/R | Max, min, avg, median, STD, skewness, kurtosis, upper quartile, avg absolute deviation, range, variance | •LR | 5 h | 60 min | MIMIC II ( | Accuracy(%) | |
| Ribeiro et al. ( | 3,825 | N/R | HR, RR, SpO2, SBP, DBP, MAP time series, PP, CO | N/R | Interquartile range, max, min, mean, median, skewness, kurtosis, linear slope, SD, variance, wavelet energy, cross-correlations between signals | Layered Learning (LL), the adopted classifier in each layer was a Light Gradient Boosting Machine | 60 min | 60 min | MIMIC III ( | Accuracy (%) = 75.9 ± 4.2 |
| Tang et al. ( | 30 | N/R | Patient's NE infusion rate per unit weight, ECG, ABP | 125 Hz | HR, PP, KR | •Physiology based approach (our method) | 1.3–6.67 h | 3.33–20 min | Inter-mountain Medical Center, MIMIC-III ( | Mins to 5 mmHg RMSE |
| Lee et al. ( | 3,301 | •Invasive: AP+ECG +PPG+EtCO2 vs. AP | 100 Hz | N/R | DL consisted of seven convolutional layers | 30 s | 5, 10, 15 min | VitalDB database ( | AUROC(%) | |
| Moghadam et al. ( | 1,000 | ABP, HR, SBP, DBP, Resp, SpO2; PP, MAP, CO, MAP2HR, RR | N/R | 33 scalar feature | •LR | 5 min | 30 min | MIMIC III ( | Accuracy(%) | |
| Lee et al. ( | 282 | Non invasive BP, HR, Mechanical Ventilation data, Bispectral index | N/R | Min, max, mean, std. Experiment performed 3-fold: 27, 56, 67 features then sum and dim reduction to get 98, 45, 20, 29. Best performance are 56 for 1st experiment and 20 for 2nd experiment | •RF | 4–1 min before intubation | 1 min | Soonchunhyang University Bucheon Hospital database | Accuracy (%) Raw feature vs. plus statistics features: | |
| Moghadam et al. ( | 1,000 | N/R | ABP,HR,SBP, DBP, Resp, SpO2; PP, MAP, CO, MAP2HR, RR | N/R | 33 scalar features. PCA optimize the feature set and resulted into 11 combined features | •LR | 5 min | 30 min | MIMIC III ( | Accuracy(%) |
| Xiao et al. ( | 2,866 | N/R | ABP | N/R | decompose MAP with SW and ensemble EMD, then a 3-layer auto encoder to get 50 outputs | Multiple gene expression programming classier | 2 h | 60 min | MIMIC II ( | Voting combination strategy vs. 10-fold cross validation (%) |
| Shin et al. ( | 207 | N/R | MAP | N/R | •LR: mean, slope, SD of MAP of past 5, 10, 20, 30, 45, 60 min | •Logistic regression (LR) | 60 min | 30 min | MIMIC II (“Hospital 1”) ( | •LR predicted on average 7.0 min before onset (Hospital 1) and 2.5 min before (Hospital 2) |
| Chan et al. ( | 538 | N/R | MAP, HR, SPO2 | N/R | N/R | Long Short-Term Memory, three layers each with 100 units | 10–60 min | 10–60 min | Kingston General Hospital | •Accuracy(%) = 80 |
| Angelotti et al. ( | 86 | N/R | ABP and ECG containing at least one ECG lead | N/R | SBP statistical moments; LF, HF, VLF spectral powers (for both RR and SBP); LF/HF (for both RR and SBP); Baroreflex amplitude; Baroreflex frequency | •4 classification trees | 20 min | 10 min | MIMIC III ( | AUC(%) with vs. w/o BRFX: |
| Pathinaru-pothi et al. ( | 30 | N/R | MAP | N/R | Use MAP severity quantizer, Consensus motif extractor, SVM based prediction engine for feature extraction | SVM | 15 min | 2.75 h | MIMIC II ( | F1 score (%) = 82 |
| Kim et al. ( | 2,291 | N/R | MAP | N/R | N/R | Collision Frequency Locality Sensitive Hashing | 5 h | 60 min | MIMIC II ( | Accuracy (%) in the range (93, 96) |
| Hamano et al. ( | 100 | N/R | MAP, EtCO2, MAC, HR, SpO2, and body temperature | N/R | Each variable is mapped | Spiking neural networks | 15 min | 5 min | OR of a tertiary hospital, Auckland NZ | 37.6% of the experiments had an SNR above 0, which means better prediction than the naive method |
| Jiang et al. ( | 2,866 | N/R | MAP | 1 Hz | 55 feature incl peak, mode, skewness, kurtosis, and Shannon entropy from original time series, first 9 IMFs and last IMF | Multi GP | 2 h | 60 min | MIMIC II ( | Accuracy (%) = 82.9 in the training set and 79.9 in the testing set |
| Ghosh et al. ( | 50 | N/R | MAP | N/R | A gap-constrained sequential contrast pattern P is required (1) Positive Support: countP (D+, g) >= alpha (2) Negative Support: countP (D-, g) < = delta | Sequential pattern mining | 30, 60 min | 60, 120 min | MIMIC II ( | Accuracy (%) single mode performance with 10 symbols vs. multi mode performance with 15 symbols: |
| Ghosh et al. ( | 528 | N/R | MAP | N/R | Utilize sequential contrast patterns as features to build classification models | SVM | 60, 90 min | 30, 60 min | MIMIC II ( | Accuracy (%) = 85.8 |
| Kim et al. ( | 2,291 | N/R | MAP | N/R | N/R | LSH with two variants, the bit sampling based (L1LSH), the random projection based (E2LSH) | 5 h | 60 min | MIMIC II ( | Accuracy (%) |
| Jiang et al. ( | 2,866 | N/R | MAP | 1 Hz | EMD to decompose MAP into 77 IMFs. Statistical features: min, mean, max, median, variance, max instantaneous freq, HF/LF energy ration | •Multi GP | 2 h | 60 min | MIMIC II ( | Accuracy (%) |
| Fan et al. ( | 1,599 | N/R | ABP | 125 Hz | EMD to extract 77 features then group to 30. Extracted features: min,mean,max, median,variance. Calculated features: max instantaneous freq, HF/LF. Also, the 12th percentile, skewness, kurtosis, mode of the last IMF | RF based on GP | 30 min | N/R | MIMIC II ( | Accuracy (%) = 77.6 |
| Kim et al. ( | 2,291 | N/R | ABP | 125 Hz | The first and second differences, 20-min variance and slope | •Dynamic Bayesian network | 30 min | 30, 60 min | MIMIC II( | Accuracy (%) |
| Jiang et al. ( | 110 | N/R | MAP | N/R | EMD was used to calculate MAP time series and BW of the AM, FM, power of IMF | GP | 2 h | 60 min | MIMIC II ( | Accuracy (%) = 83.4 in the training set and 80.6 in the testing set |
| Zhang et al. ( | 12 | N/R | MAP, HR, SBP, and DBP | N/R | MAP, HR, SBP, and DBP | ANN with one hidden layer | 30 min | 60 min | MIMIC II ( | Median Absolute Difference between the predicted and actual HI was 0.070, ranged from 0.012 to 0.175 |
| Sun et al. ( | 2,863 | N/R | MAP | 1 Hz | The 2 cluster centers, x1Mean and x2Mean, the 2 cluster ratios, x1Ratio and x2Ratio, the average of 15 min MAP signal before T0 | SVM | 60 min | 60 min | MIMIC II ( | •Accuracy = 81.2% |
| Janghorbani et al. ( | 95 | N/R | HR, SAP, DAP, MAP | N/R | •LR: 10% DAP, mean MAP, max ECO; | •LR | 30 min | 60 min | MIMIC II ( | Accuracy (%) |
| Rocha et al. ( | 311 | N/R | MAP | 125 Hz | N/R | Neural network multi-models | 12 h | 60 min | MIMIC II ( | •Sensitivity = 82.8% |
| Sun et al. ( | 1,500 | N/R | SBP, DBP, MAP, SpO2, HR | N/R | top-10 wavelet coefficients as the features | Locally Supervised Metric Learning (LSML) | 60 min | 60 min | MIMIC II ( | Accuracy (%) = 85.51 |
| Lee et al. ( | 1,357 | N/R | HR, SBP, DBP, MAP | N/R | Mean, median, SD, variance, interquartile range, skewness, kurtosis, linear regression slope, and relative energies in different spectral bands. A total of 45 features, whose space dim was reduced | Feed-forward, three-layer artificial neural networks (ANNs) | 30 min | 1–4 h | MIMIC II( | Accuracy (%) |
| Afsar ( | 60 | N/R | ABP | 125 Hz | SBP and Area under SBP wave along with the 1st, 3rd, and 6th principle component averaged over beats in each 60 s interval | Linear support vector machine (LSVC) | 1.5 h | 60 min | MIMIC II ( | Accuracy (%) |
| Wang et al. ( | 70 | N/R | MAP | N/R | db3 as wavelet mother function to decompose the MAP signal at three levels to get the LF coefficient cA3 and HF coefficients cD1, cD2, and cD3. Then extract median and maximum | SVM | 60 min | 60 min | MIMIC II ( | Accuracy (%) = 90 |
| Rocha et al. ( | 110 | N/R | ABP | N/R | The filtered signals are down-sampled by 2 and the results are called approximation and detail coefficients | Feed-forward neural Networks with two hidden layers | 2 h | 60 min | MIMIC II ( | •Sensitivity = 94.7% |
| Fournier et al. ( | 60 | N/R | ECG, PAP, ABP, central venous pressure, HR, RR, SpO2, CO, and alarms annotations | N/R | Use KL divergence to identify the most discriminative features | Nearest neighbors (NN) | 30 min | MIMIC II ( | Accuracy (%) = 80 | |
| Jousset et al. ( | 50 | N/R | MAP | N/R | N/R | SVM | 2 h | 60 min | MIMIC II ( | Accuracy (%) = 75 |
| Chiarugi et al. ( | 60 | N/R | ECG,ABP | 125 Hz | HR, SBP, mean ABP (ABPM), DBP, MAP | Decision tree | 10 h | 2 h | MIMIC II ( | Accuracy (%) 91.7 in the training set and 75 on test set |
| Henriques et al. ( | 50 | N/R | ABP | N/R | N/R | Generalized regression neural network | 6 h | 60 min | MIMIC II ( | Accuracy (%) 100 for test set A, 92.5 for test set B |
Overview of the included papers in this study. STD/SD, standard deviation; LR, logistic regression; SVM, support vector machine; RF, random forest; ABP, arterial blood pressure; HR, heart rate; SBP, systolic blood pressure; DBP, diastolic blood pressure; MAP, mean arterial pressure; PP, pulse pressure; CO, cardiac output; NE, norepinephrine; KNN, K-nearest neighbors; SW, sliding window; EMD, empirical mode decomposition; GP, genetic programming; IMF, intrinsic mode functions; GA, genetic algorithm.
Figure 2Number of publications by year.
Figure 3Prediction accuracy based on the type of algorithm. LSH is the most accurate algorithm, although only a few studies used it. SVM is the algorithm that was most often studied, and RF is the least accurate algorithm. RF, random forest; GP, genetic programming; DT, decision tree; KNN, K-nearest neighbors; SVM, support vector machine; LR, logistic regression; ANN, artificial neural network; LSH, locality sensitive hashing.
Summary of the prediction window.
|
|
|
|
|---|---|---|
| Post-operative AHE | 10 | 1 |
| 30 | 4 | |
| 60 | 21 | |
| 120 | 2 | |
| 165 | 1 | |
| Intra-operative AHE | 1 | 1 |
| 5 | 2 | |
| 10 | 1 | |
| 15 | 1 | |
| AHE during medication | 3–20 | 1 |
| 30 | 1 |
Most studies focused on predicting post-operative AHE. A 60 min prediction time length was the most common. Fewer studies aimed to predict AHE during an operation or when taking medication.
Figure 4(Left) The methodology used to extract the features was very diverse; no single methodology accounts for more than half of the studies. Statistics (e.g., maximum, minimum, mean values of the raw data) are the extracted features most often studied, followed by clinical equations (apply raw data to the equation to calculate some of the derived information, e.g., cardiac output, MAP to HR ratio). (Right) Graph showing how many features were extracted to predict AHE. The number varies greatly among the studies, with a single feature extraction being the most common.
Figure 5Evaluation metrics. The way in which the performance of machine learning algorithms is evaluated varies from study to study. Accuracy is the most common metric; it was adopted by almost 80% of the studies. AC, accuracy; SE, sensitivity; SP, specificity; PPV, positive predictive value; AUC, area under the ROC curve; F1, F1-score; NPV, negative predictive value; MCC, Matthews correlation coefficient.