| Literature DB >> 31439010 |
Duncan Shillan1,2, Jonathan A C Sterne1,2, Alan Champneys3, Ben Gibbison4,5,6.
Abstract
BACKGROUND: Intensive care units (ICUs) face financial, bed management, and staffing constraints. Detailed data covering all aspects of patients' journeys into and through intensive care are now collected and stored in electronic health records: machine learning has been used to analyse such data in order to provide decision support to clinicians.Entities:
Keywords: Artificial intelligence; Intensive care unit; Machine learning; Routinely collected data
Mesh:
Year: 2019 PMID: 31439010 PMCID: PMC6704673 DOI: 10.1186/s13054-019-2564-9
Source DB: PubMed Journal: Crit Care ISSN: 1364-8535 Impact factor: 9.097
Fig. 1PRISMA 2009 flow diagram of study review process and exclusion of papers. From [11]
Number and proportion of papers according to the aim of study and number of patients analysed
| Number of patients analysed | |||||||
|---|---|---|---|---|---|---|---|
| Aim of study | Number (%) of papers with this aima | < 100 | 100–1000 | 1000–10,000 | 10,000–100,000 | 100,000–1,000,000 | Number not reported |
| Predicting complications | 79 (30.6%) | 23 (29.1%) | 26 (32.9%) | 17 (21.5%) | 8 (10.1%) | 3 (3.8%) | 2 (2.5%) |
| Predicting mortality | 70 (27.1%) | 11 (15.7%) | 19 (27.1%) | 19 (27.1%) | 18 (25.7%) | 1 (1.4%) | 2 (2.9%) |
| Improving prognostic models/risk scoring system | 43 (16.7%) | 8 (18.6%) | 16 (37.2%) | 8 (18.6%) | 8 (18.6%) | 2 (4.7%) | 1 (2.3%) |
| Classifying sub-populations | 29 (11.2%) | 11 (37.9%) | 8 (27.6%) | 6 (20.7%) | 2 (6.9%) | 0 (0.0%) | 2 (6.9%) |
| Alarm reduction | 21 (8.14%) | 9 (42.9%) | 5 (23.8%) | 7 (33.3%) | 0 (0.0%) | 0 (0.0%) | 0 (0.0%) |
| Predicting length of stay | 18 (6.98%) | 3 (16.7%) | 7 (38.9%) | 5 (27.8%) | 3 (16.7%) | 0 (0.0%) | 0 (0.0%) |
| Predicting health improvement | 17 (6.59%) | 5 (29.4%) | 10 (58.8%) | 2 (11.8%) | 0 (0.0%) | 0 (0.0%) | 0 (0.0%) |
| Determining physiological thresholds | 16 (6.20%) | 10 (62.5%) | 4 (25.0%) | 1 (6.2%) | 0 (0.0%) | 0 (0.0%) | 1 (6.2%) |
| Improving upon previous methods | 5 (1.94%) | 2 (40.0%) | 1 (20.0%) | 1 (20.0%) | 1 (20.0%) | 0 (0.0%) | 0 (0.0%) |
| Detecting spurious recorded values | 3 (1.16%) | 1 (33.3%) | 2 (66.7%) | 0 (0.0%) | 0 (0.0%) | 0 (0.0%) | 0 (0.0%) |
| Total (accounting for duplicates) | 258 | 72 (27.9%) | 84 (32.6%) | 55 (21.3%) | 35 (13.6%) | 6 (2.33%) | 6 (2.33%) |
aWhere papers had more than one aim, all aims were recorded, so percentages may total more than 100
Number and proportion of papers according to the type of machine learning used and number of patients analysed (for prediction studies only)
| Number of patients analysed | |||||||
|---|---|---|---|---|---|---|---|
| Type of machine learning | Number (%) of papers with this typea | < 100 | 100–1000 | 1000–10,000 | 10,000–100,000 | 100,000–1,000,000 | Number not reported |
| Neural network | 72 (42.6%) | 14 (19.4%) | 27 (37.5%) | 20 (27.8%) | 9 (12.5%) | 2 (2.8%) | 0 (0.0%) |
| Support vector machine | 40 (23.7%) | 12 (30.0%) | 15 (37.5%) | 8 (20.0%) | 4 (10.0%) | 1 (2.5%) | 0 (0.0%) |
| Classification/decision trees | 35 (20.7%) | 6 (17.1%) | 11 (31.4%) | 10 (28.6%) | 5 (14.3%) | 1 (2.9%) | 2 (5.7%) |
| Random forest | 21 (12.4%) | 1 (4.8%) | 9 (42.9%) | 5 (23.8%) | 4 (19.0%) | 2 (9.5%) | 0 (0.0%) |
| Naive Bayes/Bayesian networks | 19 (11.2%) | 4 (21.1%) | 5 (26.3%) | 6 (31.6%) | 2 (10.5%) | 1 (5.3%) | 1 (5.3%) |
| Fuzzy logic/rough set | 12 (7.1%) | 3 (25.0%) | 5 (41.7%) | 2 (16.7%) | 1 (8.3%) | 0 (0.0%) | 1 (8.3%) |
| Other techniquesb | 28 (16.7%) | 2 (7.1%) | 10 (35.7%) | 8 (28.6%) | 7 (25.0%) | 1 (3.6%) | 0 (0.0%) |
| Total (accounting for duplicates) | 169 | 37 (21.9%) | 56 (33.1%) | 42 (24.9%) | 26 (15.4%) | 4 (2.37%) | 4 (2.37%) |
aPapers can have more than one approach—percentages may total more than 100
bOther techniques (number of studies): causal phenotype discovery (1), elastic net (1), factor analysis (1), Gaussian process (2), genetic algorithm (1), hidden Markov models (1), InSight (4); JITL-ELM (1), k-nearest neighbour (3), Markov decision process (1), particle swarm optimization (1), PhysiScore (1), radial domain folding (1), sequential contrast patterns (1), Superlearner (4), switching linear dynamical system (1), Weibull-Cox proportional hazards model (1), method not described (2)
Fig. 2Number of papers published according to the sample size and year of publication
Fig. 3Number of papers published according to the type of machine learning and year of publication
Number and proportion of papers according to outcome predicted and approach to validation (for prediction studies only)
| Approach to validationb | |||||||
|---|---|---|---|---|---|---|---|
| Outcome predicted | Total papersa | Validated | Independent data | Leave-P-out | Randomly selected subset | Otherb | |
| Complications | 79 (46.7%) | 73 (92.4%) | 5 (6.85%) | 5 (6.85%) | 33 (45.2%) | 30 (41.1%) | 0 (0%) |
| Mortality | 70 (41.4%) | 68 (97.1%) | 5 (7.35%) | 3 (4.41%) | 33 (48.5%) | 27 (39.7%) | 0 (0%) |
| Length of stay | 18 (10.7%) | 18 (100%) | 3 (16.7%) | 1 (5.56%) | 4 (22.2%) | 10 (55.6%) | 1 (5.6%) |
| Health improvement | 17 (10.1%) | 16 (94.1%) | 0 (0%) | 1 (6.25%) | 5 (31.2%) | 10 (56.2%) | 0 (0%) |
| Total (accounting for duplicates) | 169 | 161 (94.1%) | 10 (6.2%) | 8 (5%) | 71 (44.1%) | 71 (44.1%) | 1 (0.6%) |
aPapers can have more than one approach, so percentages may total more than 100
b“Other” techniques (number of studies): a comparison between ML and decisions made by clinicians (1)
Number and proportion of papers according to outcome predicted and measure of predictive accuracy reported (for studies that validated predictions)
| Measure of predictive accuracy reporteda | ||||||
|---|---|---|---|---|---|---|
| Outcome predicted | Total papers | AUC and accuracy/sensitivity/specificity | AUC only | Accuracy/sensitivity/specificity only |
| Otherb |
| Complication | 73 (45.3%) | 24 (32.9%) | 17 (23.3%) | 28 (38.4%) | 4 (5.5%) | |
| Mortality | 68 (42.2%) | 16 (23.5%) | 31 (45.6%) | 18 (26.5%) | 3 (4.4%) | |
| Length of stay | 18 (11.1%) | 2 (11.8%) | 3 (16.7%) | 5 (27.8%) | 8 (44.4%) | 1 (5.6%) |
| Health improvement | 16 (10%) | 1 (6.3%) | 3 (18.8%) | 11 (68.8%) | 1 (6.3%) | |
| Total | 161 | 43 (26.7%) | 54 (33.5%) | 62 (38.5%) | 8 (5.0%) | 9 (5.6%) |
aPapers can have more than one approach, so percentages may total more than 100. The total of these columns does not account for duplicates as papers can fluctuate how they discuss different results
b“Other” measures of predictive accuracy (number): congruence of ML and clinician’s decisions (1), Matthews correlation coefficient (1), mean absolute differences between observed and predicted (1), mean error rate (1), MSE as loss function (1), Pearson correlation between estimate and actual (1), ratio of wins vs loses against logistic regression (1), rules developed by ML (1)
Fig. 4Boxplots showing the distribution of AUC scores according to the size of dataset, for all studies and separately for studies predicting mortality and complications. Numbers displayed are the median AUC for each group. A cross indicates the AUC of one of the 10 papers using independent test data. We did not plot results for studies predicting the length of stay and health improvement because the numbers of such studies were small
Fig. 5Comparison of AUC scores found in complication or mortality prediction papers according to the technique used to produce them. A line of equality is also provided