Literature DB >> 28791144

Machine learning landscapes and predictions for patient outcomes.

Ritankar Das1, David J Wales2.   

Abstract

The theory and computational tools developed to interpret and explore energy landscapes in molecular science are applied to the landscapes defined by local minima for neural networks. These machine learning landscapes correspond to fits of training data, where the inputs are vital signs and laboratory measurements for a database of patients, and the objective is to predict a clinical outcome. In this contribution, we test the predictions obtained by fitting to single measurements, and then to combinations of between 2 and 10 different patient medical data items. The effect of including measurements over different time intervals from the 48 h period in question is analysed, and the most recent values are found to be the most important. We also compare results obtained for neural networks as a function of the number of hidden nodes, and for different values of a regularization parameter. The predictions are compared with an alternative convex fitting function, and a strong correlation is observed. The dependence of these results on the patients randomly selected for training and testing decreases systematically with the size of the database available. The machine learning landscapes defined by neural network fits in this investigation have single-funnel character, which probably explains why it is relatively straightforward to obtain the global minimum solution, or a fit that behaves similarly to this optimal parameterization.

Entities:  

Keywords:  energy landscapes; machine learning; neural networks; patient mortality

Year:  2017        PMID: 28791144      PMCID: PMC5541539          DOI: 10.1098/rsos.170175

Source DB:  PubMed          Journal:  R Soc Open Sci        ISSN: 2054-5703            Impact factor:   2.963


Introduction

There is an increasing need in the hospital setting for methods to predict the mortality risk and decompensation likelihood of patients [1]. Early identification of patient deterioration can assist provider organizations to properly manage and treat patients in a timely manner, to improve outcomes and allocate limited operational and financial resources efficiently [2,3]. With the rise in prevalence of electronic health records (EHRs) for clinical practice (over 75% of USA hospitals have a basic EHR system [4], while approximately 92% of UK hospitals use some form of EHR [5]) comes the opportunity to obtain clinically relevant patient data, such as vital signs and laboratory results, for patient instability forecasting-based decision support [6]. Additionally, with EHRs, patient health information is all centralized in one place, increasing ease of analysis with computational tools. In this study, our aim is to predict the likelihood of mortality for critical care patients using approaches and analyses that build on the theory of potential energy landscapes in molecular science. Patient records were obtained from the Multiparameter Intelligent Monitoring in Intensive Care (MIMIC) III clinical database, which is a comprehensive dataset that contains clinical information of patients admitted to intensive care units [7]. Previous studies spanning epidemiology, clinical decision rules development, and electronic patient monitoring [8] have employed this database to investigate a broad range of areas in addition to mortality forecasting, including sepsis [9-11], acute pancreatitis [12] and extubation failure (the inability to sustain spontaneous breathing after removal of an artificial airway) [13]. Prior work on mortality prediction largely involves the use of clinical decision support systems (CDSSs), which are information systems created for the purpose of improving clinical decision-making [14]. While CDSSs have great potential to transform clinical care, there are some implementation and design issues for present systems [15]. Most current prediction models are based on a synthesis of patient-specific information and physiological parameters [16]. This formulation means that the models make the assumption that risk factors are independent of one another, which makes the patient mortality prediction less sensitive, and therefore less accurate. Among the most popular CDSSs are the modified early warning score (MEWS), the sequential organ failure assessment (SOFA) and the simplified acute physiology score (SAPS II), which have only moderate accuracy, and have not proved to be especially beneficial for clinical use [2,17]. Our approach in the present contribution is to apply the conceptual and computational tools of potential energy landscape theory to the landscape defined by neural network fits to patient data for prediction of mortality. For molecules and condensed matter, the underlying potential energy surface in a specific electronic state encodes its dynamics, thermodynamics and structure. The lowest energy structure is the global minimum of the potential energy surface. Location of this global minimum can often be accomplished efficiently using global optimization techniques. In this study, likely candidates for global minima of neural network fits are located using basin-hopping global optimization [18-20]. For the neural network fitting, the cost function that is minimized is derived from the difference between the predicted and actual patient outcomes. Generally in the case of non-convex cost functions, there will be multiple local minima [21], and we can employ all the potential energy landscape methodologies to investigate emergent properties of the resulting machine learning landscape. Here, we illustrate how some of these tools from the potential energy landscape framework can be applied to analyse machine learning landscapes defined by a cost function involved in fitting to training data. Our aim is to showcase how energy landscape approaches can be applied to time series of medical data, and offer new perspectives for future research that may lead to improved predictive methods.

Machine learning fits and predictions

The first fitting functions considered were three-layer neural networks (figure 1) containing input, output and hidden nodes [22]. We employed a single hidden layer in this initial survey in view of successful previous applications [23], and a bias was added to the sum of weights used in the activation function for each hidden node, , and each output node, wbo [22]. The inputs correspond to patient data for a selected set of vital signs and laboratory measurements, including values from one or more of the 48 h time windows. We, therefore, consider a variable number of inputs, Nin, for each patient, data point α, represented as , with the complete input dataset written as X={x1,…,x}.
Figure 1.

A three-layer neural network with four inputs, five hidden nodes and two outputs. The training variables are the link weights, and , and the bias weights, w and wbo.

A three-layer neural network with four inputs, five hidden nodes and two outputs. The training variables are the link weights, and , and the bias weights, w and wbo. There are Nout=2 outputs, 0 and 1, corresponding to death in hospital and survival, which are calculated as for each patient dataset x and network parameters between hidden node j and output i, and between input k and hidden node j, and bias weights wbh and wbo. To reduce the effect of outliers, the two outputs were transformed into softmax probabilities as To train each network, we minimize the cost (objective) function, ENN(W;X), with respect to the variables , , and wbo, written collectively as a vector of weights W. Basin-hopping global optimization was used [18-20] to search for the global minimum, and all the distinct minima obtained during these searches were saved for later comparison. In this approach, we take steps between local minima of the cost function, accepting or rejecting moves according to a simple Metropolis criterion [24] based upon the change in cost function, scaled by a parameter that plays the role of temperature. Downhill moves are always accepted, and the probability of accepting an uphill move depends on the fictitious temperature [18-20]. For the machine learning landscapes considered in the present work, locating the global minimum is straightforward, and the choice of basin-hopping parameters is not critical. A customized L-BFGS optimization routine was employed for local minimization, based on the limited memory version [25,26] of the quasi-Newton Broyden [27], Fletcher [28], Goldfarb [29], Shanno [30], BFGS procedure. Analytic first and second derivatives (including the regularization terms defined below) were programmed for both EQ(W;X) and ENN(W;X) in the public domain GMIN and OPTIM codes for exploration of the corresponding machine learning landscapes. Some further details are provided in §3, and a review of the energy landscape perspective in the context of machine learning has recently appeared [31]. For each patient, α, we know the actual outcome c(α)=0 or 1, and the cost function was written as which includes an L2 regularization term weighted by a specified coefficient λ>0. This term is designed to reduce overfitting, biasing against large values for individual variables. It also changes the zero eigenvalue, which would otherwise result from a uniform shift in all the wbo, to 2λ. The quality of predictions obtained from fits corresponding to local minima of ENN(W;Xtrain) is judged from the corresponding values of ENN(W;Xtest), and especially from the area under the curve (AUC) values calculated from the probabilities . We have compared selected neural network results with those obtained from a quadratic function of the inputs, which is a convex function: which has Nout(1+Nin(Nin+3)/2) variables that we again write collectively as W. Probabilities were defined as with cost function

Machine learning and energy landscapes

In several recent contributions, we have shown how the energy landscape framework developed for molecular and condensed matter can be applied to cost functions that support multiple minima (non-convex functions) in machine learning [32,33]. Further discussion can be found in an overview article [31], which attempts to draw together various research strands that may be related using the ideas and computational tools of energy landscape theory [34]. Details of the methods employed to explore the machine learning landscape can be found in earlier work; these techniques are well established for applications to atomistic (and coarse-grained model) systems, and here we simply summarize the present calculations. To identify the global minimum for the neural network fits, we employed basin-hopping [18-20] as implemented in the GMIN program [35], with a modified L-BFGS minimization algorithm [25] (limited memory quasi-Newton Broyden [36], Fletcher [37], Goldfarb [38], Shanno [39] approach). The convergence condition for each minimization was a root mean square gradient below 10−6 reduced units, with tighter convergence of 10−10 for the distinct minima at the end of each run. This threshold is sufficiently tight to distinguish different local minima by simply comparing the cost function value. Finding the global minimum for these machine learning landscapes is generally straightforward, and so a large fixed step size and temperature parameter in the accept/reject step were employed to obtain a survey of local minima. Relatively short runs of 1000 basin-hopping steps were used in each case, guided by tests in previous work [32,33]. The present results surveyed a wide range of input data and alternative fits, requiring nearly 40 000 basin-hopping global optimization runs for the neural network fits, and over 10 000 local minimizations for the convex quadratic function. These tests correspond to different combinations of patient data items, different regularization parameters λ, different time windows for the patient data, and hidden layers with alternative numbers of hidden nodes. Knowledge of local minima is sufficient to gauge the predictive power of different fitting functions, but it does not define a landscape. To achieve this more detailed level of description, we must understand how the local minima are connected by transition states (saddle points of index one [40]). This analysis was performed for selected test cases, one of which is reported in more detail below. The results are not only of fundamental interest for comparison with molecular landscapes but also provide insight into the location of different minima by local minimization, and the efficiency of global optimization. All the transition state searches used the doubly-nudged [41,42] elastic band [43,44] approach to identify candidates for accurate refinement with hybrid eigenvector-following [45,46]. We employed the OPTIM [47] program to characterize transition states and pathways, and the PATHSAMPLE [48] program to organize the overall exploration of each connected landscape. The ability to visualize the landscape has played a key role in identifying molecular and condensed matter systems with the ability to self-organize, and distinguishing them from glass-formers [34]. Here, we employ disconnectivity graphs [49,50], where efficient relaxation to the global minimum is associated with single-funnel landscapes [34]. Most of the machine learning landscapes we have investigated so far seem to fall into this category [32,33], and a further example is presented in §5. Once again, the machine learning landscapes we have characterized in the present investigation appear to be funnelled, which explains why global optimization, or locating low-lying minima of the cost function, is relatively easy.

Input data

In total, 53 211 patient records were employed from the MIMIC III clinical database [51], which consists of anonymized data for patients at the Beth Israel Deaconess Medical Center (BIDMC) in Boston. The Institutional Review Boards of BIDMC and the Massachusetts Institute of Technology waived the requirement for individual patient consent, as our study does not impact clinical care and all data were deidentified. Each record analysed from MIMIC III contains a collection of up to 33 different clinical measurements, including vital signs, for a 48 h period of hospitalization. For each patient, there is also a record of two possible outcomes, namely death in hospital and early discharge. In this study, we aimed to predict death in hospital, corresponding to 3879 patients. Of these records, 458 were removed, since the recorded outcomes were contradictory, with early discharge also reported. Vital sign and laboratory measurements without times were removed (1.6% and 1.2%, respectively). The number of patients with records of each type and the abbreviations employed are summarized in table 1.
Table 1.

The 33 distinct types of vital sign and laboratory measurements contained in the MIMIC III database. In each case, the abbreviation is listed together with the number of patients, in increasing order.

datadescriptionno. patients
HCThaematocrit46
troponintroponin792
cholesterolcholesterol2721
ALPalkaline phosphatase7861
FiO2fraction of inspired oxygen11 343
PaO2partial pressure of oxygen in blood12 666
PaCO2partial pressure of carbon dioxide in blood12 674
albuminalbumin12 787
ALTalanine aminotransferase14 959
bilirubinbilirubin17 019
lactatelactate17 437
ADBPambulatory diastolic blood pressure17 975
ASBPambulatory systolic blood pressure17 977
HCO3bicarbonate18 467
SpO2peripheral capillary oxygen saturation21 745
pHpH22 417
urineurine output23 147
SaO2oxygen saturation of arterial blood28 849
INRinternational normalized ratio31 266
NIDBPnon-invasive diastolic blood pressure34 384
NISBPnon-invasive systolic blood pressure34 432
Mgmagnesium35 425
BUNblood urea nitrogen37 032
creatininecreatinine37 044
TEMPtemperature37 082
GCSGlasgow coma scale37 270
RRrespiratory rate37 283
glucoseglucose39 133
Nasodium39 376
Kpotassium39 564
WBCwhite blood cell count41 445
plateletsplatelets41 971
HRheart rate43 740
The 33 distinct types of vital sign and laboratory measurements contained in the MIMIC III database. In each case, the abbreviation is listed together with the number of patients, in increasing order. To treat data obtained at irregular time intervals, the measurements were averaged over the 48 time windows corresponding to consecutive hours, ordered backwards in time from the most recent results (index 1). Empty entries were set to the earliest non-zero average available. For each data type, an average value was calculated over all patients for every time range considered, and the inputs were scaled to shift the mean to unity in each case. This ‘feature scaling’ puts the inputs into a standardized range and the local minimizations that are run in the subsequent fitting exhibit better convergence properties. Training and testing were performed by randomly dividing the patients into two disjoint sets of equal size (with one more entry for testing when the total number of patients was odd). For each local minimum α of ENN(W;Xtrain) or EQ(W;Xtrain), the quality of the predictions was judged from the area under the receiver operating characteristic (ROC) curve obtained with the testing data as ENN(W;Xtest) or EQ(W;Xtest). The ROC curve is a plot of the true positive rate, T, against the false positive rate, F, as a function of the threshold probability, P, for making a certain classification. P is the threshold at which the output probability pNN0(W;X) or pQ0(W;X) is considered large enough to predict that outcome 0 (death in hospital) will occur. Hence, where Θ is the Heaviside step function, δ is the Kronecker delta and c(d) is the outcome (class label, 0 or 1) for patient d. The area under the curve is obtained by numerical integration of The AUC value is interpreted as the probability that the prediction will discriminate correctly between two patients chosen at random from the sets with different outcomes. AUC values between 0.7 and 0.8, 0.8 and 0.9, and 0.9 and 1 are often described as fair, good and excellent, respectively.

Results

An extensive set of tests was first run for the 33 individual medical data items (referred to as measurements below) using neural network fits over various time ranges, with 3, 4, 5 and 6 hidden nodes and λ=10−5. In total, 13 299 global optimization runs were performed for these combinations. Table 2 summarizes the time interval corresponding to the maximum AUC achieved in testing for each of the hidden node values. This survey revealed that changing the number of hidden nodes did not have much effect. The results where the highest AUC values approaching 0.7 are achieved generally correspond to readings taken near the end of the 48 h period. The effect of varying the time interval was investigated systematically for combinations of input data where better AUC values are achieved, as discussed below. Hence, further details of the results for single data types are omitted for brevity. The main purpose of this initial survey was to guide the selection of input data for improved predictions.
Table 2.

AUC values for test set predictions of patient outcome using one vital sign or laboratory data item and neural network fits for 3, 4, 5 and 6 hidden nodes. These are the results for the highest AUC values for testing, sorted for 6 hidden nodes. Fitting and testing were performed for data obtained in each hour, ranging from the last (index 1) to the first (index 48), for three consecutive hours 1–3, 2–4, …, 46–48, (again counting backwards in time, so that hour 1 is the last entry, i.e. the most recent), and for hours 1–2, 1–4, 1–6, 1–8, 1–10, 1–12 and 1–24. For the longest range 1–24, the network with 6 hidden nodes was not considered. In each case, the regularization parameter λ was fixed at 10−5. The maximum AUC values are italicized.

hidden nodes3
4
5
6
data itemstimeAUCtimeAUCtimeAUCtimeAUCNtraindataNtestdata
troponin34–360.48511–130.51234–360.50334–360.489396396
HCT1–80.5001–80.5001–80.5001–80.5002323
GCS1–240.5901–240.5821–240.5741–120.51618 63518 635
pH1–240.5381–240.5341–240.53439–390.52611 20811 209
SaO239–390.53239–390.53239–390.53239–390.53214 42414 425
Mg10–120.5441–240.54510–120.54410–120.54417 71217 713
ALT18–200.56648–480.56547–470.56216–180.55174797480
ALP20–220.58623–250.56123–250.55648–480.55139303931
cholesterol1–80.5541–240.5551–100.5551–100.55313601361
SpO248–480.5443–50.5563–50.5573–50.55810 87210 873
bilirubin13–150.5804–60.5691–40.56841–410.56285098510
K4–60.5521–120.5611–120.5611–120.56619 78219 782
TEMP1–30.5701–30.5681–30.5691–30.57018 54118 541
NISBP1–120.5761–60.5781–80.5751–80.57817 21617 216
Na1–30.5821–30.5821–30.5821–30.58219 68819 688
WBC1–240.6061–240.6411–240.6421–40.58420 72220 723
FiO2 1–100.6221–240.6121–120.61327–270.59056715672
platelets1–240.6321–240.6311–240.6381–120.59820 98520 986
albumin41–410.58035–370.59436–380.60623–250.60263936394
lactate1–100.55331–310.59631–310.60331–310.60287188719
PaO2 46–460.60546–460.60446–460.60446–460.60463336333
NIDBP1–20.6051–20.6061–20.6051–20.60517 19217 192
HR1–30.6031–30.6041–240.6111–30.60721 87021 870
glucose1–30.5981–120.6211–120.6201–120.61919 56619 567
INR1–30.61920.62020.61920.61915 63315 633
PaCO2 20.62720.62720.62720.62763376337
creatinine10.6271–240.6328–80.6288–80.62818 52218 522
RR42–440.63044–440.62940–420.62940–420.62918 64118 642
ASBP1–80.6341–100.6351–60.6411–60.63989888989
ADBP3–50.6463–50.6353–50.6403–50.64689878988
urine1–30.6641–30.6661–30.6591–30.66811 57311 574
BUN1–30.6971–30.6971–30.6961–30.69718 51618 516
HCO3 2–40.7002–40.7002–40.7002–40.70092339234
AUC values for test set predictions of patient outcome using one vital sign or laboratory data item and neural network fits for 3, 4, 5 and 6 hidden nodes. These are the results for the highest AUC values for testing, sorted for 6 hidden nodes. Fitting and testing were performed for data obtained in each hour, ranging from the last (index 1) to the first (index 48), for three consecutive hours 1–3, 2–4, …, 46–48, (again counting backwards in time, so that hour 1 is the last entry, i.e. the most recent), and for hours 1–2, 1–4, 1–6, 1–8, 1–10, 1–12 and 1–24. For the longest range 1–24, the network with 6 hidden nodes was not considered. In each case, the regularization parameter λ was fixed at 10−5. The maximum AUC values are italicized. Of the 33 different vital sign and laboratory measurements, 22 were considered in the next phase of calculations, where all 231 combinations of two distinct measurements were used as input data. This subset was selected based on the performance of the individual inputs and the number of data records available. Neural networks with 3, 5 and 7 hidden nodes were considered for time windows corresponding to the last hour (index 1) and the last two, three, six and twelve hours (ranges 1–2, 1–3, 1–6 and 1–12). A total of 3465 basin-hopping global optimization runs were therefore performed, all for λ=10−5. The maximum AUC values observed for any time window with each of the three hidden layers are collected in table 3 for the best pairs of input data, together with the number of patients in the training and test sets. The best AUC values increase from 0.7 to 0.8 when two measurements are considered in the fits, and again there is little dependence on the number of hidden nodes. There is no significant advantage in using memory of measurements taken before the final data collection in the last hour.
Table 3.

AUC values for test set predictions of patient outcome using two vital sign or laboratory data items and neural network fits for 3, 5 and 7 hidden nodes. A total of 231 different combinations of three data items were considered; these are the results for the highest AUC values, sorted on the AUC values for 7 hidden nodes. Results were obtained for time ranges of 1 and 1–2, 1–3, 1–6 and 1–12 h using λ=10−5, and the highest AUC value is reported in each case, together with the corresponding time range. The maximum AUC values obtained are highlighted in italics for each choice of hidden nodes.

hidden nodes3
5
7
data itemstimeAUCtimeAUCtimeAUCNtraindataNtestdata
urine INR1–120.7081–60.7011–30.70083488348
glucose BUN1–120.7041–60.6991–60.70018 36518 365
WBC BUN10.70210.70410.70318 30818 308
pH BUN10.70310.70310.70310 34710 348
INR BUN1–20.70310.70410.70415 54215 542
RR BUN1–120.7091–60.7071–20.70418 27818 279
platelets BUN10.7071–20.7081–20.70518 36218 363
urine HR1–120.7061–120.7101–30.70511 57111 572
NISBP BUN1–60.7081–20.70610.70616 91416 914
FiO2 BUN10.71110.71310.70746264626
bilirubin BUN10.71110.70910.70772457246
urine FiO210.66810.68210.71113971398
NIDBP BUN1–30.7121–30.71210.71116 89116 892
SpO2 BUN10.71210.71310.71210 62210 623
HCO3 BUN1–60.7141–20.7121–30.71292299230
creatinine BUN1–20.7161–20.7181–20.71718 51318 513
urine bilirubin10.72210.72210.72146584658
PaCO2 BUN1–20.7231–20.72210.72262056205
HR BUN1–60.7241–20.7231–20.72418 44318 443
HCO3 ASBP1–30.7191–30.7251–60.72437913792
BUN ADBP1–60.7381–30.7461–30.73788858885
urine BUN1–30.7361–30.7411–30.73898309831
pH GCS10.73610.7361–30.73910 24310 243
glucose GCS10.7371–60.74310.74118 36718 368
BUN ASBP1–60.7451–60.7481–60.74588868886
NISBP GCS10.75110.7481–20.74717 16717 167
GCS bilirubin1–20.7621–30.7591–30.75070887089
NIDBP GCS10.75110.7521–20.75217 14317 144
HR GCS10.7611–20.75910.76018 58218 582
platelets GCS1–60.7511–30.76410.76118 17418 175
WBC GCS1–30.75410.7641–30.76318 09618 096
SpO2 GCS1–20.7591–30.7651–30.7671083310834
PaCO2 GCS10.77110.77110.77062276228
RR GCS10.7661–30.7711–20.77018 56018 561
TEMP GCS1–60.7651–60.7691–20.77018 45918 459
GCS ADBP1–30.77610.77610.77389418942
PaO2 GCS10.77510.77810.77762246225
GCS ASBP10.77710.78110.78089428943
GCS creatinine1–30.77310.78310.78018 24418 244
INR GCS10.78310.78110.78415 51215 512
GCS FiO2 10.77910.78210.78745354536
urine GCS10.79310.79510.79398699870
HCO3 GCS10.7981–20.7961–30.79491669166
GCS BUN1–30.79510.80110.80218 23818 239
AUC values for test set predictions of patient outcome using two vital sign or laboratory data items and neural network fits for 3, 5 and 7 hidden nodes. A total of 231 different combinations of three data items were considered; these are the results for the highest AUC values, sorted on the AUC values for 7 hidden nodes. Results were obtained for time ranges of 1 and 1–2, 1–3, 1–6 and 1–12 h using λ=10−5, and the highest AUC value is reported in each case, together with the corresponding time range. The maximum AUC values obtained are highlighted in italics for each choice of hidden nodes. All three possible combinations were then considered for the same 22 measurements as for the doublet combinations, with time intervals 1 and 1–2 and hidden layers of 3 and 5 nodes. These 1540 triplet combinations therefore required 6160 global optimization runs. In this case, the results were compared with the quadratic fitting function defined in equation (2.4), for time ranges 1–2, 1–3, 1–6 and 1–12, with λ=10−5. The results for the best combinations, sorted on the neural net values, and the time interval and number of hidden nodes they were achieved for, are collected in table 4. We see that using three data types raises the best AUC values to around 0.85, which is clearly a worthwhile improvement. The correlation between the best neural network and quadratic fit predictions is analysed further below.
Table 4.

AUC values for test set predictions of patient outcome using three vital sign or laboratory data items. A total of 1540 different combinations of three data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values. For the quadratic fitting function, results were obtained for time ranges of 1 and 1–2, 1–3, 1–6 and 1–12 h. For the neural network fits, we considered time ranges of 1 and 1–2 with 3 and 5 hidden nodes (N). The highest AUC values obtained for the test data amongst all the local minima obtained in training are reported in each case for λ=10−5 and the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function.

neural network fit
quadratic fit
data itemstimeNhAUCtimeAUCNtraindataNtestdata
WBC urine GCS130.8001–30.78995699570
platelets HCO3 GCS1–250.8001–60.74991119112
INR GCS ADBP150.8011–20.72380918092
urine GCS ADBP150.8011–20.80655205521
GCS creatinine ASBP150.80210.70588518851
INR GCS ASBP150.8021–20.71280928093
PaO2 FiO2 creatinine1–230.80210.649393393
TEMP GCS BUN150.8031–20.71518 12118 121
pH GCS BUN1–250.8031–20.73110 15310 154
GCS BUN bilirubin1–250.8041–120.76970757075
RR GCS BUN150.8061–20.72318 21818 219
urine GCS FiO2 130.80610.808348348
urine platelets GCS150.80610.79596459645
INR GCS BUN1–250.8071–20.72515 44815 448
PaO2 GCS BUN130.8071–20.80561546154
NIDBP GCS BUN150.8071–30.72216 85616 856
urine INR GCS130.8071–30.80783308331
GCS creatinine BUN150.80810.71618 23618 237
WBC GCS BUN150.80810.72418 05218 053
PaCO2 GCS ADBP130.80910.80450015002
platelets GCS BUN150.80910.71818 10418 105
urine GCS ASBP130.8101–30.81155205520
PaCO2 GCS BUN150.81110.80861576158
SpO2 GCS BUN150.8111–30.80010 59810 598
NISBP GCS BUN150.81410.72016 87816 879
GCS FiO2 ASBP130.81410.80429342935
urine HCO3 GCS1–230.81510.803943943
GCS BUN ADBP150.8171–30.74588478848
HR GCS BUN150.8181–20.73318 23818 239
HCO3 GCS BUN150.81910.78691629163
SpO2 GCS FiO2 130.81910.829627627
GCS FiO2 ADBP130.82110.79129342934
urine GCS BUN1–250.8221–30.82196729673
GCS BUN ASBP150.8221–30.75988488849
GCS FiO2 BUN130.8231–120.82044814481
PaCO2 GCS FiO2 1–230.82710.839413413
HCO3 GCS ASBP1–230.8301–30.78537713771
HCO3 GCS ADBP130.83910.79137703770
PaO2 GCS FiO2 130.85310.848413413
AUC values for test set predictions of patient outcome using three vital sign or laboratory data items. A total of 1540 different combinations of three data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values. For the quadratic fitting function, results were obtained for time ranges of 1 and 1–2, 1–3, 1–6 and 1–12 h. For the neural network fits, we considered time ranges of 1 and 1–2 with 3 and 5 hidden nodes (N). The highest AUC values obtained for the test data amongst all the local minima obtained in training are reported in each case for λ=10−5 and the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function. The number of distinct measurements was further reduced to 14 for calculations involving all possible quartet combinations. For the 1001 quartets, we considered time ranges 1 and 1–2 h, with 3, 5 and 7 hidden nodes for the neural net fits, requiring 6006 global optimization runs. In this case, we compared the quadratic fits for the same time ranges, along with regularization parameters of λ=10−5 and 10−6 (table 5). The results show that the maximum AUC values do not improve significantly from the best triplet combinations in table 4, and that there is no significant difference for the two values of the regularization parameter.
Table 5.

AUC values for test set predictions of patient outcome using four vital sign or laboratory data items. A total of 1001 different combinations of four data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values for λ=10−6. Results were obtained for two λ values, time ranges of 1 and 1–2, and 3, 5 and 7 hidden nodes (N) for the neural network fits. The highest AUC values for the test data among all the local minima obtained in training are reported in each case, along with the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function and λ value.

neural network fit
quadratic fit
λ=10−5
λ=10−6
λ=10−5
λ=10−6
data itemstimeNhAUCtimeNhAUCtimeAUCtimeAUCNtraindataNtestdata
RR INR HCO3 GCS150.8191–250.8181–20.7811–20.78177047704
platelets HCO3 GCS ASBP150.823150.8181–20.7811–20.77937643765
TEMP PaCO2 GCS BUN130.819130.81810.81110.81261576157
urine platelets PaCO2 GCS150.821150.81910.81010.81058705870
urine HCO3 GCS creatinine130.822130.82010.83610.838943943
HR GCS creatinine BUN170.820170.82010.73410.73418 23618 237
RR platelets GCS BUN170.817170.8201–20.7271–20.72718 08518 086
platelets INR GCS BUN170.823150.82010.73510.73515 41315 414
urine INR GCS BUN150.819170.82010.81210.81282838283
RR GCS creatinine BUN150.818170.8201–20.7231–20.72318 21618 217
RR HCO3 GCS BUN1–230.819150.8211–20.7941–20.79291589158
INR HCO3 GCS BUN150.823170.82210.78310.78377067706
RR PaCO2 GCS BUN130.821150.82210.81910.81961516152
platelets HR GCS BUN170.820150.82210.73710.73718 10418 105
pH GCS BUN ASBP150.8251–250.82210.77110.77177387738
HR HCO3 GCS BUN150.824170.82210.79110.79091629163
urine GCS creatinine ASBP130.825150.82310.82610.82654465446
RR HR GCS BUN170.821170.8231–20.7431–20.74118 21818 219
INR HCO3 GCS ASBP150.832150.82310.78610.78534453445
RR HCO3 GCS ASBP130.828130.82310.78010.77937683768
platelets HCO3 GCS BUN150.820170.82410.78210.78191089109
GCS creatinine BUN ASBP150.825150.82510.75710.75788478848
pH urine HCO3 GCS130.763130.82610.83510.825538538
RR GCS BUN ASBP150.828150.8261–20.7611–20.76088398839
TEMP platelets GCS BUN1–250.823150.8261–20.7251–20.72717 98717 987
HR GCS BUN ASBP150.822150.82610.75510.75488488849
urine RR GCS BUN150.828130.82710.82610.82796579658
HR HCO3 GCS ASBP130.828150.82710.79210.79237713771
urine RR HCO3 GCS130.821130.82810.84910.845943943
urine TEMP GCS BUN1–230.829150.82910.82410.82496719671
urine SpO2 GCS BUN130.828170.83010.82710.82696719672
SpO2 GCS BUN ASBP150.836150.83210.83010.83157835784
PaCO2 GCS BUN ASBP150.834130.8321–20.83210.83149474947
urine GCS creatinine BUN170.834170.83210.82810.82896719672
urine HR GCS BUN150.829150.83310.82710.82796729673
platelets GCS BUN ASBP150.834150.83310.76510.76588088808
urine PaCO2 GCS BUN150.834150.83410.83010.83058825883
INR GCS BUN ASBP150.834170.83410.77910.77880578058
urine platelets GCS BUN150.836150.83410.82310.82395949595
TEMP GCS BUN ASBP1–250.834150.83910.76910.77087338734
urine INR HCO3 GCS130.808130.84010.85810.858816816
urine GCS BUN ASBP170.844150.84010.83710.83754445444
HCO3 GCS BUN ASBP130.840130.8421–20.81310.81337703771
AUC values for test set predictions of patient outcome using four vital sign or laboratory data items. A total of 1001 different combinations of four data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values for λ=10−6. Results were obtained for two λ values, time ranges of 1 and 1–2, and 3, 5 and 7 hidden nodes (N) for the neural network fits. The highest AUC values for the test data among all the local minima obtained in training are reported in each case, along with the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function and λ value. To further test whether extended combinations of clinical and laboratory measurements might lead to improved predictions, a variety of additional inputs were considered, including up to 10 different measurements. Here, we again compared the neural network and quadratic fits, this time for λ=10−6. Results for 41 of these combinations are collected in table 6. Hidden layers of 3, 5, 7, 9 and 11 nodes were used for the neural networks, with time ranges of 1 and 1–2. The quadratic fits were run for time ranges 1, 1–2, 1–3, …, 1–12. The combination of extra measurements does not improve predictions significantly over the best quartets, which can be combined with most of the other data types to produce quintets that produce comparable predictions. The extra computational expense involved in adding more inputs, and fitting more parameters for longer time intervals and more hidden nodes, does not result in significant improvements.
Table 6.

AUC values for test set predictions of patient outcome using 5 to 10 vital sign or laboratory data items. A total of 45 different combinations of data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values. For the quadratic fitting function, results were obtained for time ranges of 1 and 1–2, 1–3, 1–6, 1–7, 1–8, 1–9, 1–10, 1–11 and 1–12 h. For the neural network fits, we considered time ranges of 1 and 1–2 with 3, 5, 7, 9 and 11 hidden nodes (N). The highest AUC values obtained for the test data among all the local minima obtained in training are reported in each case for λ=10−6 and the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function. In each case, or Ntraindata+1; the highest AUC values obtained for each fitting function are highlighted in italics.

neural network fit
quadratic fit
data itemstimeNhAUCtimeAUCNtraindata
urine platelets RR pH ASBP170.6551–40.6334836
urine platelets RR pH SpO2190.6581–20.6435938
urine HCO3 GCS ASBP HR130.67110.790483
urine platelets RR pH HCO31–230.67910.731538
urine platelets RR pH INR1–230.6791–20.6765362
urine platelets RR pH creatinine1–230.68010.6735906
urine HCO3 RR GCS ASBP BUN TEMP PaCO2 SpO2130.68310.741420
urine platelets RR pH PaCO2150.6971–20.6875881
urine platelets RR pH HR1–2110.713110.7535938
urine platelets RR pH HCO3 GCS ASBP BUN INR150.71410.789391
urine platelets RR pH TEMP1–290.7141–120.7455938
urine platelets RR pH HCO3 GCS ASBP BUN INR creatinine130.7211–30.659391
urine HCO3 GCS ASBP pH130.72310.760422
urine platelets RR pH BUN1–230.7261–20.7195906
urine HCO3 GCS ASBP RR1–2110.7311–40.76081
urine HCO3 GCS ASBP creatinine130.73310.766483
urine platelets RR pH HCO3 GCS ASBP1–230.74410.764421
urine HCO3 GCS ASBP INR130.75410.743439
urine platelets RR pH HCO3 GCS ASBP BUN1–230.75510.781421
urine platelets RR pH HCO3 GCS130.75610.782536
urine HCO3 RR GCS ASBP BUN TEMP PaCO2150.75710.764420
urine HCO3 RR GCS ASBP BUN TEMP PaCO2 SpO2 creatinine130.75710.777420
urine HCO3 GCS ASBP SpO2130.76210.767483
urine HCO3 GCS ASBP TEMP130.76310.780483
urine GCS BUN ASBP HCO3130.76410.832483
urine HCO3 RR GCS ASBP1–230.76510.803483
urine HCO3 RR GCS ASBP BUN TEMP130.77110.751483
urine HCO3 GCS ASBP platelets130.77910.827482
urine HCO3 RR GCS ASBP BUN1–2110.78810.861483
urine platelets RR pH GCS170.79410.7855921
urine HCO3 GCS ASBP PaCO2130.82010.805420
urine HCO3 GCS ASBP BUN130.82810.810483
urine GCS BUN ASBP platelets1–250.82810.8265410
urine GCS BUN ASBP PaCO2130.83110.8374796
urine GCS BUN ASBP SpO21–230.83410.8355443
urine GCS BUN ASBP HR150.83610.8335444
urine GCS BUN ASBP creatinine150.84010.8425443
urine GCS BUN ASBP TEMP1–230.8411–20.8435444
urine GCS BUN ASBP INR130.84210.8394942
urine GCS BUN ASBP pH190.8511–60.8424826
urine GCS BUN ASBP RR190.8561–70.8585437
AUC values for test set predictions of patient outcome using 5 to 10 vital sign or laboratory data items. A total of 45 different combinations of data items were considered; these are the results for the highest AUC values, sorted on the neural network AUC values. For the quadratic fitting function, results were obtained for time ranges of 1 and 1–2, 1–3, 1–6, 1–7, 1–8, 1–9, 1–10, 1–11 and 1–12 h. For the neural network fits, we considered time ranges of 1 and 1–2 with 3, 5, 7, 9 and 11 hidden nodes (N). The highest AUC values obtained for the test data among all the local minima obtained in training are reported in each case for λ=10−6 and the time range and number of hidden nodes they correspond to. The maximum AUC values obtained for any combination are highlighted in italics for each fitting function. In each case, or Ntraindata+1; the highest AUC values obtained for each fitting function are highlighted in italics. To provide some idea of the likely systematic error involved in these calculations, the runs were repeated for all the multiplet combinations reported in table 6 with 3 hidden nodes for the neural networks and measurements corresponding to the final hour (index 1). The patient data for training and testing were randomized differently for each repeat, and the mean and standard deviation of the AUC values obtained in testing were calculated for five (neural networks) or 10 (quadratic fits) independent choices. These results are collected in table 7. The standard deviation generally decreases for combinations with more patients, whether the predictions are good or not. For the best AUC values, the standard deviation is usually reassuringly small. Larger AUC values can sometimes be obtained for smaller datasets, but these are due to larger fluctuations for the smaller number of patients. In choosing the optimal combinations of patient data for predictions, we should be guided by results where the population that contributes to the measurements is as large as possible.
Table 7.

Mean and standard deviation of the AUC values for test set predictions of patient outcome using 4 to 10 vital sign or laboratory data items. A total of 45 different combinations of data items were considered for the last time interval (index 1 above) and 3 hidden nodes for the neural network fits, with λ=10−6. The statistics were obtained for five and 10 random training and testing selections for the neural network and quadratic fits, respectively.

neural network fit
quadratic fit
data itemsμσμσpatients
urine HCO3 GCS ASBP RR0.6170.0810.6690.081162
urine platelets RR pH HCO3 GCS ASBP BUN INR0.6720.0700.7220.049783
urine platelets RR pH HCO3 GCS ASBP BUN INR creatinine0.7200.0510.6880.082783
urine HCO3 GCS ASBP PaCO20.7820.0630.8000.037841
urine HCO3 RR GCS ASBP BUN TEMP PaCO20.7080.0420.7860.045841
urine HCO3 RR GCS ASBP BUN TEMP PaCO2 SpO20.7110.0370.7440.033841
urine HCO3 RR GCS ASBP BUN TEMP PaCO2 SpO2 creatinine0.7490.0130.6950.061841
urine platelets RR pH HCO3 GCS ASBP0.7280.0230.7780.017843
urine platelets RR pH HCO3 GCS ASBP BUN0.7250.0450.8040.038843
urine HCO3 GCS ASBP pH0.7640.0250.8120.029844
urine HCO3 GCS ASBP INR0.7540.0310.7820.038878
urine HCO3 GCS ASBP platelets0.7700.0210.8160.044964
urine GCS BUN ASBP HCO30.7650.0620.8270.027966
urine HCO3 GCS ASBP0.7590.0490.8260.038966
urine HCO3 GCS ASBP BUN0.7550.0730.8290.029966
urine HCO3 GCS ASBP creatinine0.7480.0150.8120.028966
urine HCO3 GCS ASBP HR0.7460.0520.8060.028966
urine HCO3 GCS ASBP SpO20.7560.0120.7960.025966
urine HCO3 GCS ASBP TEMP0.7760.0390.8100.020966
urine HCO3 RR GCS ASBP0.7510.0410.7910.036966
urine HCO3 RR GCS ASBP BUN0.7380.0670.8230.033966
urine HCO3 RR GCS ASBP BUN TEMP0.7650.0230.8060.038966
urine platelets RR pH HCO3 GCS0.7130.0440.7620.0251073
urine platelets RR pH HCO30.6130.0720.6880.0421077
urine HCO3 RR GCS0.7820.0250.8210.0201886
urine GCS BUN ASBP PaCO20.8330.0040.8340.0059592
urine GCS BUN ASBP pH0.8290.0120.8300.0069653
urine platelets RR pH ASBP0.6480.0140.6240.0099672
urine GCS BUN ASBP INR0.8420.0040.8470.0109885
urine platelets RR pH INR0.6660.0080.6620.00910 725
urine GCS BUN ASBP platelets0.8290.0080.8380.00510 821
urine GCS BUN ASBP RR0.8420.0100.8450.00810 875
urine GCS BUN ASBP creatinine0.8430.0080.8430.00610 886
urine GCS BUN ASBP SpO20.8350.0020.8390.00610 887
urine GCS BUN ASBP0.8390.0040.8360.00610 888
urine GCS BUN ASBP HR0.8340.0030.8420.00710 888
urine GCS BUN ASBP TEMP0.8370.0040.8390.00610 888
urine platelets RR pH PaCO20.6890.0060.6810.01211 762
urine platelets RR pH BUN0.7320.0080.7290.00811 812
urine platelets RR pH creatinine0.6900.0100.6730.00711 812
urine platelets RR pH GCS0.7980.0100.7900.00611 842
urine platelets RR pH TEMP0.6570.0180.6490.01811 876
urine platelets RR pH0.6650.0120.6390.01011 877
urine platelets RR pH HR0.6690.0130.6580.00611 877
urine platelets RR pH SpO20.6550.0050.6380.01311 877
Mean and standard deviation of the AUC values for test set predictions of patient outcome using 4 to 10 vital sign or laboratory data items. A total of 45 different combinations of data items were considered for the last time interval (index 1 above) and 3 hidden nodes for the neural network fits, with λ=10−6. The statistics were obtained for five and 10 random training and testing selections for the neural network and quadratic fits, respectively. Another check on the results is provided by the consistency of the predictions obtained from the neural network and quadratic fitting functions. To quantify this consistency the correlation coefficients for least-squares regression between the highest AUC values obtained in testing for the alternative fits were calculated for all four cases where common input data were considered. The resulting correlation coefficients are 0.86 for the multiplets (table 6 and figure 2a), 0.88 for the quartets with λ=10−5, (table 5 and figure 2b), 0.85 for quartets with λ=10−6, (table 5 and figure 2c) and 0.85 for the triplet combinations (table 4 and figure 2d). Here, the largest AUC value obtained in testing for any time window and number of hidden nodes (for the neural nets) was used. The quality of the predictions obtained from these alternative fitting functions exhibits quite a strong correlation. However, for the combinations of three and four different measurements, we note that there is a group of points where the best neural net prediction appears to be significantly better than the best quadratic fit result. This observation will be analysed in future work.
Figure 2.

Plots of AUC values obtained in testing for neural networks (NNs) (vertical axis) against quadratic (Q) fits (horizontal axis) with different combinations of patient measurements. (a) Multiplets (five or more vital sign or laboratory data items) λ=10−6. (b) Four data items, λ=10−5. (c) Four data items, λ=10−6. (d) Three data items, λ=10−5. The maximum AUC testing value for any time interval and any number of hidden nodes (neural net fits) is used for each combination of measurements. The best fit straight line is plotted in each case.

Plots of AUC values obtained in testing for neural networks (NNs) (vertical axis) against quadratic (Q) fits (horizontal axis) with different combinations of patient measurements. (a) Multiplets (five or more vital sign or laboratory data items) λ=10−6. (b) Four data items, λ=10−5. (c) Four data items, λ=10−6. (d) Three data items, λ=10−5. The maximum AUC testing value for any time interval and any number of hidden nodes (neural net fits) is used for each combination of measurements. The best fit straight line is plotted in each case. For one of the best combinations of input data, namely the HCO3, GCS, BUN, ASBP measurements, some further analysis of the landscape was conducted. To test the effect of using measurements from different time intervals, we employed the quadratic fitting function with λ=10−5. Training was performed for measurements corresponding to all 48 h, separately, the 46 possible blocks of three consecutive hours, and for all the measurements in the intervals 1–2, 1–3, 1–4, …, up to 1–22. The results are shown in figure 3. Here, we clearly see that it is the most recent measurements that produce the most accurate predictions in testing, and that there appears to be little or no advantage in considering earlier data.
Figure 3.

Plots of AUC values obtained in testing for the quadratic fitting function with four measurements (HCO3, GCS, BUN and ASBP) over different time windows. The three plots are for fits based upon each of the 48 h, blocks of three consecutive hours and a variable range from the last measurement including all hours backwards in time, i.e. 1, 1–2, 1–3, etc., up to 1–22. The training and testing sets included 3770 and 3771 patients, respectively, randomly reordered.

Plots of AUC values obtained in testing for the quadratic fitting function with four measurements (HCO3, GCS, BUN and ASBP) over different time windows. The three plots are for fits based upon each of the 48 h, blocks of three consecutive hours and a variable range from the last measurement including all hours backwards in time, i.e. 1, 1–2, 1–3, etc., up to 1–22. The training and testing sets included 3770 and 3771 patients, respectively, randomly reordered. A disconnectivity graph for the HCO3, GCS, BUN, ASBP combination was constructed for a neural network containing 5 hidden nodes using the measurements from the final hour, with λ=10−5. The organization that results corresponds to a single-funnel landscape, as shown in figure 4. This structure means that local minima are connected to the global minimum by pathways with relatively low downhill barriers. There is no competing low-lying minimum separated from the global minimum by a high barrier, which would constitute a secondary funnel, corresponding to a kinetic trap for a molecular system. We identify this as an unfrustrated landscape [34,52-54], associated with efficient relaxation to the global minimum. Locating the global minimum using methods such as basin-hopping [18-20] is therefore straightforward. In a molecular system, this organization would also suggest that the low-lying minima are structurally similar. For the machine learning landscape, the funnelled character and lack of frustration suggests the absence of alternative fits with similarly high predictive quality, but a qualitatively different pattern of weights in the parameter space. This analysis is consistent with previous results involving machine learning classification of outcomes for geometry optimization in a simple molecular system [32]. A more detailed understanding of how the observed structure arises could be very insightful, and clearly merits further investigation. Such efforts are in progress, and will build upon analogues of thermodynamic and kinetic properties. For example, a theoretical framework to assign heat capacity features to local minima has recently been applied to both molecular and machine learning landscapes [55].
Figure 4.

Disconnectivity graph for the machine learning landscape corresponding to the four measurements in figure 3 (HCO3, GCS, BUN and ASBP) for the final hour of data collection (index 1) with a neural network containing 5 hidden nodes, with λ=10−5. The stationary point database contains 1997 minima and 7492 transition states.

Disconnectivity graph for the machine learning landscape corresponding to the four measurements in figure 3 (HCO3, GCS, BUN and ASBP) for the final hour of data collection (index 1) with a neural network containing 5 hidden nodes, with λ=10−5. The stationary point database contains 1997 minima and 7492 transition states.

Conclusion

In this contribution, we have applied conceptual and computational tools from the potential energy landscapes framework to analyse machine learning landscapes defined by a cost function involved in fitting to training data. This approach has been used to predict patient outcomes using the MIMIC III clinical database [51]. We have compared results for neural networks with a range of different hidden nodes, as well as alternative regularization parameters. The predictive performance has been tested for various vital sign and laboratory measurements, including combinations of up to 10 different data types. We have also analysed how the predictions are affected when data from different time windows over the 48 h of collection are used for training and testing. Here, we find that the most recent measurements are generally the most useful, and including older data can degrade the results. Neural networks with between 3 and 11 hidden nodes were usually found to have similar performance, and the best AUC values obtained in testing are around 0.85 for particular combinations of three or more clinical measurements. The neural network predictions have also been compared with an alternative quadratic fitting function, which exhibits a single minimum for a given set of training data. We find a strong correlation between the AUC values obtained from these different functions. To gauge the uncertainties in the patient outcome predictions, the calculations were repeated for both fitting functions using a number of different random choices of patients for the training and testing sets. The standard deviation in the resulting AUC values is lowest when the datasets are as large as possible. Likely candidates for the global minima of the neural network fits are located using basin-hopping global optimization [18-20]. As in previous explorations of machine learning landscapes, short basin-hopping runs appear sufficient to reliably locate the lowest cost function value. To understand this observation, we have explored selected landscapes in more detail, characterizing the transition states that link local minima via steepest-descent paths, and the corresponding barriers defined by the cost function metric. These explorations produce the equivalent of a kinetic transition network [56-58] for a molecular or condensed matter system. The resulting databases of minima and transition states can be visualized using disconnectivity graphs [49,50]. One example is illustrated for a combination of four clinical measurements where a neural network with five hidden nodes exhibits good predictive properties. The underlying machine learning landscape has the appearance of a single funnel, which is the structure we associate with efficient relaxation to the global minimum, and self-organizing properties in atomistic systems [34,50,59]. Local minima are connected to the global minimum by relatively small downhill barriers. This organization probably explains why it is usually not difficult to find a fit corresponding to a low-lying minimum, with similar properties to the global minimum.
  27 in total

1.  Analysis of cooperativity and localization for atomic rearrangements.

Authors:  Semen A Trygubenko; David J Wales
Journal:  J Chem Phys       Date:  2004-10-08       Impact factor: 3.488

Review 2.  Effects of computerized clinical decision support systems on practitioner performance and patient outcomes: a systematic review.

Authors:  Amit X Garg; Neill K J Adhikari; Heather McDonald; M Patricia Rosas-Arellano; P J Devereaux; Joseph Beyene; Justina Sam; R Brian Haynes
Journal:  JAMA       Date:  2005-03-09       Impact factor: 56.272

Review 3.  Energy landscapes: some new horizons.

Authors:  David J Wales
Journal:  Curr Opin Struct Biol       Date:  2010-01-22       Impact factor: 6.809

4.  Effects of computer-based clinical decision support systems on physician performance and patient outcomes: a systematic review.

Authors:  D L Hunt; R B Haynes; S E Hanna; K Smith
Journal:  JAMA       Date:  1998-10-21       Impact factor: 56.272

5.  Decoding heat capacity features from the energy landscape.

Authors:  David J Wales
Journal:  Phys Rev E       Date:  2017-03-22       Impact factor: 2.529

6.  Prediction of extubation failure for neonates with respiratory distress syndrome using the MIMIC-II clinical database.

Authors:  Arthur Mikhno; Colleen M Ennett
Journal:  Conf Proc IEEE Eng Med Biol Soc       Date:  2012

7.  Energy landscapes for a machine-learning prediction of patient discharge.

Authors:  Ritankar Das; David J Wales
Journal:  Phys Rev E       Date:  2016-06-17       Impact factor: 2.529

8.  Effect of introducing the Modified Early Warning score on clinical outcomes, cardio-pulmonary arrests and intensive care utilisation in acute medical admissions.

Authors:  C P Subbe; R G Davies; E Williams; P Rutherford; L Gemmell
Journal:  Anaesthesia       Date:  2003-08       Impact factor: 6.955

9.  High-performance detection and early prediction of septic shock for alcohol-use disorder patients.

Authors:  Jacob Calvert; Thomas Desautels; Uli Chettipally; Christopher Barton; Jana Hoffman; Melissa Jay; Qingqing Mao; Hamid Mohamadlou; Ritankar Das
Journal:  Ann Med Surg (Lond)       Date:  2016-05-10

10.  Prediction of Sepsis in the Intensive Care Unit With Minimal Electronic Health Record Data: A Machine Learning Approach.

Authors:  Thomas Desautels; Jacob Calvert; Jana Hoffman; Melissa Jay; Yaniv Kerem; Lisa Shieh; David Shimabukuro; Uli Chettipally; Mitchell D Feldman; Chris Barton; David J Wales; Ritankar Das
Journal:  JMIR Med Inform       Date:  2016-09-30
View more
  1 in total

1.  Mortality prediction model for the triage of COVID-19, pneumonia, and mechanically ventilated ICU patients: A retrospective study.

Authors:  Logan Ryan; Carson Lam; Samson Mataraso; Angier Allen; Abigail Green-Saxena; Emily Pellegrini; Jana Hoffman; Christopher Barton; Andrea McCoy; Ritankar Das
Journal:  Ann Med Surg (Lond)       Date:  2020-10-03
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.