| Literature DB >> 32324754 |
Kenneth D Roe1,2, Vibhu Jawa1,3, Xiaohan Zhang4, Christopher G Chute1,2,4,5, Jeremy A Epstein5, Jordan Matelsky6, Ilya Shpitser1,3, Casey Overby Taylor1,2,4,5,7.
Abstract
Incorporating expert knowledge at the time machine learning models are trained holds promise for producing models that are easier to interpret. The main objectives of this study were to use a feature engineering approach to incorporate clinical expert knowledge prior to applying machine learning techniques, and to assess the impact of the approach on model complexity and performance. Four machine learning models were trained to predict mortality with a severe asthma case study. Experiments to select fewer input features based on a discriminative score showed low to moderate precision for discovering clinically meaningful triplets, indicating that discriminative score alone cannot replace clinical input. When compared to baseline machine learning models, we found a decrease in model complexity with use of fewer features informed by discriminative score and filtering of laboratory features with clinical input. We also found a small difference in performance for the mortality prediction task when comparing baseline ML models to models that used filtered features. Encoding demographic and triplet information in ML models with filtered features appeared to show performance improvements from the baseline. These findings indicated that the use of filtered features may reduce model complexity, and with little impact on performance.Entities:
Mesh:
Year: 2020 PMID: 32324754 PMCID: PMC7179831 DOI: 10.1371/journal.pone.0231300
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
A hypothetical case showing mutual information (MI) calculation.
| Drug B | No change (Lab A) | Increased (Lab A) | Decreased (Lab A) |
| Survived | 12 | 11 | 12 |
| Died | 22 | 22 | 21 |
| Drug C | No change (Lab A) | Increased (Lab A) | Decreased (Lab A) |
| Survived | 25 | 8 | 5 |
| Died | 12 | 23 | 27 |
Fig 1Summary of data acquisition, data pre-processing, and data analysis steps.
Precision for top k prescription and procedure triplets.
| Triplet event | ||||
|---|---|---|---|---|
| Prescription | 67.0% | 40.0% | 30.0% | 20.0% |
| Procedure | 0.0% | 0.0% | 0.0% | 20.0% |
Summary of dataset.
| Data | Total |
|---|---|
| # Admissions in the MIMIC-III (v1.4) database | 58 576 |
| # Admissions with asthma medication | 8359 |
| # Admissions with visit duration > = 48 hours | 7777 |
| # Training set of patients for the model | 6222 |
| # Testing set of patients the model | 1555 |
Trained logistic regression model parameters.
| Features | Data subsets | Number of coefficients | Number of non-zero coefficients |
|---|---|---|---|
| 2 features | Labs | 480 | 123 |
| 4 features | Labs | 960 | 344 |
| 8 features | Labs | 1920 | 418 |
| 16 features | Labs | 3840 | 770 |
| 32 features | Labs | 7680 | 894 |
| 42 features | Labs | 10080 | 1058 |
| Labs+demo | 10224 | 619 | |
| Labs+demo+events | 11568 | 684 | |
| Labs+demo+events+triples | - | - | |
| 11 filtered features | Labs | 2640 | 614 |
| Labs+demo | 2784 | 448 | |
| Labs+demo+events | 4128 | 609 | |
| Labs+demo+events+triples | 47568 | 686 |
Trained gradient boosting model parameters.
| Features | Data subsets | Number of coefficients | Number of non-zero coefficients |
|---|---|---|---|
| 2 features | Labs | 480 | 148 |
| 4 features | Labs | 960 | 365 |
| 8 features | Labs | 1920 | 612 |
| 16 features | Labs | 3840 | 452 |
| 32 features | Labs | 7680 | 778 |
| 42 features | Labs | 10080 | 732 |
| Labs+demo | 10224 | 384 | |
| Labs+demo+events | 11568 | 374 | |
| Labs+demo+events+triples | - | - | |
| 11 filtered features | Labs | 2640 | 431 |
| Labs+demo | 2784 | 330 | |
| Labs+demo+events | 4128 | 468 | |
| Labs+demo+events+triples | 95136 | 995 |
Logistic regression, gradient boosting, and neural network: AUC of top k features.
| Features | Data subsets | Logistic regression | Gradientboosting | Neural networks |
|---|---|---|---|---|
| 2 features | Labs | 0.54 | 0.60 | 0.57 |
| 4 features | Labs | 0.55 | 0.63 | 0.55 |
| 8 features | Labs | 0.56 | 0.63 | 0.57 |
| 16 features | Labs | 0.60 | 0.66 | 0.61 |
| 32 features | Labs | 0.64 | 0.68 | 0.64 |
| 42 features | Labs | 0.64 | 0.69 | 0.63 |
| Labs+demo | 0.73 | 0.74 | 0.67 | |
| Labs+demo+events | 0.73 | 0.75 | 0.65 | |
| Labs+demo+events+triples | - | - | - | |
| 11 filtered features | Labs | 0.64 | 0.68 | 0.62 |
| Labs+demo | 0.72 | 0.73 | 0.67 | |
| Labs+demo+events | 0.73 | 0.74 | 0.68 | |
| Labs+demo+events+triples | 0.72 | 0.74 | 0.65 |
AUC = area under the receiver operating characteristic curve
Fig 2ROC curves for logistic regression classifiers.
Fig 3ROC curves for gradient boosting classifiers.
Fig 4ROC curves for neural network classifiers.
Fig 5ROC curves for logistic regression classifiers.
Fig 6ROC curves for gradient boosting classifiers.
Fig 7ROC curves for neural network classifiers.