| Literature DB >> 30462745 |
Eugene Jeong1, Namgi Park1,2, Young Choi1,3, Rae Woong Park1,3, Dukyong Yoon1,3.
Abstract
BACKGROUND: The importance of identifying and evaluating adverse drug reactions (ADRs) has been widely recognized. Many studies have developed algorithms for ADR signal detection using electronic health record (EHR) data. In this study, we propose a machine learning (ML) model that enables accurate ADR signal detection by integrating features from existing algorithms based on inpatient EHR laboratory results.Entities:
Mesh:
Year: 2018 PMID: 30462745 PMCID: PMC6248973 DOI: 10.1371/journal.pone.0207749
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Study schematic.
Existing CERT, CLEAR, and PACE algorithms are applied to the Ajou University Hospital EHR dataset to extract features for ML algorithms for combining ADR signals. EU-SPC and SIDER databases are used to define the ADR reference dataset.
Features derived from three ADR signal detection algorithms.
| Algorithm | Features |
|---|---|
| CERT | 1. Average of laboratory test results before drug exposure |
| CLEAR | 19. Average of laboratory test results in a risk group |
| PACE | 44. Prescription counts on three days before the clinical event occurred |
SE, standard error; PRR, proportional reporting ratio; ROR, proportional odds ratio; IC, information component; SD, standard deviation; EB05, the lower bound of the 90% confidence interval for the Empiric Bayes Geometric Mean (EBGM)
Performance of ML models and previous ADR signal detection methods.
| Criterion for signaling | Sensitivity | Specificity | PPV | NPV | F1-measure | AUROC | |
|---|---|---|---|---|---|---|---|
| RF | Probability > 0.5 | 0.671 | 0.780 | 0.727 | 0.732 | 0.696 | 0.816 |
| SVM | Probability > 0.5 | 0.569 | 0.796 | 0.709 | 0.680 | 0.629 | 0.737 |
| L1LR | Probability > 0.5 | 0.593 | 0.756 | 0.679 | 0.682 | 0.631 | 0.741 |
| NN† | Probability > 0.5 | 0.793 | 0.619 | 0.645 | 0.777 | 0.709 | 0.795 |
| CEART400 | p < 0.05 | 0.868 | 0.100 | 0.455 | 0.467 | 0.597 | 0.559 |
| CLEAR | p < 0.05, OR > 1 | 0.674 | 0.413 | 0.496 | 0.596 | 0.571 | 0.559 |
| PACE | PCI < 0.667 | 0.081 | 0.897 | 0.406 | 0.529 | 0.135 | 0.520 |
| CCP2 | (CERT: p < 0.05 or | 0.075 | 0.908 | 0.405 | 0.540 | 0.127 | 0.518 |
| CCP3 | CERT: p < 0.05 | 0.074 | 0.920 | 0.453 | 0.526 | 0.127 | 0.475 |
| CHI | p < 0.05 | 0.486 | 0.517 | 0.466 | 0.537 | 0.476 | 0.563 |
| PRR | PRR-1.96SE > 1 | 0.463 | 0.573 | 0.485 | 0.551 | 0.473 | 0.525 |
| ROR | ROR-1.96SE > 1 | 0.563 | 0.483 | 0.486 | 0.560 | 0.522 | 0.563 |
| YULE | Yule’s Q-1.96SE > 1 | 0.350 | 0.680 | 0.487 | 0.546 | 0.407 | 0.522 |
| BCPNN | IC-2SD >0 | 0.508 | 0.521 | 0.479 | 0.549 | 0.493 | 0.517 |
| GPS | EB05>2 | 0.010 | 1 | 1 | 0.538 | 0.020 | 0.524 |
RF, random forest; SVM, support vector machine; L1LR, L1 regularized logistic regression; NN, neural network with three hidden layers; CCP2, PCI is less than 0.667 and one of the criteria of CERT and CLEAR is fulfilled; CCP3, PCI is less than 0.667 and all criteria of CERT and CLEAR are fulfilled; CHI, χ2 test; PRR, proportional reporting ratios; ROR, reporting odds ratio; YULE, Yule’s Q; BCPNN, Bayesian Confidence Neural Network; GPS, gamma Poisson shrinker; PPV, positive predictive value; and NPV, negative predictive value
†Average ± standard deviation of the performance results from 10 experiments with tenfold cross-validation
‡Performance results on the whole dataset using their own criteria
Summary table of Tukey’s HSD post-hoc test results among ML models.
| (I) ML model | (J) ML model | Mean difference | Std. error | Sig. | 95% Confidence interval | |
|---|---|---|---|---|---|---|
| Lower bound | Upper bound | |||||
| NN | RF | -0.024 | 0.005 | <0.01 | -0.037 | -0.010 |
| SVM | 0.055 | 0.005 | <0.01 | 0.042 | 0.069 | |
| L1LR | 0.051 | 0.005 | <0.01 | 0.038 | 0.064 | |
| RF | NN | 0.024 | 0.005 | <0.01 | 0.010 | 0.037 |
| SVM | 0.040 | 0.005 | <0.01 | 0.026 | 0.054 | |
| L1LR | 0.075 | 0.005 | <0.01 | 0.061 | 0.088 | |
| SVM | NN | -0.055 | 0.005 | <0.01 | -0.069 | -0.042 |
| RF | -0.040 | 0.005 | <0.01 | -0.054 | -0.026 | |
| L1LR | -0.004 | 0.005 | 0.860 | -0.017 | 0.009 | |
| L1LR | NN | -0.050 | 0.005 | <0.01 | -0.064 | -0.038 |
| RF | -0.075 | 0.005 | <0.01 | -0.088 | -0.061 | |
| SVM | 0.005 | 0.005 | 0.860 | -0.009 | 0.017 | |
Fig 2AUROCs of ML algorithms and original ADR signal detection algorithms.
The best AUROCs (A) and average AUROCs (B) of each algorithm are shown among 10 experiments with tenfold cross-validation. No significant difference is observed in the ML models; however, the AUROCs of the ML models are much larger than those of the original methods.
Fig 3Important features in random forest algorithm.
The importance of features is expressed in terms of Gini importance by color during 10 experiments with tenfold cross-validation. The blue color implies more importance and the yellow color, less importance. The top 10 important features are marked by red boxes.
Fig 4Coefficients calculated in logistic regression.
The coefficients of features are expressed in color during 10 experiments with tenfold cross-validation. Red color indicates positive coefficients and blue color, negative coefficients.