| Literature DB >> 30695057 |
Dhruv Grover1, Sebastian Bauhoff2, Jed Friedman3.
Abstract
Independent verification is a critical component of performance-based financing (PBF) in health care, in which facilities are offered incentives to increase the volume of specific services but the same incentives may lead them to over-report. We examine alternative strategies for targeted sampling of health clinics for independent verification. Specifically, we empirically compare several methods of random sampling and predictive modeling on data from a Zambian PBF pilot that contains reported and verified performance for quantity indicators of 140 clinics. Our results indicate that machine learning methods, particularly Random Forest, outperform other approaches and can increase the cost-effectiveness of verification activities.Entities:
Mesh:
Year: 2019 PMID: 30695057 PMCID: PMC6350980 DOI: 10.1371/journal.pone.0211262
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Overview of data from Zambia pilot.
| Quarter | ||||
|---|---|---|---|---|
| Percent over-reporting | 18.6 | 15 | 22.9 | 20 |
| Count | 140 | 140 | 140 | 140 |
| Quarter 1 | 100 | 57.7 | 42.3 | 42.3 |
| Quarter 2 | 71.4 | 100 | 66.7 | 47.6 |
| Quarter 3 | 34.4 | 43.8 | 100 | 43.8 |
| Quarter 4 | 39.3 | 35.7 | 50 | 100 |
Distribution of facilities that over-report.
| N | Percent | |
|---|---|---|
| Never | 81 | 57.9 |
| One quarter | 32 | 22.9 |
| Two quarters | 12 | 8.6 |
| Three quarters | 9 | 6.4 |
| All four quarters | 6 | 4.3 |
Fig 1(a) Average ROC curves and their corresponding AUC values for random forest, logistic regression, support vector machine and naïve Bayes classifiers using cross-validation with training data. (b) Precision-Recall curves.
Average performance metrics of the four machine learning classifiers.
The best metrics in each column are shown in bold.
| Model | Accuracy | Precision | Recall | F1-score | AUC |
|---|---|---|---|---|---|
| Logistic Regression | 0.584 | 0.627 | 0.713 | 0.509 | 0.748 |
| Naïve Bayes | 0.552 | 0.523 | 0.628 | 0.425 | 0.629 |
| SVM | 0.647 | 0.691 | 0.651 | 0.815 | |
| Random Forest | 0.851 |
Prediction accuracy performance of different approaches.
| Approach | Prediction of over-reported event | |||
|---|---|---|---|---|
| Q1 | Q2 | Q3 | Q4 | |
| SRS | 18.77% | 14.98% | 22.56% | 20.04% |
| SRS with district stratification | 18.83% | 15.21% | 23.22% | 19.9% |
| SRS of offenders & non-offenders | - | 34.5% | 36.5% | 27.87% |
| SRS of only offenders | - | 44.5% | 42.19% | 38.81% |
| Logistic Regression | 58.42% | 32.84% | 31.28% | 34.76% |
| Naïve Bayes | 55.24% | 46.15% | 32.05% | 41.3% |
| SVM | 64.75% | 58.02% | 49% | 52.26% |
| Random Forest | 86.6% | 89.18% | 84.92% | 77.31% |
| Random Forest with district | 87.84% | 86.19% | 81.99% | 76.96% |
| Random Forest with intervention | 85.08% | 82.29% | 77.83% | 73.08% |
Note: Accuracy is calculated as average of 1,000 independent sampling without replacement iterations for SRS, and 10-fold cross-validation for supervised learning.