| Literature DB >> 31093548 |
Melissa Assel1, Daniel D Sjoberg1, Andrew J Vickers1.
Abstract
BACKGROUND: A variety of statistics have been proposed as tools to help investigators assess the value of diagnostic tests or prediction models. The Brier score has been recommended on the grounds that it is a proper scoring rule that is affected by both discrimination and calibration. However, the Brier score is prevalence dependent in such a way that the rank ordering of tests or models may inappropriately vary by prevalence.Entities:
Keywords: Brier score; Concordance index; Mean squared error; Net benefit; Prediction modeling; Sensitivity; Specificity
Year: 2017 PMID: 31093548 PMCID: PMC6460786 DOI: 10.1186/s41512-017-0020-3
Source DB: PubMed Journal: Diagn Progn Res ISSN: 2397-7523
Fig. 1Calibration plot for various continuous prediction models of differing degrees of miscalibration. All prediction models have an AUC of 0.75 for predicting an event with prevalence 20%. The prediction models include the following: a well-calibrated prediction model, a model that is miscalibrated such that it overestimates risk, a prediction model that underestimates risk, and a prediction model that more severely underestimates risk
Performance characteristics of binary tests and continuous prediction models with various degrees of miscalibration. All values given were calculated directly from the formulae in the text and independently verified using a simulation approach (Appendix)
| Net benefit | |||||||
|---|---|---|---|---|---|---|---|
| Test | Specificity | Sensitivity | AUC | Brier score | Threshold: 5% | Threshold: 10% | Threshold: 20% |
| Binary tests | |||||||
| Assume all negative | 100% | 0% | 0.500 | 0.2000 | 0.0000 | 0.0000 | 0.0000 |
| Assume all positive | 0% | 100% | 0.500 | 0.8000 | 0.1579 | 0.1111 | 0.0000 |
| Highly specific | 95% | 50% | 0.725 | 0.1400* | 0.0979 | 0.0956 | 0.0900 |
| Highly sensitive | 50% | 95% | 0.725 | 0.4100* | 0.1689 | 0.1456 | 0.0900 |
| Continuous prediction models | |||||||
| Well calibrated | – | – | 0.75 | 0.1386 | 0.1595 | 0.1236 | 0.0716 |
| Overestimating risk | – | – | 0.75 | 0.1708 | 0.1583 | 0.1160 | 0.0423 |
| Underestimating risk | – | – | 0.75 | 0.1540 | 0.1483 | 0.0986 | 0.0413 |
| Severely underestimating risk | – | – | 0.75 | 0.1760 | 0.0921 | 0.0372 | 0.0076 |
AUC, Brier score, and net benefit for various threshold probabilities corresponding to binary tests and continuous prediction models with various degrees of miscalibration predicting an outcome with prevalence of 20%, as shown in Fig. 1. Higher values of AUC and net benefit are desirable whereas lower values of the Brier score are desirable
*Method 1 calculation: binary test is considered to produce probabilities of 1 and 0 for a positive and negative test, respectively
†Method 2 calculation: binary test is considered to produce probabilities of the positive predictive value and 1 − negative predictive value for a positive and negative test, respectively