| Literature DB >> 33716360 |
David Benkeser1, Maya Petersen2, Mark J van der Laan2,3.
Abstract
When predicting an outcome is the scientific goal, one must decide on a metric by which to evaluate the quality of predictions. We consider the problem of measuring the performance of a prediction algorithm with the same data that were used to train the algorithm. Typical approaches involve bootstrapping or cross-validation. However, we demonstrate that bootstrap-based approaches often fail and standard cross-validation estimators may perform poorly. We provide a general study of cross-validation-based estimators that highlights the source of this poor performance, and propose an alternative framework for estimation using techniques from the efficiency theory literature. We provide a theorem establishing the weak convergence of our estimators. The general theorem is applied in detail to two specific examples and we discuss possible extensions to other parameters of interest. For the two explicit examples that we consider, our estimators demonstrate remarkable finite-sample improvements over standard approaches.Entities:
Keywords: AUC; cross-validation; estimating equations; machine learning; prediction; targeted minimum loss-based estimation
Year: 2019 PMID: 33716360 PMCID: PMC7954141 DOI: 10.1080/01621459.2019.1668794
Source DB: PubMed Journal: J Am Stat Assoc ISSN: 0162-1459 Impact factor: 5.033