Literature DB >> 29780658

The Selective Labels Problem: Evaluating Algorithmic Predictions in the Presence of Unobservables.

Himabindu Lakkaraju1, Jon Kleinberg2, Jure Leskovec1, Jens Ludwig3, Sendhil Mullainathan4.   

Abstract

Evaluating whether machines improve on human performance is one of the central questions of machine learning. However, there are many domains where the data is selectively labeled in the sense that the observed outcomes are themselves a consequence of the existing choices of the human decision-makers. For instance, in the context of judicial bail decisions, we observe the outcome of whether a defendant fails to return for their court appearance only if the human judge decides to release the defendant on bail. This selective labeling makes it harder to evaluate predictive models as the instances for which outcomes are observed do not represent a random sample of the population. Here we propose a novel framework for evaluating the performance of predictive models on selectively labeled data. We develop an approach called contraction which allows us to compare the performance of predictive models and human decision-makers without resorting to counterfactual inference. Our methodology harnesses the heterogeneity of human decision-makers and facilitates effective evaluation of predictive models even in the presence of unmeasured confounders (unobservables) which influence both human decisions and the resulting outcomes. Experimental results on real world datasets spanning diverse domains such as health care, insurance, and criminal justice demonstrate the utility of our evaluation metric in comparing human decisions and machine predictions.

Entities:  

Year:  2017        PMID: 29780658      PMCID: PMC5958915          DOI: 10.1145/3097983.3098066

Source DB:  PubMed          Journal:  KDD        ISSN: 2154-817X


  13 in total

1.  Matching methods for causal inference: A review and a look forward.

Authors:  Elizabeth A Stuart
Journal:  Stat Sci       Date:  2010-02-01       Impact factor: 2.901

2.  Doubly robust estimation in missing data and causal inference models.

Authors:  Heejung Bang; James M Robins
Journal:  Biometrics       Date:  2005-12       Impact factor: 2.571

3.  The prevention and treatment of missing data in clinical trials.

Authors:  Roderick J Little; Ralph D'Agostino; Michael L Cohen; Kay Dickersin; Scott S Emerson; John T Farrar; Constantine Frangakis; Joseph W Hogan; Geert Molenberghs; Susan A Murphy; James D Neaton; Andrea Rotnitzky; Daniel Scharfstein; Weichung J Shih; Jay P Siegel; Hal Stern
Journal:  N Engl J Med       Date:  2012-10-04       Impact factor: 91.245

4.  HUMAN DECISIONS AND MACHINE PREDICTIONS.

Authors:  Jon Kleinberg; Himabindu Lakkaraju; Jure Leskovec; Jens Ludwig; Sendhil Mullainathan
Journal:  Q J Econ       Date:  2017-08-26

5.  Multiple imputation to account for missing data in a survey: estimating the prevalence of osteoporosis.

Authors:  Andrew Kmetic; Lawrence Joseph; Claudie Berger; Alan Tenenhouse
Journal:  Epidemiology       Date:  2002-07       Impact factor: 4.822

6.  Teacher-child relationships and academic achievement: a multilevel propensity score model approach.

Authors:  Meghan P McCormick; Erin E O'Connor; Elise Cappella; Sandee G McClowry
Journal:  J Sch Psychol       Date:  2013-05-24

7.  An Introduction to Propensity Score Methods for Reducing the Effects of Confounding in Observational Studies.

Authors:  Peter C Austin
Journal:  Multivariate Behav Res       Date:  2011-06-08       Impact factor: 5.923

8.  Dealing with missing data in a multi-question depression scale: a comparison of imputation methods.

Authors:  Fiona M Shrive; Heather Stuart; Hude Quan; William A Ghali
Journal:  BMC Med Res Methodol       Date:  2006-12-13       Impact factor: 4.615

9.  Data-driven decisions for reducing readmissions for heart failure: general methodology and case study.

Authors:  Mohsen Bayati; Mark Braverman; Michael Gillam; Karen M Mack; George Ruiz; Mark S Smith; Eric Horvitz
Journal:  PLoS One       Date:  2014-10-08       Impact factor: 3.240

10.  A propensity score matched comparison of different insulin regimens 1 year after beginning insulin in people with type 2 diabetes.

Authors:  N Freemantle; B Balkau; P D Home
Journal:  Diabetes Obes Metab       Date:  2013-07-16       Impact factor: 6.577

View more
  4 in total

1.  Predicting high-risk opioid prescriptions before they are given.

Authors:  Justine S Hastings; Mark Howison; Sarah E Inman
Journal:  Proc Natl Acad Sci U S A       Date:  2020-01-14       Impact factor: 11.205

2.  From What to How: An Initial Review of Publicly Available AI Ethics Tools, Methods and Research to Translate Principles into Practices.

Authors:  Jessica Morley; Luciano Floridi; Libby Kinsey; Anat Elhalal
Journal:  Sci Eng Ethics       Date:  2019-12-11       Impact factor: 3.525

3.  Preventing rather than punishing: An early warning model of malfeasance in public procurement.

Authors:  Jorge Gallego; Gonzalo Rivero; Juan Martínez
Journal:  Int J Forecast       Date:  2020-07-18

4.  Predicting neurological recovery with Canonical Autocorrelation Embeddings.

Authors:  Maria De-Arteaga; Jieshi Chen; Peter Huggins; Jonathan Elmer; Gilles Clermont; Artur Dubrawski
Journal:  PLoS One       Date:  2019-01-28       Impact factor: 3.240

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.