Literature DB >> 33981842

Humans in the Loop: Incorporating Expert and Crowd-Sourced Knowledge for Predictions Using Survey Data.

Anna Filippova1, Connor Gilroy2, Ridhi Kashyap3, Antje Kirchner4,5, Allison C Morgan6, Kivan Polimis7, Adaner Usmani8, Tong Wang9.   

Abstract

Survey data sets are often wider than they are long. This high ratio of variables to observations raises concerns about overfitting during prediction, making informed variable selection important. Recent applications in computer science have sought to incorporate human knowledge into machine-learning methods to address these problems. The authors implement such a "human-in-the-loop" approach in the Fragile Families Challenge. The authors use surveys to elicit knowledge from experts and laypeople about the importance of different variables to different outcomes. This strategy offers the option to subset the data before prediction or to incorporate human knowledge as scores in prediction models, or both together. The authors find that human intervention is not obviously helpful. Human-informed subsetting reduces predictive performance, and considered alone, approaches incorporating scores perform marginally worse than approaches that do not. However, incorporating human knowledge may still improve predictive performance, and future research should consider new ways of doing so.

Entities:  

Keywords:  Fragile Families Challenge; machine learning; missing data; prediction; surveys

Year:  2019        PMID: 33981842      PMCID: PMC8112737          DOI: 10.1177/2378023118820157

Source DB:  PubMed          Journal:  Socius        ISSN: 2378-0231


Social science survey data sets are often wider than they are long. Resource limitations demand that surveys ask many questions of the minimum number of respondents needed for statistical analyses. Moreover, social scientists are often interested in hard-to-reach populations, accentuating the need to ask many questions of few respondents. These difficulties characterize the Fragile Families and Child Wellbeing Study (FFCWS), which follows a cohort of nearly 5,000 children born in large U.S. cities between 1998 and 2000, roughly three quarters of them to unmarried parents (Reichman et al. 2001). The study collects a wealth of information about this disadvantaged group, including children’s physical and mental health, cognitive function, schooling, and living and family conditions. Overall, the FFCWS data set contains nearly 13,000 variables. The breadth of the variables contained in the FFCWS data set presents opportunities for a prediction task such as the Fragile Families Challenge (FFC). The FFC asked participants to use a data set containing variables collected from the child’s birth until year 9, and some training data from year 15, to predict six outcomes in the year 15 data: grade point average (GPA) and grit of the child,[1] material hardship and eviction of the family, layoff of the primary caregiver, and whether the primary caregiver participated in a job skills program. Although there is considerable information on each child, there are few children in the data set. As a result, new problems arise. Specifically, the high ratio of variables to observations increases the possibility of overfitting, that is, of fitting a complex model to statistical noise in a way that yields less useful out-of-sample predictions. In this article, we explore whether human-informed variable selection and parameter tuning can help solve this problem. Machine-learning (ML) methods have been increasingly applied to data with a high ratio of variables to observations to help with these same tasks (so-called feature selection). They provide ways to effectively use vast amounts of information contained in high-dimensional data sets (Donoho 2017). In contrast to substantive social science approaches, ML methods are less concerned with theoretical informativeness and favor data-driven predictive performance. Social scientists, on the other hand, usually draw on knowledge about the underlying data-generating process linking variables to outcomes. Increasingly, a number of applications in computer science have sought to incorporate human knowledge into ML methods (e.g., Branson et al. 2010). However, applications of these “human-in-the-loop” approaches are rare in the social sciences. In this article, we implement a human-in-the-loop approach to the FFC’s prediction tasks. We surveyed a scholarly community of social scientists as well as an anonymous community of laypeople to elicit their beliefs about which variables in the FFCWS data set would best predict each of the six outcomes. We used the information from these surveys in different ways. First, we subsetted the FFCWS data set preemptively, using either the variables identified by these surveys or a preexisting set of variables identified by the Fragile Families team. Second, we used information on scores assigned to particular variables to assign weights in the ML method. In effect, our ML approach was more likely to use variables with higher scores. We contrasted these human-in-the-loop approaches to a data-driven ML approach making use of the full data set of nearly 13,000 variables. The article proceeds as follows. First we outline how we elicited scholarly expertise and lay judgments. To use the extensive collection of variables in the FFCWS for our modeling approaches, we needed to address the issue of missing values in the data set. Next we describe how we addressed missingness. Thereafter we describe the models used, present results, and conclude.

Using Expert and Crowd-Sourced Knowledge

There might be several ways to collect knowledge about the predictors of the outcomes in the FFC. One could screen publications or conduct interviews with individuals familiar with the FFCWS. We leveraged computational tools to retrieve insights from scholars in three steps. First, we used Amazon Mechanical Turk (MTurk) to retrieve the contact information on every author who had published using the FFCWS (786 authors). Then, we administered online surveys to each author to identify relevant predictors of each outcome. Expert surveys have been used for a variety of predictive or forecasting tasks, from projections of fertility, mortality, and immigration (Billari, Graziani, and Melilli 2012; Bijak and Wiśniowski 2010) to measuring the quality of democracy (Pemstein et al. 2015) and to school planning (Raftery et al. 2012). Experienced researchers carry a wealth of knowledge about the relationships between variables and outcomes in these data, not all of which is published. By surveying researchers, we hoped to recover knowledge that was otherwise inaccessible at relatively low cost and over no more than few days. We also fielded the same survey to a comparison sample of laypeople that we crowd-sourced using MTurk. To elicit expert and lay beliefs, we used a wiki survey. We chose this to maximize accessibility, efficiency, and open-ness to new knowledge (Salganik and Levy 2015). We asked participants to choose which of two randomly selected predictors were likely to best predict a given outcome. These predictors were initially drawn from a list of 27 predictors suggested by a group of researchers familiar with the FFC, but participants were given the opportunity to add candidate predictors to the list (which would then be voted on by subsequent participants). As we explain in the Appendix, these predictors were higher-level concepts rather than specific variables. We used the data from the online surveys to generate an ordered list of candidate predictors; we scored each variable as the number of times it was voted for divided by the number of times it appeared in a pair. Further details about the surveys are included in Appendix A. Overall, 104 of 786 sampled experts participated, generating 2,651 votes. Seven hundred laypeople participated in our MTurk surveys, generating 27,221 votes. We used the variables identified through the expert and MTurk surveys in two different ways for our predictions. First, we used it to subset the data. Together, the expert and MTurk surveys yielded 68 higher-level concepts, which we associated with 271 variables from the FFCWS data set. We took these 271 variables as a single, wiki survey–generated subset.[2] Second, we used the rankings generated by the expert and MTurk surveys directly, as information passed to an ML algorithm. In this case, this yielded two approaches rather than one: one that used expert scores and one that used lay scores. Details are provided in the section “Models.”

Imputation

Because most ML approaches require a numeric and complete data set, processing the FFCWS data to handle missingness was a crucial step in preparing variables for modeling. To appreciate the extent of this problem, note that all observations had some missingness on some variables, which implies that there would have been no observations left with listwise deletion. Data were missing for different reasons, including unwillingness to respond, “don’t know” responses, logical skips, panel attrition, anonymization of sensitive information, and error. Roughly 74 percent of the data were missing in a way that posed problems for prediction (Figure 1). In a complex study such as this, the problems posed by missingness are particularly acute. We thus explored different imputation approaches with trade-offs in terms of efficiency and effectiveness (Appendix B). Because our different imputation strategies make different assumptions, we produced five distinctly imputed data sets on the basis of three unique approaches.
Figure 1.

Missing data.

Note: The Fragile Families and Child Wellbeing Study data set includes 12,943 variables for 4,232 observations. Within this data set, 74 percent of our observations are missing completely at random, missing at random, or missing not at random.

Models

We modeled the six outcomes with regularized regression. Regularization is an ML technique that can improve prediction on new data by avoiding overfitting on training data (James et al. 2017). Regularized models can be fit with large numbers of variables and relatively few observations. Regularized regression biases or shrinks model coefficients toward zero, relative to their maximum likelihood estimators, by applying a penalty to the likelihood function. Each nonzero coefficient has an associated cost. Absent other information, this cost is the same for every variable. If outside information warrants, however, the penalty can be relaxed for specific variables. The human knowledge of variable rankings captured through the scores from our survey is precisely this kind of information, and we drew on these scores to relax the penalties for the associated variables to differing degrees. For each scored variable, the global shrinkage parameter λ, which determines the overall degree of regularization, was multiplied by a local, variable-specific penalty factor ranging between 0 and 1. The wiki survey scores, ranging from 0 to 100, were mapped onto penalty factors in an inverse linear fashion to determine an appropriate local penalty factor for each variable. For instance, a score of 100 mapped to a penalty factor of 0, producing an unpenalized coefficient, while a score of 0 mapped to a penalty factor of 1 and full application of the global shrinkage parameter λ. Depending on the model, variables without scores were either treated as having a wiki survey score of 0 or else excluded entirely. Although simpler, this approach takes inspiration from Bayesian approaches to global and local shrinkage (Carvalho, Polson, and Scott 2009; Lee et al. 2010; Piironen and Vehtari 2016). We fit linear regressions for the continuous outcomes (GPA, grit, and material hardship) and logistic regressions for the binary outcomes (eviction, layoff, and job training). We used the implementation of regularized regression, with an “elasticnet” penalty, from the glmnet R package (Friedman, Hastie, and Tibshirani 2010). Appendix C describes the statistical and mathematical details of our models.

Results

In sum, we explored a total of 25 different approaches to prediction, distinguished by choices made at the following stages: (1) how we imputed missing observations, (2) whether we subsetted the data set prior to prediction and in what way, and (3) whether we incorporated outside knowledge into our modeling and in what way. As discussed, we considered five types of imputed data sets, three approaches to subsetting (no subsetting, subsetting to the variables identified by our wiki survey, and subsetting to the constructed variables identified by the Fragile Families team[3]), and three approaches to incorporating scores (expert scores, MTurk scores, and no scores). There were thus 45 possible permutations across these methods; of these, we focused on 25. Limitations of time and other resources narrowed the models we could run. For instance, the multiple imputation (MI) method we chose could not be run on the full data set of 13,000 variables using available computational resources. These 25 approaches can be compared in terms of mean squared error (Figure 2). However, because we did not fill the permutation space, it is complicated to rank the performance of choices at any given stage. In an unfilled permutation space, an unrestricted comparison of any set of choices does not hold all other strategies constant. For example, the fact that we used mean imputation with six subsetting and scoring approaches, but MI with only three, skews any comparison of the five imputation choices. Because the analytic choices we made affect our predictions, this kind of comparison is invalid. Therefore, when considering the best strategy in any given dimension, we restrict ourselves to that part of the permutation space in which we can compare across the relevant choices (Figure 3). We identify the best approach as the choice which minimizes the average or median mean squared error (MSE) across all other approaches and outcomes (Figure 4). This illustrates the relative rankings of these approaches, but the differences in performance also vary in magnitude. Therefore, we also illustrate the improvement made by any given approach, which we calculate as the average percentage improvement in MSE relative to the outcome-specific baseline MSE (Figure 5).[4,5]
Figure 2.

MSE’s from approaches relevant to human-in-the-loop rankings.

Note: The mean squared errors (MSEs) from the approaches when evaluating which “human-in-the-loop” strategy performed best are shown. As explained in “Results,” because the permutation space is not filled, this comparison must be conducted on a restricted set of seven approaches (i.e., each “human-in-the-loop” approach fit to a data set imputed by linear regression accounting for variable types). Here we plot the scores from that restricted set by outcome.

Figure 3.

Permutation space of possible and relevant approaches.

Note: The choice of five different imputation strategies, three different subsetting strategies, and three different score incorporation strategies yields a permutation space of 45 different approaches. Computational resources were available to explore only 25. Moreover, as explained in “Results,” the fact that not all 45 approaches were explore made it necessary to restrict to a subset of the explored approaches when comparing choices made in one or more dimensions. This figure shows the relevant set to which the analysis was restricted when addressing one of the four questions asked in this article. Those questions are given in the facet titles on the right-hand side of the figure.

Figure 4.

Rankings by lowest average and median MSE.

Note: The figure illustrates the lessons the results yield when researchers are confronted with any one of four questions: how to impute, whether to subset, whether to score, and whether to involve humans in the loop. Approaches are ranked by both average and median mean squared error (MSE), across outcomes. See “Results” for a complete discussion.

Figure 5.

Average percentage reduction in MSE.

Note: Figure 4 gives the relevant rankings, but it does not convey how much these choices matter. Here, the average percentage reduction in mean squared error (MSE) is plotted relative to baseline, across outcomes, for the same four questions that surface in the main article.

In what follows we consider what our results suggest for four different questions: (1) how to impute, (2) whether to subset, (3) whether to incorporate scores, and finally (4) whether it makes sense to include humans in the loop at all (whether by informed subsetting, or scoring, or both).[6]

Imputation

How should researchers approach issues of missingness? Overall, our results suggest that MI is best. If researchers have the computational power to pursue this approach, they should. Note, though, that by the metric of average MSE, the next best strategy is simple mean imputation and that the dividends to MI are not obviously enormous (Figure 4a). MI results in a 4.94 percent reduction in MSE relative to baseline, on average, whereas mean imputation results in a 4.61 percent reduction (Figure 5a). So, where resource constraints are an issue, mean imputation may be a viable alternative.[7] Also, regression-based imputation methods do not clearly outperform simple mean imputation, which is noteworthy given their additional computational costs.

Subsetting

Does it make sense to preemptively subset the data before modeling? Most social science researchers who use these data no doubt do, because it is impossible for humans to make much sense of thousands of variables. It is thus tempting to do the same in a prediction exercise of this kind. Yet our results suggest that human-informed subsetting worsens rather than improves predictive performance. Of all of the results in this article, this is the clearest: human-informed subsetting discards useful information. In this domain, human loses to machine. Interestingly, the two strategies that involve subsetting are not clearly distinguishable in terms of their predictive performance. By average MSE, it seems preferable to subset to the variables from our wiki survey, but by median MSE, the constructed variables fare better. In one sense, this is as encouraging as it is surprising. The constructed variables represent the considered judgment of people with experience in the field and with the FFCWS, whereas the wiki survey variables were selected in a few days and at low cost by an anonymous community of experts and laypeople. Of course, the wiki survey was fielded within the context of the FFC with the clearly assigned task of identifying predictors for the outcomes, whereas the constructed variables were not generated explicitly for this prediction task. Nevertheless, we find there is not much to distinguish them, and if anything, the wiki survey variables perform better (Figure 5b).

Scoring

Is it useful to incorporate human knowledge into the modeling process, as described earlier? Not really, according to either of the metrics we use to rank approaches. Whether measured by average or median MSE, approaches that ignore scores altogether outperform approaches that use expert or lay scores. For advocates of an approach that marries the powers of machines to human wisdom, this is disheartening. However, there are at least two caveats. First, the differences in performance are very small. On average, as Figure 5c shows, approaches that do not use scores reduce MSE relative to baseline by about 5.39 percent, compared with 5.29 percent and 5.25 percent for experts and MTurk users, respectively. Second, as we argue below, our approach to knowledge incorporation was ad hoc. As long as it is possible to imagine better ways of incorporating human knowledge into the loop, future research should consider them.

Humans in the Loop?

Does all this suggest that there is no role for humans in the loop? Not entirely. By average MSE, the best approach overall is one that does not subset and does not incorporate outside knowledge. Yet, again, the differences between this and the next best (and, indeed, the third best approach) are slight: a reduction of 7.67 percent versus 7.59 percent. Furthermore, if ranked by lowest median MSE, our best performing approach does enlist humans: one that incorporates expert scores while not subsetting the data set (Figure 4d). The discrepancy between average and median MSE rankings is explained by the very poor performance of a no subsetting and no scores approach in predicting the layoff of a child’s primary caregiver. This may suggest that outside information is useful for some outcomes but not others. Furthermore, one possible interpretation of this result is that strategies using expert scores are more robust to bad performance in a single outcome. What is clear from our results is that if humans are to enter the loop, it ought not to be by preemptively subsetting the data but rather by incorporating their wisdom into an approach that still leverages ML to extract information from the full data set. Making use of the full data set may not always be possible, as exemplified by the computational constraints we faced for generating a full data set with MI. However when it is possible, it can usefully augment prediction. Our approach incorporating scores on the basis of expert surveys fared better as a human-in-the-loop strategy. Although neither our approach to generating scores from the wiki surveys nor incorporating them in the models is dispositive, we believe that such approaches with further refinements may hold promise for human-in-the-loop strategies. In short, although there is obviously important information that only machines pick up, strategies that incorporate human knowledge to tune parameters in a model merit further exploration.[8]

Conclusions

In this article, we considered different ways of tackling a difficulty faced by researchers seeking to use survey data sets for prediction, namely, that the large ratio of variables to observations makes informed variable selection difficult. To tackle this problem, we proposed a low-cost way to mine a scholarly community for insights. We considered ways to use this information to subset a data set preemptively or at the modeling stage (or both together). What did we find? First, our results do not recommend preemptively subsetting the data. This is common practice in social science research, which is understandable, because social scientists are often more concerned with description and explanation rather than prediction, and humans cannot make any theoretical sense of thousands of variables. But for prediction purposes, this approach obviously discards useful information. Approaches that relied on this strategy fared worse than approaches that did not. Second, we find some evidence that human insight may be useful if fed into ML approaches at the modeling stage, but we have not demonstrated this beyond doubt. By average MSE, our best performing approach is one that neither subsets nor incorporates scores. But it does not perform much better than one that incorporates expert scores, and this approach actually ranks best by median MSE across outcomes. What, then, is the future of humans in the loop? We believe that future research should consider at least two types of improvements to our approach. First, the response rate of our expert survey was low: improving this would make it much easier to compare the dividends of surveying experts rather than laypeople. We expect that experts bring knowledge that laypeople do not, but our results do not clearly demonstrate this. Second, future work should consider alternative ways to incorporate human knowledge into ML models. We did so in an ad hoc way, but better formalization of our intuition and better use of the scores in modeling will surely help in deciding the place of humans in the loop, going forward. In closing, this project considered whether approaches from the tradition of informative, human-centered modeling can be usefully combined with ML techniques. We found that their combination is not always profitable but also that their judicious combination may yet be useful.
Table A1.

Predictors and Resulting Variables.

PredictorVariables
Number of times father has missed work{f2b30a, f3b22}
Child’s IQ{hv5_ppvtpr}
Foreign-born mother*{m3h1b, m3h1a}
Parents are in cohabiting partnership*{cf2cohm}
Private school{p5l1a, p5l23}
Parents’ substance abuse{cm3drug_case, cf3alc_case, cm3alc_case, cf3drug_case, m4j21, m5g20, f5g20, f4j21}
Number of books at home{f5k14e, m4b27, f4b27, m5k14e}
Existing number of siblings*{cm3kids, cf4kids, cf3kids, cm1kids, cm4kids, cm5kids, cf5kids, cf2kids, cf1kids, cm2kids}
Father’s interest in sports or entertainment{f5k14b}
Home schooling{p5l1a}
Mother’s mental health{f4c38, f5b31x}
Child’s perseverance{k5g1e, k5g1d, k5g1a, k5g1b}
Father’s nonstandard work hours*{f4k16a, f5k17l, f3k17a, f5i16a, f2k18a}
Child’s birth weight*{cm1lbw}
Families on block known well{m4i0, m4i0l, f4i0l, p5m1, m4i0n2, m4i0n3}
Child’s physical disability{hv3a2}
Number of child’s emergency room visits{p5h10, f2b8, m2b8, hv3a9, hv4a14}
Domestic violence{m4a8b_7, f3d7n1, m3d7n1, f5f26b2_10, f4a8b_7, m3a8b_7, m3d7p, f5f28a_10, m5f28a_10, m3d7m, m3d7n, m3d7o, m3e23q, m3d9p1, m3e23o, m3d9p, f3d9i, f3d9h, f3d9m, f3d9o, f3d9n, f3d9p, m5b30a1_10, f3a8b_7, m3d9n, m3d9o, m3d9m, m3e23p, m3d9i, m3d9n1, m5f26b2_10, f3d9n1, f3d7o, f3d7n, f3d7m, m3d7p1, m3e23p1, f5b26x_10}
Household income*{cf1hhinc, cm5hhinc, cm2hhinc, cf5hhinc, cf3hhinc, cm1hhinc, cf4hhincb, cf2hhinc, cf5hhincb, cf3hhincb, cf4hhinc, cf2hhincb, cm4hhinc, cm3hhinc}
Child’s gender*{cm1bsex, hv4sex_child}
Mother’s employment{m2k12, m2k8, m3k15, m4k15}
Income-to-poverty ratio*{cf3povco, cf1inpov, cf4povcob, cm3povco, cf3p…
Mother’s education*{cm5edu}
Teacher quality*{t5g7}
Child’s participation in sports{p5i1b, m5k14b}
Grandparents are present in household{cm3gmom, cm3gdad}
Number of parental romantic relationships*{m3a13, f3a13, m5a10, f5a101, f5a10, m5a101}
Child’s exposure to someone smoking{hv4a24, p5q3cr, p5h15c}
Mother’s incarceration*{m3i29}
Child’s race*{m1h3, f1h3}
Mother’s substance abuse{cm3drug_case, cm3alc_case}
Mother’s age at childbirth*{cm2fbir, cm1age}
Child’s participation in chores{f3b32d, f3b4d, f3c3d, f3e18d, f5k14a, m3b32d, m3b4d, m3c3d, m3e18d, m5k14a, p5i1a, p5i31a, p5i40a}
Teacher says child works independently{t5b2c}
Domestic abuse in family{m4a8b_7, f3d7n1, m3d7n1, f5f26b2_10, f4a8b_7, m3a8b_7, m3d7p, f5f28a_10, m5f28a_10, m3d7m, m3d7n, m3d7o, m3e23q, m3d9p1, m3e23o, m3d9p, f3d9i, f3d9h, f3d9m, f3d9o, f3d9n, f3d9p, m5b30a1_10, f3a8b_7, m3d9n, m3d9o, m3d9m, m3e23p, m3d9i, m3d9n1, m5f26b2_10, f3d9n1, f3d7o, f3d7n, f3d7m, m3d7p1, m3e23p1, f5b26x_10}
Divorce or separation*{m2a8c}
Father’s age at childbirth*{cf1age}
Father’s education*{cf5edu}
Mother has chronic illness{m5g2a_107}
Parent’s mental health{m5g2a_101}
Amount of parental involvement in school{m4i0d, f4i0d, m4i0}
Mother’s nonstandard work hours*{m3k16a, m2k13a, m5k17l, m4k16a, m5i16a}
Parent’s chronic illness{f5a3a1_9, m5g2a_107}
Parent impulsivity{p5q3an}
Number of books in the home{f5k14e, m4b27, f4b27, m5k14e}
School quality*{t5g4_104, m4i0, m4i0b, f5k5d, p5l1a, f5k5e, p5l13f, t5c7f}
Foreign-born father*{f3h1a, f3h1b}
Father’s sense of familial responsibility{p5i37, p5i32b, f5k15, m2e4b, n5c3f, m2c3b, f2b17a, f2b17b, f2b17c, f2b17d, f2b17e, f2b17f, f2b17g, f2b17h, p5i32a, p5i32c, n5c3e}
Availability of extended family{cf4gdad, cf1gdad, cm1gmom, cm2gdad, cm5gmom, cm4gmom, cf5gmom, cf3gmom, cf3gdad, cm5gdad, cf5gdad, cm4gdad, cm3gmom, cf1gmom, cm1gdad, cm3gdad, cf4gmom, cm2gmom, cf2gmom, cf2gdad}
Father’s incarceration*{f3i29}
Mother’s prenatal smoking*{m1g4}
Family on welfare{m3i8c1, f3i8c1, m5f8c1, m4i8c1, m5f8b1, m3i8b1, m4i8b1, m2h9c1, m2h9b1, f5f8c1, f1k2a, m1j2b, f2h8c1, f4i8c1}
Child’s learning disability{kind_a13}
Child makes friends easily{t5b1h}
Mother’s multiple job holding*{m3k18, m3k17, m2k14a}
Father’s substance abuse{cf3alc_case, cf3drug_case}
Parents have savings account{m5j6h}
Father’s multiple job holding*{f2k19a, f3k18}
Father absent at time of birth*{m1a6}
Household size*{cf5adult, cm4adult, cm3kids, cm5kids, cf1adul…
Parent’s religion{f3r1, m3r1}
Neighborhood crime{t5f4a}
Father’s unemployment{m2c33, m3c41}
Childcare center enrollment{m4b13}
Never-married mother*{cm5relf}
Child’s health{f5a3i_10, f5a6g02_10, m5a3a1_10, f5a3a1_10, m5a6g01_10, m5a6g03_10, f5a6g03_10, m5a3i_10, m5a6g02_10, f5a6g01_10}
Grandparents in the household{cm3gmom, cm3gdad}
Multigenerational household{cf4gdad, cm4gdad, cf4gmom, cm4gmom}

Ideas with which each survey was seeded.

Table A2.

Participation in Wiki Survey by Outcome.

ExpertsMTurk Users
OutcomeVotesVotersVotesVoters
Grade point average530245,130137
Grit29994,419110
Material Hardship777315,533127
Eviction980323,741113
Layoff3243,964115
Job Training3344,434129
Table D1.

Holdout Scores.

ImputationSubsettingScoresEvictionGPAGritJob TrainingLayoffMaterial Hardship
MIWiki surveyedExperts0.054430.364800.252850.180840.163200.02268
MIWiki surveyedMTurk users0.054710.370150.249980.181140.163340.02265
MIWiki surveyedNo scores0.054650.365880.250130.182060.163600.02256
MIConstructedNo scores0.055330.377280.248390.182370.162650.02265
LASSONo subsettingNo scores0.054910.354550.247540.179090.166110.02056
LASSOWiki surveyedExperts0.054970.370640.255090.182200.163990.02309
LASSOWiki surveyedMTurk users0.055460.375120.250640.182090.165740.02318
LASSOWiki surveyedNo scores0.055390.367500.251390.183180.165700.02300
LASSOConstructedNo scores0.055460.377930.249760.183670.164670.02320
LM-untypedWiki surveyedExperts0.055070.370890.252920.182460.165580.02329
LM-untypedWiki surveyedMTurk users0.055460.375400.250410.182780.167040.02331
LM-untypedWiki surveyedNo scores0.055370.367810.251710.183770.165730.02310
LMNo subsettingExperts0.054540.354250.247770.178100.163790.02012
LMNo subsettingMTurk users0.054620.354440.246640.179620.164310.02000
LMNo subsettingNo scores0.054540.355370.244680.177970.165600.01999
LMWiki surveyedExperts0.054810.362790.253100.183550.164570.02263
LMWiki surveyedMTurk users0.054640.364390.250780.183060.164340.02264
LMWiki surveyedNo scores0.054920.363870.250620.185090.164620.02267
LMConstructedNo scores0.055460.377970.249390.181950.162550.02282
MeanNo subsettingExperts0.054460.354490.249130.177710.163750.02000
MeanNo subsettingMTurk users0.054480.352580.246290.179090.164270.01993
MeanNo subsettingNo scores0.054240.351820.245660.177040.164730.01985
MeanWiki surveyedExperts0.054930.364570.253240.182900.164260.02267
MeanWiki surveyedMTurk users0.054770.364770.250640.183110.164440.02268
MeanWiki surveyedNo scores0.054980.363530.250330.183710.164810.02269

Note: Each row displays mean squared errors from 1 of the 25 different strategies we used. The first column describes the imputation strategy used. The second describes whether we subsetted the data and, if so, to which set. The third describes which set of scores were used, if any. GPA = grade point average; LASSO = least absolute shrinkage and selection operator; LM = linear model; MI = multiple imputation.

Table D2.

Original Holdout Scores.

ImputationSubsettingScoresEvictionGPAGritJob TrainingLayoffMaterial Hardship
MIWiki surveyedExperts0.05450.36420.25220.18170.16370.0226
MIWiki surveyedMTurk users0.05470.37000.24980.18130.16430.0226
MIWiki surveyedNo scores0.05470.36450.25060.18270.16440.0225
MIConstructedNo scores0.05480.37610.24800.18260.16270.0227
LASSONo subsettingNo scores0.05430.35870.24530.17950.16720.0203
LASSOWiki surveyedExperts0.05550.37790.25330.18250.16720.0249
LASSOWiki surveyedMTurk users0.05550.37290.25370.18220.16700.0248
LASSOWiki surveyedNo scores0.05540.37210.25170.18320.16680.0241
LASSOConstructedNo scores0.05550.38030.25160.18400.16620.0236
LM-untypedWiki surveyedExperts0.05500.37340.25330.18220.16420.0231
LM-untypedWiki surveyedMTurk users0.05550.38040.25050.18200.16560.0232
LM-untypedWiki surveyedNo scores0.05540.37180.25040.18340.16580.0230
LMNo subsettingExperts
LMNo subsettingMTurk users
LMNo subsettingNo scores0.05450.35510.24450.17790.16540.0200
LMWiki surveyedExperts0.05490.36330.25310.18240.16490.0226
LMWiki surveyedMTurk users0.05470.36440.25090.18310.16430.0226
LMWiki surveyedNo scores0.05490.36390.25060.18510.16460.0226
LMConstructedNo scores0.05550.37640.24950.18180.16350.0228
MeanNo subsettingExperts
MeanNo subsettingMTurk users
MeanNo subsettingNo scores0.05480.35210.24710.17730.16480.0199
MeanWiki surveyedExperts
MeanWiki surveyedMTurk users0.05480.37080.25090.18380.16460.0228
MeanWiki surveyedNo scores0.05500.36580.25070.18460.16500.0227

Note: Each row displays mean squared errors (MSEs) from our original submission to the Fragile Families Challenge. To explore an additional four strategies (which are blank here), we obtained new holdout MSEs after the original challenge had closed. These are shown in Table D1. We include these original MSEs as a reference for the interested reader. GPA = grade point average; LASSO = least absolute shrinkage and selection operator; LM = linear model; MI = multiple imputation.

  4 in total

1.  Regularization Paths for Generalized Linear Models via Coordinate Descent.

Authors:  Jerome Friedman; Trevor Hastie; Rob Tibshirani
Journal:  J Stat Softw       Date:  2010       Impact factor: 6.440

2.  Bayesian probabilistic population projections for all countries.

Authors:  Adrian E Raftery; Nan Li; Hana Ševčíková; Patrick Gerland; Gerhard K Heilig
Journal:  Proc Natl Acad Sci U S A       Date:  2012-08-20       Impact factor: 11.205

3.  Stochastic population forecasts based on conditional expert opinions.

Authors:  F C Billari; R Graziani; E Melilli
Journal:  J R Stat Soc Ser A Stat Soc       Date:  2012-04       Impact factor: 2.483

4.  Wiki surveys: open and quantifiable social data collection.

Authors:  Matthew J Salganik; Karen E C Levy
Journal:  PLoS One       Date:  2015-05-20       Impact factor: 3.240

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.