Literature DB >> 32229555

Measuring the predictability of life outcomes with a scientific mass collaboration.

Matthew J Salganik1, Ian Lundberg2, Alexander T Kindel2, Caitlin E Ahearn3, Khaled Al-Ghoneim4, Abdullah Almaatouq5,6, Drew M Altschul7, Jennie E Brand3,8, Nicole Bohme Carnegie9, Ryan James Compton10, Debanjan Datta11, Thomas Davidson12, Anna Filippova13, Connor Gilroy14, Brian J Goode15, Eaman Jahani16, Ridhi Kashyap17,18,19, Antje Kirchner20, Stephen McKay21, Allison C Morgan22, Alex Pentland6, Kivan Polimis23, Louis Raes24, Daniel E Rigobon25, Claudia V Roberts26, Diana M Stanescu27, Yoshihiko Suhara6, Adaner Usmani28, Erik H Wang27, Muna Adem29, Abdulla Alhajri30, Bedoor AlShebli31, Redwane Amin32, Ryan B Amos26, Lisa P Argyle33, Livia Baer-Bositis34, Moritz Büchi35, Bo-Ryehn Chung36, William Eggert37, Gregory Faletto38, Zhilin Fan39, Jeremy Freese34, Tejomay Gadgil40, Josh Gagné34, Yue Gao41, Andrew Halpern-Manners29, Sonia P Hashim26, Sonia Hausen34, Guanhua He42, Kimberly Higuera34, Bernie Hogan43, Ilana M Horwitz44, Lisa M Hummel34, Naman Jain25, Kun Jin45, David Jurgens46, Patrick Kaminski29,47, Areg Karapetyan48,49, E H Kim34, Ben Leizman26, Naijia Liu27, Malte Möser26, Andrew E Mack27, Mayank Mahajan26, Noah Mandell50, Helge Marahrens29, Diana Mercado-Garcia44, Viola Mocz51, Katariina Mueller-Gastell34, Ahmed Musse52, Qiankun Niu32, William Nowak53, Hamidreza Omidvar54, Andrew Or26, Karen Ouyang26, Katy M Pinto55, Ethan Porter56, Kristin E Porter57, Crystal Qian26, Tamkinat Rauf34, Anahit Sargsyan58, Thomas Schaffner26, Landon Schnabel34, Bryan Schonfeld27, Ben Sender59, Jonathan D Tang26, Emma Tsurkov34, Austin van Loon34, Onur Varol60,61, Xiafei Wang62, Zhi Wang61,63, Julia Wang26, Flora Wang59, Samantha Weissman26, Kirstie Whitaker64,65, Maria K Wolters66, Wei Lee Woon67, James Wu68, Catherine Wu26, Kengran Yang54, Jingwen Yin39, Bingyu Zhao69, Chenyun Zhu39, Jeanne Brooks-Gunn70,71, Barbara E Engelhardt26,36, Moritz Hardt72, Dean Knox27, Karen Levy73, Arvind Narayanan26, Brandon M Stewart2, Duncan J Watts74,75,76, Sara McLanahan1.   

Abstract

How predictable are life trajectories? We investigated this question with a scientific mass collaboration using the common task method; 160 teams built predictive models for six life outcomes using data from the Fragile Families and Child Wellbeing Study, a high-quality birth cohort study. Despite using a rich dataset and applying machine-learning methods optimized for prediction, the best predictions were not very accurate and were only slightly better than those from a simple benchmark model. Within each outcome, prediction error was strongly associated with the family being predicted and weakly associated with the technique used to generate the prediction. Overall, these results suggest practical limits to the predictability of life outcomes in some settings and illustrate the value of mass collaborations in the social sciences.
Copyright © 2020 the Author(s). Published by PNAS.

Entities:  

Keywords:  life course; machine learning; mass collaboration; prediction

Mesh:

Year:  2020        PMID: 32229555      PMCID: PMC7165437          DOI: 10.1073/pnas.1915006117

Source DB:  PubMed          Journal:  Proc Natl Acad Sci U S A        ISSN: 0027-8424            Impact factor:   11.205


Social scientists studying the life course have described social patterns, theorized factors that shape outcomes, and estimated causal effects. Although this research has advanced scientific understanding and informed policy interventions, it is unclear how much it translates into an ability to predict individual life outcomes. Assessing predictability is important for three reasons. First, accurate predictions can be used to target assistance to children and families at risk (1, 2). Second, predictability of a life outcome from a person’s life trajectory can indicate social rigidity (3), and efforts to understand differences in predictability across social contexts can stimulate scientific discovery and improve policy-making (4). Finally, efforts to improve predictive performance can spark developments in theory and methods (5). In order to measure the predictability of life outcomes for children, parents, and households, we created a scientific mass collaboration. Our mass collaboration—the Fragile Families Challenge—used a research design common in machine learning but not yet common in the social sciences: the common task method (6). To create a project using the common task method, an organizer designs a prediction task and then recruits a large, diverse group of researchers who complete the task by predicting the exact same outcomes using the exact same data. These predictions are then evaluated with the exact same error metric that exclusively assesses their ability to predict held-out data: data that are held by the organizer and not available to participants. Although the structure of the prediction task is completely standardized, participants are free to use any technique to generate predictions. The common task method produces credible estimates of predictability because of its design. If predictability is higher than expected, the results cannot be dismissed because of concerns about overfitting (7) or researcher degrees of freedom (8). Alternatively, if predictability is lower than expected, the results cannot be dismissed because of concerns about the limitations of any particular researcher or method. An additional benefit of the common task method is that the standardization of the prediction task facilitates comparisons between different methodological and theoretical approaches. Our mass collaboration builds on a long-running, intensive data collection: the Fragile Families and Child Wellbeing Study (hereafter the Fragile Families study). In contrast to government administrative records and digital trace data that are often used for prediction, these data were created to enable social science research. The ongoing study collects rich longitudinal data about thousands of families, each of whom gave birth to a child in a large US city around the year 2000 (9). The study was designed to understand families formed by unmarried parents and the lives of children born into these families. The Fragile Families data—which have been used in more than 750 published journal articles (10)—were collected in six waves: child birth and ages 1, 3, 5, 9, and 15. Each wave includes a number of different data collection modules. For example, the first wave (birth) includes survey interviews with the mother and father. Over time, the scope of data collection increased, and the fifth wave (age 9 y) includes survey interviews with the mother, the father, the child’s primary caregiver (if not the mother or father), the child’s teacher, and the child (Fig. 1).
Fig. 1.

Data collection modules in the Fragile Families study. Each module is made up of ∼10 sections, where each section includes questions about a specific topic (e.g., marriage attitudes, family characteristics, demographic characteristics). Information about the topics included in each module is presented in . During the Fragile Families Challenge, data from waves 1 to 5 (birth to age 9 y) were used to predict outcomes in wave 6 (age 15 y).

Data collection modules in the Fragile Families study. Each module is made up of ∼10 sections, where each section includes questions about a specific topic (e.g., marriage attitudes, family characteristics, demographic characteristics). Information about the topics included in each module is presented in . During the Fragile Families Challenge, data from waves 1 to 5 (birth to age 9 y) were used to predict outcomes in wave 6 (age 15 y). Each data collection module is made up of ∼10 sections, where each section includes questions about a specific topic. For example, the interview with the mother in wave 1 (birth) has sections about the following topics: child health and development, father–mother relationships, fatherhood, marriage attitudes, relationship with extended kin, environmental factors and government programs, health and health behavior, demographic characteristics, education and employment, and income. The interview with the child in wave 5 (age 9 y) has questions about the following topics: parental supervision and relationship, parental discipline, sibling relationships, routines, school, early delinquency, task completion and behavior, and health and safety. In addition to the surveys, interviewers traveled to the child’s home at waves 3, 4, and 5 (ages 3, 5, and 9 y) to conduct an in-home assessment that included psychometric testing (e.g., Peabody Picture Vocabulary Test, Woodcock–Johnson Passage Comprehension Test, etc.), biometric measurements (e.g., height, weight, etc.), and observations of the neighborhood and home. More information about the Fragile Families data are included in . When we began designing the Fragile Families Challenge, data from waves 1 to 5 (birth to age 9 y) were already available to researchers. However, data from wave 6 (age 15 y) were not yet available to researchers outside of the Fragile Families team. This moment where data have been collected but are not yet available to outside researchers—a moment that exists in all longitudinal surveys—creates an opportunity to run a mass collaboration using the common task method. This setting makes it possible to release some cases for building predictive models while withholding others for evaluating the resulting predictions. Wave 6 (age 15 y) of the Fragile Families study includes 1,617 variables. From these variables, we selected six outcomes to be the focus of the Fragile Families Challenge: 1) child grade point average (GPA), 2) child grit, 3) household eviction, 4) household material hardship, 5) primary caregiver layoff, and 6) primary caregiver participation in job training. We selected these outcomes for many reasons, three of which were to include different types of variables (e.g., binary and continuous), to include a variety of substantive topics (e.g., academics, housing, employment), and to include a variety of units of analysis (e.g., child, household, primary caregiver). All outcomes are based on self-reported data. describes how each outcome was measured in the Fragile Families study. In order to predict these outcomes, participants had access to a background dataset, a version of the wave 1 to 5 (birth to age 9 y) data that we compiled for the Fragile Families Challenge. For privacy reasons, the background data excluded genetic and geographic information (11). The background data included 4,242 families and 12,942 variables about each family. The large number of predictor variables is the result of the intensive and long-term data collection involved in the Fragile Families study. In addition to the background data, participants in the Fragile Families Challenge also had access to training data that included the six outcomes for half of the families (Fig. 2). Similar to other projects using the common task method, the task was to use data collected in waves 1 to 5 (birth to age 9 y) and some data from wave 6 (age 15 y) to build a model that could then be used to predict the wave 6 (age 15 y) outcomes for other families. The prediction task was not to forecast outcomes in wave 6 (age 15 y) using only data collected in waves 1 to 5 (birth to age 9 y), which would be more difficult.
Fig. 2.

Datasets in the Fragile Families Challenge. During the Fragile Families Challenge, participants used the background data (measured from child’s birth to age 9 y) and the training data (measured at child age 15 y) to predict the holdout data as accurately as possible. While the Fragile Families Challenge was underway, participants could assess the accuracy of their predictions in the leaderboard data. At the end of the Fragile Families Challenge, we assessed the accuracy of the predictions in the holdout data.

Datasets in the Fragile Families Challenge. During the Fragile Families Challenge, participants used the background data (measured from child’s birth to age 9 y) and the training data (measured at child age 15 y) to predict the holdout data as accurately as possible. While the Fragile Families Challenge was underway, participants could assess the accuracy of their predictions in the leaderboard data. At the end of the Fragile Families Challenge, we assessed the accuracy of the predictions in the holdout data. The half of the outcome data that was not available for training was used for evaluation. These data were split into two sets: leaderboard and holdout. During the Fragile Families Challenge, participants could assess their predictive accuracy in the leaderboard set. However, predictive accuracy in the holdout set was unknown to participants—and organizers—until the end of the Fragile Families Challenge. All predictions were evaluated based on a common error metric: mean squared error (). includes more information about the background, training, leaderboard, and holdout data. We recruited participants to the Fragile Families Challenge through a variety of approaches including contacting colleagues, working with faculty who wanted their students to participate, and visiting universities, courses, and scientific conferences to host workshops to help participants get started. Ultimately, we received 457 applications to participate from researchers in a variety of fields and career stages (). Participants often worked in teams. We ultimately received valid submissions from 160 teams. Many of these teams used machine-learning methods that are not typically used in social science research and that explicitly seek to maximize predictive accuracy (12, 13). While the Fragile Families Challenge was underway (March 5, 2017 to August 1, 2017), participants could upload their submissions to the Fragile Families Challenge website. Each submission included predictions, the code that generated those predictions, and a narrative explanation of the approach. After the submission was uploaded, participants could see their score on a leaderboard, which ranked the accuracy of all uploaded predictions in the leaderboard data (14). In order to take part in the mass collaboration, all participants provided informed consent to the procedures of the Fragile Families Challenge, including agreeing to open-source their final submissions (). All procedures for the Fragile Families Challenge were approved by the Princeton University Institutional Review Board (no. 8061). As noted above, participants in the Fragile Families Challenge attempted to minimize the mean squared error of their predictions on the holdout data. To aid interpretation and facilitate comparison across the six outcomes, we present results in terms of , which rescales the mean squared error of a prediction by the mean squared error when predicting the mean of the training data (). is bounded above by 1 and has no lower bound. It provides a measure of predictive performance relative to two reference points. A submission with = 0 is no more accurate than predicting the mean of the training data, and a submission with is perfectly accurate.

Results

Once the Fragile Families Challenge was complete, we scored all 160 submissions using the holdout data. We discovered that even the best predictions were not very accurate: of about 0.2 for material hardship and GPA and about 0.05 for the other four outcomes (Fig. 3). In other words, even though the Fragile Families data included thousands of variables collected to help scientists understand the lives of these families, participants were not able to make accurate predictions for the holdout cases. Further, the best submissions, which often used complex machine-learning methods and had access to thousands of predictor variables, were only somewhat better than the results from a simple benchmark model that used linear regression (continuous outcomes) or logistic regression (binary outcomes) with four predictor variables selected by a domain expert: three variables about the mother measured at the child’s birth (race/ethnicity, marital status, and education level) and a measure of the outcome—or a closely related proxy—measured when the child was 9 y (Fig. 3) (). Finally, we note that our procedure—using the holdout data to select the best of the 160 submissions and then using the same holdout data to evaluate that selected submission—will produce slightly optimistic estimates of the performance of the selected submission in new holdout data, but this optimistic bias is likely small in our setting ().
Fig. 3.

Performance in the holdout data of the best submissions and a four variable benchmark model (). A shows the best performance (bars) and a benchmark model (lines). Error bars are 95% confidence intervals (). B–D compare the predictions and the truth; perfect predictions would lie along the diagonal. E–G show the predicted probabilities for cases where the event happened and where the event did not happen. In B–G, the dashed line is the mean of the training data for that outcome.

Performance in the holdout data of the best submissions and a four variable benchmark model (). A shows the best performance (bars) and a benchmark model (lines). Error bars are 95% confidence intervals (). B–D compare the predictions and the truth; perfect predictions would lie along the diagonal. E–G show the predicted probabilities for cases where the event happened and where the event did not happen. In B–G, the dashed line is the mean of the training data for that outcome. Beyond identifying the best submissions, we observed three important patterns in the set of submissions. First, teams used a variety of different data processing and statistical learning techniques to generate predictions (). Second, despite diversity in techniques, the resulting predictions were quite similar. For all outcomes, the distance between the most divergent submissions was less than the distance between the best submission and the truth (). In other words, the submissions were much better at predicting each other than at predicting the truth. The similarities across submissions meant that our attempts to create an ensemble of predictions did not yield a substantial improvement in predictive accuracy (). Third, many observations (e.g., the GPA of a specific child) were accurately predicted by all teams, whereas a few observations were poorly predicted by all teams (Fig. 4). Thus, within each outcome, squared prediction error was strongly associated with the family being predicted and weakly associated with the technique used to generate the prediction ().
Fig. 4.

Heatmaps of the squared prediction error for each observation in the holdout data. Within each heatmap, each row represents a team that made a qualifying submission (sorted by predictive accuracy), and each column represents a family (sorted by predictive difficulty). Darker colors indicate higher squared error; scales are different across subfigures; order of rows and columns are different across subfigures. The hardest-to-predict observations tend to be those that are very different from the mean of the training data, such as children with unusually high or low GPAs (). This pattern is particularly clear for the three binary outcomes—eviction, job training, layoff—where the errors are large for families where the event occurred and small for the families where it did not.

Heatmaps of the squared prediction error for each observation in the holdout data. Within each heatmap, each row represents a team that made a qualifying submission (sorted by predictive accuracy), and each column represents a family (sorted by predictive difficulty). Darker colors indicate higher squared error; scales are different across subfigures; order of rows and columns are different across subfigures. The hardest-to-predict observations tend to be those that are very different from the mean of the training data, such as children with unusually high or low GPAs (). This pattern is particularly clear for the three binary outcomes—eviction, job training, layoff—where the errors are large for families where the event occurred and small for the families where it did not.

Discussion

The Fragile Families Challenge provides a credible estimate of the predictability of life outcomes in this setting. Because of the design of the Fragile Families Challenge, low predictive accuracy cannot easily be attributed to the limitations of any particular researcher or approach; hundreds of researchers attempted the task, and none could predict accurately. However, the Fragile Families Challenge speaks directly to the predictability of life outcomes in only one setting: six specific outcomes, as predicted by a particular set of variables measured by a single study for a particular group of people. Predictability is likely to vary across settings, such as for different outcomes, over different time gaps between the predictors and outcomes, using different data sources, and for other social groups (). Nonetheless, the results in this specific setting have implications for scientists and policymakers, and they suggest directions for future research. Social scientists studying the life course must find a way to reconcile a widespread belief that understanding has been generated by these data—as demonstrated by more than 750 published journal articles using the Fragile Families data (10)—with the fact that the very same data could not yield accurate predictions of these important outcomes. Reconciling this understanding/prediction paradox is possible in at least three ways. First, if one measures our degree of understanding by our ability to predict (8, 15), then the results of the Fragile Families Challenge suggest that our understanding of child development and the life course is actually quite poor. Second, one can argue that prediction is not a good measure of understanding and that understanding can come from description or causal inference. For example, in the study of racial disparities, researchers may build understanding by carefully describing the black–white wealth gap (16), even if they are not able to accurately predict the wealth of any individual. Third, one can conclude that the prior understanding is correct but incomplete because it lacks theories that explain why we should expect outcomes to be difficult to predict even with high-quality data. Insights for how to construct such theories may come from research on the weather (17), the stock market (18), and other phenomena (19–22) where unpredictability is an object of study. Policymakers using predictive models in settings such as criminal justice (23) and child-protective services (24) should be concerned by these results. In addition to the many serious legal and ethical questions raised by using predictive models for decision-making (23–26), the results of the Fragile Families Challenge raise questions about the absolute level of predictive performance that is possible for some life outcomes, even with a rich dataset. Further, the results raise questions about the relative performance of complex machine-learning models compared with simple benchmark models (26, 27). In the Fragile Families Challenge, the simple benchmark model with only a few predictors was only slightly worse than the most accurate submission, and it actually outperformed many of the submissions (). Therefore, before using complex predictive models, we recommend that policymakers determine whether the achievable level of predictive accuracy is appropriate for the setting where the predictions will be used, whether complex models are more accurate than simple models or domain experts in their setting (26–28), and whether possible improvement in predictive performance is worth the additional costs to create, test, and understand the more complex model (26). Ideally, these assessments would be carried out with government administrative data used in policy settings because the properties of these data likely differ from the properties of the Fragile Families data, but legal and privacy issues make it difficult for researchers to access many types of administrative data (29). In addition to providing estimates of predictability in a single setting, the Fragile Families Challenge also provides the building blocks for future research about the predictability of life outcomes more generally. The predictions and open-sourced submissions from participants provide a data source for future study with the Fragile Families sample (). The Fragile Families Challenge also provides a template for one type of mass collaboration in the social sciences (30, 31). There are currently many longitudinal studies happening around the world, all with different study populations, measurement characteristics, and research goals. Each of these studies could serve as the basis for a mass collaboration similar to the Fragile Families Challenge. Progress made in these future mass collaborations might also reveal other social research problems that we can solve better collectively than individually.
  12 in total

Review 1.  Clinical versus actuarial judgment.

Authors:  R M Dawes; D Faust; P E Meehl
Journal:  Science       Date:  1989-03-31       Impact factor: 47.728

2.  Experimental study of inequality and unpredictability in an artificial cultural market.

Authors:  Matthew J Salganik; Peter Sheridan Dodds; Duncan J Watts
Journal:  Science       Date:  2006-02-10       Impact factor: 47.728

3.  Common sense and sociological explanations.

Authors:  Duncan J Watts
Journal:  AJS       Date:  2014-09

Review 4.  Prediction and explanation in social systems.

Authors:  Jake M Hofman; Amit Sharma; Duncan J Watts
Journal:  Science       Date:  2017-02-02       Impact factor: 47.728

Review 5.  Beyond prediction: Using big data for policy problems.

Authors:  Susan Athey
Journal:  Science       Date:  2017-02-02       Impact factor: 47.728

Review 6.  Predicting armed conflict: Time to adjust our expectations?

Authors:  Lars-Erik Cederman; Nils B Weidmann
Journal:  Science       Date:  2017-02-02       Impact factor: 47.728

7.  HUMAN DECISIONS AND MACHINE PREDICTIONS.

Authors:  Jon Kleinberg; Himabindu Lakkaraju; Jure Leskovec; Jens Ludwig; Sendhil Mullainathan
Journal:  Q J Econ       Date:  2017-08-26

8.  Prediction Policy Problems.

Authors:  Jon Kleinberg; Jens Ludwig; Sendhil Mullainathan; Ziad Obermeyer
Journal:  Am Econ Rev       Date:  2015-05

9.  Predicting history.

Authors:  Joseph Risi; Amit Sharma; Rohan Shah; Matthew Connelly; Duncan J Watts
Journal:  Nat Hum Behav       Date:  2019-06-03

10.  PSYCHOLOGY. Estimating the reproducibility of psychological science.

Authors: 
Journal:  Science       Date:  2015-08-28       Impact factor: 47.728

View more
  14 in total

1.  Machine learning uncovers the most robust self-report predictors of relationship quality across 43 longitudinal couples studies.

Authors:  Samantha Joel; Paul W Eastwick; Colleen J Allison; Ximena B Arriaga; Zachary G Baker; Eran Bar-Kalifa; Sophie Bergeron; Gurit E Birnbaum; Rebecca L Brock; Claudia C Brumbaugh; Cheryl L Carmichael; Serena Chen; Jennifer Clarke; Rebecca J Cobb; Michael K Coolsen; Jody Davis; David C de Jong; Anik Debrot; Eva C DeHaas; Jaye L Derrick; Jami Eller; Marie-Joelle Estrada; Ruddy Faure; Eli J Finkel; R Chris Fraley; Shelly L Gable; Reuma Gadassi-Polack; Yuthika U Girme; Amie M Gordon; Courtney L Gosnell; Matthew D Hammond; Peggy A Hannon; Cheryl Harasymchuk; Wilhelm Hofmann; Andrea B Horn; Emily A Impett; Jeremy P Jamieson; Dacher Keltner; James J Kim; Jeffrey L Kirchner; Esther S Kluwer; Madoka Kumashiro; Grace Larson; Gal Lazarus; Jill M Logan; Laura B Luchies; Geoff MacDonald; Laura V Machia; Michael R Maniaci; Jessica A Maxwell; Moran Mizrahi; Amy Muise; Sylvia Niehuis; Brian G Ogolsky; C Rebecca Oldham; Nickola C Overall; Meinrad Perrez; Brett J Peters; Paula R Pietromonaco; Sally I Powers; Thery Prok; Rony Pshedetzky-Shochat; Eshkol Rafaeli; Erin L Ramsdell; Maija Reblin; Michael Reicherts; Alan Reifman; Harry T Reis; Galena K Rhoades; William S Rholes; Francesca Righetti; Lindsey M Rodriguez; Ron Rogge; Natalie O Rosen; Darby Saxbe; Haran Sened; Jeffry A Simpson; Erica B Slotter; Scott M Stanley; Shevaun Stocker; Cathy Surra; Hagar Ter Kuile; Allison A Vaughn; Amanda M Vicary; Mariko L Visserman; Scott Wolf
Journal:  Proc Natl Acad Sci U S A       Date:  2020-07-27       Impact factor: 11.205

Review 2.  Integrating explanation and prediction in computational social science.

Authors:  Jake M Hofman; Duncan J Watts; Susan Athey; Filiz Garip; Thomas L Griffiths; Jon Kleinberg; Helen Margetts; Sendhil Mullainathan; Matthew J Salganik; Simine Vazire; Alessandro Vespignani; Tal Yarkoni
Journal:  Nature       Date:  2021-06-30       Impact factor: 49.962

3.  What failure to predict life outcomes can teach us.

Authors:  Filiz Garip
Journal:  Proc Natl Acad Sci U S A       Date:  2020-04-01       Impact factor: 11.205

4.  Is there evidence for sensitive periods in emotional development?

Authors:  Kristina Woodard; Seth D Pollak
Journal:  Curr Opin Behav Sci       Date:  2020-06-23

5.  Developmental signs of ADHD and autism: a prospective investigation in 3623 children.

Authors:  Matti Cervin
Journal:  Eur Child Adolesc Psychiatry       Date:  2022-06-24       Impact factor: 4.785

Review 6.  Seven steps toward more transparency in statistical practice.

Authors:  Eric-Jan Wagenmakers; Alexandra Sarafoglou; Sil Aarts; Casper Albers; Johannes Algermissen; Štěpán Bahník; Noah van Dongen; Rink Hoekstra; David Moreau; Don van Ravenzwaaij; Aljaž Sluga; Franziska Stanke; Jorge Tendeiro; Balazs Aczel
Journal:  Nat Hum Behav       Date:  2021-11-11

7.  Enhancing long-term forecasting: Learning from COVID-19 models.

Authors:  Hazhir Rahmandad; Ran Xu; Navid Ghaffarzadegan
Journal:  PLoS Comput Biol       Date:  2022-05-19       Impact factor: 4.779

8.  Social relationships and children's perceptions of adversity.

Authors:  Karen E Smith; Seth D Pollak
Journal:  Child Dev Perspect       Date:  2021-08-24

9.  Responsible Use of Open-Access Developmental Data: The Adolescent Brain Cognitive Development (ABCD) Study.

Authors:  Cortney Simmons; May I Conley; Dylan G Gee; Arielle Baskin-Sommers; Deanna M Barch; Elizabeth A Hoffman; Rebekah S Huber; William G Iacono; Bonnie J Nagel; Clare E Palmer; Chandni S Sheth; Elizabeth R Sowell; Wesley K Thompson; B J Casey
Journal:  Psychol Sci       Date:  2021-05-27

10.  Insect morphometry is reproducible under average investigation standards.

Authors:  Sándor Csősz; Bernhard Seifert; István Mikó; Brendon E Boudinot; Marek L Borowiec; Brian L Fisher; Matthew Prebus; Jayanthi Puniamoorthy; Jean-Claude Rakotonirina; Nicole Rasoamanana; Roland Schultz; Carolyn Trietsch; Jonah M Ulmer; Zoltán Elek
Journal:  Ecol Evol       Date:  2020-12-08       Impact factor: 2.912

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.