Literature DB >> 36124257

Assessment of the predictive power of a causal variable: An application to the Head Start impact study.

Sun Yeop Lee1, Rockli Kim2,3, Justin Rodgers4, S V Subramanian4,5.   

Abstract

In a study attempting to estimate a causal effect of a causal variable, an assessment of the predictive power of the causal variable can shed light on the heterogeneity around its average effect. Using data from the Head Start Impact Study, a randomized controlled trial of the Head Start, a nation-wide early childhood education program in the United States, we provide a parallel comparison between measures of average effect and predictive power of the Head Start on five cognitive outcomes. We observed that one year of the Head Start increased scores for all five outcomes, with effect sizes ranging from 0.12 to 0.19 standard deviations. Percent variation explained by the Head Start ranged from 0.56 to 1.62%. For binary versions of the outcomes, the overall pattern remained; the Head Start on average improved the outcomes by meaningful magnitudes. In contrast, in a fully adjusted model, the Head Start only improved area under the curve (AUC) by less than 1% and its influence on the variance of predicted probabilities was negligible. The Head-Start-only model only achieved AUC ranging from 50.22 to 55.24%. Negligible predictive power despite the significant average effect suggests that the heterogeneity in effects may be large. The average effect estimates may not generalize well to different populations or different Head Start program settings. Assessment of the predictive power of a causal variable in randomized data should be a routine practice as it can provide helpful information on the causal effect and especially its heterogeneity.
© 2022 The Author(s).

Entities:  

Year:  2022        PMID: 36124257      PMCID: PMC9482140          DOI: 10.1016/j.ssmph.2022.101223

Source DB:  PubMed          Journal:  SSM Popul Health        ISSN: 2352-8273


Introduction

The Head Start, one of the largest and the only federally funded early childhood education program in the United States, aims to enhance school readiness of children of low-income families by providing educational, health, and social services (Head Start & Early Head Start, 2020). In 2002, the Head Start Impact Study (HSIS), a randomized controlled trial (RCT) of the Head Start, was designed to test the causal effect of the Head Start on children's developmental outcomes, including cognition, social-emotional measures, health, and parenting (Puma et al., 2010). Official reports of the HSIS reported one- or two-year positive causal effects of the Head Start, especially for cognitive outcomes, although they faded away in a few years (Puma et al., 2010, 2012). Follow-up studies that conducted subgroup analyses additionally found that the effects were larger and more pronounced for certain subgroups, such as children with low cognitive abilities or Spanish as primary language (Bitler et al., 2014), or children who would have received home-based care had they not enrolled in the Head Start (Feller et al., 2016; Zhai et al., 2014). Existing literature on the heterogeneity of the Head Start effect is extensive, but most focused on average effect, taking a mean-centric approach (Lee et al., 2021). Little attention has been directed to an assessment of the predictive power measured by various metrics such as percent variation explained or area under the curve (AUC). Predictive power measures the capacity of a variable to correctly identify or estimate outcomes in independent data and is generally assessed during predictive model development. In a study attempting to estimate a causal effect of a causal variable (e.g., a random assignment to the Head Start), an assessment of the predictive power of the causal variable can shed light on the heterogeneity around its average effect. In observational settings, exposures (i.e., birthweight) with well-established average associations (or effects) have often been observed to have low predictive power, suggesting that they are not necessarily suitable for individual classification (e.g., medical screening, eligibility criteria for social programs and policies) (Kim et al., 2018; Merlo et al., 2017; Swaminathan et al., 2020); a strong effect on average does not necessarily mean strong predictive power (Pepe et al., 2004; Varga et al., 2020; Wald et al., 1999). In an RCT setting, an evaluation of the predictive power is rarely considered and is not a regular analytic practice, despite the complementary nature of assessments of average effect and predictive power (Merlo et al., 2017; Shmueli, 2010). As individual causal effects are expected to be heterogeneous (Kravitz et al., 2004; Plewis, 2002), the magnitude of this heterogeneity should be examined. It has implications when deciding whether to scale up, terminate, or tailor to specific populations (Cintron et al., 2022; Subramanian et al., 2018). Therefore, using the HSIS data, we evaluated the Head Start in terms of both average effect (e.g., effect size, odds ratio (OR)) and the predictive power (e.g., percent variation explained, AUC, variance of predicted probabilities). Specifically, we estimated average effects on and predictive power for five outcomes (i.e., Peabody Picture Vocabulary Test (PPVT), Letter-Word Identification, Applied Problem, Spelling, and Pre-Academic Skill) after one year of the Head Start. Since the HSIS has already been extensively analyzed in terms of average effects, here we only focused on outcomes with previous reports of statistically significant one-year average effects (Lee, Rodgers, et al., 2022). Therefore, the estimates for average effects are reproductions of previous results. Predictive power of a known causal variable is the inferential target of interest in this study.

Methods

Data

The HSIS was designed to evaluate the effectiveness of the Head Start on children's cognitive, behavioral, social-emotional, and health outcomes (Puma et al., 2010). The Head Start programs provide early childhood education and social services to children (i.e., aged <1–5 years) and their families through local non-profit and for-profit community agencies such as centers and schools. The HSIS implemented a multi-stage sampling procedure for recruiting participants. The sampling procedure first categorized the initial 1715 programs into 161 geographic clusters and 25 strata based on region, state-level childcare policy, race/ethnicity, and urbanicity. Next, one cluster was randomly selected from each stratum, excluding programs that were closed, merged, or saturated and grouping those with small sample sizes. These programs were then stratified by type and local contextual characteristics. Then, three programs were randomly chosen from each stratum. Lastly, centers were randomly chosen from the final set of programs. This multi-stage sampling procedure resulted in a final sample of 4442 children at 378 centers within 84 programs. This sample was then followed over time from 2002 to 2008, with measurements collected at baseline and annual to biennial follow-ups on a host of developmental and related measures. Detailed explanations of the sampling procedure and other study protocols are available in the official HSIS reports (Puma et al., 2010, 2011).

Treatment

The Head Start intervention included educational, health, nutritional, and social services with the goal of improving school readiness and child development. All Head Start centers must adhere to the Head Start Performance Standards, which are federally regulated to ensure the comprehensiveness and quality of the services provided by the centers (Puma et al., 2010). Thus, the treatment is a mixture of various services with the nation-wide, pre-specified standards. The assignment to the Head Start was randomized within each Head Start center, offering the assigned children to participate in the Head Start. To potentially benefit as many children as possible with the program, the randomization was intentionally designed to yield a higher proportion of children in the treatment group. The treatment of interest specifically represents the assignment to one year of the Head Start in the baseline year. Like any RCT, noncompliance to the treatment/control assignment occurred; 12% of the control group enrolled in the Head Start, and 19% of the treatment group did not enroll in the Head Start.

Outcomes

Children were assessed on a multitude of developmental outcomes over the course of the HSIS, commencing during children's preschool years (ages 3–4 years) in the baseline year of 2002, with follow-ups in Spring 2003, Spring 2004, Spring 2005, Spring 2006, and Spring 2008 (3rd grade). In this study, we focused on outcomes with previous reports of statistically significant average effects after one year of the Head Start: PPVT, Letter-Word Identification, Applied Problem, Spelling, and Pre-Academic Skill. Outcomes without average effects would not have meaningful predictive power. Predictive power of the Head Start for the outcomes with average effects is of our interest. Cognitive outcomes were measured by one-on-one child assessments for 45–60 min. PPVT measures receptive vocabulary in standard English (Cronbach's α = 0.62–0.84). Letter-Word Identification measures the ability to identify letters and words from a picture or isolated letters and words (α = 0.82–0.94). Spelling measures the ability to correctly spell spoken words (α = 0.70–0.94). Applied Problem measures an ability to analyze and solve math problems (α = 0.85–0.90). Pre-Academic Skill is a composite measure of Letter-Word Identification, Applied Problems, and Spelling (α = 0.67–0.85).

Covariates

While the HSIS was an RCT, the HSIS official reports recommended covariate adjustment to enhance statistical precision and adjust for any systematic bias at baseline (Puma et al., 2010, 2011). Therefore, for the average effect assessment, we adjusted for children's sociodemographic variables and study-related variables. Sociodemographic variables included gender (male, female), race/ethnicity (White/other, Black, Hispanic), primary language at baseline (English, Spanish), special needs (yes, no), primary caregiver's age (continuous), teen mom at birth (yes, no), living with a single parent (yes, no), recent immigrant parents (yes, no), parents' marital status (not married, married, separated/divorced/widowed), parental education level (less than high school, high school graduates, beyond high school), urbanicity (urban, rural), household risk (low, moderate, high). Study-related variables included age cohort (age 3, age 4) and baseline measures of the outcomes in this study.

Statistical analyses

We employed two distinct analytic approaches: 1) an average effect assessment and 2) a predictive power assessment. All analyses were performed in R (version 4.1.1) (R Core Team, 2020). Three-level multilevel linear regressions were specified in order to account for the complex sampling design (level-3: program; level-2: center; level-1: child). For average effect, we estimated the fixed effect parameter estimate of the Head Start, adjusting for covariates and random effects. The average effect was presented as both a raw score and Cohen's d effect size (Cohen, 1992). For predictive power, we estimated child-level variances of an outcome in a full model with the Head Start (i.e., Model 1) and a full model without the Head Start (i.e., Model 2). Model 1 was specified as,where is an outcome variable for child in center in program , is a vector of child-level covariates, and is an indicator variable for the treatment group (i.e., the Head Start). Total variance is partitioned into the program-level (), the center-level (), the child-level (). Model 2 is equivalent to Model 1 excluding the treatment group indicator variable. We report the percent change between the child-level variances of the two models as percent variation explained by the Head Start. Binary versions of the outcomes were also utilized so that our analyses could be performed in binary outcome scenarios. The outcomes were dichotomized at the 25th, 50th, and 75th percentiles and were indicators for scoring above the varying thresholds. Three-level multilevel logistic regressions with (i.e., Model 3) and without the Head Start (i.e., Model 4) were estimated to assess the average effects and predictive power improvements measured by AUC contributed to the Head Start. Model 3 was specified as, Additionally, variances of predicted probabilities from Model 3 and 4 were compared. In general, a better-calibrated model would have more extreme predictions, hence the greater variance. To visualize, predicted probabilities for the 50th percentile threshold dichotomized outcomes were compared side by side in histograms. Lastly, a logistic regression with only the Head Start as an independent variable (i.e., Model 5) was run to estimate independent predictive power (i.e., AUC) of the Head Start.

Results

At baseline, there was a total sample size of 4442 children participating in the study, of which 2646 were assigned to the Head Start group and 1796 were assigned to the control group (Table 1). A slightly higher proportion of the participants were Hispanics/others (36.0%) than White (33.7%) and Black (30.3%). About a quarter (25.7%) used Spanish as a primary language. Approximately half (50.4%) of children lived with a single parent, 38.0% had mothers who did not graduate from high school, and approximately one-fifth (19.2%) were recent immigrants. Out of the 4442 children, 81–82% of them composed the final analytic sample depending on the availability of data on the outcomes and covariates (Table 2).
Table 1

Sample characteristics at baseline by the treatment and control groups.

OverallControlHead StartMissing
N444217962646
Age cohort (%)32449 (55.1)985 (54.8)1464 (55.3)0
41993 (44.9)811 (45.2)1182 (44.7)
Gender (%)male2239 (50.4)912 (50.8)1327 (50.2)0
Race/ethnicity (%)White1496 (33.7)623 (34.7)873 (33.0)0
Black1348 (30.3)536 (29.8)812 (30.7)
Hispanic & others1598 (36.0)637 (35.5)961 (36.3)
Primary language (%)English3301 (74.3)1345 (74.9)1956 (73.9)0
Spanish1141 (25.7)451 (25.1)690 (26.1)
Parental education (%)more1274 (28.7)505 (28.1)769 (29.1)0
high school1481 (33.3)592 (33.0)889 (33.6)
less1687 (38.0)699 (38.9)988 (37.3)
Single parent (%)2239 (50.4)907 (50.5)1332 (50.3)0
Recent immigrant (%)855 (19.2)337 (18.8)518 (19.6)0
Marital status (%)married1972 (44.4)806 (44.9)1166 (44.1)0.1
separated & divorced &widowed724 (16.3)290 (16.1)434 (16.4)
never1742 (39.2)699 (38.9)1043 (39.4)
Special needs (%)570 (12.8)204 (11.4)366 (13.8)0
Teen mom (%)752 (16.9)330 (18.4)422 (15.9)0
Urban (%)3746 (84.3)1513 (84.2)2233 (84.4)0
Household risk (%)low3383 (76.2)1399 (77.9)1984 (75.0)0
moderate741 (16.7)277 (15.4)464 (17.5)
high318 (7.2)120 (6.7)198 (7.5)
Caregiver's age (mean (SD))28.91 (7.34)28.65 (7.06)29.08 (7.52)0
Table 2

Measures of average effect and predictive power on continuous outcomes.



Average effect
Predictive power

Sample size (follow-up rate)
Model 1 regression coefficient (95% CI); Cohen's d
Model 1 child-level variance
Model 2 child-level variance
Percent variation explained by Head Start
Outcome
 PPVT3621 (82%)5.66 (4.05, 7.26); 0.14557.74565.291.34
 Letter-Word Identification3627 (82%)5.17 (3.78, 6.55); 0.19416.47423.131.57
 Applied Problem3601 (81%)3.38 (1.93, 4.84); 0.12454.99457.720.60
 Spelling3635 (82%)3.02 (1.72, 4.31); 0.12365.38367.430.56
 Pre-Academic Skill3594 (81%)3.82 (2.81, 4.83); 0.17218.65222.241.62

Note. CI, confidence intervals., PPVT, Peabody Picture Vocabulary Test.

Model 1: a full model with the Head Start.

Model 2: a full model without the Head Start.

Sample characteristics at baseline by the treatment and control groups. Measures of average effect and predictive power on continuous outcomes. Note. CI, confidence intervals., PPVT, Peabody Picture Vocabulary Test. Model 1: a full model with the Head Start. Model 2: a full model without the Head Start. In a series of multilevel linear regressions adjusting for the selected covariates and random effects, one year of Head Start increased scores for PPVT (), Letter-Word Identification (), Applied Problem (), Spelling (), and Pre-Academic Skill () (Table 2). Percent variation explained by the Head Start for PPVT, Letter-Word Identification, Applied Problem, Spelling, and Pre-Academic Skill were 1.34%, 1.57%, 0.60%, 0.56%, and 1.62%, respectively. When being run at varying thresholds for dichotomizing outcomes, a series of multilevel logistic regressions adjusting for the covariates and random effects had an overall pattern of the Head Start increasing the odds of scoring high on cognitive outcomes with odds ratios ranging from 1.16 to 1.66 for PPVT, 1.47 to 1.77 for Letter-Word Identification, 1.07 to 1.34 for Applied Problems, 1.24 to 1.47 for Spelling, and 1.54 to 1.58 for Pre-Academic Skill (Table 3). In contrast, the Head Start did not meaningfully contribute to improvement in AUC with the difference ranging from 0.01 to 0.18% for PPVT, 0.13–0.49% for Letter-Word Identification, 0.00–0.12% for Applied Problems, 0.04–0.22% for Spelling, and 0.14–0.23% for Pre-Academic Skill. Moreover, the variance of predicted probabilities was negligibly affected by the addition of the Head Start in the model (Table 3; Fig. 1). In a logistic regression that only included the Head Start, the AUC for discriminating children with high cognitive scores ranged from 50.22 to 53.16% for PPVT, 53.98–55.24% for Letter-Word Identification, 50.66–52.02% for Applied Problem, 52.11–53.02% for Spelling, and 53.22–53.46% for Pre-Academic Skill.
Table 3

Measures of average effect and predictive power on binary outcomes (i.e., “high” scores).


Average effect
Predictive power

Model 3
Model 3
Model 4
Model 4 → Model 3
Model 5

OR (95% CI)
AUC (%)
AUC (%)
Improvement in AUC contributed to Head Start (%)
Change in the variance of predicted probability
AUC (%)
Outcome
PPVT
 >0.251.66 (1.36, 2.02)88.5888.400.180.01853.16
 >0.501.42 (1.18, 1.71)89.9089.780.130.00651.63
 >0.751.16 (0.92, 1.45)92.5492.530.010.00150.22
Letter-Word Identification
 >0.251.62 (1.38, 1.90)80.4479.950.490.03354.56
 >0.501.77 (1.50, 2.09)83.7283.060.650.03455.24
 >0.751.47 (1.20, 1.80)87.0586.920.130.01053.98
Applied Problem
 >0.251.34 (1.11, 1.60)85.6285.500.120.00852.02
 >0.501.28 (1.08, 1.52)85.8785.760.120.00551.56
 >0.751.07 (0.86, 1.34)90.1490.140.000.00050.66
Spelling
 >0.251.47 (1.24, 1.74)84.5684.340.220.01353.02
 >0.501.30 (1.09, 1.54)85.4985.360.130.00552.12
 >0.751.24 (1.01, 1.53)88.4288.380.040.00452.11
Pre-Academic Skill
 >0.251.58 (1.30, 1.90)86.4286.210.220.01753.22
 >0.501.56 (1.30, 1.87)88.2187.980.230.01153.23
 >0.751.54 (1.23, 1.92)91.3391.200.140.00953.46

Note. OR, odds ratio. CI, confidence intervals. AUC, area under the curve.

Model 3: a full model with the Head Start.

Model 4: a full model without the Head Start.

Model 5: a simple model with only the Head Start.

Fig. 1

Comparison of predicted probabilities with and without the Head Start. Predicted probabilities for “high” scores from Model 3 (with the Head Start; right) and Model 4 (without the Head Start; left) are shown. Probabilities are on the x-axis, and densities are the y-axis.

Measures of average effect and predictive power on binary outcomes (i.e., “high” scores). Note. OR, odds ratio. CI, confidence intervals. AUC, area under the curve. Model 3: a full model with the Head Start. Model 4: a full model without the Head Start. Model 5: a simple model with only the Head Start. Comparison of predicted probabilities with and without the Head Start. Predicted probabilities for “high” scores from Model 3 (with the Head Start; right) and Model 4 (without the Head Start; left) are shown. Probabilities are on the x-axis, and densities are the y-axis.

Discussion

Using the HSIS data, we present findings in parallel for two distinct evaluation metrics: average effect and predictive power. Across the outcomes with meaningfully sized average effects after one year of the Head Start, we found that the predictive power was consistently small, regardless of whether the outcomes were continuous or binary. The negligible predictive power was also observed consistently across varying binary thresholds. Average effect assessments on continuous outcomes were reproductions of previous studies. After one year of the Head Start, those who were assigned to the Head Start scored higher on a range of cognitive outcomes than those who were assigned to the control group. Across the continuous outcomes, the effect size was less than 0.2 in Cohen's d. While the effect size of 0.2 is considered as “small” based on Cohen's simple typology (Cohen, 1992), it can be considered “large” if the cost of the Head Start is taken into account in a cost-effectiveness framework (Harris, 2009; Ludwig & Phillips, 2008). In the binary versions of the outcomes, the Head Start had consistent, positive effects across varying thresholds of dichotomization. For PPVT and Applied Problem, the Head Start effect was larger when the children were grouped by lower thresholds, which was in agreement with previous findings in which those with lower baseline cognitive ability benefited more from the Head Start (Bitler et al., 2014; Lee, Rodgers, et al., 2022). To clarify, estimating average effect of the Head Start in this study was not to question conclusions from previous studies (Bitler et al., 2014; Chor, 2018; Feller et al., 2016; Lipscomb et al., 2013; McCoy et al., 2016; Miller et al., 2016; Zhai et al., 2014), but to provide a parallel comparison to the predictive power assessment. In contrast to the meaningfully sized effect on average, the predictive power assessment showed a weak ability of the Head Start to predict outcomes after one year of the study. In continuous outcomes, the Head Start explained far less than 2% of the total between-child variance for each outcome. While there are no standard guidelines, one report has suggested using Cohen's typology in which 2% is considered “small” (Cohen, 1992; Lorah, 2018). In binary scenarios, the results are clearer. Regardless of the thresholds, the Head Start did not have any meaningful improvement in AUC with less than 1%. In addition, AUC of the Head Start alone was nearly identical to 50%, meaning that the Head Start predicts the outcomes almost at random. Furthermore, the predicted probabilities were only negligibly affected in terms of their variances. The distributions of the predicted probabilities were almost unchanged when the Head Start was considered, reflecting its limited influence on the predictive power. Our findings collectively showed that the Head Start had some average effects of significance but negligible predictive power, suggesting that the heterogeneity in individual effects may be large. Indeed, many studies have explored heterogeneous effects (Lee et al., 2021), and several studies found that large amounts of the heterogeneity was unexplained by measured variables in the HSIS data (Ding et al., 2016, 2019; Lee, Rodgers, et al., 2022). Such heterogeneity may be due to the heterogeneous populations included in the HSIS or heterogeneous implementations of the Head Start programs across the United States. Either way, the large heterogeneity in effects may mean that the average effect estimates are not generalizable to different populations or different Head Start program settings. The present study has some limitations. First, measurement errors in the outcomes may inflate estimates of variance, erroneously decreasing the predictive power. However, such an error would not explain the level of predictive power we observed in this study. Second, we did not adjust for noncompliance of the treatment status. Previous works has already thoroughly investigated the Head Start effects adjusting for noncompliance. We did not have an aim at replicating this, and our intent-to-treat estimates are still unbiased and can achieve our study objective of comparing predictive power to average effect. If noncompliance was adjusted, the average effect and predictive power may have been slightly higher, but based on previous simulations on the relationships between average effect and predictive power (Pepe et al., 2004; Wald et al., 1999), our conclusion would not change. Using the RCT data of the Head Start, we found that while the Head Start had significant effects on cognitive outcomes on average, it did not have an ability to predict the outcomes well. This indicates that the heterogeneity in the individual effects across children is quite large. While the Head Start has been labeled as cost-effective and indeed has positive effects on cognitive outcomes on average, the magnitude of heterogeneity in individual effects should also be assessed. Furthermore, tailoring the program to specific subgroups or settings may be important in the case of the Head Start. Assessment of the predictive power of a causal variable in randomized data should be a routine practice as it can provide helpful information on the causal effect and especially its heterogeneity. Beyond the HSIS data, the predictive power of treatments, interventions, and programs that are well-known to be effective on average should be assessed to gain a better understanding of their effects.

Funding

This work was supported by the (grant ID: 75602).

Author contributions

RK and SVS conceptualized and designed the study. SYL contributed to the conceptualization of the study, led interpretation of the data, conducted the final analyses, and wrote the manuscript. JR contributed to the initial analyses as well as to writing the first draft of the manuscript. All authors approved of the final draft.

Ethical statement

The HSIS data are not collected specifically for this study and no one on the study team has access to identifiers linked to the data. These activities do not meet the regulatory definition of human subject research. As such, an Institutional Review Board (IRB) review is not required. The Harvard Longwood Campus IRB allows researchers to self-determine when their research does not meet the requirements for IRB oversight via guidance online regarding when an IRB application is required using an IRB Decision Tool.

Declaration of competing interest

The authors have no competing interests.
  18 in total

Review 1.  Limitations of the odds ratio in gauging the performance of a diagnostic, prognostic, or screening marker.

Authors:  Margaret Sullivan Pepe; Holly Janes; Gary Longton; Wendy Leisenring; Polly Newcomb
Journal:  Am J Epidemiol       Date:  2004-05-01       Impact factor: 4.897

2.  Evidence-based medicine, heterogeneity of treatment effects, and the trouble with averages.

Authors:  Richard L Kravitz; Naihua Duan; Joel Braslow
Journal:  Milbank Q       Date:  2004       Impact factor: 4.911

3.  Association is not prediction: A landscape of confused reporting in diabetes - A systematic review.

Authors:  Tibor V Varga; Kristoffer Niss; Angela C Estampador; Catherine B Collin; Pope L Moseley
Journal:  Diabetes Res Clin Pract       Date:  2020-10-15       Impact factor: 5.602

4.  The "average" treatment effect: A construct ripe for retirement. A commentary on Deaton and Cartwright.

Authors:  S V Subramanian; Rockli Kim; Nicholas A Christakis
Journal:  Soc Sci Med       Date:  2018-04-19       Impact factor: 4.634

5.  Does Head Start differentially benefit children with risks targeted by the program's service model?

Authors:  Elizabeth B Miller; George Farkas; Greg J Duncan
Journal:  Early Child Res Q       Date:  2016 1st Quarter

6.  Multigenerational Head Start Participation: An Unexpected Marker of Progress.

Authors:  Elise Chor
Journal:  Child Dev       Date:  2016-11-21

7.  Head Start's impact is contingent on alternative type of care in comparison group.

Authors:  Fuhua Zhai; Jeanne Brooks-Gunn; Jane Waldfogel
Journal:  Dev Psychol       Date:  2014-10-20

8.  Contribution of socioeconomic factors to the variation in body-mass index in 58 low-income and middle-income countries: an econometric analysis of multilevel data.

Authors:  Rockli Kim; Ichiro Kawachi; Brent A Coull; S V Subramanian
Journal:  Lancet Glob Health       Date:  2018-07       Impact factor: 26.763

Review 9.  Long-term effects of head start on low-income children.

Authors:  Jens Ludwig; Deborah A Phillips
Journal:  Ann N Y Acad Sci       Date:  2007-10-22       Impact factor: 5.691

10.  Differential Effectiveness of Head Start in Urban and Rural Communities.

Authors:  Dana Charles McCoy; Pamela A Morris; Maia C Connors; Celia J Gomez; Hirokazu Yoshikawa
Journal:  J Appl Dev Psychol       Date:  2016 Mar-Apr
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.