Literature DB >> 27774035

More reliable inference for the dissimilarity index of segregation.

Rebecca Allen1, Simon Burgess2, Russell Davidson3, Frank Windmeijer2.   

Abstract

The most widely used measure of segregation is the so-called dissimilarity index. It is now well understood that this measure also reflects randomness in the allocation of individuals to units (i.e. it measures deviations from evenness, not deviations from randomness). This leads to potentially large values of the segregation index when unit sizes and/or minority proportions are small, even if there is no underlying systematic segregation. Our response to this is to produce adjustments to the index, based on an underlying statistical model. We specify the assignment problem in a very general way, with differences in conditional assignment probabilities underlying the resulting segregation. From this, we derive a likelihood ratio test for the presence of any systematic segregation, and bias adjustments to the dissimilarity index. We further develop the asymptotic distribution theory for testing hypotheses concerning the magnitude of the segregation index and show that the use of bootstrap methods can improve the size and power properties of test procedures considerably. We illustrate these methods by comparing dissimilarity indices across school districts in England to measure social segregation.

Entities:  

Keywords:  Bootstrap methods; Dissimilarity index; Hypothesis testing; Segregation

Year:  2015        PMID: 27774035      PMCID: PMC5054828          DOI: 10.1111/ectj.12039

Source DB:  PubMed          Journal:  Econom J        ISSN: 1368-4221            Impact factor:   4.571


INTRODUCTION

Segregation remains a major topic of research in a number of contexts, such as neighbourhoods, workplaces and schools. Researchers study segregation by poverty status, by sex and by ethnicity, among other characteristics. Almost always, these studies are comparative in some way; for example, arguing that ethnic segregation in neighbourhoods is higher in one city than another, or that segregation by sex in some occupation has changed over time. There is often also an implicit or explicit causal model in mind, and the difference in segregation is associated with some behavioural process. However, the inferential framework for segregation indices is underdeveloped, a fact that limits the progress that can be made. In this paper, we propose an approach to strengthen this framework. It is central to our approach to think of segregation as the outcome of a process of assignment. This includes the assignment of people to neighbourhoods, workers to jobs, or pupils to schools. In general, this allocation is likely to be the result of the interlocking decisions of different agents rather than of a dictator. This perspective offers a number of advantages. First, it ties the outcome to a set of processes that can be analysed and estimated. Second, it makes it clear that the observed outcome is one of a set of possible outcomes, and so naturally leads on to a framework for statistical inference. Third, the connection with the underlying processes makes explicit that it is this systematic or behaviour‐based segregation that is the object of interest in terms of analysing the causes of segregation. There is a large body of literature concerning the measurement of segregation, with a number of indices in use, all with differing properties. The most widely used measure of segregation is the dissimilarity index, D, defined below; see Duncan and Duncan (1955). It is now widely understood that this measure also reflects randomness in the allocation of individuals to units (i.e. it measures deviations from evenness, not systematic segregation). Furthermore, the impact of randomness on D depends on the nature of the context (made precise below). This makes one of the prime tasks in the measurement of segregation difficult, namely making statements on true differences in segregation between cities, school districts, industries or time periods. For example, the overall proportion of the minority group influences this because a very small minority group is more likely to be unevenly distributed across units by chance, compared to a larger minority group. This problem is particularly acute with small unit sizes. This is easy to see in the following example. Consider a large population, half male and half female. Suppose they are assigned to work in two very large firms. A random assignment process would produce an outcome close to a 50 : 50 male–female split in each firm and an estimated D of about zero. However, if they were allocated to many firms of size 2, then a random assignment procedure would lead to many all‐female firms, many all‐male firms and many mixed firms, and a high value for D. The high value reflects a strong deviation from evenness despite pure randomness. Others have noted the problem of small unit size in the measurement of segregation; see, e.g. Carrington and Troske (1997). They proposed an adjustment to segregation indices that has since been used by researchers measuring workplace segregation where small units are particularly likely (e.g. Hellerstein and Neumark, 2008) and measuring school segregation (e.g. Söderström and Uusitalo, 2010). In comparing segregation across areas or time, small unit bias should be of concern to researchers for two reasons. First, the size of the bias will differ across comparison areas, potentially leading to an incorrect ranking of levels of segregation across areas. Second, the presence of small unit bias makes a correlation between measured segregation index values and a potentially causal variable, say X, difficult to interpret. It will impact on the estimated effect of X on measured segregation, even if the parameters of the problem (unit size, minority fraction and population) do not vary across areas. In addition, it is likely that the bias as a function of these parameters will be correlated with X, making the true relationship between X and D difficult to identify. The variable X could, for example, be income differentials. If one were to investigate racial segregation in schools in an area, then one explanation of racial segregation, as indicated by a high value of D, could be income inequality between the two groups. Income inequality could be the cause of neighbourhood and hence school segregation. If there is no income inequality between the two groups in the area, then this could be indicative of behaviour due to other preferences. In this paper, we propose an inferential framework for the canonical segregation measure, D, based on an underlying statistical model. This set‐up is related to, but different from, that used by Ransom (2000). He derives (asymptotic) inference procedures for D by specifying the sampling variation of a multinomial distribution. In Section 2., we specify the assignment problem in a very general way, and set out the difference in assignment probabilities that underlies the resulting segregation. From this, we derive bias adjustments to D in Section 3., and a likelihood ratio test for the presence of any systematic segregation in Section 4.. One of our bias adjustments is based on a simple bootstrap bias correction; other adjustments use the asymptotic normal distribution of the observed frequencies. Following Ransom (2000), in Sections 5. and 6., we further develop the asymptotic distribution theory for testing hypotheses concerning the magnitude of the segregation index and we show that use of bootstrap methods can improve the size and power properties of test procedures considerably. As in Ransom (2000), our asymptotic distribution theory relies on the number of units being fixed, with unit sizes going to infinity. Our results indicate that our methods work well in settings such as our analysis of social segregation in English schools, where the average number of units (schools) in the local authorities are about 55, with the average number of pupils per school equal to 38. Rathelot (2012) recently proposed a Beta‐Binomial mixture model to describe segregation and has shown that it performs well in a setting with many small units (i.e. under asymptotics where the number of units goes to infinity); see also d'Haultfœuille and Rathelot (2011). In Section 7., we present a brief discussion of the measure proposed by Rathelot (2012) and also the one proposed by Carrington and Troske (1997). In Section 8., we illustrate our methods in an example of social segregation in schools in England. We conclude in Section 9..

STATISTICAL FRAMEWORK

Underlying an assignment of individuals to units is an allocation process. This might be purely random, or it may be influenced by the actions of agents, including those whose allocation we are studying, as well as others. This systematic allocation process will, in general, reflect the preferences and constraints of both the individual (such as preferences for racial composition of neighbourhood or ability to pay for houses in a particular neighbourhood) and of the unit to match with particular individuals (such as a firm's desire for highly educated workers or school admissions procedures that favour children of parents of a particular religious denomination). Typically, the research question is about characterizing segregation arising from this behaviour. Our notation is as follows. There are units , nested within an area. Individuals either have or do not have a characteristic measurable on a dichotomous scale, . This could be black ethnicity, female sex or poverty status. The number of individuals in the area of status c is denoted . Individuals are assigned to units and we observe the resulting allocations, individuals in unit j having status c. The total number of individuals in unit j is . There are many indices used to measure segregation; see Duncan and Duncan (1955), Massey and Denton (1988), and White (1986) for an overview. The formula for each provides an implicit definition of segregation. Massey and Denton (1988) characterize segregation along five dimensions: evenness (dissimilarity), exposure (isolation), concentration (the amount of physical space occupied by the minority group), clustering (the extent to which minority neighbourhoods abut one another) and centralization (proximity to the centre of the city). Throughout this paper, we use the index of dissimilarity (denoted D), the most popular unevenness index in the literature. However, our analysis can be extended to other unevenness segregation indices. The formula for the index of dissimilarity D in the area, which is bounded by 0 (no segregation) and 1, is given by Duncan and Duncan (1955) as1 The basis for an allocation procedure is a set of probabilities , which specify the probability that an individual is assigned to unit j, , conditional on the individual being of status c. We define systematic segregation as being present when there exists j such that . We can see the relationship between D and the probabilities of the underlying allocation process by noting that the fractions , , are estimates of these probabilities. With , the index of dissimilarity (2.1) is just one half of . Formally, the allocation process is as follows. An area population of n individuals, with a given proportion with status , is allocated to J units, according to the probabilities . Each individual is allocated independently of the others. All implicit dependences of group formations are captured by the allocation probabilities . The outcomes of this process are the allocations and . Clearly, unit sizes are not fixed in this set‐up as they are equal to and are therefore determined by the stochastic allocation. The expected unit sizes are given by . We can now interpret the index of dissimilarity as an estimator for the population quantity: It is clear that if and only if for all . From the allocation process described above, we can estimate the conditional probabilities by maximum likelihood. As the allocations are two independent multinomial distributions, the log‐likelihood function, given the observed allocations, is given by Clearly, the maximum likelihood estimates are given by and , , exactly the same as the estimates that appear in D. Ransom (2000) proposed the following statistical model for a random sample of size n: where is the joint probability of observing an individual with status c and in unit j in the sample. Mora and Ruiz‐Castillo (2007), and references therein, consider a similar set‐up for an information index of multigroup segregation. Ransom (2000, p. 458) notes that this model is not appropriate when the population is observed, because then are known. The parameters are not those that enter the segregation index D pop, which are the conditional probabilities . Our model is applicable even when we observe the complete, finite population, because randomness is achieved by the random allocation process to units. Our statistical model is for a finite population of fixed size , with parameters , , , and is given by The logarithm of this expression is just the log‐likelihood (2.2). In the remainder of the paper, we focus on this particular model. Our design is particularly well suited for our analysis of social segregation in schools in England. The provision of education in England is organized at the district, or local authority (LA), level. Pupils within an LA have to choose a school within that district with certain limitations due, for example, to catchment area requirements. School (cohort) sizes vary substantially within an LA and school cohort sizes vary over time owing to changing demand and size of the cohort population. A different model would apply if the unit sizes are assumed fixed, as well as the population size n and minority fraction . In this case, the allocation mechanism is determined by the probability that an individual has status c conditional on being in unit j instead of the other way round. However, no matter whether unit sizes are random or fixed, D is still an estimator of D pop if, instead of the full population, we obtain a random sample drawn from it.

Bias

As D is an estimator of D pop, we define the bias of D as , where the expectation is taken over the independent multinomial distributions with probabilities , , . For given population size n and minority proportion p, we have The value of is a function of the underlying conditional probabilities, summarized by D pop, and of unevenness generated by the randomness of the allocation process. As has been well documented in the literature, see for instance Carrington and Troske (1997), D can be severely upward biased when unit sizes are small and there is no systematic segregation, that is, for all j and . Intuitively, this bias arises because D is the sum of the absolute value of differences between the minority and majority proportions in a unit. Suppose that in unit j, . Sampling variation in the estimated proportions will almost surely lead to non‐zero estimated differences, especially if unit sizes are small. Because the dissimilarity index sums the absolute values of these differences, this will lead to an upward bias in the index. For a small number of units J and small unit sizes, we can calculate the expected value of D analytically. Figure 1 graphs the bias for , , , and for various values of D pop. These values of D pop are obtained by setting according to a scheme discussed in Section 3.. The expected unit sizes are the same for the four units (i.e. 5 when , 10 when and 15 when ). The small‐unit bias is apparent in the figure. When expected unit sizes are equal to 5, is equal to 0.56 when . The graph also shows that the bias is a decreasing function of increasing systematic segregation (D pop) and a decreasing function of expected unit size.
Figure 1

Bias , , , equal expected unit sizes.

Bias , , , equal expected unit sizes.

BIAS CORRECTION

The purpose of a bias correction to D is to reduce the upward bias of the estimate of D pop, as highlighted in Figure 1. We first consider a bootstrap bias correction, as described by Hall (1992) and Davison and Hinkley (1997) among many others. Given an observed allocation, a new sample is generated with the same sample size n and minority proportion p, but using the observed conditional probabilities and for the allocation process. Note that none of the bootstrap data‐generating processes used in this paper involves resampling. The value for D in this bootstrap sample is denoted . Repeating this B times, we can calculate The population value of the segregation measure in the bootstrap sample is D itself, and so a measure of the bias of D is given by . A bootstrap bias‐corrected estimate of D pop is then obtained as This type of bias correction works well if the bias is constant for different values of D pop. This is clearly not the case here, because the biases as displayed in Figure 1 are much larger for smaller values of D. This bias correction is therefore not expected to work well for small unit sizes combined with small values of D pop. What turns out to be a more effective way of reducing the bias is a modified maximum‐likelihood approach. As with J fixed, almost surely, and the limiting distribution of is normal with expectation 0 and variance . It follows that D is consistent for D pop, and that it is asymptotically normal with an asymptotic variance that can be computed using the delta method, as pursued in Section 5.. For bias reduction, rather than working with the full log‐likelihood (2.2), we proceed as though , , were actually distributed according to their asymptotic normal distribution. However, the asymptotic normality of D is not of the usual kind where the limiting distribution of would be normal with expectation zero. We have seen that . The fact that D depends on the absolute values of the differences means that the expectation of the limiting distribution is not zero whenever the true value of D pop is either zero or is such that as . This implies that asymptotic inference of the usual sort is not valid, because the non‐zero expectation acts like a non‐centrality parameter. However, we can still use the asymptotic normal distribution of as an approximation in computing the asymptotic bias. The finite‐sample bias is given by where, in each term of the sum, the random variables and are independent. Let . The distribution of is the so‐called folded normal distribution; see Leone et al. (1961). Let θ denote . Then, the density of Z is given by Here, ϕ is the standard normal density. In order to derive the bias of D, we replace Z by , where approximately, with The asymptotic approximation to the density of  is then , where , the ‘true’ value of θ. Thus, asymptotically, the data‐dependent quantity  is sufficient for , and so an approximate maximum‐likelihood estimate of  is the value of  that maximizes the approximate density. It can be shown that, for , the maximum occurs at , and that the maximizing tends to as . Let the maximizing be denoted by . The cut‐off imposed by the function n is at . Because , we can define another estimator of D pop, where the notation ‘dc’ stands for density‐corrected. We show in the next sections that the proposed bias‐correction procedures reduce enough of the bias to make reasonable inferences about levels of segregation, provided unit sizes are not too small. Where unit sizes are very small, we show in Section 4. that the observed level of segregation can rarely statistically be distinguished from evenness. Thus, we suggest that in these cases the data are inappropriate for making inferences about segregation. In the Appendix, we consider two other plausible methods of bias reduction, but simulations (not presented here) show that they are less effective than the density‐correction method.

Monte Carlo simulations

In this section, we evaluate the performance of the bias adjustments for estimating levels of segregation. To do this, we follow the Duncan and Duncan (1955) approach of generating a level of unevenness between no segregation and complete segregation using a single parameter, . This parameter maps to a set of parabolic segregation curves via the formula:2 This formula, combined with the constraint of equal expected unit sizes, fixes the conditional allocation probabilities for both groups. An allocation is then generated by assigning n 1 and n 0 individuals to the J units using these calculated conditional probabilities. This process is repeated 5,000 times for each n, p and D pop combinations over the following parameter space: number of units, J, is fixed at 50; unit sizes are equal in expectation, with expected unit size 10, 30 or 50; proportion of individuals, p, equal to 0.05, 0.2 or 0.35; systematic segregation generator, q, varies such that the values of D pop are equal to 0, 0.056, 0.127, 0.225, 0.382 or 0.634. For the bootstrap bias correction, is calculated from (3.1) using 250 bootstrap samples. Tables 1(a), 1(b), 1(c) present the bias and root mean squared error (rmse) of D and D bc from (3.2), and D dc from (3.4). The tables show that, where the minority proportion is very small, , unit sizes are small (e.g. ) and systematic segregation is very low (e.g. ), observed segregation incorrectly suggests that a highly segregating process underlies the allocation, with . The bias corrections, although reducing the bias, cannot fully get rid of it, the smallest bias being obtained with the density‐corrected estimator, . At the other extreme, where the minority proportion is large (e.g. ), unit sizes are large (e.g. ) and systematic segregation is high (e.g. ), no correction is needed, because the mean value of observed segregation is only slightly different from D pop. However, in much social science data, the phenomenon of interest tends to have moderate (D pop around 0.1–0.4) rather than very high levels of segregation. In this range, the bias corrections tend to work well and are necessary, provided that p and are not both simultaneously very small. For example, when the minority proportion is 10% and unit sizes are expected to be 30, if underlying segregation is 0.225, the observed index of segregation would be upward biased by 0.093, whereas the density‐corrected estimator would successfully reduce this bias to just 0.005.
Table 1(a)

Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop

D pop
00.0560.1270.2250.3820.634
E[nj]=10 biasrmsebiasrmsebiasrmsebiasrmsebiasrmsebiasrmse
p=0.05
D 0.600.610.550.550.480.490.400.400.290.290.150.15
D bc 0.480.490.430.430.370.370.290.300.200.200.0970.11
D dc 0.400.410.350.350.290.290.210.220.130.140.0580.086
p=0.10
D 0.410.420.360.360.300.300.230.240.150.160.0710.084
D bc 0.260.270.210.220.150.170.0970.120.0430.0770.0090.058
D dc 0.260.270.210.220.150.160.0940.110.0400.0720.0110.056
p=0.20
D 0.310.310.260.260.200.210.150.150.0890.0970.0390.053
D bc 0.190.200.140.150.0900.110.0460.0670.0110.051−0.0020.044
D dc 0.170.180.120.130.0700.0820.0240.052−0.0090.050−0.0150.047
p=0.35
D 0.260.260.210.210.160.160.110.110.0630.0720.0260.042
D bc 0.160.160.110.120.0630.0740.0270.0500.0040.043−0.0020.038
D dc 0.150.150.0950.100.0480.0600.0090.041−0.0130.045−0.0120.040

Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250.

Table 1(b)

Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop

D pop
00.0560.1270.2250.3820.634
E[nj]=30 biasrmsebiasrmsebiasrmsebiasrmsebiasrmsebiasrmse
p=0.05
D 0.330.340.280.280.220.230.160.170.0990.110.0440.057
D bc 0.210.210.160.160.100.110.0550.0740.0150.054−0.0030.046
D dc 0.180.180.130.130.0720.0840.0240.054−0.0090.053−0.0150.049
p=0.10
D 0.240.240.190.190.140.140.0930.0980.0520.0610.0220.036
D bc 0.140.150.0950.100.0510.0630.0190.0430.0000.040−0.0030.034
D dc 0.130.140.0840.0890.0380.0510.0050.038−0.0100.041−0.0080.035
p=0.20
D 0.180.180.130.130.0880.0900.0540.0590.0290.0390.0120.026
D bc 0.110.110.0600.0650.0240.0380.0050.031−0.0010.031−0.0010.025
D dc 0.0990.100.0510.0560.0140.030−0.0060.031−0.0080.032−0.0040.026
p=0.35
D 0.150.150.100.110.0650.0680.0380.0440.0200.0300.0080.021
D bc 0.0900.0920.0450.0500.0140.0290.0020.027−0.0010.026−0.0000.021
D dc 0.0830.0860.0380.0430.0050.024−0.0070.027−0.0060.027−0.0030.021

Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250.

Table 1(c)

Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop

D pop
00.0560.1270.2250.3820.634
E[nj]=50 biasrmsebiasrmsebiasrmsebiasrmsebiasrmsebiasrmse
p=0.05
D 0.260.260.210.210.150.160.110.110.0600.0690.0260.040
D bc 0.150.160.110.110.0610.0720.0240.0480.0030.042−0.0030.035
D dc 0.150.150.0980.100.0520.0630.0160.042−0.0030.041−0.0050.035
p=0.10
D 0.190.190.140.140.0930.0960.0580.0630.0310.0410.0130.027
D bc 0.110.110.0640.0700.0270.0400.0070.032−0.0010.031−0.0010.026
D dc 0.100.110.0560.0610.0170.032−0.0030.031−0.0070.032−0.0040.027
p=0.20
D 0.140.140.0930.0940.0570.0600.0330.0380.0170.0270.0080.020
D bc 0.0820.0850.0390.0440.0110.0260.0010.024−0.0010.0230.0000.020
D dc 0.0760.0780.0320.0370.0030.023−0.0060.025−0.0050.024−0.0020.020
p=0.35
D 0.120.120.0720.0730.0410.0440.0230.0290.0120.0220.0050.016
D bc 0.0690.0710.0270.0320.0050.021−0.0000.020−0.0000.020−0.0000.017
D dc 0.0640.0660.0210.027−0.0020.020−0.0060.022−0.0030.020−0.0010.017

Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250.

The bias‐corrected estimator D bc is dominated in both bias and rmse by the density‐corrected estimator D dc in almost all experiments, except for the cases of high D pop values and larger minority proportions, in which the bias and rmse of both corrected estimates are small.

TESTS OF NO SYSTEMATIC SEGREGATION

To complement the bias‐corrected estimators of D, we provide a test for no systematic segregation. We consider two alternative methods to test whether we can reject the hypothesis that the level of segregation observed was generated by randomness alone, . It is common in the literature to run a randomization procedure to generate the distribution of D under the null of no systematic segregation – see, e.g., Boisso et al. (1994) – and D is compared to this distribution. Here, we generate the distribution of D under the null of no systematic segregation by creating B samples generated using the restricted conditional probabilities and calculating D in each sample, which we denote . The null hypothesis is then rejected at level α if , where I(.) is the indicator function. Alternatively, following the statistical model developed in Section 2., we can employ a likelihood ratio test for the hypothesis with test statistic which follows an asymptotic distribution. This asymptotic distribution is for large n and fixed J, and therefore for large unit sizes. For large J and/or small unit sizes, the asymptotic approximation can be expected to be poor, as we originally found in our simulation results discussed below. Therefore, we also utilize a bootstrap procedure to improve the size properties of the test. Let be the value of the likelihood ratio test in a sample generated from . Then the null hypothesis of no systematic segregation is rejected at level α if . Table 2 presents the test results for and and , for various values of D pop and minority proportions p. The number of Monte Carlo replications was 10,000 with 599 bootstrap samples. The size and power properties of the two tests are virtually identical. They have good size properties for all minority proportions p, with overall dominating the randomization test. The tests fail to reject the null for small values of D pop combined with small minority proportions p, exactly the circumstances in which the bias corrections do not remove much of the bias of D. Clearly, any calculation of D and the bias‐corrected estimators should be accompanied by and/or bootstrapped tests. If these tests fail to reject, no further inference should be pursued.
Table 2

Rejection frequencies of D randomization and likelihood ratio tests, for , level

D pop
p Test00.0560.1270.2250.3820.634
E[nj]=10
0.05 D* 0.0960.1040.1310.2370.6190.998
LR 0.0660.0740.0980.1940.5940.999
0.10 D* 0.0560.0690.1120.3070.8781.000
LR 0.0690.0830.1320.3620.9191.000
0.20 D* 0.0670.0860.1920.6180.9991.000
LR 0.0620.0800.1830.6060.9981.000
0.35 D* 0.0650.0900.2690.8271.0001.000
LR 0.0530.0770.2320.7911.0001.000
E[nj]=30
0.05 D* 0.0600.0710.1650.5340.9921.000
LR 0.0510.0670.1600.5460.9951.000
0.10 D* 0.0560.0860.2850.8821.0001.000
LR 0.0540.0800.2750.8771.0001.000
0.20 D* 0.0570.1170.5530.9971.0001.000
LR 0.0500.1080.5370.9971.0001.000
0.35 D* 0.0550.1470.7751.0001.0001.000
LR 0.0500.1380.7771.0001.0001.000

Rejection frequencies reported for 10,000 replications. Number of bootstrap repetitions 599.

Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250. Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250. Bias and rmse of D and bias‐corrected estimators for , and combinations of p and D pop Bias and rmse reported for 5,000 replications. Number of bootstrap repetitions 250. Rejection frequencies of D randomization and likelihood ratio tests, for , level Rejection frequencies reported for 10,000 replications. Number of bootstrap repetitions 599.

INFERENCE ON D

Having established that the bias corrections work well for a large part of the parameter space, we next develop reliable inference procedures such as 95% confidence intervals and Wald test statistics for equivalence of segregation in different areas. Inference based on a bias‐corrected estimator is, of course, expected to work well only in that part of the parameter space where the bias corrections work well (i.e. where the tests of no systematic segregation reject the null strongly, as indicated in Table 2). We start by deriving the asymptotic distribution of D given our statistical framework, following the procedures as developed in Ransom (2000). Under the data‐generating process as described in Section 2., for , with ; ; , the estimated conditional probabilities , are asymptotically normally distributed, as As and , the limiting distribution of D can then be obtained via the delta method: where is a J‐vector with rth element , where p sign if and if .3 This follows from and Clearly, this derivation is valid only when is not in a root‐n neighbourhood of zero, as discussed in Section 3.. The asymptotic distribution of D is then approximated by or, equivalently, which can form the basis for constructing confidence intervals and Wald test statistics for hypotheses of the form . If we denote by and the estimated counterparts of and , and substitute the observed fractions for , the Wald test is then computed as and converges in distribution to the distribution under the null. Clearly, we do not expect this approximation to work well when δ, group sizes and/or minority proportions are small, if only on account of the upward bias of D as established in the previous sections. However, the Wald test statistic W is asymptotically pivotal in the sense that its limiting distribution is not a function of nuisance parameters. We can therefore use bootstrap P‐values, which may result in an improvement in the finite‐sample behaviour of the test; see, e.g., Hall (1992), Davison and Hinkley (1997), Davidson and MacKinnon (2000), and Davidson (2009). If we denote the Wald statistic in the bth bootstrap sample as , calculated as then the bootstrap P‐value is given by . This bootstrap procedure is equivalent to a symmetric two‐tailed test for the t‐statistic. Let τ denote the t‐statistic that is the signed square root of the Wald statistic (5.3). Let denote the signed square root of (5.4). Then, a test that does not assume symmetry can be based on the equal‐tail bootstrap P‐value Alternatively, we can base the inference directly on any of the bias‐corrected estimators of D pop. In order to estimate the variance of the bias‐corrected estimators, we again perform a bootstrap procedure. For example, denoting the bootstrap estimate of the variance of D dc by , the Wald test based on D dc is then calculated as and this is again compared to the distribution. Figure 2 shows P‐value plots for testing the true hypothesis , for , and . The Wald test that is based on the asymptotic normal distribution of D and uses critical values is denoted W, whereas the Wald test that uses the bootstrap critical values is denoted W pb. The test based on the equal‐tail bootstrap P‐value for the t‐test is denoted T pb. The Wald statistic that uses the density‐corrected estimator and its bootstrap variance estimate is denoted W dc. The results shown are for 10,000 Monte Carlo replications. Note that 599 bootstrap samples per replication are drawn for the calculation of the variances and the bootstrap distribution of the Wald test.
Figure 2

P‐value plot, , , , .

P‐value plot, , , , . The first column of numbers in Table 3 presents the bias and rmse for the various estimates. There is an 11% upward bias in D, but D bc is unbiased. D dc has a small downward bias of 2.8%. As is clear from Figure 2, the asymptotic Wald test based on W using the critical values does not have good size properties. It rejects the true null too often (e.g. at 5% nominal size, it rejects the null in 18.6% of the replications). In contrast, using the P‐values from the bootstrap distribution of the Wald statistic improves the size behaviour considerably. At the 5% level, the rejection frequency is now reduced to 6.9%. Using the equal‐tailed bootstrap P‐values for the t‐test also improves on the size performance of the asymptotic Wald statistic, but it performs less well than W pb. W dc has the same size properties as W pb.
Table 3

Bias and rmse of D and bias‐corrected estimators

D pop =0.292 D pop =0.292 D pop =0.382
E[nj]=30,p=0.30 E[nj]=20,p=0.10 E[nj]=30,p=0.30
biasrmsebiasrmsebiasrmse
D 0.0310.0380.1060.1110.0220.032
D bc −0.0000.0270.0200.051−0.0010.026
D dc −0.0080.029−0.0000.046−0.0070.028

in all designs. Results from 10,000 Monte Carlo replications.

Bias and rmse of D and bias‐corrected estimators in all designs. Results from 10,000 Monte Carlo replications. Figure 3 shows the P‐value plot for a similar design, but now for smaller expected group sizes and a smaller minority proportion, . The bias of D in this case is 0.106, or 36%, that of D bc is around 0.020, or 6.5%, while D dc is unbiased.
Figure 3

P‐value plot, , , , .

P‐value plot, , , , . The size distortions of the test statistics are now more severe. The asymptotic Wald test is severely size distorted, with a 68% rejection rate at the 5% level. The Wald and asymmetric t‐tests using the bootstrap P‐values behave much better, with T pb behaving somewhat better. At the 5% level, the rejection frequencies for these tests are 10% and 9.0%, respectively. Here, W dc has the best size performance of all tests; it rejects the true null 7.0% of the time at the 5% level and is the only test where the size properties remain the same as those seen in Figure 2. There is a one‐to‐one correspondence between the P‐value plots as depicted in Figures 2 and 3 and the coverage properties of the confidence intervals associated with the particular test statistics. Using the normal approximation, % confidence intervals associated with the asymptotic Wald and W bc tests are constructed as and respectively, where is the quantile of the standard normal distribution. For the bootstrap Wald test, the associated confidence interval is given by where is the quantile of the distribution of the bootstrap repetitions . The equal‐tailed bootstrap t‐test has the corresponding confidence interval given by where and are, respectively, the and quantiles of the distribution of the bootstrap repetitions . For the example with and as described above, the observed rejection frequencies of 68%, 9.8%, 9.0% and 7.0% for the W, W pb, T pb and W dc tests, respectively, translate into coverage probabilities of 32%, 90.2%, 91% and 93% of the associated 95% confidence intervals. Given the upward bias of D, this leads to an interesting observation concerning the confidence interval based on the bootstrap Wald test W pb. As the size and associated coverage properties of this test are reasonably good, but as the confidence interval is symmetric around the upward biased D, this suggests that the Wpb‐based confidence interval will be quite large. Table 4 shows the averages of the lower and upper limits and lengths of the 95% confidence intervals associated with W pb, T pb, and W dc respectively. This confirms that the W pb‐based confidence interval is, on average, much wider than those based on W dc and T pb. Whereas the lower limit is quite similar to those of the other two confidence intervals, its upper limit is much higher, as expected owing to the symmetry around the upward biased D. Clearly, W pb can therefore have poor power properties when D has substantial bias.
Table 4

Average lower limit, upper limit and length of 95% confidence intervals

TestLower limitUpper limitLength
W pb 0.2280.5680.340
T pb 0.2120.3780.166
W dc 0.2090.3740.167

, , , .

Average lower limit, upper limit and length of 95% confidence intervals , , , . In principle, a likelihood ratio test of the hypothesis is possible. In practice, obtaining the maximized log‐likelihood under that constraint is a largely intractable problem, because the constraint is not differentiable with respect to and  when they are equal. The test results presented here show that inference can be based on the W pb, T pb and W dc tests when the sample size, the value of D pop and the minority proportion are such that the bias corrections work reasonably well, although, as Figures 2 and 3 show, some size distortions occur also for these tests. We next consider the case where , and . From Tables 1(a)– (c), it is clear that the bias‐corrected estimators remain biased in this case, with D dc having a bias of 0.15. Table 2 also shows that the tests for no systematic segregation only reject the null around 12% of the time. The rejection frequencies at the 5% level for the W, W pb, T pb and W dc tests are 100%, 42%, 26% and 80%, respectively, indicating that W dc in this case is severely oversized, as expected, whereas T pb performs best, but is still oversized substantially. When we increase to 30, the bias of D dc is reduced to 0.038, and the tests reject the null of no systematic segregation around 28% of the time. The rejection frequencies at the 5% level for W, W pb, T pb and W dc are now 99%, 35%, 25% and 16%, respectively, showing a substantial improvement of the performance of W dc, whereas the performance of T pb is similar to before. At the other end of the scale, when we set , and , all tests work well with rejection frequencies at the 5% level of 6.6%, 5.2%, 6.5% and 5.2% for W, W pb, T pb and W dc, respectively.

TESTS FOR EQUALITY OF SEGREGATION

A researcher may well be interested in determining whether segregation has changed significantly within an area over time, or whether segregation in one area is significantly different from that in another similar, or perhaps neighbouring, area. We consider the performances of the test statistics for comparing the two hypothetical areas for which the results were simulated above. Area 1 has , and , whereas Area 2 has , and . To study the size properties of the tests for the null hypothesis we set the two area population segregation measures as before. Given the area‐specific conditional allocation probabilities, the allocations in the areas are determined independently and therefore the Wald test is asymptotically distributed. The Wald test based on, e.g., the density‐corrected estimates is defined as whereas the bootstrap P‐values for the W pb test are based on the distribution of the bootstrap repetitions of where and are calculated from independent bootstrap repetitions. The bootstrap P‐values for the T pb test are obtained in an equivalent way. As an example, the bias of as an estimate for can be obtained from the results as presented in Table 3 and is equal to −0.075. As the covariance between D 1 and D 2 is equal to zero, the rmse can be calculated as . The equivalent numbers for are −0.007 and 0.054 for the bias and rmse, respectively. Note that this rmse calculation is not exact, because the Monte Carlo sample covariance between the two area segregation measures is not exactly equal to zero, but this difference is negligible. Figure 4 depicts the P‐value plots for the true null of equal population segregation measures D pop in the two areas. The asymptotic Wald test again over‐rejects substantially, 28.2% at the 5% level. The W dc test displays the best size properties in this case, rejecting 8.6% of the time at the 5% level, followed by T pb and then W pb. The rejection probabilities for these latter tests at the 5% level are 11.3% and 13.4%, respectively.
Figure 4

P‐value plot, , size properties.

P‐value plot, , size properties. We next turn to the power properties of these tests when the two population segregation measures D pop, 1 and D pop, 2 are not equal. We keep D pop, 2 equal to 0.292, but increase D pop, 1 to 0.382. The estimation results for this design are presented in the third column of Table 3. As discussed above, because D 2 is substantially biased upwards, we expect the W pb test to have low power. This is confirmed by the P‐value plots in Figure 5. The standard Wald test has power below nominal size, but the bootstrap‐based Wald test W pb completely fails to reject the null of equal segregation. In contrast, the Wald tests based on the bias‐corrected estimator and T pb show reasonable power properties, with T pb having the most power to detect this deviation from the null, although it has not been size‐adjusted. The P‐value plots, not shown here, for the true null that are very similar to those in Figure 4. Clearly, these results combined show that for simple hypothesis testing, W dc and T pb are the test procedures with reasonably good size and power properties in the settings we have considered.
Figure 5

P‐value plot, , power properties.

P‐value plot, , power properties.

DISCUSSION

Recently, Rathelot (2012) and d'Haultfœuille and Rathelot (2011) have also considered the problem of measuring segregation when units are small. In their set‐up, the number of individuals in units is small at around 5 or 10, whereas the number of units is large. Indeed, the methods in these papers rely on large‐number‐of‐units asymptotics. Both papers show that the parametric method proposed by Rathelot (2012) performs well in the estimation of the dissimilarity index and other measures of inequality, such as the Gini and the Theil indices. In this set‐up, the number of individuals in unit j, , is drawn from a given, unknown distribution. Then, the number of individuals in unit j having status c, , is distributed as Binomial, and is an unbiased estimate of . The parametric method of Rathelot (2012) is then to assume that is distributed as a mixture of Beta distributions, leading to the Beta‐Binomial model. In the simulations in Rathelot (2012), it is shown that estimates of the dissimilarity index from this Beta‐Binomial model have better properties in terms of smaller bias than the bootstrap bias correction in the setting of a large number of small units. We have tried to analyse the behaviour of this estimator in our set‐up, but found in the simulations of our design of Section 3.1. that the Beta‐Binomial estimation procedure often did not converge, making it difficult to compare the performance of this estimator with the others in our set‐up.4 For larger minority fractions and values of D pop, convergence of the ML estimator is easier to obtain in our design. Table 5 shows the estimation results for D beta1, which is the Beta‐Binomial estimator with a one‐component Beta distribution, for , , , where we obtained valid Monte Carlo results for and larger. The results show that in this design, D beta1 behaves similarly to D dc for , but has a substantially larger bias and/or rmse for larger values of D pop. For this design, increasing the number of mixtures of beta distributions to two does not change the results, because in almost all cases this model converges to the one‐component model. The Beta‐Binomial estimator is not consistent for a fixed number of units and unit sizes going to infinity. For example, when we increase the expected sample size to , the bias and rmse of D beta1 increases to −0.042 and 0.044, respectively, when . The bias and rmse of D itself in that case are 0.009 and 0.018, respectively.
Table 5

Bias and rmse of D beta1 and D ct

D pop
00.0560.1270.2250.3820.634
biasrmsebiasrmsebiasrmsebiasrmsebiasrmsebiasrmse
D 0.1800.1800.1300.1300.0880.0900.0540.0590.0290.0390.0120.026
D dc 0.0990.1000.0510.0560.0140.030−0.0060.031−0.0080.032−0.0040.026
D beta1 −0.0120.037−0.0200.037−0.0330.042−0.0540.060
D ct −0.0320.075−0.0670.089−0.0850.091−0.1000.110−0.0990.100−0.0640.085

, , . No results reported for D beta1 in first two columns owing to the convergence problems of the estimator.

Bias and rmse of D beta1 and D ct , , . No results reported for D beta1 in first two columns owing to the convergence problems of the estimator. Another bias‐corrected measure is the one proposed by Carrington and Troske (1997), which has been widely used in school segregation (Söderström and Uusitalo, 2010) and occupational segregation (Hellerstein and Neumark, 2008) research. Carrington and Troske (1997) argue that segregation indices can be modified to take into account the underlying value under no systematic segregation, when , . They propose a modified segregation index that measures the (economic) extent to which a sample deviates from the expected value of D under no systematic segregation, denoted , which can be calculated as in Section 4.. They argue that their new index of systematic dissimilarity does not depend on the margins in the area and is therefore a better means of comparing the extent to which systematic segregation exists. Their measure, denoted D ct here, is defined as and hence . D ct can be interpreted as the extent to which the area is more dissimilar than random allocation would imply, expressed as a fraction of the maximum amount of such excess dissimilarity that could possibly occur. implies that the allocation of individuals in the area is equivalent to no systematic segregation. It is worth noting that D ct is almost identical to the index proposed by Winship (1977), which was criticized by Falk et al. (1978) and then partially withdrawn by Winship (1978). The problem with D ct is that it is not entirely clear what it is intended to achieve. D ct is always lower than D pop, but tends to D pop for large unit sizes. In our simulations, for many values of the parameters, D ct underestimates D pop by a larger amount than D itself overestimates D pop. As an illustration of this, Table 5 presents simulation results for , and . The problem with D ct is that decomposition of a segregation index into one part produced by systematic segregation and another part produced by randomness cannot be done additively. In the previous Monte Carlo designs, all unit sizes were equal in expectation. We next perform the simulations with designs like those of the first two columns in Table 3, but for unequal expected unit sizes. In the first design, labelled Design 1 in Table 6, we split the 50 units into half with a smaller expected group size and half with a larger expected group size. For the example containing 1500 individuals, these sizes are approximately 10 and 50; for the second example of 1000 individuals, these are approximately 7 and 33. In the second design, labelled Design 2 in the table, all expected unit sizes are different, within the aforementioned ranges in steps of just under one individual. Tables 6 and 7 present the estimation and testing results. The results are all very similar to those with equal expected unit sizes, as presented in Sections 5. and 6..
Table 6

Results for designs with unequal expected unit sizes

n=1500, p=0.3 n=1000, p=0.1
biasrmsebiasrmse
Design 1
D 0.0220.0320.0830.091
D bc −0.0030.0270.0180.050
D dc −0.0100.0290.0090.047
Design 2
D 0.0270.0350.0940.100
D bc −0.0020.0270.0160.049
D dc −0.0090.0280.0050.045
Rejection frequencies for tests of H0:D pop =0.292
Nominal size0.100.050.100.05
Design 1
W 0.1920.1160.5960.457
W pb 0.1060.0560.2000.119
T pb 0.1620.1000.1750.100
W dc 0.1350.0780.1270.071
Design 2
W 0.2460.1540.7260.596
W pb 0.1150.0640.1670.096
T pb 0.1610.0950.1670.095
W dc 0.1340.0750.1340.076

, ; 10,000 Monte Carlo replications, 599 bootstrap repetitions.

Table 7

Test results for

Design 1Design 2
Size W W pb T pb W dc W W pb T pb W dc
0.100.2930.2210.1930.1390.3520.2080.1860.138
0.050.1840.1400.1190.0790.2340.1250.1180.081

See Table 6.

Results for designs with unequal expected unit sizes , ; 10,000 Monte Carlo replications, 599 bootstrap repetitions. Test results for See Table 6.

SOCIAL SEGREGATION IN SCHOOLS

In this section, we illustrate our inference procedures with an empirical application relating to social segregation in primary schools in England. The dichotomous measure is an indicator of poverty based on eligibility for free school meals (FSMs). This context is useful as it naturally produces small unit sizes, and shows a range of minority proportions and overall populations across different LAs. We use administrative data collected by the Department for Children, Families and Schools, and made available to researchers as part of the National Pupil Database on pupils aged 10/11 in English primary schools in 2006. Measurement of school segregation using this data set has been carried out by many researchers; see Allen and Vignoles (2008), Burgess et al. (2006), and Gibbons and Telhaj (2006). Using the tools developed above, we can assess whether the small unit sizes and/or small minority populations lead to incorrect inferences about differences in segregation across areas. We provide two cases. First, we compare two similar pairs of LAs, showing that quite small differences in their characteristics imply different outcomes of inference; these are North‐East Lincolnshire and North Lincolnshire, and Blackburn and Oldham. Second, we compare all the different LAs in inner‐city London, and consider which pairwise comparisons yield significant differences. Table 8 shows the descriptive statistics and the dissimilarity indices of the LAs. North‐East Lincolnshire and North Lincolnshire have almost the same number of pupils, 2005 and 2011 respectively, but differ in the number of schools, 46 and 57 respectively, and consequently also average cohort size. They also differ in the percentages of children eligible for FSMs, 21% and 13% respectively. The dissimilarity index for North‐East Lincolnshire is 0.43, higher than that of North Lincolnshire, which has an index of 0.36. Blackburn and Oldham differ rather more in size, but have closer average unit sizes, and slightly higher percentages of children eligible for FSMs.
Table 8

Key parameters of primary schools across English LAs

NumberNumberAverage%
LA nameof pupilsof schoolscohort sizeFSM D
North‐East Lincolnshire20054644210.43
North Lincolnshire20115735130.36
Blackburn21055141260.34
Oldham29908635210.47
Camden13944134420.23
Greenwich26666640360.29
Hackney21945441430.22
Hammersmith and Fulham11773930450.30
Islington18454838410.26
Kensington and Chelsea8812733360.32
Lambeth24286040400.24
Lewisham28337040290.30
Southwark29297241360.21
Tower Hamlets27036840610.20
Wandsworth21246035270.29
Westminster13363934390.33
Key parameters of primary schools across English LAs Are the school allocations in North‐East Lincolnshire more segregated than those in North Lincolnshire? Table 9 shows that the observed D marginally overstates the level of segregation in each LA, but the bias‐corrected estimates of D pop do not alter the ranking. Table 9 further presents the various test procedures and confidence intervals as described in the previous section. Here, we generate 999 bootstrap samples. The LR test for no systematic segregation clearly rejects for both LAs, with both bootstrap P‐values equal to 0. The rejection of the null of equal segregation in North‐East Lincolnshire and North Lincolnshire depends on the test statistics employed. Using the test statistics W dc and T pb, we reject the null of equal segregation in the two LAs at the 5% and 1% levels, respectively.
Table 9

Bias‐corrected dissimilarity indices, confidence intervals and test statistics for North‐East and North Lincolnshire

North‐East LincolnshireNorth Lincolnshire
D 0.4330.364
D bc 0.4200.322
D dc 0.4160.334
LR‐test, bootstrap P‐value00
CI‐W [0.386–0.481][0.306–0.421]
CI‐W pb [0.380–0.487][0.275–0.452]
CI‐T pb [0.371–0.466][0.265–0.371]
CI‐W dc [0.367–0.465][0.278–0.390]
H0:D pop , NEL =D pop , NL , P‐values
W 0.067
W pb 0.114
T pb 0.000
W dc 0.032

CI are 95% confidence intervals. Number of bootstrap repetitions 999.

Bias‐corrected dissimilarity indices, confidence intervals and test statistics for North‐East and North Lincolnshire CI are 95% confidence intervals. Number of bootstrap repetitions 999. Table 10 shows test statistics for Blackburn and Oldham. In this example, we can reject, with a high degree of confidence, the null of equal segregation in these areas. This greater confidence than in the Lincolnshire example is possible, despite similar segregation levels, because the LAs are slightly larger and the minority proportions are higher.
Table 10

Bias‐corrected dissimilarity indices, confidence intervals and test statistics for Blackburn and Oldham

BlackburnOldham
D 0.3420.472
Ddc 0.3060.446
LR test, bootstrap P‐value00
CI‐Tpb [0.288–0.362][0.420–0.485]
CI‐Wdc [0.263–0.348][0.410–0.483]
H0:D pop , Blackburn =D pop , Oldham , P‐values
T pb 0.000
W dc 0.000

CI are 95% confidence intervals. Number of bootstrap repetitions 999.

Bias‐corrected dissimilarity indices, confidence intervals and test statistics for Blackburn and Oldham CI are 95% confidence intervals. Number of bootstrap repetitions 999. For our second illustration, Table 11 compares observed and density‐corrected segregation levels across the 12 LAs in Inner London. The density correction makes little differences to the ranking of segregation levels, with just Wandsworth and Greenwich switching positions. Results for the tests of equivalence of D pop in Table 12 show that the LAs can be approximately divided into three groups, with possible multiple membership, where the tests do not reject the null of equal segregation. These groups are: Tower Hamlets, Southwark and Hackney, with the lowest level of segregation; Hackney, Camden and Lambeth, with medium level of segregation; Wandsworth, Greenwich, Hammersmith and Fulham, Lewisham, Kensington and Chelsea, and Westminster with the highest level of segregation. Islington is a medium‐segregation LA with some overlap with the group of highest‐segregation LAs.
Table 11

Bias‐corrected dissimilarity indices and confidence intervals for Inner London

D D dc LR(P)CI–W dc CI–T pb
Tower Hamlets0.1970.1620[0.126–0.192][0.125–0.198]
Southwark0.2060.1650[0.137–0.201][0.128–0.202]
Hackney0.2190.1840[0.154–0.225][0.142–0.226]
Camden0.2310.1880[0.153–0.236][0.140–0.237]
Lambeth0.2400.2090[0.172–0.241][0.170–0.248]
Islington0.2570.2310[0.183–0.258][0.188–0.273]
Wandsworth0.2900.2430[0.219–0.292][0.200–0.286]
Greenwich0.2860.2510[0.226–0.291][0.213–0.288]
Hammersmith and Fulham0.3030.2640[0.226–0.323][0.208–0.319]
Lewisham0.3040.2740[0.244–0.312][0.235–0.312]
Kensington and Chelsea0.3170.2960[0.231–0.347][0.230–0.361]
Westminster0.3280.3020[0.257–0.347][0.252–0.352]

CI are 95% confidence intervals. Number of bootstrap repetitions 999.

Table 12

P‐values for tests of equivalence of D pop for Inner London

SouHacCamLamIslWanGreHamLewKenWes
Tower0.7250.2020.1760.0400.0140.0030.0000.0010.0000.0000.000
Hamlets0.8990.4260.3890.0810.0160.0050.0010.0030.0000.0000.000
Southwark0.3680.3100.0940.0300.0000.0000.0000.0000.0000.000
0.5020.4540.1070.0220.0070.0010.0040.0000.0010.000
Hackney0.8750.4660.2340.0160.0000.0060.0000.0020.000
0.8970.3880.1230.0540.0190.0240.0020.0050.000
Camden0.6310.3380.0200.0160.0080.0020.0040.000
0.5110.1950.0980.0450.0440.0070.0090.001
Lambeth0.5550.0460.0340.0280.0020.0140.002
0.4570.2500.1280.1110.0200.0250.004
Islington0.2140.1260.0980.0380.0640.008
0.6900.4890.3510.1420.1010.033
Wandsworth0.8530.5870.3760.3140.124
0.7950.5610.3000.1870.081
Greenwich0.6550.4940.3500.150
0.6960.3980.2390.106
Hammersmith0.9110.6530.396
and Fulham0.7760.4670.318
Lewisham0.6630.388
0.5710.382
Kensington0.779
and Chelsea0.880

Top and bottom rows are for T pb and W dc, respectively. Number of bootstrap repetitions 999.

Bias‐corrected dissimilarity indices and confidence intervals for Inner London CI are 95% confidence intervals. Number of bootstrap repetitions 999. P‐values for tests of equivalence of D pop for Inner London Top and bottom rows are for T pb and W dc, respectively. Number of bootstrap repetitions 999.

CONCLUSIONS

To make statements about the true underlying degree of segregation, or to understand the processes causing segregation, it is desirable to measure the level of systematic segregation. However, where minority proportions and unit sizes are small, the level of segregation observed by researchers in their data is known to be significantly greater than systematic segregation. Furthermore, because the size of the bias of observed segregation over systematic segregation is known to be a function of minority proportion, unit sizes and systematic segregation, differences in any of these parameters between areas or over time may lead to incorrect inferences. In this paper, we have proposed and tested procedures for adjusting the dissimilarity index of segregation for this bias. Our corrections work well, provided both the minority proportion and unit size are not very small. Where very small minority proportions and unit sizes render our corrections useless, we show that levels of segregation are often not statistically distinguishable from zero. We have developed and tested our statistical framework using the index of dissimilarity, D, but it can, in principle, be extended to other segregation indices. From our statistical framework, we have developed tests for a null of no systematic segregation and a null of equal segregation in two areas, and we have established confidence intervals for levels of systematic segregation. In tests using unit sizes, minority proportions and underlying segregation levels similar to those encountered by social scientists, the Wald statistics using the bootstrap variance estimate for the bias‐corrected estimators and the test based on the equal‐tail bootstrap P‐value for the t‐test (T pb) are found to perform best. The methods proposed in this paper provide a framework for more reliable inference as to levels of segregation, which will aid the further investigation of the causes of segregation. Replication Files Click here for additional data file.
  1 in total

1.  Segregation and diversity measures in population distribution.

Authors:  M J White
Journal:  Popul Index       Date:  1986
  1 in total
  5 in total

1.  Translation and cultural adaptation of the Greek version of the confusion assessment method diagnostic algorithm and the nursing delirium screening scale and their inter-rater reliability: A prospective cohort study.

Authors:  M P Ntalouka; M Bareka; A G Brotis; A Chalkias; K Stamoulis; A Flossos; P Tzimas; E Arnaoutoglou
Journal:  Hippokratia       Date:  2020 Jan-Mar       Impact factor: 0.471

2.  Actual Racial/Ethnic Disparities in COVID-19 Mortality for the Non-Hispanic Black Compared to Non-Hispanic White Population in 353 US Counties and Their Association with Structural Racism.

Authors:  Michael Siegel; Isabella Critchfield-Jain; Matthew Boykin; Alicia Owens; Taiylor Nunn; Rebeckah Muratore
Journal:  J Racial Ethn Health Disparities       Date:  2021-08-30

3.  Measuring racial segregation in health system networks using the dissimilarity index.

Authors:  Andrea M Austin; Donald Q Carmichael; Julie P W Bynum; Jonathan S Skinner
Journal:  Soc Sci Med       Date:  2019-09-25       Impact factor: 4.634

4.  Ethnic Residential Segregation: A Multilevel, Multigroup, Multiscale Approach Exemplified by London in 2011.

Authors:  Kelvyn Jones; Ron Johnston; David Manley; Dewi Owen; Chris Charlton
Journal:  Demography       Date:  2015-12

5.  Inequality and income segregation in Brazilian cities: a nationwide analysis.

Authors:  José Firmino de Sousa Filho; Gervásio F Dos Santos; Roberto F Silva Andrade; Aureliano S Paiva; Anderson Freitas; Caio Porto Castro; Amélia A de Lima Friche; Sharrelle Barber; Waleska T Caiaffa; Maurício L Barreto
Journal:  SN Soc Sci       Date:  2022-09-10
  5 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.