Literature DB >> 32051873

Effect sizes of the differences between means without assuming variance equality and between a mean and a constant.

Satoshi Aoki1.   

Abstract

Effect sizes of the difference, or standardized mean differences, are widely used for meta-analysis or power-analysis. However, common effect sizes of the difference such as Cohen's d or Hedges' d assume variance equality that is fragile and is often violated in practical applications. Based on Welch's t tests, we defined a new effect size of the difference between means, which did not assume variance equality, thereby providing a more accurate value for data with unequal variance. In addition, we presented the unbiased estimator of an effect size of the difference between a mean and a known constant. An R package is also provided to compute these effect sizes with their variance and confidence interval.
© 2020 The Author(s).

Entities:  

Keywords:  Cohen's d; Effect size; Hedges' d; Heteroscedasticity; Mathematics; Psychology; Standardized mean difference

Year:  2020        PMID: 32051873      PMCID: PMC7002853          DOI: 10.1016/j.heliyon.2020.e03306

Source DB:  PubMed          Journal:  Heliyon        ISSN: 2405-8440


Introduction

Effect sizes of the difference or, more precisely, standardized mean differences between two groups, are widely used to estimate the magnitude of effect independent of the sample size [1], to conduct meta-analysis [2], or to conduct power-analysis [3]. The American Educational Research Association (AERA) or the American Psychological Association (APA) strongly recommend effect sizes are reported in the corresponding fields [4], [5]. Furthermore, the misuse and misunderstanding of p-value have become public [6], and use of effect sizes is spreading beyond pedagogy and psychology, where effect sizes have developed, into areas such as in biology [1]. In spite of such importance, the classical effect sizes of the difference assume variance equality (homoscedasticity), which is hard to assume practically or is even expected to be violated a priori in clinical data [7]. While Bonett [8] defined a confidence interval of an effect size estimator which did not assume homoscedasticity, its parameter was not defined. This problem of variance inequality (heteroscedasticity) has been long debated [9], [10]. In addition, the unbiased estimator of an effect size of the difference between a mean and a constant was undefined. To solve these problems, based on Welch's t test [11], [12], we defined an effect size of the difference between means that does not assume homoscedasticity and calculated the unbiased estimator of an effect size of the difference between a mean and a constant. Effect size of the difference was developed by Cohen [13], who studied in the field of psychology. Cohen [3], [13] defined the effect size as a parameter for two independently and normally distributed populations, and : which is expressed as d in the original articles [3], [13]. Note that both populations share the common variance . The estimator of this parameter was represented as in [3]. However, we refer to this estimating statistic as g to distinguish it from the other d we introduce later. The statistic g is defined as where and, for , Here, , , and are the mean of the sample, the sample (random variable), and the sample size of group 1, respectively, while , , and are those of group 2. For the denominator, this effect size uses the pooled standard deviation, which suggests the most precise population variance under the assumption of equal variance [14]. In the field of pedagogy, Glass [2] suggested another effect size of the difference, independently of Cohen's works. He defined it as “the mean difference on the outcome variable between treated and untreated subjects divided by the within group standard deviation,” where “the within group standard deviation” corresponds to the standard deviation of the untreated group. He clearly distinguished the treated (experimental) group from the untreated (control) group, and there was no assumption regarding the two groups. His effect size was subsequently formulated and named Glass' Δ by Hedges [14], which is where is the mean of the variable in the experimental group, is that in the control group, and is the unbiased standard deviation of the control group. Hedges [14] also defined the δ (1) and the g (2) independently of Cohen. Furthermore, Hedges [14] indicated that g (2) is biased from δ (1), making it unsuitable for analyses that do not treat the entire population. The unbiased estimator of δ (1) is defined as in [14] and d in [15]. In this study, we call it d, which is Using the gamma function, the correction coefficient J is defined as The effect sizes g (2) and d (5) are used in various fields of science, but they assume homoscedasticity just like Student's t-test [16], [17]. When this assumption of homoscedasticity is violated, Grissom [9] recommended the use of Glass's Δ (4) instead of d (5). However, Glass's Δ (4) and d (5) have different meaning because of the difference in denominator. Therefore, Glass's Δ (4) cannot substitute for d (5) in a strict sense. Behavior of g (2), Δ (4), and d (5) under heteroscedasticity was studied in [10], although the justification for using effect size parameter α, that they defined, to measure the statistic bias under heteroscedasticity was not shown. Bonett [8] in psychology proposed a confidence interval (CI) of effect size which does not assume homoscedasticity. First, he defined a general effect size estimator where , is a sample mean, and . Concerning effect size of the difference between two means, substituting , and gives Then, he assumed its corresponding parameter and its CI. The CI was calculated using approximation of CI [18] and variance of the estimator which was approximately calculated without assuming homoscedasticity. The parameters estimated by (7) or (8) were not formulated. Namely, he defined the CI for heteroscedasticity without defining a parameter, and this can be a problem. When the estimator does not always correspond to a single parameter, the CI of an undefined parameter loses its consistency in what to estimate, and heteroscedasticity or difference of sample sizes can change the correspondence between an estimator and a parameter (see section 5.2). Although his CI was effective relative to the other CIs in his simulation experiment where the parameter was given a value, what the value meant could change depending on the variance and sample size, and the change could not be expected since the parameter was not formulated. It should be noted, Cohen [3] also defined a parameter of an effect size of the difference between a mean and a constant for a normally distributed population and a known constant C as Cohen [3] originally referred to this as , but we refer to this as γ (9) to clearly distinguish it from d (5). Cohen [3] also defined a biased estimator of an effect size for a normally distributed population with the sample value (), the sample mean , and a known constant C as The is the square root of (3). Cohen [3] originally referred to this as , but we refer this to for the reason described above. To the best of my knowledge, the unbiased estimator of γ (9) has not been shown. There are other effect sizes of the difference that do not assume normality or independence. Since their assumption is different from that of effect size we focus on, we do not treat them in detail and briefly introduce them. Dunlap et al. [19] invented effect size of the difference between two correlated paired groups. Algina et al. [20] proposed robust effect size of the difference, which is based on g (2) using 20% trimmed mean and 20% Winsorized variance assuming that samples are taken from an observing population and another contaminating population.

Theory

An effect size of the difference between means without assuming homoscedasticity

First, we define the parameter of an effect size of the difference between means for two independently and normally distributed populations and as where r is a non-negative real number. This parameter is not generalization of δ (1) and is different from it. Then, suppose two independently and normally distributed populations with the samples () and (), and the sample mean and . Based on the statistic , the so-called Welch's t [11], [12], a biased estimator of (11) is defined as where is the same as (3), and Finally, e, the unbiased estimator of (11), is Therefore, Here, r corresponds to the ratio . J is the correction coefficient that is defined in equation (6). The degree of freedom f is approximately calculated using the Welch-Satterthwaite equation [11], [21] as The variance of e (15) is Although this effect size is derived from the difference, we refer to it as e not d. This is because Cohen's d (2) and Hedges' d (5) already exist, and more d would cause further confusion. The proof of the bias correction and variance derivation does not assume homoscedasticity (see the Appendix). In addition, e (15) is a consistent estimator of (11) at the same time. See the Appendix for the proof of the consistency.

An effect size of the difference between a mean and a known constant

Using (10), the unbiased estimator of the effect size parameter γ (9) is defined for a normally distributed population with the sample value (), the sample mean , and a known constant C as Therefore, The correction coefficient J (6) is the same as the one used above. The variance of c is See the Appendix for proofs of the bias correction and the derivation of the variance. In addition, c (17) is a consistent estimator of γ (9) (see the Appendix for the proof). When interested in constants rather than variables, defined as can be used instead of c.

Confidence intervals of effect sizes

In terms of the effect sizes of the difference, the CI based on a noncentral t variate is not directly given by a formula [22]. The CI is derived from that of noncentral parameters of noncentral t-distribution, which is in turn obtained by some searching method. The CI based on the biased effect sizes are given as: and where is the noncentral parameter that gives the upper limit of cumulative probability (e.g., 0.975 cumulative probability for 95% CI) for noncentral t-distribution with the corresponding t value (see the discussion section) and the degree of freedom, and is that which gives the lower limit (e.g., 0.025 cumulative probability for 95% CI), and and are the same as (14) and (10). The CIs based on the unbiased estimator of the effect sizes are given by multiplying the corresponding correction coefficient J (6) of the corresponding degree of freedom to the above intervals. The CI by Bonett [8] is calculated using variance of the estimator which is approximately calculated without assuming homoscedasticity and approximate assumption of CI [18]. Therefore, it is not necessary to apply Bonett's CI to e (15) or c (17), because the derivation of their CIs does not assume homoscedasticity, and their exact CIs can be calculated without approximation.

Calculation method

I developed a new package es.dif for R [23]. It enables the statistics d (5), e (15), c (17), their biased statistics, variance, and CI based on the two samples or their mean, variance, and sample size to be computed. In this package, approximation of J (6) [14] is not employed unless its degree of freedom exceeds 342, when the gamma function returns values that are too large to be treated in R. The CI is obtained by binary search. The remainder of this section presents some examples of the package. First, the following script calculates d (5), e (15), their variances and 95% CIs for data 1 () and data 2 (). Using options of the function, you can change the type I error rate for the CI, calculate biased effect sizes, and output results in the vector style. For example, (10) with 99% CI in the vector style is calculated by this script. In the vector-style output, the four values in the vector show the effect size, its variance, and lower and higher limits of the CI. In addition, this package includes functions that can output effect sizes from the (estimated) parameters and the sample sizes. The following scripts compute d (5) and e (15) for two populations, and with the sample size 5 and 10, respectively. These types of functions also have the options for the type I error rate, the biased effect size, and the vector-style output. > library(es.dif)

> data1<-c(0,1,2,3,4)

> data2<-c(0,0,1,2,2)

> es.d(data1,data2)

 [,1] [,2]

[1,] "Hedges' d:" "0.682379579593354"

[2,] "variance:" "0.484026380702367"

[3,] "CI:" "[ -0.503527216375147 ,

1.82938058482178 ]"

> es.e(data1,data2)

 [,1] [,2]

[1,] "Unbiased e:" "0.668264936033828"

[2,] "variance:" "0.506830833214916"

[3,] "CI:" "[ -0.50334965496395 ,

1.7965317007171 ]" > library(es.dif)

> data1<-c(0,0,1,2,2)

> data2<-c(2)

> es.c(data1,data2,alpha=0.01,

unbiased=FALSE,vector_out=TRUE)

[1] -1.0000000 0.9292037 -2.5390625 0.5778885 > library(es.dif)

> mean1<-1

> mean2<-0

> var1<-2

> var2<-1

> n1<-5

> n2<-10

> es.para.d(mean1,mean2,var1,var2,n1,n2)

 [,1] [,2]

[1,] "Hedges' d:" "0.82286529714397"

[2,] "variance:" "0.349443397657368"

[3,] "CI:" "[ -0.248827687382689 ,

1.86616833367494 ]"

> es.para.e(mean1,mean2,var1,var2,n1,n2)

 [,1] [,2]

[1,] "Unbiased e:" "0.674259756444758"

[2,] "variance:" "0.41613476136966"

[3,] "CI:" "[ -0.354146439977423 ,

1.65626025590509 ]"

Application & simulation

While the situation to use c (17) is clearly different, the e (15) and d (5) have a similar application range in practice. Therefore, we prepared an example of the applications in which the sample variances are not equal. Table 1 shows well-known data of three Iris species by Fisher [24], which can also be checked in R [23] using a command “iris”. Note that only the petal width of I. setosa has fewer significant digits. For this data, we calculated d (5), e (15), the ratio of d (5) to e (15), and the ratio of the standard deviations of the two comparing data. Theoretically, e (15) is a more precise estimator of its own parameter than d (5) under this heteroscedasticity.
Table 1

Measured characteristics (in centimeters) of three Iris species shown in Fisher (1936).

Iris setosa
Iris versicolor
Iris virginica
S.L.S.W.P.L.P.W.S.L.S.W.P.L.P.W.S.L.S.W.P.L.P.W.
5.13.51.40.27.03.24.71.46.33.36.02.5
4.93.01.40.26.43.24.51.55.82.75.11.9
4.73.21.30.26.93.14.91.57.13.05.92.1
4.63.11.50.25.52.34.01.36.32.95.61.8
5.03.61.40.26.52.84.61.56.53.05.82.2
5.43.91.70.45.72.84.51.37.63.06.62.1
4.63.41.40.36.33.34.71.64.92.54.51.7
5.03.41.50.24.92.43.31.07.32.96.31.8
4.42.91.40.26.62.94.61.36.72.55.81.8
4.93.11.50.15.22.73.91.47.23.66.12.5
5.43.71.50.25.02.03.51.06.53.25.12.0
4.83.41.60.25.93.04.21.56.42.75.31.9
4.83.01.40.16.02.24.01.06.83.05.52.1
4.33.01.10.16.12.94.71.45.72.55.02.0
5.84.01.20.25.62.93.61.35.82.85.12.4
5.74.41.50.46.73.14.41.46.43.25.32.3
5.43.91.30.45.63.04.51.56.53.05.51.8
5.13.51.40.35.82.74.11.07.73.86.72.2
5.73.81.70.36.22.24.51.57.72.66.92.3
5.13.81.50.35.62.53.91.16.02.25.01.5
5.43.41.70.25.93.24.81.86.93.25.72.3
5.13.71.50.46.12.84.01.35.62.84.92.0
4.63.61.00.26.32.54.91.57.72.86.72.0
5.13.31.70.56.12.84.71.26.32.74.91.8
4.83.41.90.26.42.94.31.36.73.35.72.1
5.03.01.60.26.63.04.41.47.23.26.01.8
5.03.41.60.46.82.84.81.46.22.84.81.8
5.23.51.50.26.73.05.01.76.13.04.91.8
5.23.41.40.26.02.94.51.56.42.85.62.1
4.73.21.60.25.72.63.51.07.23.05.81.6
4.83.11.60.25.52.43.81.17.42.86.11.9
5.43.41.50.45.52.43.71.07.93.86.42.0
5.24.11.50.15.82.73.91.26.42.85.62.2
5.54.21.40.26.02.75.11.66.32.85.11.5
4.93.11.50.15.43.04.51.56.12.65.61.4
5.03.21.20.26.03.44.51.67.73.06.12.3
5.53.51.30.26.73.14.71.56.33.45.62.4
4.93.11.50.16.32.34.41.36.43.15.51.8
4.43.01.30.25.63.04.11.36.03.04.81.8
5.13.41.50.25.52.54.01.36.93.15.42.1
5.03.51.30.35.52.64.41.26.73.15.62.4
4.52.31.30.36.13.04.61.46.93.15.12.3
4.43.21.30.25.82.64.01.25.82.75.11.9
5.03.51.60.65.02.33.31.06.83.25.92.3
5.13.81.90.45.62.74.21.36.73.35.72.5
4.83.01.40.35.73.04.21.26.73.05.22.3
5.13.81.60.25.72.94.21.36.32.55.01.9
4.63.21.40.26.22.94.31.36.53.05.22.0
5.33.71.50.25.12.53.01.16.23.45.42.3
5.03.31.40.25.72.84.11.35.93.05.11.8
5.03.41.50.25.92.84.31.36.63.05.62.0
0.350.380.170.10.520.310.470.200.640.320.550.27

Note: S.L. = sepal length; S.W. = sepal width; P.L. = petal length; P.W. = petal width. The last two rows show the average and the standard deviation of the corresponding column.

Measured characteristics (in centimeters) of three Iris species shown in Fisher (1936). Note: S.L. = sepal length; S.W. = sepal width; P.L. = petal length; P.W. = petal width. The last two rows show the average and the standard deviation of the corresponding column. The calculated result is shown in Table 2. When considering their significant digits, the comparing pair of the sepal length of I. setosa and I. virginica showed the different effect size of d (5) and e (15) (in bold in Table 2). Even though most pairs showed identical values of d (5) and e (15), the result revealed that violation of the assumption of homoscedasticity in d (5) can affect the result even in two significant digits.
Table 2

Calculated effect sizes of the difference for the data shown in Table 1.

Chara.Taxaded/esd ratio
S.L.1 vs 2-2.1-2.11.0010290.682893
1 vs 3-3.1-3.01.0021850.554334
2 vs 3-1.1-1.11.0003280.811744
S.W.1 vs 21.81.81.0002851.214233
1 vs 31.21.21.0002121.181483
2 vs 3-0.64-0.641.0000060.973028
P.L.1 vs 2-7.8-7.81.0045100.369243
1 vs 3-9.9-9.91.0052560.314392
2 vs 3-2.5-2.51.0001970.851450
P.W.1 vs 2-7-71.0023180.542139
1 vs 3-8-81.0042220.390349
2 vs 3-2.9-2.91.0007810.720017

Note: Chara. = characteristics; S.L. = sepal length; S.W. = sepal width; P.L. = petal length; P.W. = petal width; Taxa = compared taxa; 1 = I. setosa; 2 = I. versicolor; 3 = I. virginica; d = effect size d(5); e = effect size e(15). These effect sizes are shown in the original significant digits. d/e = the ratio of d(5) to e(15); sd ratio = the ratio of the standard deviations of the compared data. Note that reverse comparisons, such as 2 vs 1, were also conducted, but omitted from this table because their effect sizes are the opposites of the original values, and d/e and sd ratio are the inverses of the original ones.

Calculated effect sizes of the difference for the data shown in Table 1. Note: Chara. = characteristics; S.L. = sepal length; S.W. = sepal width; P.L. = petal length; P.W. = petal width; Taxa = compared taxa; 1 = I. setosa; 2 = I. versicolor; 3 = I. virginica; d = effect size d(5); e = effect size e(15). These effect sizes are shown in the original significant digits. d/e = the ratio of d(5) to e(15); sd ratio = the ratio of the standard deviations of the compared data. Note that reverse comparisons, such as 2 vs 1, were also conducted, but omitted from this table because their effect sizes are the opposites of the original values, and d/e and sd ratio are the inverses of the original ones. Fig. 1 shows the ratio of d (5) to e (15) plotted against the ratio of standard deviations of the comparing data. This figure shows that the similar two standard deviations give similar d (5) and e (15). In other words, the more different two standard deviations encourage the use of e (15) over d (5) more.
Figure 1

Plotted graph of Table 2.

Plotted graph of Table 2. To examine the nature of d (5) and e (15), we also conducted a simulation study. In addition to d (5) and e (15), Bonett's statistic (8) was also included as a reference, although its accuracy cannot be discussed because of the lack of the parameter definition. The above effect sizes and their width of 95% CI were calculated for 100,000 Monte Calro replications from and for each condition, and they were represented by their average values. The population means were fixed to 1 and 0. The sample sizes were changed from 10 to 30 by 10. The population standard deviation was fixed to 1 and was changed 1 to 10 by 1. However, some redundant data were omitted from the result. The calculation was conducted using es.dif R package shown above and metafor R package [25]. The R source code used for the simulation was shown in the Appendix. Table 3 shows the result of the simulation. When the sample size ratio was conserved under , e (15) gave more similar and concordant values than d (5). For example, e (15) for under were 0.142, 0.140 and 0.141, whereas the corresponding d (5) were 0.148, 0.143 and 0.143. This is the nature and advantage of e (15) which is designed to estimate the same parameter under heteroscedasticity and the same sample size ratio. The width of CI was narrowest for d (5) under , and e (15) had the second narrowest. Under , e (15), e (15) and (8) had the narrowest CI under , and , respectively. The narrowest CIs of e (15) were followed by d (5), whereas what followed the narrowest CIs of (8) was not fixed. It was shown that e (15) had wider situation under which it had the narrowest or second narrowest CI than d (5) or (8). Bonett's statistic (8) equaled to d (5) under as their definition. Under and , e (15) was closer to (8) than d (5). This might imply relative accuracy of (8) over d (5) under heteroscedasticity.
Table 3

Comparison of effect sizes in simulation.

n1n2σ1σ2d.ESd.Par.e.ESe.Par.B.ESB.Par.d.CIe.CIB.CI
1010111.0001.0000.9951.0001.000U.D.1.8231.8281.911
1010140.355N.C.0.3440.3430.355U.D.1.7221.6911.834
1010170.210N.C.0.2010.2000.210U.D.1.7131.6681.825
10101100.148N.C.0.1420.1410.148U.D.1.7101.6611.822
1020110.9981.0000.9971.0001.002U.D.1.5791.6041.646
1020140.303N.C.0.4080.4080.346U.D.1.4931.5031.310
1020170.175N.C.0.2430.2430.203U.D.1.4871.4881.274
10201100.124N.C.0.1730.1710.143U.D.1.4861.4831.264
1030110.9991.0001.0001.0001.007U.D.1.4821.5361.551
1030140.284N.C.0.4580.4590.343U.D.1.4121.4271.096
1030170.164N.C.0.2780.2770.202U.D.1.4081.4161.046
10301100.114N.C.0.1960.1970.141U.D.1.4071.4111.032
2010111.0021.0001.0011.0001.006U.D.1.5801.6051.647
2010140.431N.C.0.3020.3020.360U.D.1.5231.4621.830
2010170.262N.C.0.1750.1740.213U.D.1.5151.4431.843
20101100.186N.C.0.1230.1220.151U.D.1.5121.4381.846
2020110.9991.0000.9981.0000.999U.D.1.3031.3051.333
2020140.347N.C.0.3420.3430.347U.D.1.2341.2271.277
2020170.204N.C.0.2000.2000.204U.D.1.2271.2141.268
20201100.143N.C.0.1400.1410.143U.D.1.2251.2111.266
2030111.0011.0001.0001.0001.001U.D.1.1891.1991.215
2030140.318N.C.0.3790.3780.346U.D.1.1251.1291.050
2030170.184N.C.0.2220.2220.201U.D.1.1201.1191.032
20301100.130N.C.0.1570.1570.142U.D.1.1191.1151.027
3010110.9981.0000.9991.0001.005U.D.1.4821.5361.551
3010140.486N.C.0.2850.2860.362U.D.1.4471.3781.833
3010170.304N.C.0.1660.1640.216U.D.1.4421.3601.854
30101100.212N.C.0.1130.1150.148U.D.1.4401.3551.859
3020110.9991.0000.9991.0001.000U.D.1.1891.1991.215
3020140.386N.C.0.3160.3160.349U.D.1.1331.1201.268
3020170.229N.C.0.1840.1830.205U.D.1.1271.1081.269
30201100.161N.C.0.1290.1290.144U.D.1.1251.1051.269
3030111.0021.0001.0011.0001.002U.D.1.0671.0691.084
3030140.347N.C.0.3430.3430.347U.D.1.0111.0101.037
3030170.204N.C.0.2020.2000.204U.D.1.0061.0001.029
30301100.143N.C.0.1410.1410.143U.D.1.0050.9981.027

Note: d = effect size d(5); e = effect size e(15); B = effect size (8); Par. = parameter of effect size; CI = width of confidence interval; N.C. = not calculable; U.D. = undefined. The narrowest CI in each row is underlined.

Comparison of effect sizes in simulation. Note: d = effect size d(5); e = effect size e(15); B = effect size (8); Par. = parameter of effect size; CI = width of confidence interval; N.C. = not calculable; U.D. = undefined. The narrowest CI in each row is underlined.

Discussion

Correspondence of effect sizes and t tests

Comparison of t tests and the effect sizes of the difference except (8) shows the clear correspondence between them (Table 4). Statistic d (5) corresponds to the unpaired two-sample t test [16], [17], whose statistic is the basis of g (2). Statistic (12) uses the statistic (13) of Welch's t test [12], which aims to test two means with unequal variances, and (10) uses the same statistic as the one-sample t test [17]. Considering this, it is natural that power analyses should be conducted, using the corresponding pair of the effect size and t test. In other words, power analyses of Student's one-sample t test, Student's unpaired two-sample t test, and Welch's t test should be conducted based on the c statistic (17), d (5), and the e statistic (15), respectively. Co-use of non-corresponding t-test and effect size causes inconsistence of the assumption about the population(s).
Table 4

Correspondence of assumptions, t values, and effect sizes of the difference.

One sample & a constantTwo samples under homoscedasticityTwo samples under heteroscedasticity
As.NormalityNormality, Independence, & HomoscedasticityNormality & Independence
tt=Y1¯Cs12/(n11)t=Y1¯Y2¯Spooled/n˜t=Y1¯Y2¯s12/n1+s22/n2
ESc=Y1¯Cs1Jd=Y1¯Y2¯SpooledJe=Y1¯Y2¯(s12/n1+s22/n2)n˜J

Note: As. = assumption; t = t value; ES = effect size. The degree of freedom of J is omitted for the space and must be calculated corresponding degree of freedom.

Correspondence of assumptions, t values, and effect sizes of the difference. Note: As. = assumption; t = t value; ES = effect size. The degree of freedom of J is omitted for the space and must be calculated corresponding degree of freedom.

Influence of sample size on effect size

In this subsection, the relationship between the effect sizes of the difference and sample sizes is described. The value of g (2), a biased estimator of the effect size of the difference under homoscedasticity, is independent of the sample sizes when the assumption of homoscedasticity () is fulfilled. When , it depends on the ratio as implied in [9]. This is because g (2) is no longer an estimator of δ (1) under , and it will be a biased estimator of the other parameter , which is Note that even d (5) cannot be the unbiased estimator of when , because g (2) is not distributed as non-central t variate in this situation. Even if and vary, g (2) roughly estimates the same parameter, given the ratio q is fixed. Next, the (12) is a biased estimator of (11), but (11) equals to the other parameters in the particular situation. When , , and (12) equals to g (2), and is independent of the sample sizes. When and , . In this case, (12) equals to g (2) and is also independent of the sample sizes. While d (5) is a biased estimator of , e (15) is its unbiased estimator. Therefore, usage of e (15) is always preferable to d (5) in this situation. When and , (12) depends on the rate . Therefore, strictly speaking, multiple s can be comparable only when the sample size ratio r is identical. The effect size estimator (8) did not have a defined parameter, but when and , (8) equals to g (2) and (12), and is independent of sample size. Under and , (8) also equals to g (2) and suffers from the same problem as it. Under and , the value of (8) is no longer the same as g (2), and precise discussion on its behavior is hinderred by the lack of its parameter definition. When trying to consider (8) as a noncentral t-variate like the other effect sizes, its degree of freedom should be about , and and should affect the degree of freedom under . Unlike g (2) or (12), (10) is always independent of the sample size. The behavior of the unbiased estimator of the effect sizes (d (5), e (15), and c (17)) are almost identical to those that are biased, but they slightly increase as the sample sizes become large. This is because of the correction coefficient J (6), and its behavior is illustrated in detail in [14]. In summary, in terms of the effect size of the difference between two means, usage of e (15) is preferable to d (5) or (8), and e (15) can be the remedy for application of effect size of the difference under heteroscedasticity. However, when the ratio of the two sample sizes cannot be set as uniform under heteroscedasticity, neither d (5) nor e (15) can be precisely compared. This is a form of the Behrens-Fisher problem, which cannot be solved strictly.

Potential applications of the new effect sizes

The effect size e (15) has a vast applicable range covering all kinds of natural and social sciences. This is because e (15) corresponds to Welch's t test, whose use is nowadays encouraged over Student's t test (e.g., [26]). The effect size e (15) is the best option, especially when the ratio of the sample sizes of two groups can be fixed. The effect size c (17) has a relatively narrower range regarding the application. In comparison of paired two groups (the difference in pairs vs. 0) and in some simulation studies (result of simulation vs. the optimal value) or physics (result of experiment vs. physical constant), an effect size of the constant may be needed.

Declarations

Author contribution statement

Satoshi Aoki: Conceived and designed the analysis; Analyzed and interpreted the data; Contributed analysis tools or data; Wrote the paper.

Funding statement

This study was partly supported by National Bioresource Project from AMED, grant number 16km0210053j0005.

Competing interest statement

The authors declare no conflict of interest.

Additional information

Supplementary content related to this article has been published online at https://CRAN.R-project.org/package=es.dif.
  8 in total

Review 1.  Heterogeneity of variance in clinical data.

Authors:  R J Grissom
Journal:  J Consult Clin Psychol       Date:  2000-02

2.  Review of assumptions and problems in the appropriate conceptualization of effect size.

Authors:  R J Grissom; J J Kim
Journal:  Psychol Methods       Date:  2001-06

3.  The statistical power of abnormal-social psychological research: a review.

Authors:  J COHEN
Journal:  J Abnorm Soc Psychol       Date:  1962-09

4.  An alternative to Cohen's standardized mean difference effect size: a robust parameter and confidence interval in the two independent groups case.

Authors:  James Algina; H J Keselman; Randall D Penfield
Journal:  Psychol Methods       Date:  2005-09

Review 5.  Effect size, confidence interval and statistical significance: a practical guide for biologists.

Authors:  Shinichi Nakagawa; Innes C Cuthill
Journal:  Biol Rev Camb Philos Soc       Date:  2007-11

6.  The generalisation of student's problems when several different population variances are involved.

Authors:  B L WELCH
Journal:  Biometrika       Date:  1947       Impact factor: 2.445

7.  Confidence intervals for standardized linear contrasts of means.

Authors:  Douglas G Bonett
Journal:  Psychol Methods       Date:  2008-06

8.  The accuracy of effect-size estimates under normals and contaminated normals in meta-analysis.

Authors:  Philomena Marfo; G A Okyere
Journal:  Heliyon       Date:  2019-06-11
  8 in total
  3 in total

Review 1.  Five decades of misunderstanding in the social Hymenoptera: a review and meta-analysis of Michener's paradox.

Authors:  Robert L Jeanne; Kevin J Loope; Andrew M Bouwma; Erik V Nordheim; Michael L Smith
Journal:  Biol Rev Camb Philos Soc       Date:  2022-03-25

2.  Smaller farm size and ruminant animals are associated with increased supply of non-provisioning ecosystem services.

Authors:  Johan O Karlsson; Pernilla Tidåker; Elin Röös
Journal:  Ambio       Date:  2022-04-16       Impact factor: 6.943

3.  The Sound of a Circular City: Towards a Circularity-Driven Quietness.

Authors:  Aggelos Tsaligopoulos; Stella Sofia Kyvelou; Michalis Chiotinis; Aimilia Karapostoli; Eleftheria E Klontza; Demetris F Lekkas; Yiannis G Matsinos
Journal:  Int J Environ Res Public Health       Date:  2022-09-27       Impact factor: 4.614

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.