Literature DB >> 28694745

Model Averaging with AIC Weights for Hypothesis Testing of Hormesis at Low Doses.

Steven B Kim1, Nathan Sanders1.   

Abstract

For many dose-response studies, large samples are not available. Particularly, when the outcome of interest is binary rather than continuous, a large sample size is required to provide evidence for hormesis at low doses. In a small or moderate sample, we can gain statistical power by the use of a parametric model. It is an efficient approach when it is correctly specified, but it can be misleading otherwise. This research is motivated by the fact that data points at high experimental doses have too much contribution in the hypothesis testing when a parametric model is misspecified. In dose-response analyses, to account for model uncertainty and to reduce the impact of model misspecification, averaging multiple models have been widely discussed in the literature. In this article, we propose to average semiparametric models when we test for hormesis at low doses. We show the different characteristics of averaging parametric models and averaging semiparametric models by simulation. We apply the proposed method to real data, and we show that P values from averaged semiparametric models are more credible than P values from averaged parametric methods. When the true dose-response relationship does not follow a parametric assumption, the proposed method can be an alternative robust approach.

Entities:  

Keywords:  Akaike information criterion; hormesis; hypothesis testing; model averaging; model misspecification

Year:  2017        PMID: 28694745      PMCID: PMC5495511          DOI: 10.1177/1559325817715314

Source DB:  PubMed          Journal:  Dose Response        ISSN: 1559-3258            Impact factor:   2.658


Introduction

In toxicology, hormesis is known to be a biphasic dose–response relationship with a stimulatory response at low doses and an inhibitory response at high doses.[1,2] In this article, we focus on statistical hypothesis testing for hormesis when the outcome of interest is binary. The null hypothesis is the absence of hormesis denoted by H 0, and the alternative hypothesis is the presence of hormesis denoted by H 1. For illustration, 2 models for H 0 and 2 models for H 1 are shown in Figure 1.
Figure 1.

Dose–response relationships. The top figures show 2 hypothetical monotonic dose–response relationships (null hypothesis, denoted by H 0), and the bottom figures show 2 hypothetical hormetic dose–response relationships (alternative hypothesis, denoted by H 1).

Dose–response relationships. The top figures show 2 hypothetical monotonic dose–response relationships (null hypothesis, denoted by H 0), and the bottom figures show 2 hypothetical hormetic dose–response relationships (alternative hypothesis, denoted by H 1). Mathematically speaking, we do not have a sign change in the slope of a dose–response curve at low doses when H 0 is true. The slope is entirely positive at low doses (top left in the figure), or the zero slope (ie, flat line) becomes a positive slope after passing some threshold dose point (top right in the figure). On the other hand, we have 1 sign change in the slope at low doses when H 1 is true. The starting slope is negative, and the slope becomes positive at some dose point (bottom left and bottom right in the figure). When H 0 is true, for a given significance level α, we want the probability of rejecting H 0 to be at α or below. When H 1 is true, we want the statistical power as high as possible. We can increase the statistical power by increasing the number of experimental doses and/or the sample size inside the hormetic range. In such a case, various statistical methods are available including polynomial, fractional polynomial modeling, splines, and nonparametric smoothing techniques.[3] However, having such an ideal experimental design is not always possible in practice. Although we prefer statistical methods that require weak assumptions, we need to borrow a parametric assumption to gain statistical power in small-sample studies. Several useful parametric models used in dose–response assessments are equipped in Benchmark Dose Model software.[4] These models can be modified to quadratic or other alternative forms in order to model a nonmonotonic dose–response relationship.[5-7] A parametric model yields a high statistical power when it is correctly specified, but it can lead to a very low statistical power when it is misspecified. The loss of statistical power due to model misspecification will be shown in this study. There are 2 main reasons. First, the observed data points cannot be adequately modeled by the parametric structure.[3] Second, in the case of model misspecification, data points at high doses make situations even worse because they have too much contribution in the parameter estimation. Such data points are called high-leverage points.[8] We have 2 aims in this article. First, we explicitly address the impact of model misspecification and high-leverage points in parametric modeling when we perform hypothesis testing for hormesis at significance level α. Second, we propose averaging multiple semiparametric models using the weights calculated by Akaike information criterion (AIC).[9] In both frequentist and Bayesian frameworks, model averaging methods have been widely discussed in cancer risk assessments, particularly for the estimation of benchmark dose. Model averaging allows us to reduce the impact of model misspecification and to account for model uncertainty.[7,10-13] In this article, we present the simulation study to compare parametric methods and the proposed semiparametric method. For illustration, we apply the methods to some data discussed in Calabrese and Baldwin[14] which seem to show evidence for hormesis at various degrees. In the application, we show that P values calculated from parametric models can be misleading, and P values calculated from the semiparametric method better match with observed dose–response trend.

Statistical Methods

In this section, we review a logistic regression model in “Logistic Regression Model” subsection We base our discussion on the logistic model among many parametric models because of its popularity, and the same discussion can be carried out for another form of parameterization. We briefly review the model averaging method based on the AIC in “Model Averaging in Parametric Models (L3 and L4)” subsection. We then discuss the application of model averaging to semiparametric models in “Application of Model Averaging to Semiparametric Models” subsection. More mathematical detail is included in the Appendix. The following notation is used in the section. Let x ≥ 0 denote the jth fixed experimental dose for j = 1,…, J, where J is the total number of experimental doses. Without loss of generality, let x 1 = 0 be the control dose and x 1 < x 2 <…< x. Let Y denote the binary random variable, where Y = 1 if the ith experimental unit treated at dose x shows a toxic outcome and Y = 0 otherwise. Let π denote the unknown probability of observing Y = 1 (ie, the probability of observing a toxic outcome at does x). Let n denote the number of experimental units observed at dose x, and let N = Σ denote the total sample size for the experiment.

Logistic Regression Model

To model a potential “J-shaped” dose–response relationship, a logistic regression model with the quadratic term is briefly discussed by May and Bigelow.[3] They considered the log dose, but it is not an important distinction in our discussion. Throughout the article, the logistic regression model is referred to as the L3 model, where L3 stands for the logistic regression model with 3 parameters β0, β1, and β2. The quadratic term allows a hormetic dose–response curve at low doses, and similar parameterizations are discussed by Bogen[6] and Kim et al[7] using different link functions. In the quadratic form, hypothesis testing for hormesis is simply formulated as H 0: β1 ≥ 0 versus H 1: β1 < 0. In a large sample, we can make inference for β1 based on the maximum likelihood estimator for β1, but we usually do not observe such a large sample size particularly inside the hormetic range in practice.[3] In this article, we base our inference on bootstrapping to approximate the sampling distribution of the maximum likelihood estimator for β1.[15] Our focus is on the consequence of wrong implementation of this model. The L3 model is an efficient statistical strategy when it is correctly specified. On the other hand, when the model is misspecified, it can lead us to a misleading result even in a large sample. Under model misspecification, observations made at high experimental doses have too much contribution in the estimation of β1 (known as high-leverage points), and it can significantly decrease statistical power. Motivated by a power transformation described by Tukey transformation,[16] the L3 model can be modified as where β3 > 0. This parameterization is referred to as the L4 model, where L4 stands for the logistic regression with 4 parameters β0, β1, β2, and β3. A similar parameterization was discussed by Kim et al.[17] Compared to the L3 model, the L4 model is more flexible while it maintains the same form of hypothesis testing H 0: β1 ≥ 0 versus H 1: β1 < 0. As shown in Figure 2, the L3 model is limited to a symmetric hormetic zone only (figure on the left), and the L4 model is able to generate an asymmetric hormetic zone by adding the power parameter β3 (figures in the middle and on the right). Note that the L3 model is a special case of the L4 model when β3 = 1.
Figure 2.

Hormetic dose–response relationships generated by the logistic regression models. The left figure is the L3 model with β 0 = −1.5, β 1 = −5, and β 2 = 10, the middle figure is the L4 model with β 0 = −1.5, β 1 = −5, β 2 = 10, and β 3 = 0.5, and the right figure is the L4 model with β 0 = −1.5, β 1 = −5, β 2 = 10, and β 3 = 2.

Hormetic dose–response relationships generated by the logistic regression models. The left figure is the L3 model with β 0 = −1.5, β 1 = −5, and β 2 = 10, the middle figure is the L4 model with β 0 = −1.5, β 1 = −5, β 2 = 10, and β 3 = 0.5, and the right figure is the L4 model with β 0 = −1.5, β 1 = −5, β 2 = 10, and β 3 = 2.

Model Averaging in Parametric Models (L3 and L4)

The advantage of the L3 model is efficiency (high statistical power) when the true hormetic zone is symmetric and the true dose–response relationship follows the quadratic form. The disadvantage is a loss of statistical power when the model is misspecified. It is also sensitive to high-leverage points. The impact of model misspecification and high-leverage points can be increased particularly when the experiment is poorly designed. On the other hand, the advantage of the L4 model is flexibility, and it can maintain relatively high statistical power when the true hormetic zone is asymmetric. The disadvantage of L4 is a loss of statistical power due to overparameterization when the true hormetic dose–response relationship can be adequately modeled by the L3 model. To compromise the characteristics of L3 and L4, we may consider the model averaging method based on the AIC.[9] The method of AIC model averaging is often used to account for model uncertainty and to reduce the impact of model misspecification when a single model is used for inference.[10,18,19] Let w L3 and w L4 denote the AIC weight calculated from the observed data, where w L3 + w L3 = 1 (see Appendix). When a fitted model has a higher value of the likelihood function than the other fitted model, it receives a higher weight. In addition, it penalizes an additional model parameter to balance between goodness of fit and overparameterization. When we have 2 bootstrap distributions from L3 and L4, we can consider a mixture distribution according the AIC weights. Since both L3 and L4 determine hormesis by the sign of β1, we can perform the hypothesis testing H 0: β1 ≥ 0 versus H 1: β1 < 0 based on the mixture distribution of estimated β1. At significance level α = .05, we reject H 0 in favor of H 1 when the (1 − α)th quantile of the distribution is below 0. Throughout the discussion, this model averaging method is referred to as MAP (model averaging with the parameter models). When H 0 is true, the type I error rate under MAA is fairly close to the minimum of the type I error rate under L3 and the type I error rate under L4 (see Appendix and Simulation Result). Averaging β1 in L3 and β1 in L4 is a meaningless procedure when our goal is parameter estimation. However, our goal is hypothesis testing not parameter estimation. Though β1 in L3 and β1 in L4 have different meanings, both β1 < 0 in L3 and β1 < 0 in L4 have the same meaning (presence of hormesis) in the context of our hypothesis testing. Under each model using bootstrapping, we may seek evidence for β1 < 0 through the bootstrap distribution of estimated β1. Therefore, when a mixture bootstrap distribution of estimated β1 is mostly negative (regardless of magnitude), it may be regarded as evidence for hormesis.

Application of Model Averaging to Semiparametric Models

Recall that we let π denote the probability of a toxic outcome at dose x. In general, we need evidence for π1 > π2 to reject H 0 in favor of H 1. In this section, we introduce a proposed model averaging method which does not require a link function between the probability of a toxic outcome and the dose. Let M 1 denote the saturated model with J free parameters, π1, π2,…, π, so each π is estimated by the observed proportion at dose x. Let M 2 denote a model with J − 1 free parameters by the condition π2 = π3, so π2 is estimated by the observed proportion at the 2 doses x 2 and x 3. Let M 3 denote a model with J − 2 free parameters by the condition π2 = π3 = π4, so π2 is estimated by the observed proportion at the 3 doses x 2, x 3, and x 4. Using general notation, let M denote the model with J − k + 1 free parameters such that π2 =…= π +1. We consider up to M −2 and obtain the AIC-weights w 1,…, w −2 by maximizing the log-likelihood function (see Appendix for detail). When we have more experimental units inside the hormetic range, the model averaging will become more robust regardless of the values of x 1,…, x because the model structure depends on the order x 1,…, x and estimated π1,…, π. It does not assume a particular shape of dose–response curve. For each M, we can obtain the bootstrap distribution of estimated π1 − π2, then we test for hypothesis testing H 0: π1 − π2 ≤ 0 versus H 1: π1 − π2 > 0 based on the mixture of the J − 2 bootstrap distributions weighted by the AIC weights. At significance level α, we reject H 0 in favor of H 1 when the αth quantile of the mixture distribution exceeds 0. Throughout the discussion, this model averaging method is referred to as MASP (model averaging with semiparametric models).

Results

In this section, we compare the operating characteristics of the 4 aforementioned models: L3, L4, MAP, and MASP. We consider 16 simulation scenarios with 5 scenarios under H 0 and 11 scenarios under H 1 (see Simulation Design). We summarize the simulation results by the probability of rejecting H 0 under each scenario and for each model (see Simulation Result). Then, we apply the 4 models to the data discussed in Calabrese and Baldwin[14] which provided some degree of evidence for hormesis at low doses (see Application).

Simulation Design

To control noise in the simulation, for all 16 scenarios, we assumed J = 6 experimental doses geometrically spaced as x 1 = 0, x 2 = 0.0625, x 3 = 0.125, x 4 = 0.25, x 5 = 0.5, and x 6 = 1, and we assumed n = 50 for each dose group so that N = 300 is the total sample size. For the case of H 0, we generated data under the logistic models L3 or L4 (scenarios 1-5). For the case of H 1, we generated data under L3 (scenarios 6 and 7), L4 (scenarios 8-13) and neither (scenarios 14-16). The scenarios under the parametric structures are shown in Figure 3 (scenarios 1-13). For these parametric scenarios, the parameter values are presented in Table 1. For scenarios 14 to 16, we broke the parametric structures, so both L3 and L4 are misspecified models in the 3 scenarios. We made pointwise assumptions π1 = 0.2, π2 = 0.09, π3 = 0.1, π4 = 0.2, π5 = 0.4, and π6 = 0.6 in scenario 14; 0.2, 0.09, 0.07, 0.2, 0.4, and 0.6 in scenario 15, respectively; and 0.2, 0.07, 0.08, 0.2, 0.4, and 0.6 in scenario 16, respectively. Each scenario was simulated 1000 times, and 2000 bootstrap samples were used per simulated sample. We fixed the significance level at α = .05, and the probability of rejecting H 0 was recorded for each method under each scenario.
Figure 3.

Simulation scenarios generated by the logistic models (scenarios 1-13). The parameter values are provided in Table 1.

Table 1.

Parameter Values for Scenarios 1 to 13 Under the Logistic Model.

Scenario12345678910111213
β0−1.3−1.3−1.3−1.3−2.3−1.39−1.39−1.39−1.39−1.39−1.39−1.39−1.39
β100000−7.33−10.02−5.18−4.62−7.09−6.31−7.33−10.02
β20131314.6620.047.335.8210.027.9514.6620.04
β3011.5.511.5.33.5.33.5.5
Simulation scenarios generated by the logistic models (scenarios 1-13). The parameter values are provided in Table 1. Parameter Values for Scenarios 1 to 13 Under the Logistic Model.

Simulation Result

Table 2 provides the simulation results. When H 0 was true in scenarios 1 to 5, the L3 model violated α = .05 in some scenarios at mild degree, and the L4 model violated α = .05 at serious degree in scenario 2. The estimated type I error probability was .082 which is difficult to believe that it just happened by chance for 1000 replications of the scenario. In each scenario, the model averaging method MAP rejected H 0 with a probability between the resulting probabilities in L3 and L4 with an anticipated result. The model averaging method MASP obeyed α = .05 in the 5 null scenarios.
Table 2.

Simulation Results, the Probability of Rejecting H 0 Based on 1000 Simulated Data per Scenario.a

ScenarioTrue ModelL3L4MAP MASP
1L3/L4.055.044.042.044
2L3/L4.056.082.061.037
3L3/L4.019.029.018.020
4L3/L4.066.038.043.025
5L3/L4.000.003.000.019
6L3.984.764.894.279
7L3.998.845.948.400
8L4.277.570.467.474
9L4.135.474.327.470
10L4.387.706.641.673
11L4.145.635.517.689
12L4.318.525.446.475
13L4.453.768.660.670
14.000.180.107.438
15.000.200.144.580
16.000.268.200.666

aL3 represents logistic regression model with the 3 parameters β0, β1, and β2; L4, logistic regression model with the 4 parameters β0, β1, β2, and β3; MAP, model averaging with the parametric models L3 and L4; MASP, model averaging with the semiparametric models.

Simulation Results, the Probability of Rejecting H 0 Based on 1000 Simulated Data per Scenario.a aL3 represents logistic regression model with the 3 parameters β0, β1, and β2; L4, logistic regression model with the 4 parameters β0, β1, β2, and β3; MAP, model averaging with the parametric models L3 and L4; MASP, model averaging with the semiparametric models. In scenarios 6 and 7, when H 1 was true under the L3 model, the L4 model led to slightly lower statistical powers than the L3 model due to overparameterization. The MAP model yielded statistical power between the results in M3 and M4 as anticipated. On the other hand, the MASP yielded substantially lower statistical powers in the 2 scenarios. When the true dose–response curve is generated under the simple L3 model, the L3 model outperformed which is not surprising. In scenarios 8 to 13, when H 1 was true under the L4 model, the L3 model could not tolerate model misspecification by showing substantially lower statistical powers due to the inflexibility. The L4 model mostly showed outperformance because the scenarios belong to its own parameterization. In these 6 scenarios, the MASP model was consistently more powerful than the MAp model despite the contribution of the true L4 model to MAP. In scenarios 9 to 11, the MASP model showed comparable results to the results from the true L4 model. We now turn our focus on scenarios 14, 15, and 16. Recall that these 3 scenarios did not belong to any of L3, L4, MAP, and MASP. The M3 model could not reject H 0 even once, the L4 model showed statistical powers of .180, .200, and .268, respectively, and the MAP model showed statistical powers of .107, .144, and .200, respectively. On the other hand, the MASP showed statistical powers of .438, .580, and .666, respectively. The take-home message is clear. The parametric L3 and L4 models sometimes showed outperformance within their own parameterizations (statistical power of .984 and .998 from L3 in scenarios 6 and 7, respectively; statistical power of .570, .474, .706, .525, and .768 from L4 in scenarios 8, 9, 10, 12 and 13, respectively; see Table 2), but they performed poorly when the truth was not under the model (statistical power of 0 from L3 in scenarios 14, 15, and 16; statistical power of .180, .200, and .268 from L4 in scenarios 14, 15, and 16 which are less than one half when compared to .438, .580, and .666 from MASP in the respective scenarios; see Table 2). This is a general phenomenon in statistics. On the other hand, the MASP model was relatively less sensitive, and the model averaging among the semiparametric model (MASP) showed greater statistical power than the model averaging among the parametric models (MAP) in the scenarios except when L3 was the true model (see scenarios 8-16 in Table 2).

Application

In this section, we apply the 4 models (L3, L4, MAP, and MASP) to some of binary data discussed in Calabrese and Baldwin[14] and compare P values from the 4 models. Calabrese and Baldwin[14] discussed the effects of saccharin on hyperplasia of the urinary bladder. The 6 experimental doses were 0, .01, .1, 1, 5, and 7.5 (% in diets). The respective observed proportions of tumor incidence were 10/73 (14%), 6/71 (8%), 4/81 (5%), 4/76 (5%), 6/64 (9%), and 19/62 (31%) for male rats (see the top left panel in Figure 4). When we implemented the L3, L4, MAP, and MASP models, the respective P values were .042, .027, .034, and .059, respectively. This is one of the cases when the parametric models adequately described the dose–response relationships, and the MASP model could not achieve the significance level α = .05. At the same experimental doses, the respective observed proportions of incidence were 3/85 (4%), 0/81 (0%), 0/81 (0%), 3/90 (3%), 5/88 (6%), and 10/76 (13%) for female rats (see the top right panel in Figure 4). When we implemented the L3, L4, MAP, and MASP models, the respective P values were .811, .070, .207, and .014, respectively. As shown in the figure (the top right panel), the data points in the observed range are not symmetric, so the L3 model was not able to model the observed nonmonotonic dose–response relationship adequately.
Figure 4.

Fitted dose–response curves under the L3 and L4 models using the maximum likelihood estimates.

Fitted dose–response curves under the L3 and L4 models using the maximum likelihood estimates. Calabrese and Baldwin[14] also discussed the effect of 3-Methylcholanthrene on pulmonary tumors in female rats. The 9 experimental doses were 0, .005, .015, .046, .137, .4, 1.2, 3.7, and 11.1 (μg). The observed proportions of tumor incidence were 15/34 (44%), 1/18 (6%), 5/19 (26%), 7/18 (39%), 6/20 (30%), 12/24 (50%), 8/11 (73%), 10/10 (100%), and 11/11 (100%), respectively (see the bottom left panel in Figure 4). Compared to the previous data set (effect of saccharin on urinary bladder), it has a smaller total sample size but a larger number of experimental doses. The L3, L4, MAP, and MASP models yielded P values of .869, .070, .224, and .003, respectively. As shown in Figure 4 (the bottom left panel), the L3 model was not flexible enough to describe the observed data, and the L4 model was quite flexible to follow the observed nonmonotonic trend though it did not achieve statistical significance. Under the MASP model, the data served as significance evidence for hormesis with a P value .003 < α = .05. As a final example, the same paper[14] discussed the effect of cadmium chloride on testicular tumors. From the 7 experimental doses 0, 1, 2.5, 5, 10, 20, and 40 (µmol/kg), the observed proportions of incidence were 8/45 (17.8%), 1/30 (3.3%), 3/29 (10.3%), 3/30 (10.0%), 4/30 (13.3%), 21/29 (72.4%), and 24/29 (82.8%), respectively (see the bottom right panel in Figure 4). Despite the strong hormetic trend, P values from L3, L4, MAP, and MASP were .99, .64, .65, and .06, respectively, and they seem to be influenced by the 2 data pints at the high doses 20 and 40 μmol/kg. The P values from the parametric methods (L3, L4, and MAP) do not seem credible, and the P value from averaging the semiparametric models seems more credible based on the observed trend before modeling. In the figure (the bottom right panel), the L3 model was not able to model the asymmetric hormetic trend, and the flexibility of the L4 model was used to chase the 2 data points at the high experimental doses rather than the data points at low doses. An interesting issue with the L4 model is discussed in the following section.

Discussion

The focus of this article is not to argue the existence of hormesis for a particular carcinogen. Our focus is a valid hypothesis testing for a hormetic effect at low doses. We presented the different characteristics of averaging parametric models and averaging semiparametric models. In conclusion, when the true hormetic relationship is under the simple L3 model, the parametric approach MAP (and individual L3 and L4) outperformed the semiparametric approach MASP, which is not surprising. On the other hand, when we compare the 2 averaging methods, MASP outperformed MAP when the true hormetic relationship is nonparametric and even when the truth is generated under the parametric L4 model. It is also shown that the parametric approaches cannot tolerate model misspecification for the hypothesis testing. To this end, when the truth does not follow a parametric form, MASP can be useful for more robust and higher statistical power. In large sample studies with many experimental doses in a hormetic range, a nonparametric method can be a more reasonable approach because it can disconnect information between doses inside a hormetic range and higher doses. May and Bigelow[3] discussed the practical challenges due to insufficient sample sizes and a lack of experimental doses (missing a potential hormetic range). In small sample studies, borrowing mathematical structure (ie, parametric models) to gain efficiency seems inevitable. As discussed in “Results,” however, it sometimes gives us a misleading result not simply due to a lack of evidence but due to model misspecification and high-leverage data points (recall the low statistical power from L3, L4, and MAP in scenarios 14-16 in Table 2). Motivated by this fact, we considered averaging semiparametric models (MASP) with AIC weights to test for a hormetic effect at low doses. When we implemented to real data, calculated P values from MASP made more sense than P values from L3, L4, and MAP particularly in the last applied example in “Application” (cadmium chloride on testicular tumors). In the simulation study, we showed pros and cons of the parametric methods (L3, L4, and MAP) and of the semiparametric method (MASP). Under the correctly specified parametric form, the statistical power from L3 was highest among the 4 methods as shown in scenarios 6 and 7 in Table 2, and MAP outperformed MASP in the 2 scenarios generated by L3. Under the wrong parametric form, we lost statistical power substantially as shown in scenarios 14 to 16 (zero statistical power from L3 in Table 2). We observed that MASP outperformed MSP in the 3 scenarios generated outside of L3 or L4. Even when the true scenario was generated by L4, MASP could yield higher statistical power than MAP as shown in scenarios 8 to 13 (see Table 2). In practice, we cannot guarantee that observed outcomes (which are based on the unknown true dose–response relationship) will follow a parametric model as seen in the cadmium chloride data, and it is not under our control. Instead, the proposed model averaging method with semiparametric models does not require the parametric form, and we learned that it is relatively less sensitive to the shape of observed dose–response trend. For the L4 model, we refitted the cadmium chloride data with the restriction 0 < β3 < 1. The restricted parameterization yielded a P value lower than α = .05 by adequately modeling the asymmetric hormetic zone. However, when we ran a simulation study under the null scenarios, it violated the significance level α = .05 seriously. In other words, the restricted parameterization tended to favor H 1 too often when H 0 is true. Additionally, we implemented the nonparametric regression method discussed in Hall and Heckman[20] in the simulation study. The advantage of a nonparametric method is insensitivity to data points at high doses, but it yielded lower statistical powers than MASP in all of the alternative scenarios (generally below .2). We also studied other methods among others. We modified the test statistic in Baraud et al,[21] which is more appropriate for the binomial data (ie, difference in the 2 probabilities, comparing the control group and the lowest nonzero dose group). It tended to be too conservative, and it did not seem appropriate for a sparse experimental doses. Similarly, we tested an umbrella shape at low doses (decreasing then increasing) using the additive constrained regression.[22] Due to the small number of dose groups, the model fitted the data exactly, and it inflated the type I error rate in null scenarios. A nonparametric method is often useful for a large sample study, but 50 experimental units per dose group with 6 experimental units do not seem sufficiently large particularly for binary outcomes. To further study L3, L4, MAP, and MASP for increased sample sizes (fixing the experimental doses), we manipulated the cadmium chloride data by doubling the sample size while maintaining the observed proportions, that is 16/90 (17.8%), 2/60 (3.3%), 6/58 (10.3%), 6/60 (10.0%), 8/60 (13.3%), 42/58 (72.4%), and 48/58 (82.8%). Recall the P values were .99, .64, .65, and .06 for L3, L4, MAP, and MASP, respectively, before the manipulation (see Application). After the manipulation, the P values changed to .999, .727, .727, and .006 for L3, L4, MAP, and MASP, respectively. When we multiplied the sample size by 10, that is 80/450 (17.8%), 10/300 (3.3%), 30/290 (10.3%), 30/300 (10.0%), 40/300 (13.3%), 210/290 (72.4%), and 240/290 (82.8%), the P values were close to 1, .943, .943, and close to 0 for L3, L4, MAP, and MASP, respectively. Then, we tested an additional simulation scenario under the assumptions (1) π1 = .178, π2 = .033, π3 = .103, π4 = .100, π5 = .133, π6 = .724, and π7 = .828 and (2) n 1 = 450, n 2 = 300, n 3 = 290, n 4 = 300, n 5 = 300, n 6 = 290, and n 7 = 290. The resulting statistical powers were near 0, .005, .005, and near 1 for L3, L4, MAP, and MASP, respectively. The results illustrate the impact of model misspecification even with large data under the parametric methods when we test for hormesis. The Generic Hockey Stick model proposed by Bogen[6] is a parametric model with the enhanced polynomial flexibility and the link function used in the linearized multistage model.[23] In sparse data, it may have the advantage of utilizing 3 parameters to allow greater flexibility than the L3 models considered in this study. However, the Fisher expected information tells us that the parameter estimation still heavily depends on high data points when we use a polynomial predictor. In practice, high-response data are removed when (1) they are irrelevant to low-dose inference and (2) they severely deviate from an assumed parametric model. We concern about too few data points after removing the data points, and it may increase the type I error rate under the null scenario. It is our future study. Based on the simulation study, we have thought that averaging L4 and MASP can balance the sensitivity (when the observed data points can be approximated by the assumed parametric structure) and the robustness (when they do not follow the assumed parametric structure). It is our current research direction.
  14 in total

1.  A parametric model for detecting hormetic effects in developmental toxicity studies.

Authors:  Daniel L Hunt; Dale Bowman
Journal:  Risk Anal       Date:  2004-02       Impact factor: 4.000

2.  Model averaging using the Kullback information criterion in estimating effective doses for microbial infection and illness.

Authors:  Hojin Moon; Hyun-Joo Kim; James J Chen; Ralph L Kodell
Journal:  Risk Anal       Date:  2005-10       Impact factor: 4.000

3.  Model uncertainty and risk estimation for experimental studies of quantal responses.

Authors:  A John Bailer; Robert B Noble; Matthew W Wheeler
Journal:  Risk Anal       Date:  2005-04       Impact factor: 4.000

Review 4.  Hormesis defined.

Authors:  Mark P Mattson
Journal:  Ageing Res Rev       Date:  2007-12-05       Impact factor: 10.895

5.  Modeling nonlinear dose-response relationships in epidemiologic studies: statistical approaches and practical challenges.

Authors:  Susanne May; Carol Bigelow
Journal:  Dose Response       Date:  2006-05-22       Impact factor: 2.658

6.  Estimation of a benchmark dose in the presence or absence of hormesis using posterior averaging.

Authors:  Steven B Kim; Scott M Bartell; Daniel L Gillen
Journal:  Risk Anal       Date:  2014-11-10       Impact factor: 4.000

7.  Generic Hockey-Stick Model for Estimating Benchmark Dose and Potency: Performance Relative to BMDS and Application to Anthraquinone.

Authors:  Kenneth T Bogen
Journal:  Dose Response       Date:  2010-10-21       Impact factor: 2.658

8.  Potential uncertainty reduction in model-averaged benchmark dose estimates informed by an additional dose study.

Authors:  Kan Shao; Mitchell J Small
Journal:  Risk Anal       Date:  2011-03-09       Impact factor: 4.000

9.  Can the concept of hormesis Be generalized to carcinogenesis?

Authors:  E J Calabrese; L A Baldwin
Journal:  Regul Toxicol Pharmacol       Date:  1998-12       Impact factor: 3.271

10.  Information-theoretic model-averaged benchmark dose analysis in environmental risk assessment.

Authors:  Walter W Piegorsch; Lingling An; Alissa A Wickens; R Webster West; Edsel A Peña; Wensong Wu
Journal:  Environmetrics       Date:  2013-05-01       Impact factor: 1.900

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.