Literature DB >> 29026280

M-test in linear models with negatively superadditive dependent errors.

Yuncai Yu1, Hongchang Hu2, Ling Liu3, Shouyou Huang2.   

Abstract

This paper is concerned with the testing hypotheses of regression parameters in linear models in which errors are negatively superadditive dependent (NSD). A robust M-test base on M-criterion is proposed. The asymptotic distribution of the test statistic is obtained and the consistent estimates of the redundancy parameters involved in the asymptotic distribution are established. Finally, some Monte Carlo simulations are given to substantiate the stability of the parameter estimates and the power of the test, for various choices of M-methods, explanatory variables and different sample sizes.

Entities:  

Keywords:  M-test; Monte Carlo simulations; NSD random sequences; asymptotic property; linear regression models

Year:  2017        PMID: 29026280      PMCID: PMC5610259          DOI: 10.1186/s13660-017-1509-6

Source DB:  PubMed          Journal:  J Inequal Appl        ISSN: 1025-5834            Impact factor:   2.491


Introduction

Consider the linear regression model: where and are real-valued responses and real-valued random vectors, respectively. The superscript ⊤ represents the transpose throughout this paper, is a p-vector of the unknown parameter, and are random errors. It is well known that linear regression models have received much attentions for their immense applications in various areas such as engineering technology, economics and social sciences. Unfortunately, there exists the problem that the classical maximum likelihood estimator for these models is sufficiently sensitive to outliers. To overcome this defect, Huber proposed the M-estimate which possesses the robustness (see Huber [1]) by minimizing where ρ is a convex function. It is obvious that many important estimates can be addressed easily. For instance, the least square (LS) estimate with , the least absolute deviation (LAD) estimate with , and the Huber estimate with , , where is the indicator function of A. Let be a minimizer of (2) and consequently is a M-estimate of β. Some excellent results as regards the asymptotic properties of with various forms of ρ have been reported in [2-5]. Most of the results rely on the independence errors. As Huber claimed in [1], the independence assumption on the errors is a serious restriction. It is practically essential and imperative to explore the case of dependent errors, which is a theoretically challenging. Under the dependence assumption of the errors, Berlinet et al. [6] proved the consistency of M-estimates for linear models with strong mixing errors. Cui et al. [7] obtained the asymptotic distributions of M-estimates for linear models with spatially correlated errors. Wu [8] investigated the weak and strong Bahadur representations of the M-estimates for linear models with stationary causal errors. Wu [9] established the strong consistency of M-estimates for linear models with negatively dependent (NA) random errors. In the following we will introduce a wide random sequence, NSD random sequence, whose definition based on the superadditive functions.

Definition 1

Hu [10] A function , is called superadditive if for all , where ‘∨’ is for a componentwise maximum and ‘∧’ is for a componentwise minimum.

Definition 2

Hu [10] A random vector is said to be NSD if where are independent random variables such that have same marginal distribution with for each t, and ϕ is a superadditive function such that the expectations in (3) exist.

Definition 3

Wang et al. [11] A sequence of random variables is called NSD if for all , is NSD. The concept of NSD random variables, which generalizes the concept of NA, was proposed by Hu [1]. In such paper, author provided several essential properties and valuable theorems for NSD. It is realized that many multivariate distributions possess the NSD property exhibited in practical examples. Compared with NA, NSD contains more widely sequence [12], i.e., NA means NSD, but not vise verse. Consequently, NSD has received an increasing attention for its enormous research value in comparison with NA both in copula theory and applications [13-19]. Specifically, a Kolmogorov and a Rosenthal inequality of NSD random variables are introduced in [16] and [17], respectively. Furthermore, Wang et al. [11] obtained the complete convergence for weighted sums of NSD random variables and investigated the complete consistency of LS estimates in the EV models. Wang et al. [19] established the strong consistency of M-estimates for linear models with NSD errors via improving the moment condition. The purpose of this paper is to investigate the M-test problem of the regression parameters in the model (1) with NSD random errors, we consider a test for the following hypothesis: where H is a known matrix with the rank q , b is a known p-vector. A sequence of the local alternatives is considered as follows: where is a known p-vector such that where , is the Euclidean norm. Denote Actually, and are the M-estimates in the restricted and unrestricted model (1), respectively. To test the hypothesis (4), we adopt M-criterion which regards as the criterion to measure the level of departure from the null hypothesis. Several classical conclusions have been presented in [20-22] when the errors are assumed to be independence, we will generalize the case to NSD random errors. Throughout this paper, let C be a positive constant. Put if is a p-vector. Let and . A random sequence is said to on -norm, , if . Denote if converges to 0 in probability and if converges to a constant in probability. The rest of the paper is organized as follows. In Section 2, the asymptotic distribution of is obtained with the NSD random errors, and the consistence estimates of the redundancy parameters λ and are constructed under the local hypothesis. Section 3 gives the theoretical proofs of main results. The simulations are presented to show the performances of parameter estimates and the M-test for the powers in Section 4, and the conclusions are given in Section 5.

Main results

In this paper, let ρ be a non-monotonic convex function on , and denote and as the right and left derivatives of the function ρ, respectively. The derivative function ψ is chosen to satisfy , for all . Now, several assumptions are listed as follows: The function exists with , and has a positive derivative λ at . , and . There exists a positive constant Δ such that for , the function is monotonic on u. . Denote , assume that for sufficiently large n, and

Remark 1

(A1)-(A4) are often applied in the asymptotic theory of M-estimate in regression models (see [20-30]). (A5) is reasonable because it is equivalent to the bound of , and here is a particular case of the condition for some , which was used in Wang et al. [19]. Those functions were mentioned in (2) whose ‘derivative’ function ψ correspond to least square (LS) estimate with , least absolute deviation (LAD) estimate with and Huber estimate with are satisfied with the above conditions.

Theorem 1

In the model (1), assume that , which is a sequence of identically distributed NSD random variables, is an uniformly integral family on L2-norm, and (A1)-(A5) hold. Then has an asymptotic non-central chi-squared distribution with p-degrees of freedom and a non-central parameter , namely, where , , . In particular, when the local alternatives , which means that the true parameters deviate from the null hypothesis slightly, then has an asymptotic central chi-squared distribution with p degrees of freedom For a given significance level α, we can determine the rejection region as follows: where , are the -quantile and -quantile of central chi-squared distribution with p degrees of freedom, respectively.

Theorem 2

Denote where , and is a sequence chosen to satisfy Under the conditions of Theorem  1, we have Under the assumption , replacing λ, by their consistent estimates and , then

Proof of theorems

It is convenient to consider the rescaled model where , , . It is easily to check that Assume that , there exists a matrix K with the rank such that and , then, for some , and can be written as Denote , , then Let . Under the null hypothesis, model (11) can be rewritten as Set , , under the local alternatives (13), Define , and satisfies Obviously, , are the M-estimates of and , respectively. Thus Next, we will state some lemmas that are needed in order to prove the main results of this paper.

Lemma 1

Hu [10] If is a NSD random sequence, we have the following properties. For any , If are non-decreasing functions, then is still a NSD random sequence. The sequence is also NSD.

Lemma 2

(Rosenthal inequality) (Shen et al. [17]) Let be a NSD random sequence with and for some , then, for all and ,

Lemma 3

Anderson et al. [31] Let D be an open convex subset of and are convex functions on D, for any , If f is a real function on D, then f is also convex, and for all compact subset , Moreover, if f is a differentiable function on D, and represent the gradient and sub-gradient of f, respectively, then (16) implies that for all

Lemma 4

Assume that is a sequence of identically distributed NSD random sequence with finite variance, and an array of real numbers is satisfied , . Then, for any real , , where , are disjoint subsets of , i refers to imaginary unit.

Proof

For a pair of NSD random variables X, Y, by the property (a) in Lemma 1, we have Denote by the joint distribution functions of , and , the marginal distribution function of X, Y, one gets Form (18), we obtain where are complex valued functions on with , . Combining (17) and (18) yields Taking , , it is easily seen that thus for any real numbers r, u Next, we proceed the proof by induction on n. Lemma 4 for is trivial and for is true by (20). Assume that the result is true for all (). For , there exist some , , such that Denote , , then Note that are non-decreasing functions, we have by the induction hypothesis that where , , . Thus, this completes the proof of Lemma 4. □

Lemma 5

Billingsley [32] If , for each j, and uniform measure ϱ is satisfied that for all , then

Lemma 6

Suppose that and satisfy the assumptions of Lemma  4. Further assume that is an uniformly integral family on -norm, then where . Without loss of generality, we suppose that for all . By (18), we have because of the negativity of . Then, for , Taking in assumption (A4), we get, for all and sufficiently large j, Therefore, for a fixed small ε, there exists a positive integer such that Denote , where denotes the integer part of x, and , , We define , , and put Note that it is easy to see that , where # stands for the cardinality of a set. Next, we will proof that satisfies the Lindeberg condition. Let , by Lemma 2, it yields Clearly, is uniform integrable since is uniform integrable. Hence, for any positive ε, is verified to satisfy the Lindeberg condition by Now taking an independence random sequence such that have same marginal distribution with for each j. Let and be the eigenfunctions of and , respectively. Choosing , we have by Lemma 4 and (21) By Levy’s theorem, obtains the asymptotic normality, applying Lemma 5, then the identically distribution property of implies that which completes the proof of Lemma 6. □

Lemma 7

In the model (1), assume that is a sequence of NSD identically distributed random variables, (A1)-(A4) are satisfied, for any positive constant δ and sufficiently large n, then where is a p-vector. Denote For fixed , it follows from (A5) that From (A1) and (22), there exist a sequence of positive numbers and such that, for sufficiently large n, In view of the monotonicity of , the summands of is also monotonous with respect to from the property (b) in Lemma 1. We divide the summands of into positive and negative two parts, by the property (c) in Lemma 1, they are still NSD. Therefore, applying Schwarz’s inequality and (22), we obtain Hence for sufficiently large n, we have Lemma 7 is proved by (23) and Lemma 3. □

Lemma 8

Under conditions of Lemma  7 and the local alternatives (5)-(6), then we see that, for any positive constant δ and sufficiently large n, where , . Take the proofs of (24) and (25) as examples, the rest, equations (26) and (27), are the same. Note that (24) can be written as On the other hand, and , hence there exists a constant such that Thus (24) and (25) follow naturally by (28) and Lemma 7. □

Lemma 9

Under the conditions of Theorem  1, as , we have The estimate of (2) can be defined essentially as the solution of the following equation: Denote , (31) can be rewritten to give By a routine argument, we shall prove that Let U be a denumerable dense subset in the unit sphere of such that Write where , . Obviously, for a given , is non-decreasing on L since ψ is non-decreasing. For any , let Thus by (32), there exists a number , as , Note that is a non-decreasing function on for given L, then Based on Lemma 8 and , one can see that On the other hand, by Schwarz’s inequality, we have Combining (34) and (35), there exists such that Applying Chebyshev’s inequality, the inequality and Lemma 8, we have From (36) and (37), it follows that Likewise, when , we obtain Thus the result (33) follows from (38) and the arbitrariness of ε. By Lemma 8 and , it follows that which implies that Consequently, (29) is proved. As defined in (15), can similarly be written as , replacing , by and , respectively, (30) is proved by . □

Proof of Theorem 1

According to (33) and Lemma 8, one gets Similarly, From (14), (39) and (40), one can see that Since , , , we see by (A4) that is bounded by In the view of and Lemma 6, Thus Theorem 1 follows immediately from (41) and (42). □

Proof of Theorem 2

Consider the model (11), without loss of generality, assume that the true parameter is equal to 0. For any , write By the monotonicity of ψ, Schwarz’s inequality and (A2), we get, for sufficiently large n, By Lemma 9, Consequently, (9) is proved. As mentioned in Chen et al. [21], in order to prove (10), it is desired to prove that Actually, by the monotonicity of , and the assumption (8), applying Lemma 7 and Lemma 2, we get On the other hand, since , This completes the proof of Theorem 2. □

Simulation

We evaluate the parameter estimates and the M-test for the powers by Monte Carlo techniques. Under the null hypothesis, the estimators of regression coefficients and redundancy parameters are derived by some M-methods such as LS method, LAD method and Huber method. Under the local alternative hypothesis, the powers of the M-test is obtained with the rejection region given by Theorem 1. In this section, the case of the NSD sequence is raised as follows: where and are positive sequences, and are negatively dependent (correspond to ) random variables with the distribution Now, we will prove that is a NSD sequence. Obviously, one may easily to check that As stated in Hu [10], the superaddictivity of ϕ is equivalent to , , if the function ϕ has continuous second partial derivatives. In which, can be chosen as a superadditive function. Note that the have same marginal distribution with for each t, by Jensen’s inequality, the sequence is proved to be NSD since Throughout the simulations, the Huber function is taken to be , . The explanatory variables are generated from two random models and all of the simulations are run for 1,000 replicates and calculate the averages of the derived estimates to avoid the randomness impact. The linear model with NSD errors is given by , , , where the NSD errors are assumed to follow a multivariate mixture of normal distribution with joint distribution , . The null hypothesis is . The sample size is taken to be , , . The joint distribution is taken to be . The explanatory variables are generated by the following two random models: I. , ; II. , , where u obeys a standard uniform distribution . Firstly, we generate a NSD sequence by the Gibbs sampling technique. Figure 1 shows the fitted distribution (full line) of NSD is close to the normal distribution, relatively speaking, the NSD distribution tends to behave a truncated distribution feature.
Figure 1

Histograms and fitted distributions of M-estimates residuals with different explanatory variables and M-methods (sample size is ).

Histograms and fitted distributions of M-estimates residuals with different explanatory variables and M-methods (sample size is ). Next, we evaluate the estimators of regression coefficients and redundancy parameters under the null hypothesis, Table 1 illustrates that the M-methods are valid (the corresponded M-estimates are close to true parameters , ) and the estimators of redundancy parameters are effective (one may easily to check that and when the convex function is taken to LS function, for other estimates, although their values are different based on different methods, the sign and significance remain the same, so the general conclusions remain the same). Additionally, with the increasing sample size, the estimations are more and more accurate. In fact, the estimations behave well though the sample size is not large (). As excepted, the fitted residual densities are close to the assumed NSD errors in Figure 2, and all of them still show a truncated distribution feature. Figure 3 checks the residuals are NSD by using the empirical distribution to approximate the distribution function, which supports the NSD errors assumption.
Table 1

The evaluations of regression coefficients and redundancy parameters

Estimates n LS LAD Huber
I II I II I II
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\hat{\beta}_{0}$\end{document}βˆ0 1001.0310.9851.0160.9781.0061.007
5000.9940.9941.0081.0061.0031.006
10001.0020.9991.0001.0061.0021.003
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\hat{\beta}_{1}$\end{document}βˆ1 1001.9832.0081.9922.0162.1312.131
5002.0032.0111.9972.0031.9961.992
10001.9971.9971.9991.9981.9961.994
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\hat{\sigma}^{2}_{n}$\end{document}σˆn2 10012.76412.6710.9840.9879.0959.100
50012.96512.9410.9970.9979.1939.206
100012.96712.9560.9980.9989.2089.291
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$\hat{\lambda}_{n}$\end{document}λˆn 1001.0001.0000.2820.2820.8250.825
5001.0001.0000.2410.2410.8220.823
10001.0001.0000.2330.2340.8220.821
Figure 2

Histograms and fitted distributions of M-estimates residuals with different explanatory variables and M-methods (sample size is ).

Figure 3

A comparison fitted distribution functions of residuals and assumed NSD errors (sample size is ).

Histograms and fitted distributions of M-estimates residuals with different explanatory variables and M-methods (sample size is ). A comparison fitted distribution functions of residuals and assumed NSD errors (sample size is ). The evaluations of regression coefficients and redundancy parameters Finally, we study the empirical significant levels and the powers of M-test. Under the local hypothesis, has an asymptotic central chi-squared distribution with two degrees of freedom by Theorem 1 and Theorem 2, we may reject the null hypothesis if the simulative value in (7). Table 2 presents the powers at significance levels and for various choices of M-methods, explanatory variables and different sample sizes , , . The result represents that the empirical significant levels are close to the nominal levels, consequently, the M-test is valid. Figure 4 illustrates that can approximate the central well by comparing the empirical distributions of with , which implies that the M-test is valid under the local alternatives.
Table 2

The powers of the M-test with NSD errors, ‘∗’ is for the nominal significant levels

n Significance levels LS LAD Huber
I II I II I II
1000.05 0.0630.0680.0820.0790.0620.062
0.01 0.0130.0110.0280.0190.0130.016
5000.05 0.0590.0570.0640.0520.0540.059
0.01 0.0090.0130.0200.0130.0090.012
10000.05 0.0560.0560.0620.0520.0480.057
0.01 0.0120.0150.0130.0110.0100.013
Figure 4

A comparison fitted distribution functions of and the central chi-squared distribution with two degrees (sample size is ).

A comparison fitted distribution functions of and the central chi-squared distribution with two degrees (sample size is ). The powers of the M-test with NSD errors, ‘∗’ is for the nominal significant levels

Conclusions

The results presented here generalize conclusions in [20-22]. In the simulations it turns out that the M-tests for the linear model with NSD errors are insensitive to different choices of M-methods and explanatory variables, therefore it shows robustness, which illustrates that the M-test is effective.
  1 in total

1.  Tests of significance in multivariate analysis.

Authors:  C R RAO
Journal:  Biometrika       Date:  1948-05       Impact factor: 2.445

  1 in total
  1 in total

1.  A difference-based approach in the partially linear model with dependent errors.

Authors:  Zhen Zeng; Xiangdong Liu
Journal:  J Inequal Appl       Date:  2018-10-01       Impact factor: 2.491

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.