Literature DB >> 35768449

Robust model selection using the out-of-bag bootstrap in linear regression.

Fazli Rabbi1, Alamgir Khalil1, Ilyas Khan2, Muqrin A Almuqrin3, Umair Khalil4, Mulugeta Andualem5.   

Abstract

Outlying observations have a large influence on the linear model selection process. In this article, we present a novel approach to robust model selection in linear regression to accommodate the situations where outliers are present in the data. The model selection criterion is based on two components, the robust conditional expected prediction loss, and a robust goodness-of-fit with a penalty term. We estimate the conditional expected prediction loss by using the out-of-bag stratified bootstrap approach. In the presence of outliers, the stratified bootstrap ensures that we obtain bootstrap samples that are similar to the original sample data. Furthermore, to control the undue effect of outliers, we use the robust MM-estimator and a bounded loss function in the proposed criterion. Specifically, we observe that instead of minimizing the penalized loss function or the conditional expected prediction loss separately, it is better to minimize them simultaneously. The simulation and real-data based studies confirm the consistent and satisfactory behavior of our bootstrap model selection procedure in the presence of response outliers and covariate outliers.
© 2022. The Author(s).

Entities:  

Mesh:

Year:  2022        PMID: 35768449      PMCID: PMC9243146          DOI: 10.1038/s41598-022-14398-1

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

A variety of models are used in statistical modeling. Often the focus is to identify the single best model, which describes the data well while being parsimonious. The model selection procedure involves fitting a set of competing models and then selecting the best model by comparing the values of their goodness-of-fit statistics, their prediction loss, or both of these two. Several studies on model selection procedures have concluded that these methods depend on maximum likelihood-type or least squares approaches[1-6] and are possibly affected by the presence of outlying observations in the data. Robust model selection methods aim to work well in situations when some of the observations are outliers and/or the error distribution is not normal. Several robust model selection procedures have been proposed in the literature. To cope with these problems in model selection, different approaches are proposed. Some of them are based on robust modifications of well-known standard criteria such as Akaike information criterion or Mallows’ criterion, or on various resampling techniques, like bootstrap or cross-validation[7-18]. The main objective of this research work is to propose a modified version of[19] for model selection in the presence of outliers. Suppose that we have a column vector of responses and is an design matrix. Let denote any subset of size from , and let is an matrix. Let denote the th row vector of the design matrix . Then the linear regression model corresponding to model α is given bywhere and are independent, and the errors are assumed to have location 0 and scale 1, is an unknown -vector of regression coefficients. Let represent a collection of candidate models. The interest here is to select a model from based on the specified properties of the corresponding fit. To fit the linear regression model, the MM-estimator of[20] was adopted, which combines excellent robustness properties along with high efficiency in the absence of outliers in the data. In model selection, three aspects are generally considered i.e., specifying an estimator, fitting models by using the specified estimator and finally, the fitted models are compared. Furthermore, for each of the models the approach can be extended by considering various types of estimators like LS-estimator, M-estimator, and MM-estimator etc. The models are indexed by and estimate by the estimator . The following two minimal requirements for a good model are discussed by[19]: it has the capability to fit the sample data and reasonably well, and it has the ability to predict future observations with great accuracy. The ability of a model to fit the sample data and is measured by applying a penalized loss function and the expected prediction loss is used to measure the ability to predict future observations. It has been found in the literature that bootstrapping a robust estimator encounters some difficulties in the presence of outliers. For robust regression, an -out-of- paired bootstrap approach is proposed by[12]. Their study findings revealed that implementing the bootstrap procedure directly to a data set containing outliers, generally, fails due to two reasons: (1) The use of , which is non-robust against outliers, and (2) the bootstrap samples may contain a high proportion of outliers as compared to the original data set. Müller and Welsh[19] addressed both of these issues by using stratified bootstrap with appropriate choice of in the presence of outliers. Their approach ensured that one can obtain bootstrap samples that are similar to the sample data. According to their approach, bootstrap samples are constructed in such a manner that the residuals distribution for each bootstrap sample will reflect the relatively same residuals distribution observed in the original data. Their strategy seems to solve the issue well in practice. Our objective in this paper is to pursue the investigation in[19] and make some refinements, by utilizing the concept of out-of-bag bootstrap to develop a robust model selection criterion which deals with outliers and heavy tailed error distributions. The out-of-bag (OOB) observations are those which are not part of the bootstrap sample. These OOB observations can be used for estimating the prediction error, yielding the so-called OOB error. This type of error is often claimed to be an unbiased estimator for the true error rate[21,22]. The rest of the paper is organized as follows: We discuss the existing robust model selection criteria in “Robust model selection criteria” section. Section “The proposed robust model selection criterion” describes a proposed robust model selection criterion. We show the performance of our modified robust criterion via simulation studies in “Simulation studies” section. We present a data examplein “Data example (Stack loss data)” section and conclude with a short discussion in “Conclusion” section.

Robust model selection criteria

In this section, we discuss the existing robust model selection criteria based on robust expected prediction loss. Consider a vector of responses and the design matrix . The conditional expected prediction loss of a model α for a given non- negative loss function is calculated bywhere is the estimator of , is a vector of future responses at X, independent of y, and is the measure of spread for a given data. Initially, this type of prediction loss was introduced by[5] as a model selection criterion by using a loss function in the least squares regression. To select a model from a set ,[19] proposed the following criterion function Following[5,19] estimated the unknown distribution of the data by usingan m-out-of-n stratified bootstrap procedure, whereas the penalized in-sample term in (3) is estimated directly. The estimated selection criteria functions are given bywhere is the bootstrap estimate of , denotes expectation with respect to the bootstrap distribution and is the number of distinct observations in the bootstrap sample which satisfies the conditions given by The criterion function in (4) was modified by[18] using the following steps: calculate and order the residuals, set the number of strata S at between 3 and 8 depending on the sample size n, set stratum boundaries of the residuals, allocate observations into different strata so that observations in the extreme tail are kept in lower or upper tail strata and other strata comprising the remaining observations, sample rows of (y,X) independently with replacement from each stratum so that total bootstrap sample of size is construct the estimator from data obtained in step (v), calculate the criterion function from n-m observations i.e., observations used to obtain are not included when calculating , repeat the steps (vi) and (vii) independent times and then estimate the modified robust expected prediction loss bywhere is the bootstrap estimate of , denotes expectation with respect to the bootstrap distribution and is the number of distinct observations in the bootstrap sample used to obtain and [-m] means that the observations are excluded from total observations when calculating . Here the focus is on the model that minimizes , or i.e.

The proposed robust model selection criterion

In this section, we propose a robust model selection procedure based on two components, a robust penalized loss function, and a modified robust expected prediction loss. We estimate the penalized in-sample term in the criterion function bywhere denotes a function of sample size . The two restrictions on function are that and as . The two restrictions on δ (n) are imposed to penalize complexity, which expresses a preference for smaller and simpler models. These conditions are satisfied by the choice . We combine (6) and (10) to estimate the robust criterion function bywhere is the bootstrap estimate of , denotes expectation with respect to the bootstrap distribution and is the number of distinct observations in the bootstrap sample. An important issue is “how large should the number of bootstrap replications K in our proposed criterion. There is no hard and fast rule for the number of bootstrap replications. However, for estimation of standard error, it is usually in the range of 25–250[23]. The first term in criterion function (11) measures the relationship between the observed sample data and ; the second term penalizes complexity (i.e., preference for smaller models), while the ability to predict future observations is measured by the last term. To use (11), we have to specify and . The robustness viewpoint is adopted for the purpose of fitting the core of the data and predicting core observations, rather than fitting and predicting the tails having atypical observations. So a bounded function is selected. Here, trimming is preferred, so that for sufficiently large |x| the function is constant. As in[11,14,18,19], the simplest function is given bywhich is quadratic near the origin and becomes constant when it is away from the origin. As in[19], we use  = 2. To measure spread we use the full model , because for residuals spread, a large model can produce a valid measure. For simplicity, we measure by the median absolute deviation (MAD) from the median multiplied by 1.483 and is given bywhere and is the estimator for . Among the models being considered, we select a model that minimizes , i.e. The optimal depends on the true model. As in[14,19], one should use /4 ≤   ≤ /2 for moderate (50 ≤   ≤ 200). If is small, is small and the parameter estimators do not converge for some bootstrap samples; but if is large, may be smaller than a fourth of . Choosing the number of strata S at between 3 and 8, depending on the sample size [24]. The penalized loss function in the proposed criterion function, given in (10), is just like a robust version of AIC proposed by[25,26]. But the main difference is due to the function and the estimator in our criterion. The penalized in-sample term in (11) is similar to the robust version of[3]. Furthermore, for the penalized in-sample term was reduced to[3] criterion.

Simulation studies

To assess and compare the finite sample performance of our proposed method with the existent model selection methods, we carried out two simulation studies, that is, one for contamination free dataset in a simulation setting 1 and the other for the contaminated data set in a simulation setting 2.

Simulation setting 1

The finite-sample performance of our proposed criterion is compared with existing model selection procedures via real dataset and simulated data set.

The Gunst and Mason data

To compare the finite sample performance of our proposed method with the existent model selection methods through the real dataset, we use the following regression modewhere are iid standard normal errors; is the column of 1’s; and the values of and are taken from the solid waste data of[27], as in[5,12,13,18,19]. We compare the estimator [expressed in (13)], with [expressed in (7)], [expressed in (8)], [expressed in (9)] and robust BIC [expressed in (14)], In the zero contamination case, the least squares estimator is used to fit the regression models. The penalty term is used in all simulations. The estimated selection probabilities for ,, and based on the LS estimator and are mentioned in Table 1, whereas the estimated selection probabilities based on the LS estimator and are given in Table 2. The results given in Tables 1 and 2 are based on  = 1000 simulations and  = 100 bootstrap samples for  = 15, 20, 25.
Table 1

Estimated selection probabilities of , , and based on the least squares estimator and .

True \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta ^{T}$$\end{document}βTModelm = 15m = 20m = 25\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{\alpha }_{40}$$\end{document}α40
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{15,40}$$\end{document}α¯15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{15{\mathbf{,}}40}}$$\end{document}α^15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{15,40}$$\end{document}α~15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{15,40}^{*}$$\end{document}α^15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{20,40}$$\end{document}α¯20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{20{\mathbf{,}}40}}$$\end{document}α^20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{20,40}$$\end{document}α~20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{20,40}^{*}$$\end{document}α^20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{25,40}$$\end{document}α¯25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{25{\mathbf{,}}40}}$$\end{document}α^25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{25,40}$$\end{document}α~25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{25,40}^{*}$$\end{document}α^25,40
(2,0,0,4,0)1,4*0.9430.9720.9280.9580.8750.9430.8760.9250.7700.9030.8140.8960.835
1,4,50.0100.0060.0130.0090.0240.0140.0300.0180.0420.0230.0380.0270.046
1,3,40.0190.0100.0210.0110.0500.0140.0440.0230.1000.0380.0750.0390.046
1,2,40.0280.0120.0360.0220.0460.0290.0460.0320.0690.0340.0600.0360.057
1,3,4,50.0000.0000.0010.0000.0010.0000.0010.0010.0050.0010.0040.0010.004
1,2,4,50.0000.0000.0000.0000.0010.0000.0010.0010.0040.0000.0050.0000.009
1,2,3,40.0000.0000.0010.0000.0030.0000.0020.0000.0080.0010.0040.0010.003
1,2,3,4,50.0000.0000.0000.0000.0000.0000.0000.0000.0020.0000.0000.0000.000
(2,0,0,4,8)1,4,5*0.9650.9780.9550.9720.9070.9480.9140.9390.8380.9100.8660.9140.877
1,3,4,50.0130.0070.0210.0100.0430.0190.0390.0250.0770.0410.0630.0400.054
1,2,4,50.0220.0150.0240.0180.0480.0310.0450.0350.0710.0450.0600.0430.063
1,2,3,4,50.0000.0000.0000.0000.0020.0020.0020.0010.0140.0040.0110.0030.006
(2,9,0,4,8)1,4,50.0130.0220.0050.0070.0020.0120.0010.0030.0000.0000.0000.0030.000
1,2,50.0010.0020.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,3,4,50.0010.0030.0010.0010.0040.0050.0010.0030.0040.0050.0010.0030.001
1,2,4,5*0.9760.9660.9790.9840.9560.9660.9580.9710.9160.9420.9300.9490.934
1,2,3,4,50.0090.0070.0150.0080.0380.0170.0400.0230.0800.0440.0690.0450.065
(2,9,6,4,8)1,3,4,50.0710.0970.0270.0490.0150.0320.0060.0120.0080.0180.0020.0080.001
1,2,4,50.0100.0200.0060.0090.0000.0030.0000.0000.0010.0030.0000.0010.000
1,2,3,50.0110.0140.0000.0030.0000.0010.0000.0000.0000.0000.0000.0000.000
1,2,3,4,5*0.9080.8690.9670.9390.9850.9640.9940.9880.9910.9790.9980.9910.999

The (*) indicates the optimal model.

Significant values are in bold.

Table 2

Estimated selection probabilities of , , and based on the least squares estimator and .

True \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta ^{T}$$\end{document}βTModelm = 15m = 20m = 25\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overset{\lower0.5em\hbox{$\smash{\scriptscriptstyle\smile}$}}{\alpha }_{40}$$\end{document}α40
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{15,40}$$\end{document}α¯15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{15{\mathbf{,}}40}}$$\end{document}α^15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{15,40}$$\end{document}α~15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{15,40}^{*}$$\end{document}α^15,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{20,40}$$\end{document}α¯20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{20{\mathbf{,}}40}}$$\end{document}α^20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{20,40}$$\end{document}α~20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{20,40}^{*}$$\end{document}α^20,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{25,40}$$\end{document}α¯25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{25{\mathbf{,}}40}}$$\end{document}α^25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{25,40}$$\end{document}α~25,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{25,40}^{*}$$\end{document}α^25,40
(2,0,0,4,0)1,4*0.8970.9710.8790.9290.8000.9270.8320.9020.6720.8460.7810.8850.839
1,4,50.0280.0100.0280.0210.0530.0210.0450.0280.0770.0460.0560.0320.043
1,3,40.0290.0100.0290.0180.0710.0180.0570.0270.1160.0540.0760.0370.047
1,2,40.0420.0090.0420.0290.0600.0310.0540.0390.0940.0460.0650.0390.055
1,3,4,50.0010.0000.0010.0010.0060.0010.0040.0010.0130.0030.0070.0020.004
1,2,4,50.0020.0000.0020.0020.0050.0020.0060.0020.0110.0020.0090.0040.009
1,2,3,40.0010.0000.0010.0000.0050.0000.0020.0010.0120.0030.0060.0010.003
1,2,3,4,50.0000.0000.0000.0000.0000.0000.0000.0000.0050.0000.0000.0000.000
(2,0,0,4,8)1,4,5*0.9340.9810.9180.9570.8600.9510.8750.9330.7810.8930.8350.9090.878
1,3,4,50.0230.0100.0320.0170.0650.0210.0570.0290.0980.0520.0750.0400.055
1,2,4,50.0430.0090.0500.0260.0680.0280.0610.0380.1000.0490.0770.0500.061
1,2,3,4,50.0000.0000.0000.0000.0070.0000.0070.0000.0210.0060.0130.0010.006
(2,9,0,4,8)1,4,50.0000.0050.0000.0000.0000.0010.0000.0000.0000.0010.0000.0010.002
1,3,4,50.0000.0000.0000.0010.0000.0000.0010.0010.0000.0010.0010.0010.001
1,2,4,5*0.9790.9890.97209820.9320.9720.9330.9680.8750.9390.9100.9510.933
1,2,3,4,50.0210.0060.0270.0170.0680.0270.0660.0310.1250.0590.0890.0470.064
(2,9,6,4,8)1,3,4,50.0080.0360.0010.0070.0020.0100.0010.0050.0000.0050.0010.0010.001
1,2,4,50.0010.0040.0000.0010.0000.0010.0000.0000.0000.0010.0000.0000.000
1,2,3,50.0000.0010.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,2,3,4,5*0.9910.9590.9990.9920.9980.9890.9990.9951.0000.9940.9990.9990.999

The results are based on 1000 Monte Carlo simulations and K = 100 bootstrap replications.

Significant values are in bold.

Estimated selection probabilities of , , and based on the least squares estimator and . The (*) indicates the optimal model. Significant values are in bold. Estimated selection probabilities of , , and based on the least squares estimator and . The results are based on 1000 Monte Carlo simulations and K = 100 bootstrap replications. Significant values are in bold. The simulation resultspresented in Tables 1 and 2 are summarized as follows: The performance of the modified model selection procedure using the least squares estimator is comparable to the existing methods , , and the BIC(). The proposed selection criterion outperforms the existent procedures in both cases, i.e., either using the squared loss function or the robust loss function . For the full model, if bootstrap sample size increases, the estimated selection probabilities also increase. For example, in the case of  = 15, the correct percent is 93.9%, whereas, for  = 25, the correct percent is 99.1%. Moreover, model selection based on the robust loss function is superior to the squared loss function. For instance, in the case when the optimal model has all the predictors, then the modified model selection procedure using the squared loss function selects the optimal model 93.9% of the time, which is less than the 99.2% obtained by using a robust loss function. Furthermore, the modified selection criterion is less dependent on bootstrap sample size as compared to the existent criteria and

Simulated data and model selection consistency

To show model selection consistency and performance of the proposed criterion on simulated data, the following regression model with  = 5 is considered.where is generated from standard normal distribution, the regression variables are generated from and added an intercept column of 1’s to produce the design matrix . To generate the response variable , we use Eq. (15). The true data generating models are: , i.e. the model had one nonzero variable, , i.e. the model had two nonzero variables, , i.e. the model had three nonzero variables and The estimated selection probabilities for , , and are calculated for  = 24 and n = 40,80,120,160, based on  = 1000 simulations with bootstrap replications of  = 50 and are tabulated in Table 3.
Table 3

Selection probabilities of , , and based on LS-estimator and .

True Modelm = 24, n = 40m = 24, n = 80m = 24, n = 120m = 24, n = 160
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,40}$$\end{document}α¯24,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,40}$$\end{document}α^24,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,40}$$\end{document}α~24,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,40}^{*}$$\end{document}α^24,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,80}$$\end{document}α¯24,80\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,80}$$\end{document}α^24,80\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,80}$$\end{document}α~24,80\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,80}^{*}$$\end{document}α^24,80\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,120}$$\end{document}α¯24,120\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,120}$$\end{document}α^24,120\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,120}$$\end{document}α~24,120\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,120}^{*}$$\end{document}α^24,120\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,160}$$\end{document}α¯24,160\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,160}$$\end{document}α^24,160\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,160}$$\end{document}α~24,160\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,160}^{*}$$\end{document}α^24,160
(1,0,0,1,0)10.0000.0020.0020.0020.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,4*0.4050.7310.6720.7850.7520.8860.8510.8970.8930.9530.9230.9480.9530.9780.9540.969
1,4,50.1580.0880.1040.0710.0650.0340.0390.0300.0380.0160.0310.0170.0120.0070.0120.010
1,3,40.1320.0730.0800.0630.0840.0360.0520.0320.0360.0160.0260.0180.0170.0080.0150.010
1,2,40.1720.0870.1020.0690.0710.0400.0460.0360.0290.0140.0180.0150.0160.0070.0190.011
1,3,4,50.0410.0060.0160.0040.0030.0010.0020.0010.0010.0000.0000.0000.0000.0000.0000.000
1,2,4,50.0480.0080.0140.0040.0150.0020.0090.0030.0020.0000.0010.0010.0010.0000.0000.000
1,2,3,40.0250.0040.0100.0020.0100.0010.0010.0010.0010.0010.0010.0010.0010.0000.0000.000
1,2,3,4,50.0190.0010.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
(1,0,0,1,1)1,50.0000.0010.0000.0010.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,40.0000.0010.0000.0010.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,4,5*0.5660.8210.7840.8650.8280.9390.8960.9380.9300.9710.9500.9690.9750.9900.9680.984
1,3,4,50.1750.0800.0920.0640.0790.0250.0490.0260.0350.0140.0270.0160.0110.0070.0160.007
1,2,4,50.2060.0930.1140.0660.0850.0350.0530.0360.0330.0140.0220.0140.0130.0030.0160.009
1,2,3,4,50.0530.0040.0100.0030.0080.0010.0020.0000.0020.0010.0010.0010.0010.0000.0000.000
(1,1,0,1,1)1,2,4,5*0.7790.9210.8940.9380.9220.9710.9530.9730.9650.9830.9750.9820.9910.9940.9870.991
1,2,3,4,50.2210.0790.1060.0620.0780.0290.0470.0270.0350.0170.0250.0180.0090.0060.0130.009

The results are based on L = 1000 MC simulations and K = 50 bootstrap replications.

Significant values are in bold.

Selection probabilities of , , and based on LS-estimator and . The results are based on L = 1000 MC simulations and K = 50 bootstrap replications. Significant values are in bold. From the simulation results presented in Table 3, we see that our proposed criterion is comparatively consistent procedure for model selection in linear regression problems.

Simulation setting 2

Simulated data from uniform distribution

In this subsection, the finite-sample performance of our modified criterion is compared with existing model selection procedures in the presence of outliers. The sample data is generated from the following modelwhere the design matrix X has columns generated as uniform on [− 1, 1]. The following six different error distributions are considered: is [3/8] outliers (i.e., [5/8]from a standard normal and [3/8] from a normal with and ); is [1/4] outliers (i.e., [3/4] from a standard normal and [1/4] from a normal with and ); is [1/8] outliers (i.e., [7/8] from a standard normal and [1/8] from a normal with and ); , the Gaussian distribution with μ = 0 and ; , the Cauchy distribution; , the slash distribution(i.e. where and ) In Table 4, the following possible models are considered:
Table 4

Estimated selection probabilities of , , and based on MM-estimator and LS-estimator.

ErrorsModelMM-estimatorLS-estimator
Simple bootstrapStratified bootstrapSimple bootstrap
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,64}$$\end{document}α¯24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,64}$$\end{document}α^24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,64}$$\end{document}α~24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,64}^{*}$$\end{document}α^24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }^{S8}_{24,64}$$\end{document}α¯24,64S8\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }^{S8}_{24,64}$$\end{document}α^24,64S8\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }^{S8}_{24,64}$$\end{document}α~24,64S8\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }^{*S8}_{24,64}$$\end{document}α^24,64S8\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{24,64}$$\end{document}α¯24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,64}$$\end{document}α^24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{24,64}$$\end{document}α~24,64\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{24,64}^{*}$$\end{document}α^24,64
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{1}$$\end{document}110.3680.3920.2670.2570.0000.0000.0010.0010.7561.0001.0001.000
1,30.0000.0000.0000.0000.0000.0000.0000.0000.2440.0000.0000.000
1,2*0.2900.3620.7220.7380.9160.9680.9970.9970.0000.0000.0000.000
1,2,30.3420.2460.0110.0050.0840.0320.0020.0020.0000.0000.0000.000
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{2}$$\end{document}210.0000.0000.0000.0000.0000.0000.0000.0000.9991.0001.0001.000
1,30.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,2*0.9350.9680.9830.9860.8860.9580.9780.9840.0010.0000.0000.000
1,2,30.0650.0320.0170.0140.01140.0420.0220.0160.0000.0000.0000.000
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{3}$$\end{document}310.0000.0000.0000.0000.0000.0000.0000.0001.0001.0001.0001.000
1,30.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,2*0.8940.9520.9470.9630.8690.9460.9430.9600.0000.0000.0000.000
1,2,30.1060.0480.0530.0370.1310.0540.0570.0400.0000.0000.0000.000
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{4}$$\end{document}410.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,30.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.0000.000
1,2*0.8910.9610.9330.9560.8590.9390.9190.9480.8690.9490.9290.958
1,2,30.1090.0390.0670.0440.1410.0610.0810.0520.1310.0510.0710.042
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{5}$$\end{document}510.0080.0160.0120.0180.0050.0130.0100.0130.7700.8230.8410.866
1,30.0000.0000.0000.0000.0010.0000.0000.0000.0010.0010.0000.000
1,2*0.9540.9740.9610.9730.9290.9660.9520.9690.2270.1750.1590.134
1,2,30.0380.0100.0270.0090.0650.0210.0380.0180.0020.0010.0000.000
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{6}$$\end{document}610.0620.1390.0810.1240.0420.0930.0670.0990.8280.8820.9090.921
1,30.0040.0050.0040.0030.0080.0050.0040.0040.0050.0020.0010.001
1,2*0.8850.8420.8820.8550.8790.8780.8920.8750.1640.1160.0900.078
1,2,30.0490.0140.0330.0180.0710.0240.0370.0220.0030.0000.0000.000

The outputs are based on 1000 MCsimulations and K = 100 bootstrap replications. The is used for all selection criteria.

Significant values are in bold.

Model (1) means, a model with intercept only; Model (1, 2) means a model having intercept and X1; Model (1, 3) means a model having intercept and X2; Model (1, 2, 3) means the full model. Estimated selection probabilities of , , and based on MM-estimator and LS-estimator. The outputs are based on 1000 MCsimulations and K = 100 bootstrap replications. The is used for all selection criteria. Significant values are in bold. Following[19], the MM-estimator of[20] is used to fit the robust regression models. For this purpose, the rlm ( ) function in R is used for estimating the regression parameters. Furthermore, the LS estimates are computed for comparison with MM-estimates. As mentioned by[28], when the proportion of extreme observations in some of the bootstrap samples is higher than in the original sample, then the bootstrap distribution may provide a very poor estimator of the distribution of the MM-estimates. To deal with this numerical instability, we use the stratified bootstrap with equal-sized strata. In this approach, bootstrap samples are constructed so that the distribution of the residuals in each bootstrap sample reflects the one observed in the original data set. The selection probabilities based on  = 1000 simulations with bootstrap replications of K = 100 are given in Table 4. From the simulation results presented in Table 4, it is clear that the modified selection procedure using the robust function and MM-estimator is robust in the presence of highly contaminated data. For example, the percent correct is 73.8% for un-stratified bootstrap, whereas the percent correct is 99.7% for stratified bootstrap under the contaminated normal situation . For all error distributions, the modified robust criterion outperforms the existence criteria. The simulation studies suggest that when errors are non-normal, then using robust regression is superior to using LS, but in the case of normal errors, robust regression is inferior to LS. Furthermore, in the presence of outliers and heavy-tailed error distributions, the modified robust criterion using MM-estimator outperforms LS-estimator by a large margin. For example, under error distribution, for MM-estimator the percent correct is 96.9%, whereas, the percent correct is 13.4% for LS-estimator. These results demonstrate that the modified robust procedure has good robustness characteristics with contaminated normal and heavy-tailed distributions, whereas the LS procedure performs very poorly in both cases. This clearly proves the lack of robustness of the LS procedure in the presence of outliers and heavy-tailed distributions. An excellent amount of improvement is obtained in the bootstrap model selection procedure by using the combined criterion as observed in the above simulation study.

Modified solid waste data of Gunst and Mason

To evaluate the performance of our proposed robust model selection method, we modified the Gunst and Mason data by planting 10% and 20% outliers. The response vector is generated aswhere is the column of 1’s; and the values of and are taken from the solid waste data of[26]. To create high-leverage points, we replace the first four to eight observations of each regressor variable value by 20. The true generating model has two non-zero predictors, i.e. and we choose the following five different error distributions to represent various deviations from normality: is 10% wild (i.e., 90% from a standard normal and 10% from a normal with, ); is 20% wild (i.e., 80% from a standard normal and 20% from a normal with ,) ; .is t(3) (i.e., t-distribution with 3 degrees of freedom); , is standard normal; , is Cauchy distribution with location = 0 and scale = 1. The selection probabilities of , , and on the basis of stratified bootstrap with the MM estimator are computed. The selection probabilities based on  = 1000 simulations with bootstrap replications of K = 50 are given in Table 5.
Table 5

Estimated selection probabilities of , , and based on MM estimator.

ErrorsTrue \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\varvec{\beta}}^{\rm T}$$\end{document}βTModel10% X-outliers20% X-outliers
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{16,40}$$\end{document}α¯16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{16{\mathbf{,}}40}}$$\end{document}α^16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{16,40}$$\end{document}α~16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{16,40}^{*}$$\end{document}α^16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{16,40}$$\end{document}α¯16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${\hat{\mathbf{\alpha }}}_{{16{\mathbf{,}}40}}$$\end{document}α^16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{16,40}$$\end{document}α~16,40\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{16,40}^{*}$$\end{document}α^16,40
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{1}$$\end{document}1(2,0,0,4,8)1,50.0020.0040.0020.0020.0080.0340.0070.018
1,4,5*0.9120.9660.9020.9400.8230.8830.8240.876
1,3,50.0050.0060.0080.0080.0340.0350.0360.033
1,2,50.0020.0040.0030.0030.0110.0160.0130.015
1,3,4,50.0360.0080.0340.0190.0510.0180.0520.027
1,2,4,50.0410.0120.0450.0260.0620.0140.0480.026
1,2,3,4,50.0020.0000.0060.0020.0110.0000.0200.005
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{2}$$\end{document}2(2,0,0,4,8)1,50.0080.0320.0170.0270.0180.0930.0290.056
1,4,5*0.6060.7930.6840.7810.5310.7140.6420.717
1,3,50.0210.0300.0330.0330.0340.0390.0380.046
1,2,50.0130.0280.0160.0210.0190.0170.0210.022
1,3,4,50.0910.0270.0720.0410.1100.0390.0850.046
1,2,4,50.2550.0890.1700.940.2820.0980.1790.109
1,2,3,4,50.0060.0010.0080.0030.0040.0000.0010.001
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{3}$$\end{document}3(2,0,0,4,8)1,50.0040.0210.0030.0070.0260.0770.0270.055
1,40.0000.0000.0000.0000.0000.0010.0010.001
1,4,5*0.8870.9290.8860.9200.8070.8150.8170.824
1,3,50.0150.0220.0190.0210.0520.0530.0570.052
1,2,50.0100.0120.0130.0170.0320.0280.0310.030
1,3,4,50.0450.0120.0420.0180.0420.0130.0300.017
1,2,4,50.0360.0.030.0320.0140.0390.0120.0340.018
1,2,3,4,50.0030.0010.0050.0030.0020.0010.0030.003
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{4}$$\end{document}4(2,0,0,4,8)1,50.0010.0040.0010.0020.0050.0190.0080.015
1,4,5*0.9210.9650.9130.9480.8630.9040.8590.897
1,3,50.0050.0070.0030.0050.0290.0350.0320.032
1,2,50.0030.0050.0050.0040.0120.0120.0090.010
1,3,4,50.0390.0110.0340.0210.0460.0170.0400.022
1,2,4,50.0290.0080.0390.0190.0380.0110.0390.020
1,2,3,4,50.0020.0000.0050.0000.0070.0010.0130.003
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in_{5}$$\end{document}5(2,0,0,4,8)1,50.04501140.0500.0810.1620.2740.1620.251
1.40.0010.0020.0010.0010.0070.0180.0090.011
1,4,5*0.8390.8200.8420.8430.6700.6110.6920.637
1,3,50.0460.0340.0440.0350.0550.0410.0560.046
1,2,50.0200.0220.0190.0220.0450.0420.0370.033
1,3,4,50.0260.0040.0170.0090.0240.0090.0160.008
1,2,4,50.0180.0030.0200.0080.00280.0050.0190.013
1,2,3,50.0020.0000.0030.0000.0070.0000.0070.000
1,2,3,4,50.0030.0010.0020.0010.0020.0000.0020.001

The results are based on 1000 Monte Carlo simulations and K = 50 bootstrap replications.

Significant values are in bold.

Estimated selection probabilities of , , and based on MM estimator. The results are based on 1000 Monte Carlo simulations and K = 50 bootstrap replications. Significant values are in bold. Table 5 demonstrates the simulation results with 10% and 20% of outliers in the covariates and five different error distributions as discussed in the simulation setting. If we look at the results, we see that the performance of our robust procedure is very good for amongst all error distributions while it does not perform very well for in the presence of x-outliers. The selection probabilities for error distribution are similar to that of . Furthermore, the selection probabilities are good for distribution (10% symmetric wild case) as compared to contamination type (20% symmetric wild case). Overall, the selection probabilities for each of the criteria decrease when the percentage of both x- and y-outliers goes up. Moreover, selection probabilities in the presence of response outliers and covariate outliers, the performance of our proposed model selection criterion based on MM-estimation is comparable to the existing criteria even when the contamination level changes from i.e., 10% to 20%.

Data example (Stack loss data)

In this section, we analyze the Stack loss data presented by[29]. This dataset consists of three explanatory variables, and it contains four outliers,namely observations 1, 3, 4, and 21.The response is the Stack loss (y) observed on  = 21 observations. The explanatory variables are theFlow of cooling air (X1), Cooling Water Temperature (X2), and Concentration of acid (X3).We applied our robust method , the existing methods, and the traditional methods on the data. Table 5 presents a summary of selected best models. Table 6 shows the classical methodsselect the full model, whereas robust criteria agreed with the importance of the two variables, X1 and X2.The best model according to our criterion includes X1 and X2.
Table 6

Selected best model for the stack loss data using a range of model selection procedures.

Selected variables\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\overline{\alpha }_{10,21}$$\end{document}α¯10,21\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{10,21}$$\end{document}α^10,21\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\tilde{\alpha }_{10,21}$$\end{document}α~10,21\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\hat{\alpha }_{10,21}^{*}$$\end{document}α^10,21AICBIC
X12.973.504.424.824.614.71
X23.023.594.585.002.192.28
X33.484.205.325.863.213.31
X1, X21.702.553.233.871.551.70
X1, X32.383.424.675.452.692.84
X2, X32.093.493.724.771.471.62
X1, X2, X31.813.053.905.281.341.54

Significant values are in bold.

Selected best model for the stack loss data using a range of model selection procedures. Significant values are in bold.

Conclusion

In this article, we have presented a novel procedure for robust model selection in linear regression. The criterion is a modification to the bootstrap model selection method based on robust estimator proposed by[19]. The simulation results reveal that the performance of model selection is improved when using the OOB error in the present studies. Moreover, the undue effect of outliers is controlled by using both a robust MM-estimator and a bounded loss function in the proposed criterion. The proposed model selection criterion can maintain their robust properties in the presence of response outliers and covariate outliers. The proposed criterion is compared with other robust model selection criteria described in previous literature. We observed that in the presence of outliers and heavy-tailed error distributions, the MM-estimator outperformed the least squares estimator by a large margin. This clearly proved the lack of robustness of the least squares procedure in the presence of outliers and in heavy-tailed distributions. Furthermore, when errors are non-normal, then robust regression is found superior to least squares, but in the case of normal errors, robust regression is found inferior to least squares. From simulation-based and real-data based results, we conclude that our modified robust model selection procedure is consistent and works well in situations where outliers are present in the data. As observed in our simulation study, an excellent amount of improvement is gained by minimizing the combined criterion, rather than minimizing the penalized loss function or the modified conditional expected prediction loss separately. Furthermore, our robust model selection criterion will perform better when the data generating model is small. Supplementary Information.
  1 in total

1.  Analysis of the Stage Performance Effect of Environmental Protection Music and Dance Drama Based on Artificial Intelligence Technology.

Authors:  Li Zeng
Journal:  J Environ Public Health       Date:  2022-09-19
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.