Literature DB >> 26290569

Genomic-Enabled Prediction of Ordinal Data with Bayesian Logistic Ordinal Regression.

Osval A Montesinos-López1, Abelardo Montesinos-López2, José Crossa3, Juan Burgueño4, Kent Eskridge5.   

Abstract

Most genomic-enabled prediction models developed so far assume that the response variable is continuous and normally distributed. The exception is the probit model, developed for ordered categorical phenotypes. In statistical applications, because of the easy implementation of the Bayesian probit ordinal regression (BPOR) model, Bayesian logistic ordinal regression (BLOR) is implemented rarely in the context of genomic-enabled prediction [sample size (n) is much smaller than the number of parameters (p)]. For this reason, in this paper we propose a BLOR model using the Pólya-Gamma data augmentation approach that produces a Gibbs sampler with similar full conditional distributions of the BPOR model and with the advantage that the BPOR model is a particular case of the BLOR model. We evaluated the proposed model by using simulation and two real data sets. Results indicate that our BLOR model is a good alternative for analyzing ordinal data in the context of genomic-enabled prediction with the probit or logit link.
Copyright © 2015 Montesinos-López et al.

Entities:  

Keywords:  Bayesian ordinal regression; GenPred; Gibbs sampler; genomic selection; logit; probit; shared data resource

Mesh:

Year:  2015        PMID: 26290569      PMCID: PMC4592994          DOI: 10.1534/g3.115.021154

Source DB:  PubMed          Journal:  G3 (Bethesda)        ISSN: 2160-1836            Impact factor:   3.154


Genomic-enabled prediction models are revolutionizing animal and plant breeding. There is some evidence that they are powerful for predicting the genomic merit of animals and plants based on high-density single-nucleotide polymorphism (SNP) marker panels and are being recommended increasingly for genomic prediction in human health (Yang and Tempelman 2012). However, most genomic-enabled prediction models assume a continuous and normally distributed phenotype. Because often this assumption is not fulfilled, researchers normally approach phenotypes in three ways: (a) they ignore the lack of normality in the phenotypes; (b) they transform the non-normal phenotype to approximate it to normality; or (c) they use generalized linear mixed models (GLMMs) to model the appropriate distribution of the phenotype (Stroup 2015). The use of the first approach is justified for large sample sizes with the central limit theorem, which states that treatment means have an approximate normal distribution if the sample size is large enough. However, there is a lot of evidence indicating that the first approach produces highly biased results for small- and moderate-sample sizes (Stroup 2012, 2015). Transformations introduced by Bartlett (1947) for non-normal data were proposed for variance-stabilization to fulfill the assumption of homogeneous variance; they are still considered standard procedures in many agricultural disciplines. Implementation with transformations is equal to that for phenotypes normally distributed (based on the linear model). However, there is mounting evidence that transformations do more harm than good for the models required by most agricultural research, because the use of the linear model with or without transformed data produces a great loss of accuracy and power (Stroup 2015), mostly in small sample sizes. Nelder and Wedderburn (1972) introduced generalized linear models, a major departure from the usual approach to non-normal data. GLMMs extend the linear model theory to accommodate non-normal data with heterogeneous variance and even correlated observations. Viewed through the GLMM lens, the pre-1990s understanding of non-normal data––still pervasive in the agricultural research community––is antiquated at best, obsolete at worst. Today, small sample investigations are providing an increasing body of evidence that GLMMs work as well in practice as they do in theory because they produce more accuracy and power than approaches (a) and (b). Also, now there are textbooks and software available for the implementation of GLMMs, although the implementation of approaches (a) and (b) is still dominant in agricultural research (Stroup 2015). In genomic-enabled prediction, the use of GLMMs is still in its early stages because their implementation is not straightforward given that the number of observations (n) is usually smaller than the number of covariates (p) included in the model. In addition, a complex dependence structure is observed among covariates (markers) and observations (lines) due to the joint involvement of biological processes and pathways. To overcome this situation in the pregenomic era, Gianola (1980, 1982) and Gianola and Foulley (1983) proposed a probit (threshold) model for dealing with ordinal categorical traits in animal breeding. This probit model was extended to deal with p >> n in the genomic era by González-Recio and Forni (2011) and Villanueva for binary trials, and by Wang and Montesinos-López for more than two ordinal categories. Also, BGLR (Bayesian generalized linear regression), software developed for genomic-enabled prediction that is able to deal with normal, binary, ordinal, and censored data (de los Campos and Perez-Rodriguez 2013; Perez-Rodriguez and de los Campos 2014), is now available. However, no GLMMs are available for genomic-enabled prediction for counts and percentage phenotypes. For modeling ordinal categorical phenotypes, the ordinal logistic regression model is often preferred over the ordinal probit model in statistical applications, because it provides regression coefficients that are more interpretable due to their connection to odds ratios (Zucknick and Richardson 2014). However, in genomic-enabled prediction (when p >> n), only the Bayesian probit model is frequently implemented, given that Bayesian methods that introduce sparseness through additional priors on the model size are very well-suited to this problem. Therefore, because of the lack of a Bayesian logistic ordinal model analogous to the Bayesian probit ordinal model that uses a data augmentation approach, the logistic model is not practical for genomic selection. Both logistic and normal distributions are symmetric with a basic, unimodal “bell curve” shape. The only difference is that the logistic distribution has a somewhat heavier tails, which means that it is less sensitive to outlying data (and hence somewhat more robust for modeling mis-specifications or erroneous data). This is another advantage of logistic regression over probit regression. Because of its easy implementation, the use of BPOR is extremely common, even though it is less robust for modeling mis-specifications and its coefficients are less interpretable. Because of the aforementioned properties of the logistic model, some researchers have proposed approximations to logit regression. For example, Bartholomew and Knott (1999) proposed , where Φ is the cumulative density function for the standard normal distribution and , whereas Camilli (1994) proposed using k = 1.702, obtained by minimizing the maximum distance between two cumulative distribution functions. Although Amemiya (1981) proposed a value of k = 1.6 and computed tables for representative values of the density function for different values of k, he did not explain why he used . More recently, Savalei (2006) obtained a value of k = 1.75 based on minimizing the Kullback-Leibler information. However, although some of these approximations do a reasonable job of approximating the logistic distribution, they are only approximations, and it goes without saying that an exact solution is preferred. In this paper, we propose a Bayesian logistic ordinal regression (BLOR) model for genomic-enabled prediction by using a data augmentation approach. We illustrate our proposed method with simulation and real data. We compare the BLOR with the Bayesian probit ordinal regression (BPOR) model with and without approximation.

Materials and Methods

Gray leaf spot (GLS) and Septoria data sets

GLS is one of the most important foliar diseases of maize worldwide. The GLS data set is composed of 278 maize lines; the ordinal trait measured in each line was GLS [1 (no disease), 2 (low infection), 3 (moderate infection), 4 (high infection), 5 (complete infection)] caused by the fungus Cercospora zeae-maydis, evaluated in three environments (México, Zimbabwe, and Colombia). These data are part of a data set previously analyzed by Crossa , González-Camacho , and Montesinos-López . Genotypes of all 278 lines were obtained using the 55k SNP Illumina platform. SNPs with >10% missing values or a minor allele frequency of ≤ 0.05 were excluded from the data. After line-specific quality control (applying the same quality control to each line separately), the maize data still contained 46,347 SNPs. On the other hand, the Septoria data set contains 268 wheat lines planted in Toluca, México, in 2010, and the trait (Septoria scores) was measured using an ordinal four-point scale. Genotypes of these lines were obtained with 45,000 genotype by sequencing (GBS), following the protocol of Poland . We kept only 13,913 GBS that had <50% missing data; after filtering for minor allele frequency, we ended up with 6787 GBS that were used in the analysis. For the implementation of the proposed model, we formed five data sets from these two real data sets (GLS and Septoria), four from the GLS data set and one from the Septoria data set. The first three data sets formed from GLS correspond to each environment in which they were evaluated for GLS; the last one was formed by pooling the data from the three environments (information from the three environments without taking into account the environments as covariates).

Bayesian logistic ordinal regression

Let (where i represents the genotype and j denotes the number of replicates or experimental units of each genotype. The total number of observations is In other words, the observed vector contains elements, and the n-dimensional vector y of all responses can be written as . The response variable represents an assignment into one of C mutually exclusive and exhaustive categories that follow an order. Therefore, the ordinal logistic regression model can be written in terms of a latent response variable as follows: where are called “liabilities”, , where denotes the logistic distribution, and the vectors (p × 1) are explanatory variables associated with the fixed effects . The random effect . In genomic-enabled prediction, Since are unobservable and can be measured indirectly by an observable ordinal variable , then can be defined by:This means that is divided by thresholds into C intervals, corresponding to C ordered categories. The first threshold, , defines the upper bound of the interval corresponding to observed outcome 1. Similarly, threshold defines the lower bound of the interval corresponding to observed outcome C. Threshold defines the boundary between the interval corresponding to observed outcomes c − 1 and c for (c = 1,2,…, C − 1). Threshold parameters are with , and Assuming that the error term of the latent response is distributed as , the cumulative response probability for the c category of the ordinal outcome is:for .Similarly, model (2) can be written as a cumulative logit model:This GLMM model is described by: (1) two distributions, one for observations in the response variable Multinomial ), where is the p × 1 vector of fixed effects, and another for the random effects, or , where b is the effect of line i; (2) linear predictor , where denotes the cth link (c = 1,2,…, C − 1) for the fixed and random effects combination, is the intercept (threshold) for the cth link, and are known row incidence vectors corresponding to fixed effects in . Because there are C categories, a total of C − 1 link functions are required to fully specify the model; and (3) link function: cumulative logit {, (c = 1,2,…, C − 1)}. Using the inverse link for this model, we can calculate as follows:Since we have latent variables distributed as and we observe if, and only if, , then the joint posterior density of the parameter vector and latent variable becomesLet’s assume a scaled independent inverse chi-square prior for , a normal prior distribution for , a normal prior distribution for , and also a prior for (Gianola 2013). Following Sorensen , the prior for the C − 1 unknown thresholds has been given as order statistics from U(, distribution,where . The fully conditional posterior distributions are provided below and details of all derivations are given in Appendix A.

Liabilities and Pólya-Gamma values

The fully conditional posterior distribution of liability is a truncated normal distribution and its density is For simplicity, ELSE is the data and the parameters, except the one in question. is a normal density with parameters as indicated in the argument, Φ is the cumulative distribution function of a normal density with mean and variance , and the fully conditional posterior distribution of ω is

Regression coefficients (β)

The fully conditional posterior of is as follows:where , . With , , , , , , . It is important to point out that if we use a prior for (improper uniform distribution), then in and we need to make 0 the term .

Polygenic effects (b)

Now the fully conditional posterior of is given as

Variance of polygenic effects

Next, the fully conditional posterior of is

Threshold effects ()

The density of the fully conditional posterior distribution of the cth threshold, , is

Variance of regression coefficients

The fully conditional posterior of is

The Gibbs sampler

The Gibbs sampler is implemented by sampling repeatedly from the following loop: Sample liabilities from the truncated normal distribution in (3). Sample values from the Pólya-Gamma distribution in (4). Sample the regression coefficients from the normal distribution in (5). Sample the polygenic effects from the normal distribution in (6). Sample the variance effect ( from the scaled inverted χ2 distribution in (7). Sample the thresholds from the uniform distribution in (8). Sample the variance of regression coefficients ( from the scaled inverted χ2 distribution in (9). Return to step 1 or terminate if chain length is adequate to meet convergence diagnostics. In the absence of polygenic effects (b), the aforementioned Gibbs sampler can be used only by ignoring steps 4 and 5. If all marker effects are taken into account in the design matrix, , with a prior for the beta regression coefficients, we end up with a threshold Bayesian ridge regression. This is a version for ordinal categorical data of the ridge estimator of Hoerl and Kennard (1970), since the posterior expectation of is equal to with pseudo-response . Another important point is that by setting each , the aforementioned Gibbs sampler for the BLOR with the logistic link is reduced to the Gibbs sampler for the BPOR with the probit link proposed by Albert and Chib (1993). This implies that the proposed BLOR model is more general and includes the Gibbs sampler for the BPOR model as a particular case.

Simulation study

The purpose of this simulation study is twofold: (1) to compare the performance of the proposed BLOR with: (a) the approximation resulting from using the estimates of the BPOR with , denoted as BLOR*, (b) with the results of using maximum likelihood estimators (MLEs) with logit link for ordinal data (MLLOR), and (c) the approximation resulting from multiplying the MLEs with probit link for ordinal data by 1.75, denoted as MLLOR*; (2) to evaluate the performance of the BLOR in the presence of outliers. We used the value 1.75 because, according to the literature review, it is the most reasonable value (Savalei 2006). To reach these two goals, two data sets were simulated. Both simulation studies were carried out with the following liability:where and , and the vectors have been drawn independently, with components following a uniform distribution within the interval [−0.1, 0.1]. The threshold parameters used were γ1= −0.8416, γ2= −0.2533, γ3= −0.2533, and γ4= −0.8416. Then the response variable was generated as follows: For simulated data set 1, we used four values of sample size n = 5, 10, 20, and 40 and all the were drawn independently from a L(0,1), whereas for simulated data set 2, we used only one sample size (n = 40) and the error terms () were obtained from two distributions [L(0,1) and a student's t distribution with four degrees of freedom, denoted as t4]. We studied four scenarios: Scenario 1, the percentage of outliers (PO) from the t4 was 5% and the remaining percentage was obtained from the L(0,1) distribution; Scenario 2: the PO from the t4 was 10%, and 90% from the L(0,1); Scenario 3: the PO from the t4 was 20%, and 80% from the L(0,1); and Scenario 4: the PO from the t4 was 30%, and 70% from the L(0,1). The MLE estimates for the ordinal regression were obtained using the polr function of the MASS package in R (R Core Team 2015). It is important to point out that the priors used for the Bayesian methods were not informative for , and for the hyperparameters for thresholds, we used and . We computed 20,000 Markov chain Monte Carlo (MCMC) samples. Bayes estimates were computed using 10,000 samples because the first 10,000 were discarded as burn-in.

Model implementation

The Gibbs sampler described previously for the BLOR model was implemented with the R-software (R Core Team 2015). Implementation was done with a Bayesian approach and MCMC through the Gibbs sampler algorithm, which samples sequentially from the full conditional distributions until it reaches a stationary process, converging with the joint posterior distribution (Gelfand and Smith 1990). In the real data, to reduce the potential impact of MCMC errors on prediction accuracy, we performed a total of 60,000 iterations with a burn-in of 20,000, so that 40,000 samples were used for inference. We did not apply thinning of the chains, following the suggestions of Geyer (1992), Maceachern and Berliner (1994), and Link and Eaton (2012), who provide justification for the ban on subsampling MCMC output for approximating simple features of the target distribution (e.g., means, variances and percentiles), since thinning is neither necessary nor desirable, and unthinned chains are more precise. It is important to point out that implementation of the BLOR model for the real data sets was done using the hyperparameters , , γmin = −1000 and γmax= −1000 for thresholds parameters, all of which were chosen to lead weakly informative priors.

Assessing prediction accuracy

We used cross-validation to estimate the prediction accuracy of the proposed models. The data set was divided into training and validation sets 10 times, with 90% of the data set used for training and 10% for testing; this was done only for the pooled data. The training set was used to fit the model and the validation set was used to evaluate the prediction accuracy of the proposed models. Since the phenotype response is ordinal categorical, we used the Brier score (Brier 1950) to measure prediction ability, which is equal towhere BS denotes the Brier Score, and d takes a value of 1 if the ordinal categorical response observed for individual i falls into category c; otherwise, d = 0. This scoring rule uses all the information contained in the predictive distribution, not just a small portion like the hit rate or the log-likelihood score. Therefore, it is a reasonable choice for comparing categorical regression models, although there are other scoring rules that also have good properties. The range of BS in equation (10) is between 0 and 2. For this reason, we divided Brier scores (BS)/2 to get the BS bounded between 0 and 1; lower scores imply better predictions. It is important to point out that we also used the BS when analyzing the full data sets. We also used the deviance information criterion (DIC) to compare Bayesian models, as suggested by Gelman ; here, the lower the DIC, the better the model.

Data availability

The two real data sets and two simulated data sets together with R codes are deposited in the link http://hdl.handle.net/11529/10254. The phenotypic data for GLS in three environments (México, Zimbabwe, and Colombia) for the 278 maize lines, the 46,347 SNPs, and the R scripts developed to fit the predictive models used in this study are given in the files PhenoGLS.RData, and MarkersGLSFinal.RDat. The repository also contains the Septoria genotypic and phenotypic data sets, in files SeptoriaGenotypic.RDat, and SeptoriaPhenotypic. RDat, respectively. The R codes to generate the simulated data sets 1 and 2, and for analyzing the real data set from Colombia are directly given in Appendices B, C, and D, respectively.

Results

In the following sections, we investigate the performance of the proposed BLOR estimator through a simulation study and with real data.

Simulated data set

In Table 1, we report average estimates obtained by all methods, along with SDs; all the results are based on 50 replications. From Table 1, it is clear that as the sample size increases, the average biases and SD decrease in all cases. This confirmed the consistency properties of all the estimates. Table 1 also shows that, in general, the point estimates of the Bayesian estimates (BLOR and BLOR*) are similar to the MLEs (MLLOR, and MLLOR*, which was approximated with the MLE with probit ordinal regression); however, the approximations (BLOR* and MLLOR*) have greater bias and SD. BLOR has less bias and SD in most of the studied parameters, producing better parameter estimates than the MLLOR (which is the correct method that use maximum likelihood with the logit link function). For this reason, the proposed BLOR is an excellent alternative. However, a more in-depth simulation study is required to ensure that these findings are valid for all possible scenarios.
Table 1

Simulated data set 1: Average values (Mean) and SD of MLEs and the Bayesian estimators, with four sample sizes (n)

ni 
ParameterTrue ValueBLORBLOR*MLLORMLLOR*
MeanSDMeanSDMeanSDMeanSD
β1−6-6.1411.935−6.7112.117−6.3802.489−6.8262.668
β2−5-4.9572.262−5.5462.726−5.5962.731−5.9272.872
β377.5502.7467.8153.1126.6592.6987.1102.885
5γ1−0.842-0.8510.190−0.9370.152−0.8830.177−0.9420.185
γ2−0.253-0.2540.154−0.2710.142−0.2770.167−0.2980.179
γ30.2530.2740.1700.3280.1710.2620.2030.2810.218
γ40.8420.8780.1710.9670.1510.8630.2110.9200.220
β1−6−6.2241.562−6.5341.650-6.1181.673−6.4801.767
β2−5-4.9871.825−5.4331.901−4.7171.500−5.0221.619
β377.3061.9717.7621.8256.6061.8367.0381.939
10γ1−0.842-0.8430.100−0.9260.147−0.8470.127−0.9070.135
γ2−0.253-0.2390.097−0.2840.131−0.2730.110−0.2960.119
γ30.2530.2760.1130.2720.1230.2330.1100.2490.120
γ40.8420.8610.1160.9200.1240.8410.1150.8970.123
β1−6−6.1221.063−6.2781.390-6.0300.936−6.4221.017
β2−5-5.0991.262−5.5381.103−5.1810.962−5.4881.019
β377.2711.1147.4791.3947.2431.1187.6691.183
20γ1−0.842−0.8490.108−0.9170.100-0.8470.073−0.9050.077
γ2−0.253-0.2520.106−0.2910.094−0.2500.069−0.2700.074
γ30.2530.2590.0910.2610.0990.2540.0680.2720.073
γ40.8420.8600.0970.8830.1060.8500.0790.9070.084
β1−6-6.0580.804−6.4670.924−6.1990.719−6.5630.783
β2−5−5.1750.817−5.2310.879−4.8040.791-5.1010.856
β377.1630.8157.6740.9587.0930.8677.5280.904
40γ1−0.842−0.8440.065−0.9110.069-0.8410.051−0.8990.055
γ2−0.253-0.2580.053−0.2780.064−0.2480.05−0.2670.056
γ30.2530.2550.0520.2670.0600.2520.0440.2710.048
γ40.8420.8560.0540.9000.071 0.8350.0470.8930.050

BLOR* and MLLOR* use the parameter estimates of BPOR and MLLOR and approximate BLOR and MLLOR with . The best model has the value for the parameter closer to the true value; these are presented in bold. MLEs, maximum likelihood estimators; BLOR, Bayesian logistic ordinal regression; MLLOR, maximum likelihood logistic ordinal regression.

BLOR* and MLLOR* use the parameter estimates of BPOR and MLLOR and approximate BLOR and MLLOR with . The best model has the value for the parameter closer to the true value; these are presented in bold. MLEs, maximum likelihood estimators; BLOR, Bayesian logistic ordinal regression; MLLOR, maximum likelihood logistic ordinal regression. Table 2 shows that the smaller the PO, the less bias there is in the parameter estimates under the four methods (two Bayesians and two under Maximum Likelihood). However, under the two Bayesian methods, the proposed BLOR showed less bias than BLOR*, and this behavior is observed for all parameters and PO under study, except when the PO is 30% and for the parameter γ2. Under the two maximum likelihood methods, the approximate method (MLLOR*) showed greater bias in all scenarios and parameters. Finally, although BLOR and MLLOR produced better results in term of bias, in most cases, MLLOR was better than BLOR.
Table 2

Simulated data set 2: average values (Mean) and SD of MLEs and Bayesian estimators, with four POs


PO 
Parameter 
True ValueBLORBLOR*MLLORMLLOR*
MeanSDMeanSDMeanSDMeanSD
β1−6-5.9940.824−6.5250.915−6.2300.723−6.6080.766
β2−5−5.0880.880−5.6780.864-5.0150.763−5.3240.815
β377.1830.8027.6070.8937.1110.6517.5350.689
5γ1−0.842−0.8620.066−0.9230.067-0.8530.050−0.9100.053
γ2−0.253−0.2630.071−0.2810.059-0.2560.048−0.2760.051
γ30.2530.2580.0680.2850.0510.2610.0400.2810.043
γ40.8420.8610.0670.9280.0580.8560.0410.9140.043
β1−6-6.1560.854−6.6700.979-6.1360.607−6.5040.647
β2−5−5.3590.899−5.5441.094-5.1490.610−5.4590.639
β377.3490.8307.7390.8567.1090.6457.5360.682
10γ1−0.842−0.8830.063−0.9250.076-0.8660.048−0.9240.050
γ2−0.253−0.2660.070−0.2670.072-0.2650.052−0.2850.056
γ30.2530.2600.0710.3130.0650.2640.0510.2840.055
γ40.8420.8680.0680.9640.0820.8610.0520.9180.055
β1−6−6.5290.730−6.8280.903-6.3450.645−6.7090.689
β2−5-5.2440.759−5.7220.877−5.2750.629−5.5890.671
β377.5250.7847.8600.7587.3440.7357.7800.772
20γ1−0.842−0.9150.065−0.9720.060-0.8830.049−0.9420.053
γ2−0.253−0.2750.059−0.2950.057-0.2600.044−0.2800.048
γ30.2530.2790.0630.2970.0590.2710.0470.2910.051
γ40.8420.9220.0600.9770.0600.8950.0530.9540.057
β1−6−6.7940.803−7.0111.013-6.5550.654−6.9160.691
β2−5−5.6520.754−5.8270.754-5.2980.634−5.5900.666
β378.0650.8948.3510.8817.4910.7527.9250.798
30γ1−0.842−0.9720.071−1.0040.075-0.8980.041−0.9560.044
γ2−0.253−0.3010.060−0.2960.072-0.2690.044−0.2890.047
γ30.2530.2790.0670.3180.0690.2860.0440.3060.047
γ40.8420.9440.0681.0230.0600.9220.0440.9810.046

The outliers were generated with a student's t distribution with four degrees of freedom. BLOR* and MLLOR* use the parameter estimates of BPOR and MLLOR and approximate BLOR and MLLOR with . The best model has the value for the parameter closer to the true value; these are presented in bold. MLEs, maximum likelihood estimators; POs, percentages of outliers; BLOR, Bayesian logistic ordinal regression; MLLOR, maximum likelihood logistic ordinal regression.

The outliers were generated with a student's t distribution with four degrees of freedom. BLOR* and MLLOR* use the parameter estimates of BPOR and MLLOR and approximate BLOR and MLLOR with . The best model has the value for the parameter closer to the true value; these are presented in bold. MLEs, maximum likelihood estimators; POs, percentages of outliers; BLOR, Bayesian logistic ordinal regression; MLLOR, maximum likelihood logistic ordinal regression.

Real data sets

Next we compared BLOR and BPOR with the real data set (GLS and Septoria data) described in the Materials and Methods. However, because with this data set the number of parameters (betas and thresholds) to be estimated is high, we compared both models with point and interval estimates of the probabilities estimated for each category on the four- and five-point ordinal scale for each data set studied. In Table 3, we compare BLOR, BPOR, and BLOR* (this is the approximate method since the estimates resulting from BPOR were used to approximate BLOR with with five data sets (three locations, the resulting pooling of the three locations and the Septoria data set) made up from the data sets described in the section Materials and Methods. First, when comparing BLOR with BLOR*, we see that there are differences in the point probability estimates, and the lower the probabilities, the greater the differences. However, in general, the widths of the credible sets are similar. Second, when comparing BLOR with BPOR, we see that the point and credible sets for each location and Septoria produced similar results; however, for the pooled data, BLOR produced estimates with narrower confidence sets than BPOR (Table 3). We observed that the estimates produced using BPOR are less accurate because wider confidence sets are produced when the data are pooled; this could be because the assumption of errors normally distributed with mean zero and variance one when the data are pooled is not fulfilled. Also, there were no differences in the BS of the three models (BLOR, BLOR*, and BPOR).
Table 3

Real data sets: GLS and Septoria data sets


Model 
Set (DIC) 
 StatisticProbability of Each Category 
BS
12345
Mean0.0250.2560.3920.1960.1320.363
ZimbabweL0.0180.2360.3680.1730.1120.363
(4150.29)U0.0340.2780.4150.2160.1490.363
Mean0.0540.4420.2670.1690.0690.350
BLORMéxicoL0.0350.3990.2280.1390.0500.349
(1313.82)U0.0750.4800.3080.2000.0920.352
Mean0.2040.2480.2590.2120.0770.390
ColombiaL0.1790.2190.2320.1870.0580.389
(2577.92)U0.2330.2790.2870.2390.0950.391
Mean0.0830.2840.3320.1980.1040.377
PooledL0.0690.2690.3150.1840.0930.377
(8327.36)U0.0930.2990.3500.2110.1190.377
Mean0.0770.1500.4320.3410.335
SeptoriaL0.0490.1120.3760.2940.333
(654.33)U0.1100.1950.4930.3880.338
Mean0.0320.2370.4160.1900.1240.363
ZimbabweL0.0250.2190.3890.1710.1100.363
(4156.86)U0.0390.2580.4430.2110.1400.365
Mean0.0570.4360.2800.1560.0700.350
BLOR*MéxicoL0.0410.3920.2420.1280.0540.350
(1315.21)U0.0750.4790.3230.1870.0890.353
Mean0.1930.2600.2790.1940.0740.390
ColombiaL0.1680.2260.2480.1680.0600.389
(2581.66)U0.2200.2920.3100.2230.0890.392
Mean0.0820.2770.3580.1840.1000.377
PooledL0.0740.2590.3410.1680.0900.377
(8339.54)U0.0910.2940.3750.2000.1090.378
Mean0.0750.1370.4570.3320.334
SeptoriaL0.0510.0980.3920.2680.330
(652.50)U0.1040.1760.5270.3930.341
Mean0.0250.2530.3920.1990.1320.363
ZimbabweL0.0170.2330.3660.1790.1130.363
(4150.18)U0.0330.2740.4160.2190.1480.363
Mean0.0540.4400.2650.1710.0700.350
BPORMéxicoL0.0360.3990.2290.1420.0510.350
(1314.70)U0.0760.4790.3040.2030.0920.352
Mean0.2060.2490.2610.2090.0750.390
ColombiaL0.1760.2210.2300.1830.0580.389
(2578.42)U0.2330.2770.2930.2330.0950.391
Mean0.0710.2560.3310.2180.1230.377
PooledL0.0420.1830.3130.1910.1040.374
(8329.84)U0.0860.2870.3500.2860.1710.389
Mean0.0750.1500.4300.3450.334
SeptoriaL0.0470.1100.3690.2820.330
(651.79)U0.1090.1910.5000.4020.339

BLOR* use the parameter estimates of the BPOR and approximate the BLOR with . Point probability estimates, credible sets for each category, DIC, and BS for the threshold Bayesian ridge regression. L and U denote lower and upper confidence sets, respectively. GLS, Gray leaf spot; DIC, deviance information criterion; BS, Brier scores; BLOR, Bayesian logistic ordinal regression; BPOR, Bayesian probit ordinal regression.

Comparing the models with DIC, we see that in Zimbabwe, BPOR produced the lowest DIC (DIC = 4150.18), whereas BLOR* produced the highest (DIC = 4156.86). However, the DIC of BLOR (DIC = 4150.29) was very close to that of BPOR. In México, the lowest DIC was for BLOR (DIC = 1313.82), whereas the highest was for BLOR* (DIC = 1315.21). In Colombia, BLOR had the lowest DIC (2577.92) and BLOR* had the highest (2581.66). In the Septoria data set, BPOR had the lowest DIC (651.79) and BLOR had the highest (654.33). Finally, in the pooled data, BLOR (8327.36) also had the lowest DIC, whereas BLOR* had the highest (8339.539). All these results shows that even approximations that did a reasonable job (BLOR*) were sometimes very far from the exact methods (BLOR and BPOR). For this reason, whenever possible, an exact model (BLOR or BPOR) should be chosen. The fact that BPOR is sometimes the best (lower DIC) implies that for these data sets, the assumption in Equation (1) is enough. Finally, in Table 4 we present the BS for the testing sets of the pooled real GLS data with 10 cross-validations and 90% of data used for the training set and 10% for the testing set. For models BLOR and BPOR, the BS are almost identical, which means that, with regard to prediction, both models had a similar performance. However, although the approximation method (BLOR*) produced a higher BS, its prediction accuracy was very close to that of BLOR and BPOR. Although we found that the three models had a similar performance regarding prediction accuracy with this data set, more in-depth research is required to validate this observation.
Table 4

GLS data set

ModelBrier Scores
MeanMinMax
BLOR0.3730.3650.381
BLOR*0.3740.3640.382
BPOR0.3730.3650.381

BLOR* uses the parameter estimates of BPOR and approximates BLOR with . Brier scores (mean, minimum and maximum; lower scores indicate better prediction) evaluated for validation samples from the pooled data. GLS, Gray leaf spot; BLOR, Bayesian logistic ordinal regression; BPOR, Bayesian probit ordinal regression.

BLOR* use the parameter estimates of the BPOR and approximate the BLOR with . Point probability estimates, credible sets for each category, DIC, and BS for the threshold Bayesian ridge regression. L and U denote lower and upper confidence sets, respectively. GLS, Gray leaf spot; DIC, deviance information criterion; BS, Brier scores; BLOR, Bayesian logistic ordinal regression; BPOR, Bayesian probit ordinal regression. BLOR* uses the parameter estimates of BPOR and approximates BLOR with . Brier scores (mean, minimum and maximum; lower scores indicate better prediction) evaluated for validation samples from the pooled data. GLS, Gray leaf spot; BLOR, Bayesian logistic ordinal regression; BPOR, Bayesian probit ordinal regression. This paper proposes a method for BLOR using the Pólya-Gamma data augmentation approach of Scott and Pillow (2013), which produces a Gibbs sampler with full conditional distributions similar to that of the BPOR model of Albert and Chib (1993). The proposed method is reduced to the BPOR model when the sampled values, , from the Pólya-Gamma distribution in Equation (4) are set to 1. This is an advantage because with the proposed model, researchers can perform an exact logistic or probit ordinal regression without having to do approximations to perform a logistic ordinal regression. The performance of the proposed method was compared with the approximation using the probit ordinal regression model in a small simulation study and real data sets (GLS and Septoria data sets) using a four- and five-point ordinal scale. On the basis of the simulation study, it is clear that the estimation of parameters using the approximation produces a considerable amount of bias and can give rise to wrong conclusions in association studies. However, we observed with the real data that, in terms of prediction ability, both models (BLOR and BPOR) have a similar performance even though we observed BLOR had lower DIC values in México, Colombia, and the pooled data. This means that when violation of the assumption in Equation (1) is not strong, any model can be used. For this reason, we observed greater accuracy (narrow confidence sets) for the BLOR model compared with the exact BPOR model (BPOR) only with the pooled data set without a covariate for location. Although with the real data we did not observe an advantage in prediction accuracy with the proposed BLOR model, it is very well documented in statistical literature that logistic ordinal regression is more robust for dealing with outlying data, because logistic distribution has heavier tails which was corroborated in terms of parameter estimates with the simulation done using simulated data set 2. For this reason, the proposed BLOR should be preferred because it is usually not practical to test if the error term in Equation (1) is or . In addition to being more robust, the proposed method also provides regression coefficients that are more interpretable because of their connection to odds ratios (Zucknick and Richardson 2014). However, this advantage does not make sense when p >> n, because the main driving force in Bayesian models in the case of p >> n is the prior and not the data (Gianola 2013). Even with this restriction, this paper unifies logistic and probit ordinal regression under a Bayesian framework and is a useful alternative for genomic-enabled prediction of ordinal categorical trials where available data sets have a larger number of parameters to estimate than observations. Also, the proposed method should be preferred over BPOR when outliers are present and not easily detected. This is especially true for multidimensional data, since many times a few outliers in a data set can have strong influence on parameter estimation and inference. Finally, it is important to point out that, to devise the method proposed in this paper, we generalized the work of Scott and Pillow (2013) for ordered categorical responses. Our method is elegant, easy to implement, and produces a unified Gibbs sampler framework useful for both the logit and the probit link. For this reason, we believe it is an appealing alternative for plant and animal researchers. Also, the proposed BLOR model can be easily extended to take into account genotype × environment interactions, which play an extremely important role in plant breeding, especially when selecting candidate genotypes.
  11 in total

1.  The use of transformations.

Authors:  M S BARTLETT
Journal:  Biometrics       Date:  1947-03       Impact factor: 2.571

2.  Accuracy of genome-wide evaluation for disease resistance in aquaculture breeding programs.

Authors:  B Villanueva; J Fernández; L A García-Cortés; L Varona; H D Daetwyler; M A Toro
Journal:  J Anim Sci       Date:  2011-07-08       Impact factor: 3.159

3.  Bayesian methods for estimating GEBVs of threshold traits.

Authors:  C-L Wang; X-D Ding; J-Y Wang; J-F Liu; W-X Fu; Z Zhang; Z-J Yin; Q Zhang
Journal:  Heredity (Edinb)       Date:  2012-10-31       Impact factor: 3.821

4.  Sire evaluation for ordered categorical data with a threshold model.

Authors:  D Gianola; J Foulley
Journal:  Genet Sel Evol       Date:  1983       Impact factor: 4.297

5.  Priors in whole-genome regression: the bayesian alphabet returns.

Authors:  Daniel Gianola
Journal:  Genetics       Date:  2013-05-01       Impact factor: 4.562

6.  A method of sire evaluation for dichotomies.

Authors:  D Gianola
Journal:  J Anim Sci       Date:  1980-12       Impact factor: 3.159

7.  A Bayesian antedependence model for whole genome prediction.

Authors:  Wenzhao Yang; Robert J Tempelman
Journal:  Genetics       Date:  2011-11-30       Impact factor: 4.562

8.  Genome-wide prediction of discrete traits using Bayesian regressions and machine learning.

Authors:  Oscar González-Recio; Selma Forni
Journal:  Genet Sel Evol       Date:  2011-02-17       Impact factor: 4.297

9.  Genome-wide regression and prediction with the BGLR statistical package.

Authors:  Paulino Pérez; Gustavo de los Campos
Journal:  Genetics       Date:  2014-07-09       Impact factor: 4.562

10.  Threshold models for genome-enabled prediction of ordinal categorical traits in plant breeding.

Authors:  Osval A Montesinos-López; Abelardo Montesinos-López; Paulino Pérez-Rodríguez; Gustavo de Los Campos; Kent Eskridge; José Crossa
Journal:  G3 (Bethesda)       Date:  2014-12-23       Impact factor: 3.154

View more
  6 in total

1.  Multi-dimensional machine learning approaches for fruit shape phenotyping in strawberry.

Authors:  Mitchell J Feldmann; Michael A Hardigan; Randi A Famula; Cindy M López; Amy Tabb; Glenn S Cole; Steven J Knapp
Journal:  Gigascience       Date:  2020-05-01       Impact factor: 6.524

Review 2.  Genome and Environment Based Prediction Models and Methods of Complex Traits Incorporating Genotype × Environment Interaction.

Authors:  José Crossa; Osval Antonio Montesinos-López; Paulino Pérez-Rodríguez; Germano Costa-Neto; Roberto Fritsche-Neto; Rodomiro Ortiz; Johannes W R Martini; Morten Lillemo; Abelardo Montesinos-López; Diego Jarquin; Flavio Breseghello; Jaime Cuevas; Renaud Rincent
Journal:  Methods Mol Biol       Date:  2022

3.  A semi-parametric Bayesian model for semi-continuous longitudinal data.

Authors:  Junting Ren; Susan Tapert; Chun Chieh Fan; Wesley K Thompson
Journal:  Stat Med       Date:  2022-03-10       Impact factor: 2.497

4.  Genomic Bayesian Prediction Model for Count Data with Genotype × Environment Interaction.

Authors:  Abelardo Montesinos-López; Osval A Montesinos-López; José Crossa; Juan Burgueño; Kent M Eskridge; Esteban Falconi-Castillo; Xinyao He; Pawan Singh; Karen Cichy
Journal:  G3 (Bethesda)       Date:  2016-05-03       Impact factor: 3.154

5.  An R Package for Bayesian Analysis of Multi-environment and Multi-trait Multi-environment Data for Genome-Based Prediction.

Authors:  Osval A Montesinos-López; Abelardo Montesinos-López; Francisco Javier Luna-Vázquez; Fernando H Toledo; Paulino Pérez-Rodríguez; Morten Lillemo; José Crossa
Journal:  G3 (Bethesda)       Date:  2019-05-07       Impact factor: 3.154

6.  Classification and Regression Models for Genomic Selection of Skewed Phenotypes: A Case for Disease Resistance in Winter Wheat (Triticum aestivum L.).

Authors:  Lance F Merrick; Dennis N Lozada; Xianming Chen; Arron H Carter
Journal:  Front Genet       Date:  2022-02-23       Impact factor: 4.599

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.