| Literature DB >> 27499558 |
Shifeng Xiong1, Bin Dai2, Jared Huling3, Peter Z G Qian3.
Abstract
We introduce an efficient iterative algorithm, intended for various least squares problems, based on a design of experiments perspective. The algorithm, called orthogonalizing EM (OEM), works for ordinary least squares and can be easily extended to penalized least squares. The main idea of the procedure is to orthogonalize a design matrix by adding new rows and then solve the original problem by embedding the augmented design in a missing data framework. We establish several attractive theoretical properties concerning OEM. For the ordinary least squares with a singular regression matrix, an OEM sequence converges to the Moore-Penrose generalized inverse-based least squares estimator. For ordinary and penalized least squares with various penalties, it converges to a point having grouping coherence for fully aliased regression matrices. Convergence and the convergence rate of the algorithm are examined. Finally, we demonstrate that OEM is highly efficient for large-scale least squares and penalized least squares problems, and is considerably faster than competing methods when n is much larger than p. Supplementary materials for this article are available online.Entities:
Keywords: Computational statistics; Design of experiments; MCP; Missing data; Optimization; Orthogonal design; SCAD; The Lasso
Year: 2016 PMID: 27499558 PMCID: PMC4974022 DOI: 10.1080/00401706.2015.1054436
Source DB: PubMed Journal: Technometrics ISSN: 0040-1706