| Literature DB >> 24465052 |
Xiaotong Shen1, Wei Pan2, Yunzhang Zhu1, Hui Zhou2.
Abstract
High-dimensional feature selection has become increasingly crucial for seeking parsimonious models in estimation. For selection consistency, we derive one necessary and sufficient condition formulated on the notion of degree-of-separation. The minimal degree of separation is necessary for any method to be selection consistent. At a level slightly higher than the minimal degree of separation, selection consistency is achieved by a constrained L0-method and its computational surrogate-the constrained truncated L1-method. This permits up to exponentially many features in the sample size. In other words, these methods are optimal in feature selection against any selection method. In contrast, their regularization counterparts-the L0-regularization and truncated L1-regularization methods enable so under slightly stronger assumptions. More importantly, sharper parameter estimation/prediction is realized through such selection, leading to minimax parameter estimation. This, otherwise, is impossible in absence of a good selection method for high-dimensional analysis.Entities:
Keywords: (p, n) versus fixed p-asymptotics; Constrained regression; difference convex programming; nonconvex regularization; parameter and nonparametric models
Year: 2013 PMID: 24465052 PMCID: PMC3898843 DOI: 10.1007/s10463-012-0396-3
Source DB: PubMed Journal: Ann Inst Stat Math ISSN: 0020-3157 Impact factor: 1.267