Literature DB >> 25620858

Tighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Time.

Zhaoran Wang1, Huanran Lu1, Han Liu1.   

Abstract

We provide statistical and computational analysis of sparse Principal Component Analysis (PCA) in high dimensions. The sparse PCA problem is highly nonconvex in nature. Consequently, though its global solution attains the optimal statistical rate of convergence, such solution is computationally intractable to obtain. Meanwhile, although its convex relaxations are tractable to compute, they yield estimators with suboptimal statistical rates of convergence. On the other hand, existing nonconvex optimization procedures, such as greedy methods, lack statistical guarantees. In this paper, we propose a two-stage sparse PCA procedure that attains the optimal principal subspace estimator in polynomial time. The main stage employs a novel algorithm named sparse orthogonal iteration pursuit, which iteratively solves the underlying nonconvex problem. However, our analysis shows that this algorithm only has desired computational and statistical guarantees within a restricted region, namely the basin of attraction. To obtain the desired initial estimator that falls into this region, we solve a convex formulation of sparse PCA with early stopping. Under an integrated analytic framework, we simultaneously characterize the computational and statistical performance of this two-stage procedure. Computationally, our procedure converges at the rate of [Formula: see text] within the initialization stage, and at a geometric rate within the main stage. Statistically, the final principal subspace estimator achieves the minimax-optimal statistical rate of convergence with respect to the sparsity level s*, dimension d and sample size n. Our procedure motivates a general paradigm of tackling nonconvex statistical learning problems with provable statistical guarantees.

Entities:  

Year:  2014        PMID: 25620858      PMCID: PMC4301447     

Source DB:  PubMed          Journal:  Adv Neural Inf Process Syst        ISSN: 1049-5258


  7 in total

1.  A penalized matrix decomposition, with applications to sparse principal components and canonical correlation analysis.

Authors:  Daniela M Witten; Robert Tibshirani; Trevor Hastie
Journal:  Biostatistics       Date:  2009-04-17       Impact factor: 5.899

2.  Tighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Time.

Authors:  Zhaoran Wang; Huanran Lu; Han Liu
Journal:  Adv Neural Inf Process Syst       Date:  2014

3.  A STRICTLY CONTRACTIVE PEACEMAN-RACHFORD SPLITTING METHOD FOR CONVEX PROGRAMMING.

Authors:  He Bingsheng; Han Liu; Zhaoran Wang; Xiaoming Yuan
Journal:  SIAM J Optim       Date:  2014-07       Impact factor: 2.850

4.  On Consistency and Sparsity for Principal Components Analysis in High Dimensions.

Authors:  Iain M Johnstone; Arthur Yu Lu
Journal:  J Am Stat Assoc       Date:  2009-06-01       Impact factor: 5.033

5.  OPTIMAL COMPUTATIONAL AND STATISTICAL RATES OF CONVERGENCE FOR SPARSE NONCONVEX LEARNING PROBLEMS.

Authors:  Zhaoran Wang; Han Liu; Tong Zhang
Journal:  Ann Stat       Date:  2014       Impact factor: 4.028

6.  MINIMAX BOUNDS FOR SPARSE PCA WITH NOISY HIGH-DIMENSIONAL DATA.

Authors:  Aharon Birnbaum; Iain M Johnstone; Boaz Nadler; Debashis Paul
Journal:  Ann Stat       Date:  2013-06       Impact factor: 4.028

7.  Analysis of population structure: a unifying framework and novel methods based on sparse factor analysis.

Authors:  Barbara E Engelhardt; Matthew Stephens
Journal:  PLoS Genet       Date:  2010-09-16       Impact factor: 5.917

  7 in total
  2 in total

1.  Tighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Time.

Authors:  Zhaoran Wang; Huanran Lu; Han Liu
Journal:  Adv Neural Inf Process Syst       Date:  2014

2.  Sparse PCA with Oracle Property.

Authors:  Quanquan Gu; Zhaoran Wang; Han Liu
Journal:  Adv Neural Inf Process Syst       Date:  2014
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.