Skip to main content

Tighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Time

Author(s): Wang, Z; Lu, H; Liu, H

Download
To refer to this page use: http://arks.princeton.edu/ark:/88435/pr16869
Full metadata record
DC FieldValueLanguage
dc.contributor.authorWang, Z-
dc.contributor.authorLu, H-
dc.contributor.authorLiu, H-
dc.date.accessioned2021-10-11T14:17:08Z-
dc.date.available2021-10-11T14:17:08Z-
dc.date.issued2014en_US
dc.identifier.citationWang, Zhaoran, Huanran Lu, and Han Liu. "Tighten after relax: Minimax-optimal sparse PCA in polynomial time." In Advances in Neural Information Processing Systems 27, pp. 3383-3391. 2014.en_US
dc.identifier.issn1049-5258-
dc.identifier.urihttp://papers.nips.cc/paper/5252-tighten-after-relax-minimax-optimal-sparse-pca-in-polynomial-time-
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/pr16869-
dc.description.abstractWe provide statistical and computational analysis of sparse Principal Component Analysis (PCA) in high dimensions. The sparse PCA problem is highly nonconvex in nature. Consequently, though its global solution attains the optimal statistical rate of convergence, such solution is computationally intractable to obtain. Meanwhile, although its convex relaxations are tractable to compute, they yield estimators with suboptimal statistical rates of convergence. On the other hand, existing nonconvex optimization procedures, such as greedy methods, lack statistical guarantees. In this paper, we propose a two-stage sparse PCA procedure that attains the optimal principal subspace estimator in polynomial time. The main stage employs a novel algorithm named sparse orthogonal iteration pursuit, which iteratively solves the underlying nonconvex problem. However, our analysis shows that this algorithm only has desired computational and statistical guarantees within a restricted region, namely the basin of attraction. To obtain the desired initial estimator that falls into this region, we solve a convex formulation of sparse PCA with early stopping. Under an integrated analytic framework, we simultaneously characterize the computational and statistical performance of this two-stage procedure. Computationally, our procedure converges at the rate of 1/t√ within the initialization stage, and at a geometric rate within the main stage. Statistically, the final principal subspace estimator achieves the minimax-optimal statistical rate of convergence with respect to the sparsity level s∗, dimension d and sample size n. Our procedure motivates a general paradigm of tackling nonconvex statistical learning problems with provable statistical guarantees.en_US
dc.format.extent3383 - 3391en_US
dc.language.isoen_USen_US
dc.relation.ispartofAdvances in Neural Information Processing Systemsen_US
dc.rightsFinal published version. Article is made available in OAR by the publisher's permission or policy.en_US
dc.titleTighten after Relax: Minimax-Optimal Sparse PCA in Polynomial Timeen_US
dc.typeConference Articleen_US
pu.type.symplectichttp://www.symplectic.co.uk/publications/atom-terms/1.0/conference-proceedingen_US

Files in This Item:
File Description SizeFormat 
MinimaxOptimalPCAPolynomialTime.pdf764.92 kBAdobe PDFView/Download


Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.