scispace - formally typeset
Open AccessJournal ArticleDOI

Sparse PCA: Optimal rates and adaptive estimation

T. Tony Cai, +2 more
- 01 Dec 2013 - 
- Vol. 41, Iss: 6, pp 3074-3110
Reads0
Chats0
TLDR
In this paper, the authors considered both minimax and adaptive estimation of the principal subspace in the high dimensional setting and established the optimal rates of convergence for estimating the subspace which are sharp with respect to all the parameters, thus providing a complete characterization of the difficulty of the estimation problem in terms of the convergence rate.
Abstract
Principal component analysis (PCA) is one of the most commonly used statistical procedures with a wide range of applications. This paper considers both minimax and adaptive estimation of the principal subspace in the high dimensional setting. Under mild technical conditions, we first establish the optimal rates of convergence for estimating the principal subspace which are sharp with respect to all the parameters, thus providing a complete characterization of the difficulty of the estimation problem in term of the convergence rate. The lower bound is obtained by calculating the local metric entropy and an application of Fano’s lemma. The rate optimal estimator is constructed using aggregation, which, however, might not be computationally feasible. We then introduce an adaptive procedure for estimating the principal subspace which is fully data driven and can be computed efficiently. It is shown that the estimator attains the optimal rates of convergence simultaneously over a large collection of the parameter spaces. A key idea in our construction is a reduction scheme which reduces the sparse PCA problem to a high-dimensional multivariate regression problem. This method is potentially also useful for other related problems.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Streaming Linear System Identification with Reverse Experience Replay

TL;DR: This work provides the first - to the best of the knowledge - optimal SGD-style algorithm for the classical problem of linear system identification aka VAR model estimation and demonstrates that knowledge of dependency structure can aid us in designing algorithms which can deconstruct the dependencies between samples optimally in an online fashion.
Journal ArticleDOI

Geometric Inference for General High-Dimensional Linear Inverse Problems

TL;DR: A unified geometric framework for the statistical analysis of a general ill-posed linear inverse model which includes as special cases noisy compressed sensing, sign vector recovery, trace regression, orthogonal matrix estimation, and noisy matrix completion is presented.
Posted Content

Near-Optimal Estimation of Simultaneously Sparse and Low-Rank Matrices from Nested Linear Measurements

TL;DR: In this paper, the authors considered the problem of estimating simultaneously low-rank and row-wise sparse matrices from nested linear measurements where the linear operator consists of the product of a linear operator and a matrix.
Posted Content

Diffusion Approximations for Online Principal Component Estimation and Global Convergence

TL;DR: The diffusion approximation tools are adopted to study the dynamics of Oja's iteration which is an online stochastic gradient method for the principal component analysis and it is shown that the Ojas iteration for the top eigenvector generates a continuous-state discrete-time Markov chain over the unit sphere.
Journal ArticleDOI

The generalized orthogonal Procrustes problem in the high noise regime

TL;DR: In this paper, a simple and efficient procedure based on invariant polynomials (effectively: the Gram matrices) is proposed to recover the signal, and the approach adapts to the noise level and is statistically optimal for both the low and high noise regimes.
References
More filters
Book

Elements of information theory

TL;DR: The author examines the role of entropy, inequality, and randomness in the design of codes and the construction of codes in the rapidly changing environment.
Book

Matrix Analysis

TL;DR: In this article, the authors present results of both classic and recent matrix analyses using canonical forms as a unifying theme, and demonstrate their importance in a variety of applications, such as linear algebra and matrix theory.
Book

An Introduction to Multivariate Statistical Analysis

TL;DR: In this article, the distribution of the Mean Vector and the Covariance Matrix and the Generalized T2-Statistic is analyzed. But the distribution is not shown to be independent of sets of Variates.
Journal ArticleDOI

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization

TL;DR: It is shown that if a certain restricted isometry property holds for the linear transformation defining the constraints, the minimum-rank solution can be recovered by solving a convex optimization problem, namely, the minimization of the nuclear norm over the given affine space.
Related Papers (5)