scispace - formally typeset
Open AccessPosted Content

Iterative Hessian sketch: Fast and accurate solution approximation for constrained least-squares

Reads0
Chats0
TLDR
This work provides a general lower bound on any randomized method that sketches both the data matrix and vector in a least-squares problem and presents a new method known as the iterative Hessian sketch, which can be used to obtain approximations to the original least- Squares problem using a projection dimension proportional to the statistical complexity of the least-Squares minimizer, and a logarithmic number of iterations.
Abstract
We study randomized sketching methods for approximately solving least-squares problem with a general convex constraint. The quality of a least-squares approximation can be assessed in different ways: either in terms of the value of the quadratic objective function (cost approximation), or in terms of some distance measure between the approximate minimizer and the true minimizer (solution approximation). Focusing on the latter criterion, our first main result provides a general lower bound on any randomized method that sketches both the data matrix and vector in a least-squares problem; as a surprising consequence, the most widely used least-squares sketch is sub-optimal for solution approximation. We then present a new method known as the iterative Hessian sketch, and show that it can be used to obtain approximations to the original least-squares problem using a projection dimension proportional to the statistical complexity of the least-squares minimizer, and a logarithmic number of iterations. We illustrate our general theory with simulations for both unconstrained and constrained versions of least-squares, including $\ell_1$-regularization and nuclear norm constraints. We also numerically demonstrate the practicality of our approach in a real face expression classification experiment.

read more

Citations
More filters
Journal ArticleDOI

Communication-Efficient Distributed Statistical Inference

TL;DR: In this paper, a communication-efficient surrogate likelihood (CSL) framework for distributed statistical inference problems is presented, which provides a communication efficient surrogate to the global likelihoods.
Journal ArticleDOI

Newton Sketch: A Near Linear-Time Optimization Algorithm with Linear-Quadratic Convergence

TL;DR: In this paper, the authors proposed a randomized second-order method for optimization known as the Newton sketch, which is based on performing an approximate Newton step using a randomly projected Hessian.
Journal ArticleDOI

Randomized sketches for kernels: Fast and optimal nonparametric regression

TL;DR: In this article, a lower bound on the minimax risk of kernel regression in terms of the localized Rademacher complexity is established. But this lower bound is not applicable to nonparametric regression.
Journal ArticleDOI

Randomized numerical linear algebra: Foundations and algorithms

TL;DR: This survey describes probabilistic algorithms for linear algebraic computations, such as factorizing matrices and solving linear systems, that have a proven track record for real-world problems and treats both the theoretical foundations of the subject and practical computational issues.
Journal ArticleDOI

Randomized Sketches of Convex Programs With Sharp Guarantees

TL;DR: This work analyzes RP-based approximations of convex programs, in which the original optimization problem is approximated by solving a lower dimensional problem, and proves that the approximation ratio of this procedure can be bounded in terms of the geometry of the constraint set.
References
More filters
Book

Elements of information theory

TL;DR: The author examines the role of entropy, inequality, and randomness in the design of codes and the construction of codes in the rapidly changing environment.
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems

TL;DR: A new fast iterative shrinkage-thresholding algorithm (FISTA) which preserves the computational simplicity of ISTA but with a global rate of convergence which is proven to be significantly better, both theoretically and practically.
Journal ArticleDOI

Atomic Decomposition by Basis Pursuit

TL;DR: Basis Pursuit (BP) is a principle for decomposing a signal into an "optimal" superposition of dictionary elements, where optimal means having the smallest l1 norm of coefficients among all such decompositions.
Journal ArticleDOI

Model selection and estimation in regression with grouped variables

TL;DR: In this paper, instead of selecting factors by stepwise backward elimination, the authors focus on the accuracy of estimation and consider extensions of the lasso, the LARS algorithm and the non-negative garrotte for factor selection.
Related Papers (5)