scispace - formally typeset
Open AccessProceedings Article

List-decodable Linear Regression

Reads0
Chats0
TLDR
The first polynomial-time algorithm for robust regression in the list-decodable setting where an adversary can corrupt a greater than 1/2 fraction of examples was given in this paper.
Abstract
We give the first polynomial-time algorithm for robust regression in the list-decodable setting where an adversary can corrupt a greater than 1/2 fraction of examples. For any \alpha < 1, our algorithm takes as input a sample {(x_i,y_i)}_{i \leq n} of n linear equations where \alpha n of the equations satisfy y_i = \langle x_i,\ell^*\rangle +\zeta for some small noise \zeta and (1-\alpha) n of the equations are {\em arbitrarily} chosen. It outputs a list L of size O(1/\alpha) - a fixed constant - that contains an \ell that is close to \ell^*. Our algorithm succeeds whenever the inliers are chosen from a certifiably anti-concentrated distribution D. In particular, this gives a (d/\alpha)^{O(1/\alpha^8)} time algorithm to find a O(1/\alpha) size list when the inlier distribution is a standard Gaussian. For discrete product distributions that are anti-concentrated only in regular directions, we give an algorithm that achieves similar guarantee under the promise that \ell^* has all coordinates of the same magnitude. To complement our result, we prove that the anti-concentration assumption on the inliers is information-theoretically necessary. To solve the problem we introduce a new framework for list-decodable learning that strengthens the ``identifiability to algorithms'' paradigm based on the sum-of-squares method.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted Content

Recent Advances in Algorithmic High-Dimensional Robust Statistics.

TL;DR: The core ideas and algorithmic techniques in the emerging area of algorithmic high-dimensional robust statistics with a focus on robust mean estimation are introduced and an overview of the approaches that have led to computationally efficient robust estimators for a range of broader statistical tasks are provided.
Proceedings Article

Reducibility and Statistical-Computational Gaps from Secret Leakage

TL;DR: This work gives the first evidence that an expanded set of hardness assumptions, such as for secret leakage planted clique, may be a key first step towards a more complete theory of reductions among statistical problems.
Proceedings Article

List decodable learning via sum of squares

TL;DR: A framework for list-decodable learning via the Sum-of-Squares SDP hierarchy is developed and an algorithm that outputs a list of linear functions such that there exists some $\hat{\ell} \in \mathcal{L}$ that is close to $\ell$.
Posted Content

Robust Linear Regression: Optimal Rates in Polynomial Time

TL;DR: The central technical contribution is to algorithmically exploit independence of random variables in the ”sum-of-squares” framework by formulating it as the aforementioned polynomial inequality.
Posted Content

Robustly Learning Mixtures of k Arbitrary Gaussians.

TL;DR: The main tools are an efficient \emph{partial clustering} algorithm that relies on the sum-of-squares method, and a novel tensor decomposition algorithm that allows errors in both Frobenius norm and low-rank terms.
References
More filters
Journal ArticleDOI

The ellipsoid method and its consequences in combinatorial optimization

TL;DR: The method yields polynomial algorithms for vertex packing in perfect graphs, for the matching and matroid intersection problems, for optimum covering of directed cuts of a digraph, and for the minimum value of a submodular set function.
Book ChapterDOI

Hierarchical mixtures of experts and the EM algorithm

TL;DR: An expectation-maximization (EM) algorithm for adjusting the parameters of the tree-structured architecture for supervised learning is presented and an online learning algorithm in which the parameters are updated incrementally is developed.
Book ChapterDOI

Sums of Squares, Moment Matrices and Optimization Over Polynomials

TL;DR: This work considers the problem of minimizing a polynomial over a semialgebraic set defined byPolynomial equations and inequalities, which is NP-hard in general and reviews the mathematical tools underlying these properties.
Journal ArticleDOI

On a lemma of Littlewood and Offord

TL;DR: In this paper, it was shown that if all the Xi are not less than 1, then the interval (1, + 1 ) contains Cn,m s u m s s l, n even.
Book ChapterDOI

Squared Functional Systems and Optimization Problems

Yu. Nesterov
TL;DR: It is proved that such cones can be always seen as a linear image of the cone of positive semidefinite matrices, and a description of the cones of univariate and non-negative trigonometric polynomials is given.
Related Papers (5)