The Fast Johnson-Lindenstrauss Transform and Approximate Nearest Neighbors
Nir Ailon,Bernard Chazelle +1 more
Reads0
Chats0
TLDR
A new low-distortion embedding of $\ell-2^d$ into $\ell_p^{O(\log n)}$ ($p=1,2$) called the fast Johnson-Lindenstrauss transform (FJLT) is introduced, based upon the preconditioning of a sparse projection matrix with a randomized Fourier transform.Abstract:
We introduce a new low-distortion embedding of $\ell_2^d$ into $\ell_p^{O(\log n)}$ ($p=1,2$) called the fast Johnson-Lindenstrauss transform (FJLT). The FJLT is faster than standard random projections and just as easy to implement. It is based upon the preconditioning of a sparse projection matrix with a randomized Fourier transform. Sparse random projections are unsuitable for low-distortion embeddings. We overcome this handicap by exploiting the “Heisenberg principle” of the Fourier transform, i.e., its local-global duality. The FJLT can be used to speed up search algorithms based on low-distortion embeddings in $\ell_1$ and $\ell_2$. We consider the case of approximate nearest neighbors in $\ell_2^d$. We provide a faster algorithm using classical projections, which we then speed up further by plugging in the FJLT. We also give a faster algorithm for searching over the hypercube.read more
Citations
More filters
Journal ArticleDOI
Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions
Alexandr Andoni,Piotr Indyk +1 more
TL;DR: An algorithm for the c-approximate nearest neighbor problem in a d-dimensional Euclidean space, achieving query time of O(dn 1c2/+o(1)) and space O(DN + n1+1c2 + o(1) + 1/c2), which almost matches the lower bound for hashing-based algorithm recently obtained.
Journal ArticleDOI
User-Friendly Tail Bounds for Sums of Random Matrices
TL;DR: This paper presents new probability inequalities for sums of independent, random, self-adjoint matrices and provides noncommutative generalizations of the classical bounds associated with the names Azuma, Bennett, Bernstein, Chernoff, Hoeffding, and McDiarmid.
Journal ArticleDOI
Approximate Nearest Neighbor: Towards Removing the Curse of Dimensionality
TL;DR: Two algorithms for the approximate nearest neighbor problem in high dimensional spaces for data sets of size n living in IR are presented, achieving query times that are sub-linear in n and polynomial in d.
Journal ArticleDOI
Compressed sensing with coherent and redundant dictionaries
TL;DR: A condition on the measurement/sensing matrix is introduced, which is a natural generalization of the now well-known restricted isometry property, and which guarantees accurate recovery of signals that are nearly sparse in (possibly) highly overcomplete and coherent dictionaries.
Posted Content
Compressed Sensing with Coherent and Redundant Dictionaries
TL;DR: In this article, a condition on the measurement/sensing matrix is introduced, which guarantees accurate recovery of signals that are nearly sparse in (possibly) highly overcomplete and coherent dictionaries.
References
More filters
Journal ArticleDOI
Monte Carlo Sampling Methods Using Markov Chains and Their Applications
TL;DR: A generalization of the sampling method introduced by Metropolis et al. as mentioned in this paper is presented along with an exposition of the relevant theory, techniques of application and methods and difficulties of assessing the error in Monte Carlo estimates.
Book
The Probabilistic Method
TL;DR: A particular set of problems - all dealing with “good” colorings of an underlying set of points relative to a given family of sets - is explored.
Proceedings ArticleDOI
Approximate nearest neighbors: towards removing the curse of dimensionality
Piotr Indyk,Rajeev Motwani +1 more
TL;DR: In this paper, the authors present two algorithms for the approximate nearest neighbor problem in high-dimensional spaces, for data sets of size n living in R d, which require space that is only polynomial in n and d.
Journal ArticleDOI
An optimal algorithm for approximate nearest neighbor searching fixed dimensions
TL;DR: In this paper, it was shown that given an integer k ≥ 1, (1 + ϵ)-approximation to the k nearest neighbors of q can be computed in additional O(kd log n) time.